query
stringlengths
7
3.85k
document
stringlengths
11
430k
metadata
dict
negatives
sequencelengths
0
101
negative_scores
sequencelengths
0
101
document_score
stringlengths
3
10
document_rank
stringclasses
102 values
InsertP a single record using an executor, and panics on error. See Insert for whitelist behavior description.
func (o *AssetRevision) InsertP(exec boil.Executor, whitelist ...string) { if err := o.Insert(exec, whitelist...); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Failure) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) InsertP(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) {\n\tif err := o.Insert(ctx, exec, columns); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (r *TaskRepository) Insert(db db.DB, Task *entities.Task) error {\n\t_, err := db.NamedExec(`\n\tINSERT INTO tasks (uuid,title,user_id,status,created_at,updated_at)\n\tVALUES (:uuid, :title, :user_id, :status, now(), now())`, Task)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error inserting task to db: %w\", err)\n\t}\n\n\treturn nil\n}", "func Insert(db gorp.SqlExecutor, i interface{}) error {\n\treturn Mapper.Insert(db, i)\n}", "func (m *MySQL) Insert(p *packet.Packet, interval time.Duration, t time.Time) {\n\t_, err := m.stmt.Exec(p.Interface, p.Bytes, p.SrcName, p.DstName, p.Hostname, p.Proto, p.SrcPort, p.DstPort, int(interval.Seconds()), t)\n\tif err != nil {\n\t\tlog.Println(\"sql err:\", err)\n\t\tlog.Println(\"Time:\", t.Unix())\n\t\tspew.Dump(p)\n\t}\n}", "func (o *RecordMeasure) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (impl *EmployeeRepositoryImpl) Insert(employee *domain.Employee) {\n\temployeeRaw := employee.ToEmployeeRaw()\n\temployeeRaw.Id = 0 // setting to 0 will trigger auto increment\n\tquery := \"INSERT INTO employees(id, first_name, last_name, created_at) VALUES (0, ?, ?, ?);\"\n\timpl.db.MustExec(query, employeeRaw.FirstName, employeeRaw.LastName, time.Now().UTC()) //MustExec will panic on error (Exec will not)\n\t// no error returned it will panic on error\n}", "func (i *Batch) Insert(data TaskArgs) (err error) {\n\tif i.isRun {\n\t\ti.insertChan <- data\n\t} else {\n\t\terr = errors.New(\"failed to insert, batch already stopped\")\n\t}\n\treturn\n}", "func (b *backend) insertUnsafe(t *entroq.Task) {\n\tif b.existsIDUnsafe(t.ID) {\n\t\tlog.Panicf(\"Duplicate insertion ID attempted: %v\", t.ID)\n\t}\n\t// If not there, put it there. We could also use append on the long-form\n\t// version of the heap (specifying the whole map, etc.) but this is more\n\t// straightforward. Just be aware that here be dragons due to the need\n\t// to take a pointer to the heap before manipulating it.\n\th, ok := b.heaps[t.Queue]\n\tif !ok {\n\t\th = new(taskHeap)\n\t\tb.heaps[t.Queue] = h\n\t}\n\titem := &hItem{task: t}\n\theap.Push(h, item)\n\tb.byID[t.ID] = item\n}", "func (o *Failure) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no failure provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(failureColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tfailureInsertCacheMut.RLock()\n\tcache, cached := failureInsertCache[key]\n\tfailureInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailureColumnsWithDefault,\n\t\t\tfailureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(failureType, failureMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(failureType, failureMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `failure` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `failure` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `failure` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, failurePrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into failure\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == failureMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for failure\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfailureInsertCacheMut.Lock()\n\t\tfailureInsertCache[key] = cache\n\t\tfailureInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Picture) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (w *Wrapper) Insert(data interface{}) (err error) {\n\tw.query = w.buildInsert(\"INSERT\", data)\n\tres, err := w.executeQuery()\n\tif err != nil || !w.executable {\n\t\treturn\n\t}\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn\n\t}\n\tw.LastInsertID = int(id)\n\treturn\n}", "func (instance *TopicTypeRepoImpl) Insert(data *entities.TopicType) (err error) {\n\tpanic(\"implement me\")\n}", "func (r *Repository) Insert(ctx context.Context, registrant Registrant) error {\n\ttx := r.db.Begin()\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\ttx.Rollback()\n\t\t}\n\t}()\n\n\tif err := tx.Error; err != nil {\n\t\treturn err\n\t}\n\n\tif err := tx.Create(&registrant).Error; err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\n\treturn tx.Commit().Error\n}", "func (o *Organism) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (c *MySQLClient) Insert(p *purchase.Purchase) error {\n\tif p.ID != 0 {\n\t\treturn fmt.Errorf(\"purchase cannot have a preexisting ID\")\n\t}\n\n\tvar err error\n\tvar buyBytes, sellBytes []byte\n\tif p.BuyOrder != nil {\n\t\tbuyBytes, err = json.Marshal(p.BuyOrder)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to marshal buy order: %v\", err)\n\t\t}\n\t}\n\tif p.SellOrder != nil {\n\t\tsellBytes, err = json.Marshal(p.SellOrder)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to marshal sell order: %v\", err)\n\t\t}\n\t}\n\n\tquery := `INSERT INTO trader_one(buy_order, sell_order) VALUES (?, ?)`\n\tctx, cancelFunc := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancelFunc()\n\tstmt, err := c.db.PrepareContext(ctx, query)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to prepare SQL statement: %v\", err)\n\t}\n\tdefer stmt.Close()\n\n\tres, err := stmt.ExecContext(ctx, jsonString(buyBytes), jsonString(sellBytes))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to insert row: %v\", err)\n\t}\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to find new ID: %v\", err)\n\t}\n\tp.ID = id\n\treturn nil\n}", "func (t TaskInstanceRepoCassandra) Insert(ctx context.Context, taskInstance TaskInstance) (err error) {\n\terr = t.insertTaskInstance(ctx, taskInstance)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = t.insertTaskInstanceStartedAt(ctx, taskInstance)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn t.insertInstancesByID(ctx, taskInstance)\n}", "func (s *DbRecorder) Insert() error {\n\tswitch s.flavor {\n\tcase \"postgres\":\n\t\treturn s.insertPg()\n\tdefault:\n\t\treturn s.insertStd()\n\t}\n}", "func (o *Asset) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func Insert(db *sql.DB, q Queryer) error {\n\tsql := prepareInsertString(q)\n\tif _, err := db.Exec(sql, prepareInsertArguments(q)...); err != nil {\n\t\treturn fmt.Errorf(\"Failed to insert records: %s\", err)\n\t}\n\treturn nil\n}", "func (o *DMessageEmbed) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (p *ProviderDAO) Insert(provider models.Provider) error {\n\terr := db.C(COLLECTION).Insert(&provider)\n\treturn err\n}", "func (c Client) Insert(entity interface{}, ptrResult interface{}) error {\n\treturn c.InsertInto(entity, ptrResult, reflect.TypeOf(entity).Name())\n}", "func (o *ClaimInList) InsertP(exec boil.Executor, columns boil.Columns) {\n\tif err := o.Insert(exec, columns); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q *Queue) insert(entities interface{}, mutex *sync.Mutex) *gorm.DB {\n\tmutex.Lock()\n\tres := q.db.Clauses(clause.OnConflict{DoNothing: true}).Create(entities)\n\tmutex.Unlock()\n\n\treturn res\n}", "func (tx *TestTX) Insert(list ...interface{}) error {\n\targs := tx.Called(list...)\n\treturn args.Error(0)\n}", "func (m MariaDB) Insert(ctx context.Context, document entity.PersonalData) (entity.PersonalData, error) {\n\tp := receive(document)\n\tsqlQuery := \"INSERT INTO person (id, name, last_name, phone, email, year_od_birth ) VALUES (?,?,?,?,?,?)\"\n\t_, err := m.Person.ExecContext(ctx, sqlQuery, p.ID, p.Name, p.LastName, p.Phone, p.Email, p.YearOfBirth)\n\tif err != nil {\n\t\treturn entity.PersonalData{}, errors.Wrap(err, \"could not exec query statement\")\n\t}\n\treturn document, nil\n}", "func (ps *PlayerStore) Insert(ctx context.Context, player store.Player) (*store.Player, error) {\n\tquery := `\n INSERT INTO players(roster_id,first_name,last_name,alias,status)\n VALUES($1,$2,$3,$4,$5)\n RETURNING *\n `\n\tdb := ps.db.GetDB()\n\tctx, cancel := ps.db.RequestContext(ctx)\n\tdefer cancel()\n\n\tvar p store.Player\n\terr := db.QueryRowContext(ctx, query,\n\t\tplayer.RosterID,\n\t\tplayer.FirstName,\n\t\tplayer.LastName,\n\t\tplayer.Alias,\n\t\tplayer.Status).\n\t\tScan(\n\t\t\t&p.PlayerID,\n\t\t\t&p.RosterID,\n\t\t\t&p.FirstName,\n\t\t\t&p.LastName,\n\t\t\t&p.Alias,\n\t\t\t&p.Status)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &p, nil\n}", "func (m *SnippetModel) Insert(title, content, expires string) (int, error) {\n\t// Create the SQL statement we want to execute. It's split over several lines\n\t// for readability - so it's surrounded by backquotes instead of normal double quotes\n\tstmt := `INSERT INTO snippets (title, content, created, expires)\n\tvalues (?, ?, UTC_TIMESTAMP(), DATE_ADD(UTC_TIMESTAMP, INTERVAL ? DAY))`\n\n\t// Use the Exec() method on the embedded connection pool to execute the statement.\n\t// The first parameter is the SQL statement followed by the table fields.\n\t// The method returns a sql.Result object which contains some basic information\n\t// about what happened when the statement was executed\n\tresult, err := m.DB.Exec(stmt, title, content, expires)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Use the LastInsertId() method on the result object to get the ID of our\n\t// newly inserted record in the snippets table.\n\tid, err := result.LastInsertId()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// The ID returned has the type int64 so we convert it to an int type before returning\n\treturn int(id), nil\n}", "func (repository *RecordCommandRepositoryCircuitBreaker) InsertRecord(data repositoryTypes.CreateRecord) (entity.Record, error) {\n\toutput := make(chan entity.Record, 1)\n\thystrix.ConfigureCommand(\"insert_record\", config.Settings())\n\terrors := hystrix.Go(\"insert_record\", func() error {\n\t\trecord, err := repository.RecordCommandRepositoryInterface.InsertRecord(data)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\toutput <- record\n\t\treturn nil\n\t}, nil)\n\n\tselect {\n\tcase out := <-output:\n\t\treturn out, nil\n\tcase err := <-errors:\n\t\treturn entity.Record{}, err\n\t}\n}", "func (t BlobsTable) Insert(ctx context.Context, id []byte, _type int32, data []byte) (*Blob, error) {\n\treturn t.driver.insert(ctx, id, _type, data)\n}", "func (o *Skin) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (c *Conn) Insert(cmd string, args ...interface{}) (rowid int64, err error) {\n\tn, err := c.ExecDml(cmd, args...)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\tif n == 0 { // No change => no insert...\n\t\treturn -1, nil\n\t}\n\treturn c.LastInsertRowid(), nil\n}", "func insert(ctx context.Context, tx *sqlx.Tx, todo *Todo) error {\n\tres, err := tx.NamedExecContext(ctx, insertTodo, todo)\n\tif err == nil {\n\t\ttodo.ID, err = res.LastInsertId()\n\t}\n\treturn err\n}", "func (b *PgTxRepository) Insert(ctx context.Context, data interface{}, lastInsertedID interface{}) error {\n\tcolumns, values := b.ExtractColumnPairs(data)\n\n\t// Prepare query\n\tqueryBuilder := sq.\n\t\tInsert(b.table).\n\t\tColumns(columns...).\n\t\tValues(values...).\n\t\tSuffix(\"returning \\\"id\\\"\").\n\t\tPlaceholderFormat(sq.Dollar)\n\n\t// Build SQL Query\n\tquery, args, err := queryBuilder.ToSql()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to build query: %w\", err)\n\t}\n\n\t// The statements prepared for a transaction by calling the transaction's Prepare or Stmt methods\n\t// are closed by the call to Commit or Rollback.\n\tstmt, err := b.tx.PreparexContext(ctx, query)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Do the insert query\n\tif lastInsertedID != nil {\n\t\t// We do QueryRowxContext because Postgres doesn't work with lastInsertedID\n\t\terr = stmt.QueryRowxContext(ctx, args...).Scan(lastInsertedID)\n\t} else {\n\t\t// Here we don't need the lastInsertedID\n\t\t_, err = stmt.ExecContext(ctx, args...)\n\t}\n\n\treturn err\n}", "func (c *Conn) Insert(ctx context.Context, i Item) (err error) {\n\t_, err = c.db.Exec(ctx, \"INSERT INTO jobs (url) VALUES ($1)\", i.URL)\n\treturn\n}", "func (dv *DataVault) InsertRecord(dvInsertRecord *record.DvInsertRecord) error {\n\tsqls, sqlErr := dvInsertRecord.GenerateMultiSQL()\n\n\tif sqlErr != nil {\n\t\treturn sqlErr\n\t}\n\n\ttransaction, beginErr := dv.Db.Begin()\n\tif beginErr != nil {\n\t\treturn beginErr\n\t}\n\n\t//TODO: test with various database vendor\n\tfor _, sql := range sqls {\n\t\texecErr := dv.execSQL(sql, transaction)\n\t\tif execErr != nil {\n\t\t\ttransaction.Rollback()\n\t\t\treturn execErr\n\t\t}\n\t}\n\n\tcommitErr := transaction.Commit()\n\tif commitErr != nil {\n\t\treturn commitErr\n\t}\n\n\treturn nil\n}", "func (self *CoreWorkload) DoInsert(db DB, object interface{}) bool {\n\tkeyNumber := self.keySequence.NextInt()\n\tdbKey := self.buildKeyName(keyNumber)\n\tvalues := self.buildValues(dbKey)\n\n\tvar status StatusType\n\tnumberOfRetries := int64(0)\n\tvar random *rand.Rand\n\tfor {\n\t\tstatus = db.Insert(self.table, dbKey, values)\n\t\tif status == StatusOK {\n\t\t\tbreak\n\t\t}\n\t\t// Retry if configured. Without retrying, the load process will fail\n\t\t// even if one single insertion fails. User can optionally configure\n\t\t// an insertion retry limit(default is 0) to enable retry.\n\t\tnumberOfRetries++\n\t\tif numberOfRetries < self.insertionRetryLimit {\n\t\t\tif random == nil {\n\t\t\t\trandom = rand.New(rand.NewSource(time.Now().UnixNano()))\n\t\t\t}\n\t\t\t// sleep for a random number between\n\t\t\t// [0.8, 1.2) * InsertionRetryInterval\n\t\t\tsleepTime := int64(float64(1000*self.insertionRetryInterval) * (0.8 + 0.4*random.Float64()))\n\t\t\ttime.Sleep(time.Duration(sleepTime))\n\t\t} else {\n\t\t\t// error inserting, not retrying any more\n\t\t\tbreak\n\t\t}\n\t}\n\treturn (status == StatusOK)\n}", "func (s *InstanceBindData) Insert(ibd *internal.InstanceBindData) error {\n\tif ibd == nil {\n\t\treturn errors.New(\"entity may not be nil\")\n\t}\n\n\tif ibd.InstanceID.IsZero() {\n\t\treturn errors.New(\"instance id must be set\")\n\t}\n\n\topKey := s.key(ibd.InstanceID)\n\n\trespGet, err := s.kv.Get(context.TODO(), opKey)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"while calling database on get\")\n\t}\n\tif respGet.Count > 0 {\n\t\treturn alreadyExistsError{}\n\t}\n\n\tdso, err := s.encodeDMToDSO(ibd)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := s.kv.Put(context.TODO(), opKey, dso); err != nil {\n\t\treturn errors.Wrap(err, \"while calling database on put\")\n\t}\n\n\treturn nil\n}", "func (s *SnippetStore) Insert(title, content string, expires int) (int, error) {\n\treturn 2, nil\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *repository) Insert(rec *Record) (int, error) {\n\tq := \"INSERT INTO records (name) VALUES ($1) RETURNING id;\"\n\n\tvar id int\n\tif err := r.db.QueryRow(q, rec.name).Scan(&id); err != nil {\n\t\treturn -1, err\n\t}\n\n\treturn id, nil\n}", "func (o *FeatureRelationship) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (sp *SalePermission) Insert(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\tvar res sql.Result\n\t// if already exist, bail\n\tif sp._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetSalePermissionTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// sql insert query, primary key must be provided\n\tsqlstr := `INSERT INTO ` + tableName +\n\t\t` (` +\n\t\t`spid, desc, action_class` +\n\t\t`) VALUES (` +\n\t\t`?, ?, ?` +\n\t\t`)`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)))\n\tif tx != nil {\n\t\tres, err = tx.Exec(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)\n\t} else {\n\t\tres, err = dbConn.Exec(sqlstr, sp.Spid, sp.Desc, sp.ActionClass)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\tsp._exists = true\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// retrieve id\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set primary key and existence\n\tsp.Spid = uint(id)\n\tsp._exists = true\n\n\treturn nil\n}", "func (tttp *TriageTimeTableProvider) Insert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif tttp._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql insert query, primary key must be provided\n\tconst sqlstr = `INSERT INTO agency_portal.TRIAGE_TIME_TABLE_PROVIDER (` +\n\t\t`TRIAGE_TIME_TABLE_UUID, PROVIDER_UUID, STATUS, STATUS_TLM, WHEN_CREATED, TLM` +\n\t\t`) VALUES (` +\n\t\t`?, ?, ?, ?, ?, ?` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, tttp.TriageTimeTableUUID, tttp.ProviderUUID, tttp.Status, tttp.StatusTlm, tttp.WhenCreated, tttp.Tlm)\n\t_, err = db.Exec(sqlstr, tttp.TriageTimeTableUUID, tttp.ProviderUUID, tttp.Status, tttp.StatusTlm, tttp.WhenCreated, tttp.Tlm)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ttttp._exists = true\n\n\treturn nil\n}", "func (table *Table) Insert(db DB, record Map) (Result, error) {\n\tc := Context{StateInsert, db, table, table, record, Field{}}\n\treturn table.execAction(c, table.OnInsert, table.DefaultInsert)\n}", "func (m *MySQL) Insert(i interface{}) (sql.Result, error) {\n\ts, _, vals, err := m.toSQL(i, INSERT)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn m.Exec(s, vals...)\n}", "func (pg *PGStorage) Insert(a *Address) error {\n\tvar err error\n\terr = pg.con.QueryRow(`\n\t\tINSERT INTO address(hash, income, outcome, ballance)\n\t\tvalues($1, $2, $3, $4)\n\t\tRETURNING ID`,\n\t\ta.Hash,\n\t\ta.Income,\n\t\ta.Outcome,\n\t\ta.Ballance).Scan(\n\t\t&a.ID)\n\treturn err\n}", "func (dao *PlayerDAO) Insert(player models.Player) error {\n\tctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)\n\tdefer cancel()\n\t_, err := db.Collection(PCOLLECTION).InsertOne(ctx, bson.D{\n\t\t{\"firstname\", player.FirstName},\n\t\t{\"lastname\", player.LastName},\n\t\t{\"nickname\", player.NickName},\n\t\t{\"skilllevel\", player.SkillLevel},\n\t})\n\treturn err\n}", "func (model *SnippetModel) Insert(title, content, expires string) (int, error) {\n\tstmt := `INSERT INTO snippets (title, content, created, expires)\n\tVALUES(?, ?, UTC_TIMESTAMP(), DATE_ADD(UTC_TIMESTAMP(), INTERVAL ? DAY))`\n\n\tresult, err := model.DB.Exec(stmt, title, content, expires)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tid, err := result.LastInsertId()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn int(id), nil\n}", "func (a *Actor) Insert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif a._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql insert query, primary key must be provided\n\tconst sqlstr = `INSERT INTO public.actor (` +\n\t\t`actor_id, first_name, last_name, last_update` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate)\n\terr = db.QueryRow(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate).Scan(&a.ActorID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ta._exists = true\n\n\treturn nil\n}", "func (data *SimpleDbType) Insert(dbMgr *mgr.DBConn) error {\n\tdb := dbMgr.Open()\n\n\tsqlStatement := `INSERT INTO test_table (\"name\", \"number\") VALUES ($1, $2) RETURNING id`\n\tid := 0\n\terr := db.QueryRow(sqlStatement, data.Name, data.Number).Scan(&id)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(\"New record ID is:\", id)\n\tdbMgr.Close()\n\treturn err\n}", "func (m *SnippetModel) Insert(title, content, expires string) (int, error) {\n\t// Start a transaction\n\t// Each action that is done is atomic in nature:\n\t// All statements are executed successfully or no statement is executed\n\ttx, err := m.DB.Begin()\n\tif err != nil {\n\t\treturn 0, nil\n\t}\n\n\t// Statement to insert data to the database\n\tstmt := `INSERT INTO snippets (title, content, created, expires)\n\tVALUES(?, ?, UTC_TIMESTAMP(), DATE_ADD(UTC_TIMESTAMP(), INTERVAL ? DAY))`\n\t// Pass in the placeholder parameters aka the ? in the stmt\n\tresult, err := tx.Exec(stmt, title, content, expires)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn 0, err\n\t}\n\t// Return the id of the inserted record in the snippets table\n\tid, err := result.LastInsertId()\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn 0, err\n\t}\n\n\t// id is an int64 to convert it to a int\n\terr = tx.Commit()\n\treturn int(id), err\n\n}", "func (t *Testzzz) Insert(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\tvar res sql.Result\n\t// if already exist, bail\n\tif t._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetTestzzzTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// sql insert query, primary key provided by autoincrement\n\tsqlstr := `INSERT INTO ` + tableName +\n\t\t` (` +\n\t\t`a, b, c` +\n\t\t`) VALUES (` +\n\t\t`?, ?, ?` +\n\t\t`)`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, t.A, t.B, t.C)))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif tx != nil {\n\t\tres, err = tx.Exec(sqlstr, t.A, t.B, t.C)\n\t} else {\n\t\tres, err = dbConn.Exec(sqlstr, t.A, t.B, t.C)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// retrieve id\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set primary key and existence\n\tt.ID = int(id)\n\tt._exists = true\n\n\treturn nil\n}", "func (o *Vote) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tvoteInsertCacheMut.RLock()\n\tcache, cached := voteInsertCache[key]\n\tvoteInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `vote` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `vote` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `vote` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, votePrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into vote\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == voteMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for vote\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tvoteInsertCacheMut.Lock()\n\t\tvoteInsertCache[key] = cache\n\t\tvoteInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (p *Project) Insert(session *xorm.Session) (int, error) {\n\taffected, err := session.Insert(p)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn 0, err\n\t}\n\treturn int(affected), nil\n}", "func insert(res http.ResponseWriter, req *http.Request) {\n\tstmt, err := db.Prepare(`INSERT INTO customer VALUES (\"james\");`)\n\tcheck(err)\n\tdefer stmt.Close()\n\tr, err := stmt.Exec()\n\tcheck(err)\n\tn, err := r.RowsAffected()\n\tcheck(err)\n\n\tfmt.Fprintln(res, \"INSERTED RECORD\", n)\n}", "func (o *Failure) InsertGP(whitelist ...string) {\n\tif err := o.Insert(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (ds *DjangoSession) Insert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase ds._exists: // already exists\n\t\treturn logerror(&ErrInsertFailed{ErrAlreadyExists})\n\tcase ds._deleted: // deleted\n\t\treturn logerror(&ErrInsertFailed{ErrMarkedForDeletion})\n\t}\n\t// insert (manual)\n\tconst sqlstr = `INSERT INTO django.django_session (` +\n\t\t`session_key, session_data, expire_date` +\n\t\t`) VALUES (` +\n\t\t`:1, :2, :3` +\n\t\t`)`\n\t// run\n\tlogf(sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate)\n\tif _, err := db.ExecContext(ctx, sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate); err != nil {\n\t\treturn logerror(err)\n\t}\n\t// set exists\n\tds._exists = true\n\treturn nil\n}", "func (fr *FetchResult) Insert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif fr._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql insert query, primary key provided by autoincrement\n\tconst sqlstr = `INSERT INTO fetch_result (` +\n\t\t`created_at, updated_at` +\n\t\t`) VALUES (` +\n\t\t`?, ?` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, fr.CreatedAt, fr.UpdatedAt)\n\tres, err := db.Exec(sqlstr, fr.CreatedAt, fr.UpdatedAt)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// retrieve id\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set primary key and existence\n\tfr.ID = uint(id)\n\tfr._exists = true\n\n\treturn nil\n}", "func Insert(mock sqlmock.Sqlmock, table string, err error, lastInsertID int64) {\n\tsql := fmt.Sprintf(\"INSERT INTO %s\", table)\n\n\tif err != nil {\n\t\tmock.ExpectExec(sql).WillReturnError(err)\n\t\treturn\n\t}\n\n\tmock.ExpectExec(sql).WillReturnResult(sqlmock.NewResult(lastInsertID, 0))\n}", "func (c *BigQueryClient) RecordExecution(ctx context.Context, at time.Time, inserted int) error {\n\tinserter := c.ExecTable.Inserter()\n\tinserter.IgnoreUnknownValues = true\n\n\tif err := inserter.Put(ctx, Execution{Timestamp: at, Inserted: inserted}); err != nil {\n\t\tputErr := err.(bigquery.PutMultiError)\n\t\tfor _, rowErr := range putErr {\n\t\t\tlog.From(ctx).Error(\"inserting row\", zap.Error(rowErr.Errors))\n\t\t}\n\n\t\treturn err\n\t}\n\treturn nil\n}", "func (oee *OtxEpubEpub) Insert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif oee._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql insert query, primary key must be provided\n\tconst sqlstr = `INSERT INTO public.otx_epub_epub (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\terr = db.QueryRow(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents).Scan(&oee.DocumentPtrID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\toee._exists = true\n\n\treturn nil\n}", "func (task *InsertRecordsTask) Insert(records []wisply.Recorder) error {\n\terr := task.insertRecords(records)\n\tif err != nil {\n\t\ttask.hasProblems(err)\n\t\treturn err\n\t}\n\tnumber := strconv.Itoa(len(records))\n\ttask.Finish(number + \" records inserted\")\n\treturn nil\n}", "func (j *JobRun) PostInsert(s gorp.SqlExecutor) error {\n\treturn j.PostUpdate(s)\n}", "func (f *FakeClient) Insert(p *purchase.Purchase) error {\n\treturn nil\n}", "func (s *State) Insert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif s._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql insert query, primary key provided by sequence\n\tconst sqlstr = `INSERT INTO public.state (` +\n\t\t`\"namespace\", \"context_id\", \"created_at\", \"state\", \"data\", \"event\", \"processing_error\"` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6, $7` +\n\t\t`) RETURNING \"id\"`\n\n\t// run query\n\tXOLog(sqlstr, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError)\n\terr = db.QueryRow(sqlstr, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError).Scan(&s.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ts._exists = true\n\n\treturn nil\n}", "func (m *MovieModel) Insert(movie *Movie) error {\n\t// Define the SQL query for inserting a new record in the movies table and returning the system generated data\n\tquery := `INSERT INTO movies (title, year, runtime, genres) \n\t\t\t\t\t\tVALUES ($1, $2, $3, $4)\n\t\t\t\t\t\tRETURNING id, created_at, version`\n\n\t// Create an args slice containing the values for the placeholder parameters from the movie struct\n\targs := []interface{}{movie.Title, movie.Year, movie.Runtime, pq.Array(movie.Genres)}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\t// Execute the query.\n\treturn m.DB.QueryRowContext(ctx, query, args...).Scan(&movie.ID, &movie.CreatedAt, &movie.Version)\n}", "func (b *Block) Insert() error {\n\treturn b.storage.Insert(b)\n}", "func (p *PartialMutation) Insert(sess sqlbuilder.SQLBuilder, structPtr interface{}, whereColumn, whereValue string, extraFields map[string]interface{}) error {\n\tif structPtr == nil || reflect.TypeOf(structPtr).Kind() != reflect.Ptr {\n\t\treturn fmt.Errorf(\"expecting a pointer but got %T\", structPtr)\n\t}\n\n\tvar (\n\t\tcolumns []string\n\t\tvalues []interface{}\n\t\terr error\n\t)\n\n\tif len(p.includeFields) > 0 {\n\t\tcolumns, values, err = p.getColumnsValuesIncluding(structPtr, p.includeFields)\n\t} else {\n\t\tcolumns, values, err = p.getColumnsValuesExcluding(structPtr, p.excludeFields)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif extraFields != nil {\n\t\tfor k, v := range extraFields {\n\t\t\tcolumns = append(columns, k)\n\t\t\tvalues = append(values, v)\n\t\t}\n\t}\n\n\tlenColumns := len(columns)\n\tlenValues := len(values)\n\tif lenColumns == 0 || lenValues == 0 {\n\t\treturn errors.New(\"query with zero columns and values\")\n\t}\n\n\tif lenColumns != lenValues {\n\t\treturn errors.New(\"columns and values length missmatch\")\n\t}\n\n\tquery := sess.InsertInto(p.table).Columns(columns...).Values(values...)\n\tres, err := query.Exec()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif n, _ := res.RowsAffected(); n == 0 {\n\t\treturn errors.E(errors.Errorf(\"operation insert can not be performed, zero rows affected, resource %s\", whereValue), errors.NotExist)\n\t}\n\n\tif _, ok := sess.(sqlbuilder.Tx); ok {\n\t\treturn nil\n\t}\n\n\treturn p.col().Find(whereColumn, whereValue).Limit(1).One(structPtr)\n}", "func (s *PetStore) Insert(record *Pet) error {\n\treturn s.Store.Insert(Schema.Pet.BaseSchema, record)\n}" ]
[ "0.6625801", "0.6542454", "0.6542454", "0.61717796", "0.6154179", "0.6146324", "0.6144909", "0.6132511", "0.6121985", "0.60695183", "0.60656875", "0.6057612", "0.60149854", "0.5995778", "0.59945893", "0.5967067", "0.5919993", "0.59197664", "0.59056336", "0.5865164", "0.58605695", "0.5859489", "0.5821754", "0.57945144", "0.5792489", "0.5785886", "0.5780079", "0.5760186", "0.5751777", "0.5742695", "0.5742048", "0.5732893", "0.5726325", "0.5718179", "0.5714456", "0.5711905", "0.57069194", "0.56778085", "0.56757456", "0.565544", "0.5616441", "0.56141", "0.56124955", "0.5590656", "0.5587524", "0.5575719", "0.557297", "0.55707514", "0.5557324", "0.55511117", "0.5547393", "0.5543792", "0.5539593", "0.5537549", "0.5534622", "0.55148894", "0.549342", "0.54900944", "0.54684246", "0.5429443", "0.54229087", "0.54208213", "0.54086804", "0.5405795", "0.54027843", "0.54027843", "0.54027843", "0.54027843", "0.54027843", "0.54027843", "0.53880495", "0.5378718", "0.5375211", "0.536166", "0.5359047", "0.5348404", "0.5344612", "0.5343369", "0.53384334", "0.53381026", "0.5330665", "0.5329898", "0.5324608", "0.53231853", "0.532286", "0.5321649", "0.53171647", "0.5312256", "0.53019744", "0.52954704", "0.52888954", "0.52825475", "0.52799296", "0.52780247", "0.52705646", "0.5260567", "0.5258845", "0.52577835", "0.52478474", "0.52470696" ]
0.5781983
26
Insert a single record using an executor. Whitelist behavior: If a whitelist is provided, only those columns supplied are inserted No whitelist behavior: Without a whitelist, columns are inferred by the following rules: All columns without a default value are included (i.e. name, age) All columns with a default, but nonzero are included (i.e. health = 75)
func (o *AssetRevision) Insert(exec boil.Executor, whitelist ...string) error { if o == nil { return errors.New("public: no asset_revision provided for insertion") } var err error if err := o.doBeforeInsertHooks(exec); err != nil { return err } nzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o) key := makeCacheKey(whitelist, nzDefaults) assetRevisionInsertCacheMut.RLock() cache, cached := assetRevisionInsertCache[key] assetRevisionInsertCacheMut.RUnlock() if !cached { wl, returnColumns := strmangle.InsertColumnSet( assetRevisionColumns, assetRevisionColumnsWithDefault, assetRevisionColumnsWithoutDefault, nzDefaults, whitelist, ) cache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, wl) if err != nil { return err } cache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, returnColumns) if err != nil { return err } if len(wl) != 0 { cache.query = fmt.Sprintf("INSERT INTO \"asset_revision\" (\"%s\") VALUES (%s)", strings.Join(wl, "\",\""), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1)) } else { cache.query = "INSERT INTO \"asset_revision\" DEFAULT VALUES" } if len(cache.retMapping) != 0 { cache.query += fmt.Sprintf(" RETURNING \"%s\"", strings.Join(returnColumns, "\",\"")) } } value := reflect.Indirect(reflect.ValueOf(o)) vals := queries.ValuesFromMapping(value, cache.valueMapping) if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, cache.query) fmt.Fprintln(boil.DebugWriter, vals) } if len(cache.retMapping) != 0 { err = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...) } else { _, err = exec.Exec(cache.query, vals...) } if err != nil { return errors.Wrap(err, "public: unable to insert into asset_revision") } if !cached { assetRevisionInsertCacheMut.Lock() assetRevisionInsertCache[key] = cache assetRevisionInsertCacheMut.Unlock() } return o.doAfterInsertHooks(exec) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Inventory) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no inventory provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(inventoryColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tinventoryInsertCacheMut.RLock()\n\tcache, cached := inventoryInsertCache[key]\n\tinventoryInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryColumnsWithDefault,\n\t\t\tinventoryColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(inventoryType, inventoryMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(inventoryType, inventoryMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `inventory` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `inventory` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `inventory` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, inventoryPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into inventory\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == inventoryMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for inventory\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tinventoryInsertCacheMut.Lock()\n\t\tinventoryInsertCache[key] = cache\n\t\tinventoryInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *InstrumentClass) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no instrument_class provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(instrumentClassColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tinstrumentClassInsertCacheMut.RLock()\n\tcache, cached := instrumentClassInsertCache[key]\n\tinstrumentClassInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassColumnsWithDefault,\n\t\t\tinstrumentClassColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"instruments\\\".\\\"instrument_class\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"instruments\\\".\\\"instrument_class\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into instrument_class\")\n\t}\n\n\tif !cached {\n\t\tinstrumentClassInsertCacheMut.Lock()\n\t\tinstrumentClassInsertCache[key] = cache\n\t\tinstrumentClassInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *CMFFamilyUserPoliciesTake) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no cmf_family_user_policies_take provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tif o.UpdatedAt.IsZero() {\n\t\t\to.UpdatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(cmfFamilyUserPoliciesTakeColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tcmfFamilyUserPoliciesTakeInsertCacheMut.RLock()\n\tcache, cached := cmfFamilyUserPoliciesTakeInsertCache[key]\n\tcmfFamilyUserPoliciesTakeInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tcmfFamilyUserPoliciesTakeAllColumns,\n\t\t\tcmfFamilyUserPoliciesTakeColumnsWithDefault,\n\t\t\tcmfFamilyUserPoliciesTakeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(cmfFamilyUserPoliciesTakeType, cmfFamilyUserPoliciesTakeMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(cmfFamilyUserPoliciesTakeType, cmfFamilyUserPoliciesTakeMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `cmf_family_user_policies_take` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `cmf_family_user_policies_take` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `cmf_family_user_policies_take` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into cmf_family_user_policies_take\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == cmfFamilyUserPoliciesTakeMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, identifierCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for cmf_family_user_policies_take\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tcmfFamilyUserPoliciesTakeInsertCacheMut.Lock()\n\t\tcmfFamilyUserPoliciesTakeInsertCache[key] = cache\n\t\tcmfFamilyUserPoliciesTakeInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *Failure) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no failure provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(failureColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tfailureInsertCacheMut.RLock()\n\tcache, cached := failureInsertCache[key]\n\tfailureInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailureColumnsWithDefault,\n\t\t\tfailureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(failureType, failureMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(failureType, failureMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `failure` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `failure` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `failure` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, failurePrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into failure\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == failureMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for failure\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfailureInsertCacheMut.Lock()\n\t\tfailureInsertCache[key] = cache\n\t\tfailureInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Weather) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"db: no weather provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(weatherColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tweatherInsertCacheMut.RLock()\n\tcache, cached := weatherInsertCache[key]\n\tweatherInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherColumnsWithDefault,\n\t\t\tweatherColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(weatherType, weatherMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(weatherType, weatherMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"prh\\\".\\\"weather\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"prh\\\".\\\"weather\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to insert into weather\")\n\t}\n\n\tif !cached {\n\t\tweatherInsertCacheMut.Lock()\n\t\tweatherInsertCache[key] = cache\n\t\tweatherInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (impl *EmployeeRepositoryImpl) Insert(employee *domain.Employee) {\n\temployeeRaw := employee.ToEmployeeRaw()\n\temployeeRaw.Id = 0 // setting to 0 will trigger auto increment\n\tquery := \"INSERT INTO employees(id, first_name, last_name, created_at) VALUES (0, ?, ?, ?);\"\n\timpl.db.MustExec(query, employeeRaw.FirstName, employeeRaw.LastName, time.Now().UTC()) //MustExec will panic on error (Exec will not)\n\t// no error returned it will panic on error\n}", "func (o *RawVisit) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no raw_visits provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rawVisitColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\trawVisitInsertCacheMut.RLock()\n\tcache, cached := rawVisitInsertCache[key]\n\trawVisitInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitColumnsWithDefault,\n\t\t\trawVisitColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"raw_visits\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"raw_visits\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitInsertCacheMut.Lock()\n\t\trawVisitInsertCache[key] = cache\n\t\trawVisitInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *Vote) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tvoteInsertCacheMut.RLock()\n\tcache, cached := voteInsertCache[key]\n\tvoteInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `vote` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `vote` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `vote` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, votePrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into vote\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == voteMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for vote\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tvoteInsertCacheMut.Lock()\n\t\tvoteInsertCache[key] = cache\n\t\tvoteInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Offer) Insert(exec boil.Executor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no offers provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(offerColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tofferInsertCacheMut.RLock()\n\tcache, cached := offerInsertCache[key]\n\tofferInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferColumnsWithDefault,\n\t\t\tofferColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(offerType, offerMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(offerType, offerMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"offers\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"offers\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to insert into offers\")\n\t}\n\n\tif !cached {\n\t\tofferInsertCacheMut.Lock()\n\t\tofferInsertCache[key] = cache\n\t\tofferInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Transaction) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no transaction provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\ttransactionInsertCacheMut.RLock()\n\tcache, cached := transactionInsertCache[key]\n\ttransactionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `transaction` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `transaction` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `transaction` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, transactionPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to insert into transaction\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for transaction\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionInsertCacheMut.Lock()\n\t\ttransactionInsertCache[key] = cache\n\t\ttransactionInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no transactions provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.UpdatedAt.Time.IsZero() {\n\t\to.UpdatedAt.Time = currTime\n\t\to.UpdatedAt.Valid = true\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\ttransactionInsertCacheMut.RLock()\n\tcache, cached := transactionInsertCache[key]\n\ttransactionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `transactions` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `transactions` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `transactions` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, transactionPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into transactions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.TransactionID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"TransactionID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.TransactionID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for transactions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionInsertCacheMut.Lock()\n\t\ttransactionInsertCache[key] = cache\n\t\ttransactionInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (p *PartialMutation) Insert(sess sqlbuilder.SQLBuilder, structPtr interface{}, whereColumn, whereValue string, extraFields map[string]interface{}) error {\n\tif structPtr == nil || reflect.TypeOf(structPtr).Kind() != reflect.Ptr {\n\t\treturn fmt.Errorf(\"expecting a pointer but got %T\", structPtr)\n\t}\n\n\tvar (\n\t\tcolumns []string\n\t\tvalues []interface{}\n\t\terr error\n\t)\n\n\tif len(p.includeFields) > 0 {\n\t\tcolumns, values, err = p.getColumnsValuesIncluding(structPtr, p.includeFields)\n\t} else {\n\t\tcolumns, values, err = p.getColumnsValuesExcluding(structPtr, p.excludeFields)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif extraFields != nil {\n\t\tfor k, v := range extraFields {\n\t\t\tcolumns = append(columns, k)\n\t\t\tvalues = append(values, v)\n\t\t}\n\t}\n\n\tlenColumns := len(columns)\n\tlenValues := len(values)\n\tif lenColumns == 0 || lenValues == 0 {\n\t\treturn errors.New(\"query with zero columns and values\")\n\t}\n\n\tif lenColumns != lenValues {\n\t\treturn errors.New(\"columns and values length missmatch\")\n\t}\n\n\tquery := sess.InsertInto(p.table).Columns(columns...).Values(values...)\n\tres, err := query.Exec()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif n, _ := res.RowsAffected(); n == 0 {\n\t\treturn errors.E(errors.Errorf(\"operation insert can not be performed, zero rows affected, resource %s\", whereValue), errors.NotExist)\n\t}\n\n\tif _, ok := sess.(sqlbuilder.Tx); ok {\n\t\treturn nil\n\t}\n\n\treturn p.col().Find(whereColumn, whereValue).Limit(1).One(structPtr)\n}", "func (o *Task) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tasks provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tif o.UpdatedAt.IsZero() {\n\t\t\to.UpdatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(taskColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\ttaskInsertCacheMut.RLock()\n\tcache, cached := taskInsertCache[key]\n\ttaskInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskColumnsWithDefault,\n\t\t\ttaskColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(taskType, taskMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"tasks\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"tasks\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskInsertCacheMut.Lock()\n\t\ttaskInsertCache[key] = cache\n\t\ttaskInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (s *DbRecorder) Insert() error {\n\tswitch s.flavor {\n\tcase \"postgres\":\n\t\treturn s.insertPg()\n\tdefault:\n\t\treturn s.insertStd()\n\t}\n}", "func (m *Mytable) Insert(ctx context.Context) *spanner.Mutation {\n\treturn spanner.Insert(\"mytable\", MytableColumns(), []interface{}{\n\t\tm.A, m.B,\n\t})\n}", "func (ps *PlayerStore) Insert(ctx context.Context, player store.Player) (*store.Player, error) {\n\tquery := `\n INSERT INTO players(roster_id,first_name,last_name,alias,status)\n VALUES($1,$2,$3,$4,$5)\n RETURNING *\n `\n\tdb := ps.db.GetDB()\n\tctx, cancel := ps.db.RequestContext(ctx)\n\tdefer cancel()\n\n\tvar p store.Player\n\terr := db.QueryRowContext(ctx, query,\n\t\tplayer.RosterID,\n\t\tplayer.FirstName,\n\t\tplayer.LastName,\n\t\tplayer.Alias,\n\t\tplayer.Status).\n\t\tScan(\n\t\t\t&p.PlayerID,\n\t\t\t&p.RosterID,\n\t\t\t&p.FirstName,\n\t\t\t&p.LastName,\n\t\t\t&p.Alias,\n\t\t\t&p.Status)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &p, nil\n}", "func Insert(stmt bolt.Stmt, data interface{}) error {\n\n\tm := structs.Map(data)\n\n\tflag := false\n\tfor _, v := range m {\n\t\tif reflect.ValueOf(v).Kind() == reflect.Map {\n\t\t\tflag = true\n\t\t}\n\t}\n\tif flag {\n\t\tm = flatMap(m, \"\")\n\t}\n\n\t// Parses data struct to a map before sending it to ExeNeo()\n\tresult, err := stmt.ExecNeo(m)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnumResult, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Printf(\"CREATED ROWS: %d\\n\", numResult)\n\treturn nil\n}", "func (o *Author) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no authors provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(authorColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tauthorInsertCacheMut.RLock()\n\tcache, cached := authorInsertCache[key]\n\tauthorInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorColumnsWithDefault,\n\t\t\tauthorColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authorType, authorMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(authorType, authorMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"authors\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"authors\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to insert into authors\")\n\t}\n\n\tif !cached {\n\t\tauthorInsertCacheMut.Lock()\n\t\tauthorInsertCache[key] = cache\n\t\tauthorInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Rental) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no rental provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rentalColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\trentalInsertCacheMut.RLock()\n\tcache, cached := rentalInsertCache[key]\n\trentalInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\trentalColumns,\n\t\t\trentalColumnsWithDefault,\n\t\t\trentalColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rentalType, rentalMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(rentalType, rentalMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `rental` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `rental` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `rental` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, rentalPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to insert into rental\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.RentalID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == rentalMapping[\"RentalID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.RentalID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to populate default values for rental\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\trentalInsertCacheMut.Lock()\n\t\trentalInsertCache[key] = cache\n\t\trentalInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *BlackCard) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no black_cards provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(blackCardColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tblackCardInsertCacheMut.RLock()\n\tcache, cached := blackCardInsertCache[key]\n\tblackCardInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tblackCardColumns,\n\t\t\tblackCardColumnsWithDefault,\n\t\t\tblackCardColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(blackCardType, blackCardMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(blackCardType, blackCardMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"black_cards\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"black_cards\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into black_cards\")\n\t}\n\n\tif !cached {\n\t\tblackCardInsertCacheMut.Lock()\n\t\tblackCardInsertCache[key] = cache\n\t\tblackCardInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (dao *HeroDao) Insert(inCols []string, vals []interface{}) (int64, int32) {\r\n\tvar err error\r\n\r\n\tif len(inCols) <= 0 || len(vals) <= 0 || len(inCols) != len(vals) {\r\n\t\terr = errors.New(\"hero dao Insert() param length differ\")\r\n\t\tlogs.Debug(err)\r\n\t\treturn 0, -1\r\n\t}\r\n\tquery := `insert ` + TbHero + ` set `\r\n\tfor idx, item := range inCols {\r\n\t\tif idx != 0 {\r\n\t\t\tquery += `,`\r\n\t\t}\r\n\t\tquery += item + `=?`\r\n\t}\r\n\tlogs.Debug(\"query:\", query)\r\n\r\n\tstmt, err := dao.Prepare(query)\r\n\tif err != nil {\r\n\t\tlogs.Error(\"user insert err0:\", err)\r\n\t\treturn 0, -1\r\n\t}\r\n\tdefer stmt.Close()\r\n\r\n\tres, err := stmt.Exec(vals...)\r\n\tif err != nil {\r\n\t\tlogs.Error(\"user insert err1:\", err)\r\n\t\treturn 0, -1\r\n\t}\r\n\r\n\tid, err := res.LastInsertId()\r\n\tif err != nil {\r\n\t\tlogs.Error(\"user insert err2:\", err)\r\n\t\treturn 0, -1\r\n\t}\r\n\tlogs.Debug(\"new user id:\", id)\r\n\treturn id, 0\r\n}", "func (o *Shelf) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no shelf provided for insertion\")\n\t}\n\to.whitelist = whitelist\n\to.operation = \"INSERT\"\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(shelfColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tshelfInsertCacheMut.RLock()\n\tcache, cached := shelfInsertCache[key]\n\tshelfInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tshelfColumns,\n\t\t\tshelfColumnsWithDefault,\n\t\t\tshelfColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(shelfType, shelfMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(shelfType, shelfMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.query = fmt.Sprintf(\"INSERT INTO `shelf` (`%s`) VALUES (%s)\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `shelf` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, shelfPrimaryKeyColumns))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into shelf\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == shelfMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for shelf\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tshelfInsertCacheMut.Lock()\n\t\tshelfInsertCache[key] = cache\n\t\tshelfInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Jet) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jets provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jetColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tjetInsertCacheMut.RLock()\n\tcache, cached := jetInsertCache[key]\n\tjetInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetColumnsWithDefault,\n\t\t\tjetColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jetType, jetMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(jetType, jetMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `jets` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `jets` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `jets` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, jetPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into jets\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for jets\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tjetInsertCacheMut.Lock()\n\t\tjetInsertCache[key] = cache\n\t\tjetInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Address) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no address provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(addressColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\taddressInsertCacheMut.RLock()\n\tcache, cached := addressInsertCache[key]\n\taddressInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\taddressColumns,\n\t\t\taddressColumnsWithDefault,\n\t\t\taddressColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(addressType, addressMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(addressType, addressMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `address` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `address` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `address` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, addressPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to insert into address\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.AddressID = uint16(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == addressMapping[\"AddressID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.AddressID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to populate default values for address\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\taddressInsertCacheMut.Lock()\n\t\taddressInsertCache[key] = cache\n\t\taddressInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Auth) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auths provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.Time.IsZero() {\n\t\to.CreatedAt.Time = currTime\n\t\to.CreatedAt.Valid = true\n\t}\n\tif o.UpdatedAt.Time.IsZero() {\n\t\to.UpdatedAt.Time = currTime\n\t\to.UpdatedAt.Valid = true\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tauthInsertCacheMut.RLock()\n\tcache, cached := authInsertCache[key]\n\tauthInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthColumnsWithDefault,\n\t\t\tauthColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authType, authMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(authType, authMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"auths\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"auths\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into auths\")\n\t}\n\n\tif !cached {\n\t\tauthInsertCacheMut.Lock()\n\t\tauthInsertCache[key] = cache\n\t\tauthInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *UsernameListing) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no username_listings provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tif queries.MustTime(o.UpdatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(usernameListingColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tusernameListingInsertCacheMut.RLock()\n\tcache, cached := usernameListingInsertCache[key]\n\tusernameListingInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tusernameListingAllColumns,\n\t\t\tusernameListingColumnsWithDefault,\n\t\t\tusernameListingColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(usernameListingType, usernameListingMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(usernameListingType, usernameListingMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"username_listings\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"username_listings\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into username_listings\")\n\t}\n\n\tif !cached {\n\t\tusernameListingInsertCacheMut.Lock()\n\t\tusernameListingInsertCache[key] = cache\n\t\tusernameListingInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (t BlobsTable) Insert(ctx context.Context, id []byte, _type int32, data []byte) (*Blob, error) {\n\treturn t.driver.insert(ctx, id, _type, data)\n}", "func (o *AuthUser) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_user provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(authUserColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tauthUserInsertCacheMut.RLock()\n\tcache, cached := authUserInsertCache[key]\n\tauthUserInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tauthUserColumns,\n\t\t\tauthUserColumnsWithDefault,\n\t\t\tauthUserColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authUserType, authUserMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(authUserType, authUserMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.query = fmt.Sprintf(\"INSERT INTO `auth_user` (`%s`) VALUES (%s)\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `auth_user` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, authUserPrimaryKeyColumns))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into auth_user\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authUserMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_user\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthUserInsertCacheMut.Lock()\n\t\tauthUserInsertCache[key] = cache\n\t\tauthUserInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Asset) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no asset provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tassetInsertCacheMut.RLock()\n\tcache, cached := assetInsertCache[key]\n\tassetInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tassetColumns,\n\t\t\tassetColumnsWithDefault,\n\t\t\tassetColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetType, assetMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(assetType, assetMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"audit\\\".\\\"asset\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"audit\\\".\\\"asset\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into asset\")\n\t}\n\n\tif !cached {\n\t\tassetInsertCacheMut.Lock()\n\t\tassetInsertCache[key] = cache\n\t\tassetInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Job) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jobs provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tif queries.MustTime(o.UpdatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jobColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tjobInsertCacheMut.RLock()\n\tcache, cached := jobInsertCache[key]\n\tjobInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobColumnsWithDefault,\n\t\t\tjobColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(jobType, jobMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"jobs\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"jobs\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into jobs\")\n\t}\n\n\tif !cached {\n\t\tjobInsertCacheMut.Lock()\n\t\tjobInsertCache[key] = cache\n\t\tjobInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *TrainingCost) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no training_costs provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tif o.UpdatedAt.IsZero() {\n\t\t\to.UpdatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(trainingCostColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\ttrainingCostInsertCacheMut.RLock()\n\tcache, cached := trainingCostInsertCache[key]\n\ttrainingCostInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostColumnsWithDefault,\n\t\t\ttrainingCostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"training_costs\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"training_costs\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostInsertCacheMut.Lock()\n\t\ttrainingCostInsertCache[key] = cache\n\t\ttrainingCostInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *Source) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no sources provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(sourceColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tsourceInsertCacheMut.RLock()\n\tcache, cached := sourceInsertCache[key]\n\tsourceInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourceColumnsWithDefault,\n\t\t\tsourceColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(sourceType, sourceMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"sources\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"sources\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to insert into sources\")\n\t}\n\n\tif !cached {\n\t\tsourceInsertCacheMut.Lock()\n\t\tsourceInsertCache[key] = cache\n\t\tsourceInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (s *Session) Insert(dest interface{}) (int64, error) {\n\ts.initStatemnt()\n\ts.statement.Insert()\n\tscanner, err := NewScanner(dest)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer scanner.Close()\n\tif s.statement.table == \"\" {\n\t\ts.statement.From(scanner.GetTableName())\n\t}\n\tinsertFields := make([]string, 0)\n\tfor n, f := range scanner.Model.Fields {\n\t\tif !f.IsReadOnly {\n\t\t\tinsertFields = append(insertFields, n)\n\t\t}\n\t}\n\ts.Columns(insertFields...)\n\tif scanner.entityPointer.Kind() == reflect.Slice {\n\t\tfor i := 0; i < scanner.entityPointer.Len(); i++ {\n\t\t\tval := make([]interface{}, 0)\n\t\t\tsub := scanner.entityPointer.Index(i)\n\t\t\tif sub.Kind() == reflect.Ptr {\n\t\t\t\tsubElem := sub.Elem()\n\t\t\t\tfor _, fn := range insertFields {\n\t\t\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tfv := subElem.Field(f.idx)\n\t\t\t\t\tval = append(val, fv.Interface())\n\t\t\t\t}\n\n\t\t\t} else {\n\t\t\t\tfor _, fn := range insertFields {\n\t\t\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tfv := sub.Field(f.idx)\n\t\t\t\t\tval = append(val, fv.Interface())\n\t\t\t\t}\n\t\t\t}\n\t\t\ts.statement.Values(val)\n\t\t}\n\n\t} else if scanner.entityPointer.Kind() == reflect.Struct {\n\t\tval := make([]interface{}, 0)\n\t\tfor _, fn := range insertFields {\n\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfv := scanner.entityPointer.Field(f.idx)\n\t\t\tval = append(val, fv.Interface())\n\t\t}\n\t\ts.statement.Values(val)\n\t} else {\n\t\treturn 0, InsertExpectSliceOrStruct\n\t}\n\tsql, args, err := s.statement.ToSQL()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\ts.logger.Debugf(\"[Session Insert] sql: %s, args: %v\", sql, args)\n\ts.initCtx()\n\tsResult, err := s.ExecContext(s.ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn sResult.RowsAffected()\n}", "func (o *HoldenAt) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no HoldenAt provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(holdenAtColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tholdenAtInsertCacheMut.RLock()\n\tcache, cached := holdenAtInsertCache[key]\n\tholdenAtInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtColumnsWithDefault,\n\t\t\tholdenAtColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"HoldenAt\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"HoldenAt\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into HoldenAt\")\n\t}\n\n\tif !cached {\n\t\tholdenAtInsertCacheMut.Lock()\n\t\tholdenAtInsertCache[key] = cache\n\t\tholdenAtInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func insert(res http.ResponseWriter, req *http.Request) {\n\tstmt, err := db.Prepare(`INSERT INTO customer VALUES (\"james\");`)\n\tcheck(err)\n\tdefer stmt.Close()\n\tr, err := stmt.Exec()\n\tcheck(err)\n\tn, err := r.RowsAffected()\n\tcheck(err)\n\n\tfmt.Fprintln(res, \"INSERTED RECORD\", n)\n}", "func (r *TaskRepository) Insert(db db.DB, Task *entities.Task) error {\n\t_, err := db.NamedExec(`\n\tINSERT INTO tasks (uuid,title,user_id,status,created_at,updated_at)\n\tVALUES (:uuid, :title, :user_id, :status, now(), now())`, Task)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error inserting task to db: %w\", err)\n\t}\n\n\treturn nil\n}", "func Insert(ctx context.Context, req *proto.CreateRequest) error {\n\tlog.Printf(\"Inside Insert()\\n\")\n\n\ttxn, err := GetDbConn().Begin()\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer txn.Rollback()\n\tlog.Printf(\"Transaction started \\n\")\n\n\tfmt.Printf(\"Creating prepared statement\\n\")\n\tstmtStr, err := txn.Prepare(\"INSERT INTO restaurant_scores(business_id, business_name, business_address, business_city, business_state, business_postal_code, business_latitude, business_longitude, business_location, business_phone_number, inspection_id, inspection_date, inspection_score, inspection_type, violation_id, violation_description, risk_category, neighborhoods_old, police_districts, supervisor_districts, fire_prevention_districts, zip_codes, analysis_neighborhoods) VALUES( $1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12, $13, $14, $15, $16, $17, $18, $19, $20, $21, $22, $23 )\")\n\t//TODO clean\n\t// stmtStr, err := txn.Prepare(\"INSERT INTO restaurant_scores(business_id, business_name, business_address, business_city, business_state, business_postal_code, business_latitude, business_longitude, business_location, business_phone_number, inspection_id, inspection_date, inspection_score, inspection_type, violation_id, violation_description, risk_category, neighborhoods_old, police_districts, supervisor_districts, fire_prevention_districts, zip_codes, analysis_neighborhoods) VALUES( 1, 2, 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' )\")\n\tif err != nil {\n\t\tfmt.Printf(\"Error in creating statement %v\", err)\n\t\treturn err\n\t}\n\tdefer stmtStr.Close()\n\tlog.Printf(\"Statement Insertd \\n\")\n\n\tlog.Printf(\"Executing the statement for business_id %v \\n\", req.GetRecord().GetBusinessId())\n\t//Keeping long statement as punch cards time has gone\n\tres, err := stmtStr.Exec(req.GetRecord().GetBusinessId(), req.GetRecord().GetBusinessName(), req.GetRecord().GetBusinessAddress(), req.GetRecord().GetBusinessCity(), req.GetRecord().GetBusinessState(), req.GetRecord().GetBusinessPostalCode(), req.GetRecord().GetBusinessLatitude(), req.GetRecord().GetBusinessLongitude(), req.GetRecord().GetBusinessLocation(), req.GetRecord().GetBusinessPhoneNumber(), req.GetRecord().GetInspectionId(), req.GetRecord().GetInspectionDate(), req.GetRecord().GetInspectionScore(), req.GetRecord().GetInspectionType(), req.GetRecord().GetViolationId(), req.GetRecord().GetViolationDescription(), req.GetRecord().GetRiskCategory(), req.GetRecord().GetNeighborhoodsOld(), req.GetRecord().GetPoliceDistricts(), req.GetRecord().GetSupervisorDistricts(), req.GetRecord().GetFirePreventionDistricts(), req.GetRecord().GetZipCodes(), req.GetRecord().GetAnalysisNeighborhoods())\n\tif err != nil {\n\t\tlog.Printf(\"Error while inserting rows %v\", err)\n\t}\n\tlog.Printf(\"INSERT done with Result = %v\\n doing commit now \\n\", res)\n\n\terr = txn.Commit()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tlog.Printf(\"Exiting Insert()\\n\")\n\treturn nil\n}", "func (o *CurrentChartDataMinutely) Insert(exec boil.Executor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no current_chart_data_minutely provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currentChartDataMinutelyColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tcurrentChartDataMinutelyInsertCacheMut.RLock()\n\tcache, cached := currentChartDataMinutelyInsertCache[key]\n\tcurrentChartDataMinutelyInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tcurrentChartDataMinutelyColumns,\n\t\t\tcurrentChartDataMinutelyColumnsWithDefault,\n\t\t\tcurrentChartDataMinutelyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"current_chart_data_minutely\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"current_chart_data_minutely\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into current_chart_data_minutely\")\n\t}\n\n\tif !cached {\n\t\tcurrentChartDataMinutelyInsertCacheMut.Lock()\n\t\tcurrentChartDataMinutelyInsertCache[key] = cache\n\t\tcurrentChartDataMinutelyInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (q *Queue) insert(entities interface{}, mutex *sync.Mutex) *gorm.DB {\n\tmutex.Lock()\n\tres := q.db.Clauses(clause.OnConflict{DoNothing: true}).Create(entities)\n\tmutex.Unlock()\n\n\treturn res\n}", "func (o *PremiumSlot) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no premium_slots provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(premiumSlotColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tpremiumSlotInsertCacheMut.RLock()\n\tcache, cached := premiumSlotInsertCache[key]\n\tpremiumSlotInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotColumnsWithDefault,\n\t\t\tpremiumSlotColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"premium_slots\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"premium_slots\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into premium_slots\")\n\t}\n\n\tif !cached {\n\t\tpremiumSlotInsertCacheMut.Lock()\n\t\tpremiumSlotInsertCache[key] = cache\n\t\tpremiumSlotInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Friendship) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no friendship provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(friendshipColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tfriendshipInsertCacheMut.RLock()\n\tcache, cached := friendshipInsertCache[key]\n\tfriendshipInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tfriendshipAllColumns,\n\t\t\tfriendshipColumnsWithDefault,\n\t\t\tfriendshipColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(friendshipType, friendshipMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(friendshipType, friendshipMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `friendship` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `friendship` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `friendship` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, friendshipPrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into friendship\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, identifierCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for friendship\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfriendshipInsertCacheMut.Lock()\n\t\tfriendshipInsertCache[key] = cache\n\t\tfriendshipInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (data *SimpleDbType) Insert(dbMgr *mgr.DBConn) error {\n\tdb := dbMgr.Open()\n\n\tsqlStatement := `INSERT INTO test_table (\"name\", \"number\") VALUES ($1, $2) RETURNING id`\n\tid := 0\n\terr := db.QueryRow(sqlStatement, data.Name, data.Number).Scan(&id)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfmt.Println(\"New record ID is:\", id)\n\tdbMgr.Close()\n\treturn err\n}", "func (o *RecordMeasure) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no record_measures provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(recordMeasureColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\trecordMeasureInsertCacheMut.RLock()\n\tcache, cached := recordMeasureInsertCache[key]\n\trecordMeasureInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\trecordMeasureColumns,\n\t\t\trecordMeasureColumnsWithDefault,\n\t\t\trecordMeasureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(recordMeasureType, recordMeasureMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(recordMeasureType, recordMeasureMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"record_measures\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"record_measures\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to insert into record_measures\")\n\t}\n\n\tif !cached {\n\t\trecordMeasureInsertCacheMut.Lock()\n\t\trecordMeasureInsertCache[key] = cache\n\t\trecordMeasureInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Employee) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no employee provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(employeeColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\temployeeInsertCacheMut.RLock()\n\tcache, cached := employeeInsertCache[key]\n\temployeeInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\temployeeAllColumns,\n\t\t\temployeeColumnsWithDefault,\n\t\t\temployeeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(employeeType, employeeMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(employeeType, employeeMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"employee\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"employee\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into employee\")\n\t}\n\n\tif !cached {\n\t\temployeeInsertCacheMut.Lock()\n\t\temployeeInsertCache[key] = cache\n\t\temployeeInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func Insert(db gorp.SqlExecutor, i interface{}) error {\n\treturn Mapper.Insert(db, i)\n}", "func (o *APIKey) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no api_keys provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(apiKeyColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tapiKeyInsertCacheMut.RLock()\n\tcache, cached := apiKeyInsertCache[key]\n\tapiKeyInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyColumnsWithDefault,\n\t\t\tapiKeyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"api_keys\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"api_keys\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into api_keys\")\n\t}\n\n\tif !cached {\n\t\tapiKeyInsertCacheMut.Lock()\n\t\tapiKeyInsertCache[key] = cache\n\t\tapiKeyInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (t TaskInstanceRepoCassandra) Insert(ctx context.Context, taskInstance TaskInstance) (err error) {\n\terr = t.insertTaskInstance(ctx, taskInstance)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = t.insertTaskInstanceStartedAt(ctx, taskInstance)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn t.insertInstancesByID(ctx, taskInstance)\n}", "func (d *Database) Insert(db DB, table string, src interface{}) error {\n\tpkName, pkValue, err := d.PrimaryKey(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif pkName != \"\" && pkValue != 0 {\n\t\treturn fmt.Errorf(\"meddler.Insert: primary key must be zero\")\n\t}\n\n\t// gather the query parts\n\tnamesPart, err := d.ColumnsQuoted(src, false)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvaluesPart, err := d.PlaceholdersString(src, false)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvalues, err := d.Values(src, false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// run the query\n\tq := fmt.Sprintf(\"INSERT INTO %s (%s) VALUES (%s)\", d.quoted(table), namesPart, valuesPart)\n\tif d.UseReturningToGetID && pkName != \"\" {\n\t\tq += \" RETURNING \" + d.quoted(pkName)\n\t\tvar newPk int64\n\t\terr := db.QueryRow(q, values...).Scan(&newPk)\n\t\tif err != nil {\n\t\t\treturn &dbErr{msg: \"meddler.Insert: DB error in QueryRow\", err: err}\n\t\t}\n\t\tif err = d.SetPrimaryKey(src, newPk); err != nil {\n\t\t\treturn fmt.Errorf(\"meddler.Insert: Error saving updated pk: %v\", err)\n\t\t}\n\t} else if pkName != \"\" {\n\t\tresult, err := db.Exec(q, values...)\n\t\tif err != nil {\n\t\t\treturn &dbErr{msg: \"meddler.Insert: DB error in Exec\", err: err}\n\t\t}\n\n\t\t// save the new primary key\n\t\tnewPk, err := result.LastInsertId()\n\t\tif err != nil {\n\t\t\treturn &dbErr{msg: \"meddler.Insert: DB error getting new primary key value\", err: err}\n\t\t}\n\t\tif err = d.SetPrimaryKey(src, newPk); err != nil {\n\t\t\treturn fmt.Errorf(\"meddler.Insert: Error saving updated pk: %v\", err)\n\t\t}\n\t} else {\n\t\t// no primary key, so no need to lookup new value\n\t\t_, err := db.Exec(q, values...)\n\t\tif err != nil {\n\t\t\treturn &dbErr{msg: \"meddler.Insert: DB error in Exec\", err: err}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *Illness) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no illness provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(illnessColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tillnessInsertCacheMut.RLock()\n\tcache, cached := illnessInsertCache[key]\n\tillnessInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tillnessAllColumns,\n\t\t\tillnessColumnsWithDefault,\n\t\t\tillnessColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(illnessType, illnessMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(illnessType, illnessMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"illness\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"illness\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into illness\")\n\t}\n\n\tif !cached {\n\t\tillnessInsertCacheMut.Lock()\n\t\tillnessInsertCache[key] = cache\n\t\tillnessInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (m MariaDB) Insert(ctx context.Context, document entity.PersonalData) (entity.PersonalData, error) {\n\tp := receive(document)\n\tsqlQuery := \"INSERT INTO person (id, name, last_name, phone, email, year_od_birth ) VALUES (?,?,?,?,?,?)\"\n\t_, err := m.Person.ExecContext(ctx, sqlQuery, p.ID, p.Name, p.LastName, p.Phone, p.Email, p.YearOfBirth)\n\tif err != nil {\n\t\treturn entity.PersonalData{}, errors.Wrap(err, \"could not exec query statement\")\n\t}\n\treturn document, nil\n}", "func insertAccountRecord(db *sqlite.Driver, account *Account) error {\n\tvar err error\n\tvar stmt = fmt.Sprintf(`insert into %s (%s, created_at, updated_at) values (?, ?, ?, ?, ?)`, tableName, insertColumns)\n\n\t_, err = db.Exec(stmt,\n\t\taccount.Name,\n\t\taccount.Credential,\n\t\taccount.PermLevel,\n\t\ttime.Now(),\n\t\ttime.Now(),\n\t)\n\treturn err\n}", "func (o *Peer) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no peers provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tif queries.MustTime(o.UpdatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(peerColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tpeerInsertCacheMut.RLock()\n\tcache, cached := peerInsertCache[key]\n\tpeerInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerColumnsWithDefault,\n\t\t\tpeerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(peerType, peerMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `peers` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `peers` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `peers` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, peerPrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to insert into peers\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == peerMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, identifierCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for peers\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpeerInsertCacheMut.Lock()\n\t\tpeerInsertCache[key] = cache\n\t\tpeerInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *Comment) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no comment provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(commentColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tcommentInsertCacheMut.RLock()\n\tcache, cached := commentInsertCache[key]\n\tcommentInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tcommentColumns,\n\t\t\tcommentColumnsWithDefault,\n\t\t\tcommentColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(commentType, commentMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(commentType, commentMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `comment` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `comment` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `comment` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, commentPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into comment\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.Index = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == commentMapping[\"Index\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.Index,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for comment\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tcommentInsertCacheMut.Lock()\n\t\tcommentInsertCache[key] = cache\n\t\tcommentInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func Create(record Record) {\r\n\tdb.QueryRow(\"insert into records values($1,$2,$3,$4,$5)\", record.Uuid, record.Mail, record.Seq, record.Pssm, record.Result)\r\n}", "func (d *dbBase) Insert(ctx context.Context, q dbQuerier, mi *modelInfo, ind reflect.Value, tz *time.Location) (int64, error) {\n\tnames := make([]string, 0, len(mi.fields.dbcols))\n\tvalues, autoFields, err := d.collectValues(mi, ind, mi.fields.dbcols, false, true, &names, tz)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tid, err := d.InsertValue(ctx, q, mi, false, names, values)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif len(autoFields) > 0 {\n\t\terr = d.ins.setval(ctx, q, mi, autoFields)\n\t}\n\treturn id, err\n}", "func (o *Ticket) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tickets provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(ticketColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tticketInsertCacheMut.RLock()\n\tcache, cached := ticketInsertCache[key]\n\tticketInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketColumnsWithDefault,\n\t\t\tticketColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(ticketType, ticketMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(ticketType, ticketMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"tickets\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"tickets\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into tickets\")\n\t}\n\n\tif !cached {\n\t\tticketInsertCacheMut.Lock()\n\t\tticketInsertCache[key] = cache\n\t\tticketInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (m *MySQL) Insert(p *packet.Packet, interval time.Duration, t time.Time) {\n\t_, err := m.stmt.Exec(p.Interface, p.Bytes, p.SrcName, p.DstName, p.Hostname, p.Proto, p.SrcPort, p.DstPort, int(interval.Seconds()), t)\n\tif err != nil {\n\t\tlog.Println(\"sql err:\", err)\n\t\tlog.Println(\"Time:\", t.Unix())\n\t\tspew.Dump(p)\n\t}\n}", "func Insert(r *http.Request, col *tiedot.Col) (id int, err error) {\n\tdata := map[string]interface{}{}\n\tdecoder := json.NewDecoder(r.Body)\n\tdecoder.Decode(&data)\n\n\tid, err = col.Insert(data)\n\treturn\n}", "func (o *Stock) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stock provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(stockColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tstockInsertCacheMut.RLock()\n\tcache, cached := stockInsertCache[key]\n\tstockInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tstockColumns,\n\t\t\tstockColumnsWithDefault,\n\t\t\tstockColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(stockType, stockMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(stockType, stockMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `stock` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `stock` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `stock` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, stockPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into stock\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.StockID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == stockMapping[\"StockID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.StockID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for stock\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tstockInsertCacheMut.Lock()\n\t\tstockInsertCache[key] = cache\n\t\tstockInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (table *Table) Insert(db DB, record Map) (Result, error) {\n\tc := Context{StateInsert, db, table, table, record, Field{}}\n\treturn table.execAction(c, table.OnInsert, table.DefaultInsert)\n}", "func (o *Board) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"rdb: no boards provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(boardColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tboardInsertCacheMut.RLock()\n\tcache, cached := boardInsertCache[key]\n\tboardInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardColumnsWithDefault,\n\t\t\tboardColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(boardType, boardMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(boardType, boardMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `boards` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `boards` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `boards` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, boardPrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to insert into boards\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == boardMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to populate default values for boards\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tboardInsertCacheMut.Lock()\n\t\tboardInsertCache[key] = cache\n\t\tboardInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (s *SQL) InsertOne(block interface{}) error {\n\tif err := s.C.Insert(block); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n\n}", "func (o *Vote) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tvoteInsertCacheMut.RLock()\n\tcache, cached := voteInsertCache[key]\n\tvoteInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"vote\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"vote\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into vote\")\n\t}\n\n\tif !cached {\n\t\tvoteInsertCacheMut.Lock()\n\t\tvoteInsertCache[key] = cache\n\t\tvoteInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Smallblog) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no smallblog provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(smallblogColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tsmallblogInsertCacheMut.RLock()\n\tcache, cached := smallblogInsertCache[key]\n\tsmallblogInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tsmallblogAllColumns,\n\t\t\tsmallblogColumnsWithDefault,\n\t\t\tsmallblogColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(smallblogType, smallblogMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(smallblogType, smallblogMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `smallblog` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `smallblog` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `smallblog` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, smallblogPrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into smallblog\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, identifierCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for smallblog\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tsmallblogInsertCacheMut.Lock()\n\t\tsmallblogInsertCache[key] = cache\n\t\tsmallblogInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (r *Entity) Insert() (result sql.Result, err error) {\n\treturn Model.Data(r).Insert()\n}", "func (o *FeatureRelationship) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no feature_relationship provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(featureRelationshipColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tfeatureRelationshipInsertCacheMut.RLock()\n\tcache, cached := featureRelationshipInsertCache[key]\n\tfeatureRelationshipInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tfeatureRelationshipColumns,\n\t\t\tfeatureRelationshipColumnsWithDefault,\n\t\t\tfeatureRelationshipColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(featureRelationshipType, featureRelationshipMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(featureRelationshipType, featureRelationshipMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"feature_relationship\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to insert into feature_relationship\")\n\t}\n\n\tif !cached {\n\t\tfeatureRelationshipInsertCacheMut.Lock()\n\t\tfeatureRelationshipInsertCache[key] = cache\n\t\tfeatureRelationshipInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func InsertIgnoreFunc(scope *gorm.Scope, columnNames, groups []string) {\n\tdefaultWithFormat(scope, columnNames, groups, \"INSERT IGNORE INTO %s (%s) VALUES %s\")\n}", "func (data *StudentMasterDb) Insert(expiryDate string) <-chan DbModelError {\n\n\tJob := make(chan DbModelError, 1)\n\tsuccessResp := map[string]string{}\n\tvar customError DbModelError\n\tif CheckPing(&customError); customError.Err != nil {\n\t\tJob <- customError\n\t\treturn Job\n\t}\n\n\t// Verify as a new User\n\tvar studentExists bool\n\tdbSP, _ := RetriveSP(\"STU_EXISTS_WITH_EMAIL\")\n\terr := Db.QueryRow(dbSP, data.PersonalEmail).Scan(&data.StakeholderID, &data.PersonalEmail, &studentExists)\n\n\tif err != nil && err != sql.ErrNoRows {\n\t\tfmt.Println(\"query operation failed\" + err.Error())\n\t\tJob <- DbModelError{\n\t\t\t\"500\", \"S1AUT001\", fmt.Errorf(\"Cannot Read Database %v \", err.Error()), successResp,\n\t\t}\n\t\treturn Job\n\t}\n\t//fmt.Printf(\" 49 %v %+v\\n \", studentExists, err)\n\n\t// Return if already exists\n\tif studentExists {\n\t\tJob <- DbModelError{\n\t\t\t\"403\", \"S1AUT002\", fmt.Errorf(\"Account exists with email: %s\", data.PersonalEmail), successResp,\n\t\t}\n\t\treturn Job\n\n\t}\n\tsID, refCode, cbError := createStuSID(data.DateOfBirth)\n\tif cbError.ErrCode != \"000\" {\n\t\tJob <- cbError\n\t\treturn Job\n\t}\n\tdata.StakeholderID = sID\n\tfmt.Println(data.StakeholderID)\n\t// Prepare Db Insert\n\tdbSP, _ = RetriveSP(\"STU_INS_NEW_USR\")\n\tstmt, err := Db.Prepare(dbSP)\n\tif err != nil {\n\n\t\tfmt.Println(\"error while inserting\" + err.Error())\n\t\tJob <- DbModelError{\n\t\t\t\"500\", \"S1AUT003\", fmt.Errorf(\"Error While registering Student %v \", err.Error()), successResp,\n\t\t}\n\t\treturn Job\n\t}\n\tdefer stmt.Close()\n\tdata.CreationDate = time.Now()\n\tdata.LastUpdatedDate = data.CreationDate\n\tresults, err := stmt.Exec(&data.StakeholderID, &data.FirstName, &data.MiddleName, &data.LastName, &data.PersonalEmail, &data.PhoneNumber, &data.AlternatePhoneNumber, &data.Gender, &data.DateOfBirth, &data.AadharNumber, &data.PermanentAddressLine1, &data.PermanentAddressLine2, &data.PermanentAddressLine3, &data.PermanentAddressCountry, &data.PermanentAddressState, &data.PermanentAddressCity, &data.PermanentAddressDistrict, &data.PermanentAddressZipcode, &data.PermanentAddressPhone, &data.PresentAddressLine1, &data.PresentAddressLine2, &data.PresentAddressLine3, &data.PresentAddressCountry, &data.PresentAddressState, &data.PresentAddressCity, &data.PresentAddressDistrict, &data.PresentAddressZipcode, &data.PresentAddressPhone, &data.UniversityName, &data.UniversityID, &data.ProgramName, &data.ProgramID, &data.BranchName, &data.BranchID, &data.CollegeID, &data.CollegeEmailID, &data.Password, &data.UniversityApprovedFlag, &data.CreationDate, &data.LastUpdatedDate, &data.AccountStatus, false, false, expiryDate, &data.Attachment, data.AttachmentName, &data.CreationDate, refCode, data.ProfilePicture)\n\tfmt.Printf(\"results: %+v \\n %+v\", results, err)\n\tif err != nil {\n\n\t\tfmt.Println(\"error while inserting\" + err.Error())\n\t\tJob <- DbModelError{\n\t\t\t\"500\", \"S1AUT004\", fmt.Errorf(\"Error While registering Student %v \", err.Error()), successResp,\n\t\t}\n\t\treturn Job\n\t}\n\n\t// Print data in Console\n\tfmt.Printf(\"line 80 %+v %+v \\n \", data, err)\n\n\tcustomError.ErrTyp = \"000\"\n\tsuccessResp[\"Phone\"] = data.PhoneNumber\n\tsuccessResp[\"StakeholderID\"] = data.StakeholderID\n\tsuccessResp[\"Email\"] = data.PersonalEmail\n\tcustomError.SuccessResp = successResp\n\n\tJob <- customError\n\n\treturn Job\n\n}", "func (c *Client) InsertReturning(ctx context.Context, v Model, cols ...string) error {\n\treturn InsertReturning(ctx, c, v, cols...)\n}", "func Insert(insertChan chan Record, conn *sql.DB, writer io.Writer) {\n\n}", "func (tbl DbCompoundTable) Insert(req require.Requirement, vv ...*Compound) error {\n\tif req == require.All {\n\t\treq = require.Exactly(len(vv))\n\t}\n\n\tvar count int64\n\t//columns := allXExampleQuotedInserts[tbl.Dialect().Index()]\n\t//query := fmt.Sprintf(\"INSERT INTO %s %s\", tbl.name, columns)\n\t//st, err := tbl.db.PrepareContext(tbl.ctx, query)\n\t//if err != nil {\n\t//\treturn err\n\t//}\n\t//defer st.Close()\n\n\tinsertHasReturningPhrase := false\n\treturning := \"\"\n\tfor _, v := range vv {\n\t\tvar iv interface{} = v\n\t\tif hook, ok := iv.(sqlgen2.CanPreInsert); ok {\n\t\t\terr := hook.PreInsert()\n\t\t\tif err != nil {\n\t\t\t\treturn tbl.logError(err)\n\t\t\t}\n\t\t}\n\n\t\tb := &bytes.Buffer{}\n\t\tio.WriteString(b, \"INSERT INTO \")\n\t\tio.WriteString(b, tbl.name.String())\n\n\t\tfields, err := constructDbCompoundInsert(b, v, tbl.Dialect(), true)\n\t\tif err != nil {\n\t\t\treturn tbl.logError(err)\n\t\t}\n\n\t\tio.WriteString(b, \" VALUES (\")\n\t\tio.WriteString(b, tbl.Dialect().Placeholders(len(fields)))\n\t\tio.WriteString(b, \")\")\n\t\tio.WriteString(b, returning)\n\n\t\tquery := b.String()\n\t\ttbl.logQuery(query, fields...)\n\n\t\tvar n int64 = 1\n\t\tif insertHasReturningPhrase {\n\t\t\trow := tbl.db.QueryRowContext(tbl.ctx, query, fields...)\n\t\t\tvar i64 int64\n\t\t\terr = row.Scan(&i64)\n\n\t\t} else {\n\t\t\tres, e2 := tbl.db.ExecContext(tbl.ctx, query, fields...)\n\t\t\tif e2 != nil {\n\t\t\t\treturn tbl.logError(e2)\n\t\t\t}\n\n\t\t\tif e2 != nil {\n\t\t\t\treturn tbl.logError(e2)\n\t\t\t}\n\n\t\t\tn, err = res.RowsAffected()\n\t\t}\n\n\t\tif err != nil {\n\t\t\treturn tbl.logError(err)\n\t\t}\n\t\tcount += n\n\t}\n\n\treturn tbl.logIfError(require.ErrorIfExecNotSatisfiedBy(req, count))\n}", "func (o *Customer) Insert(exec boil.Executor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no customers provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(customerColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tcustomerInsertCacheMut.RLock()\n\tcache, cached := customerInsertCache[key]\n\tcustomerInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tcustomerColumns,\n\t\t\tcustomerColumnsWithDefault,\n\t\t\tcustomerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(customerType, customerMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(customerType, customerMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"customers\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"customers\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into customers\")\n\t}\n\n\tif !cached {\n\t\tcustomerInsertCacheMut.Lock()\n\t\tcustomerInsertCache[key] = cache\n\t\tcustomerInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func Insert(db DB, table string, src interface{}) error {\n\treturn InsertContext(context.Background(), db, table, src)\n}", "func (d MySQLDriver) Insert(tableName string, columns string, values ...interface{}) (affected int64, id int64, err error) {\n\tdb, err := d.OpenDB()\n\tif err != nil {\n\t\tcheckErr(err)\n\t}\n\tdefer db.Close()\n\tfmt.Printf(\"tablename:%+v columns:%+v values:%+v\\n\", tableName, columns, values)\n\n\tstmt, err := db.Prepare(\"INSERT \" + tableName + \" SET \" + columns)\n\tcheckErr(err)\n\tres, err := stmt.Exec(values...)\n\tcheckErr(err)\n\tfmt.Printf(\"res:%+v \\n\", res)\n\tid, err = res.LastInsertId()\n\taffect, err := res.RowsAffected()\n\tcheckErr(err)\n\treturn affect, id, err\n}", "func (o *AuthUserUserPermission) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_user_user_permissions provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(authUserUserPermissionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tauthUserUserPermissionInsertCacheMut.RLock()\n\tcache, cached := authUserUserPermissionInsertCache[key]\n\tauthUserUserPermissionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tauthUserUserPermissionColumns,\n\t\t\tauthUserUserPermissionColumnsWithDefault,\n\t\t\tauthUserUserPermissionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authUserUserPermissionType, authUserUserPermissionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(authUserUserPermissionType, authUserUserPermissionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.query = fmt.Sprintf(\"INSERT INTO `auth_user_user_permissions` (`%s`) VALUES (%s)\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `auth_user_user_permissions` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, authUserUserPermissionPrimaryKeyColumns))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into auth_user_user_permissions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authUserUserPermissionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_user_user_permissions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthUserUserPermissionInsertCacheMut.Lock()\n\t\tauthUserUserPermissionInsertCache[key] = cache\n\t\tauthUserUserPermissionInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (t *Table) Insert(logs logs.JSON) error {\n\t// construct insert statement\n\tinsert, args := InsertTableStatement(t.Name, t.Schema, logs)\n\n\t// insert the data\n\t_, err := t.Exec(insert, args...)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"inserting records for %s table\", t.Name)\n\t}\n\treturn nil\n}", "func (db *DB) Insert(model ...interface{}) error {\n\treturn db.inner.Insert(model...)\n}", "func (o *Picture) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no pictures provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(pictureColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tpictureInsertCacheMut.RLock()\n\tcache, cached := pictureInsertCache[key]\n\tpictureInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpictureColumnsWithDefault,\n\t\t\tpictureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(pictureType, pictureMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(pictureType, pictureMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `pictures` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `pictures` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `pictures` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, picturePrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into pictures\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for pictures\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpictureInsertCacheMut.Lock()\n\t\tpictureInsertCache[key] = cache\n\t\tpictureInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func Insert(db *sql.DB, table string, attributes interface{}) (result sql.Result, err error, query string, args []interface{}) {\n\tif db == nil {\n\t\terr = errors.New(\"db can't be nil\")\n\t\treturn\n\t}\n\n\tif table == \"\" {\n\t\terr = errors.New(\"table can't be empty\")\n\t\treturn\n\t}\n\n\tif attributes == nil {\n\t\terr = errors.New(\"attributes can't be nil\")\n\t\treturn\n\t}\n\n\tcolumns, args, _, err := ReflectValue(attributes, TagKey)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tquery, err = BuildInsert(table, columns)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tresult, err = db.Exec(query, args...)\n\treturn\n}", "func (o *Currency) Insert(exec boil.Executor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no currency provided for insertion\")\n\t}\n\n\tvar err error\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\tif o.UpdatedAt.IsZero() {\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currencyColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tcurrencyInsertCacheMut.RLock()\n\tcache, cached := currencyInsertCache[key]\n\tcurrencyInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyColumnsWithDefault,\n\t\t\tcurrencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currencyType, currencyMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(currencyType, currencyMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"currency\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"currency\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into currency\")\n\t}\n\n\tif !cached {\n\t\tcurrencyInsertCacheMut.Lock()\n\t\tcurrencyInsertCache[key] = cache\n\t\tcurrencyInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Block) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no block provided for insertion\")\n\t}\n\n\tvar err error\n\n\tnzDefaults := queries.NonZeroDefaultSet(blockColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tblockInsertCacheMut.RLock()\n\tcache, cached := blockInsertCache[key]\n\tblockInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockColumnsWithDefault,\n\t\t\tblockColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(blockType, blockMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(blockType, blockMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"block\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"block\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into block\")\n\t}\n\n\tif !cached {\n\t\tblockInsertCacheMut.Lock()\n\t\tblockInsertCache[key] = cache\n\t\tblockInsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (d *Database) Insert(db DB, table string, src interface{}) error {\n\treturn d.InsertContext(context.Background(), db, table, src)\n}", "func (ses *Ses) Ins(tbl string, columnPairs ...interface{}) (err error) {\n\tses.log(_drv.Cfg().Log.Ses.Ins)\n\terr = ses.checkClosed()\n\tif err != nil {\n\t\treturn errE(err)\n\t}\n\tif tbl == \"\" {\n\t\treturn errF(\"tbl is empty.\")\n\t}\n\tif len(columnPairs) < 2 {\n\t\treturn errF(\"Parameter 'columnPairs' expects at least 2 column name-value pairs.\")\n\t}\n\tif len(columnPairs)%2 != 0 {\n\t\treturn errF(\"Variadic parameter 'columnPairs' received an odd number of elements. Parameter 'columnPairs' expects an even number of elements.\")\n\t}\n\t// build INSERT statement, params slice\n\tparams := make([]interface{}, len(columnPairs)/2)\n\tbuf := new(bytes.Buffer)\n\tbuf.WriteString(\"INSERT INTO \")\n\tbuf.WriteString(tbl)\n\tbuf.WriteString(\" (\")\n\tlastColName := \"\"\n\tfor p := 0; p < len(params); p++ {\n\t\tn := p * 2\n\t\tcolumnName, ok := columnPairs[n].(string)\n\t\tif !ok {\n\t\t\treturn errF(\"Variadic parameter 'columnPairs' expected an element at index %v to be of type string\", n)\n\t\t}\n\t\tif p == len(params)-1 {\n\t\t\tlastColName = columnName\n\t\t} else {\n\t\t\tbuf.WriteString(columnName)\n\t\t\tif p < len(params)-2 {\n\t\t\t\tbuf.WriteString(\", \")\n\t\t\t}\n\t\t}\n\t\tparams[p] = columnPairs[n+1]\n\t}\n\tbuf.WriteString(\") VALUES (\")\n\tfor n := 1; n < len(params); n++ {\n\t\tbuf.WriteString(fmt.Sprintf(\":%v\", n))\n\t\tif n < len(params)-1 {\n\t\t\tbuf.WriteString(\", \")\n\t\t}\n\t}\n\tbuf.WriteString(\")\")\n\tbuf.WriteString(\" RETURNING \")\n\tbuf.WriteString(lastColName)\n\tbuf.WriteString(\" INTO :RET_VAL\")\n\tstmt, err := ses.Prep(buf.String()) // prep\n\tif err != nil {\n\t\treturn errE(err)\n\t}\n\tdefer stmt.Close()\n\t_, err = stmt.Exe(params...) // exe\n\tif err != nil {\n\t\treturn errE(err)\n\t}\n\treturn nil\n}", "func (a *Activity) Insert(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\tvar res sql.Result\n\t// if already exist, bail\n\tif a._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetActivityTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// sql insert query, primary key provided by autoincrement\n\tsqlstr := `INSERT INTO ` + tableName +\n\t\t` (` +\n\t\t`name, activity_type_id, activity_type_code, content, activity_image, image_url, image_color, status, sort, extend, admin_id, admin_name, start_time, end_time, created_at, updated_at` +\n\t\t`) VALUES (` +\n\t\t`?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?` +\n\t\t`)`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, a.Name, a.ActivityTypeID, a.ActivityTypeCode, a.Content, a.ActivityImage, a.ImageURL, a.ImageColor, a.Status, a.Sort, a.Extend, a.AdminID, a.AdminName, a.StartTime, a.EndTime, a.CreatedAt, a.UpdatedAt)))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif tx != nil {\n\t\tres, err = tx.Exec(sqlstr, a.Name, a.ActivityTypeID, a.ActivityTypeCode, a.Content, a.ActivityImage, a.ImageURL, a.ImageColor, a.Status, a.Sort, a.Extend, a.AdminID, a.AdminName, a.StartTime, a.EndTime, a.CreatedAt, a.UpdatedAt)\n\t} else {\n\t\tres, err = dbConn.Exec(sqlstr, a.Name, a.ActivityTypeID, a.ActivityTypeCode, a.Content, a.ActivityImage, a.ImageURL, a.ImageColor, a.Status, a.Sort, a.Extend, a.AdminID, a.AdminName, a.StartTime, a.EndTime, a.CreatedAt, a.UpdatedAt)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// retrieve id\n\tid, err := res.LastInsertId()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set primary key and existence\n\ta.ID = uint(id)\n\ta._exists = true\n\n\treturn nil\n}", "func insertEntry(db *sql.DB, stamp time.Time, watts int) error {\n\tstampStr := stamp.Format(RFC3339NoZ)\n\tinsertSQLFormat := insertSQLFormat(db)\n\tinsertSQL := fmt.Sprintf(insertSQLFormat, stampStr, watts)\n\t_, err := db.Exec(insertSQL)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *<%= classedName %>) Insert() error {\n\tif _, err := orm.NewOrm().Insert(o); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *Tenant) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"dbmodel: no tenants provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tif o.UpdatedAt.IsZero() {\n\t\t\to.UpdatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(tenantColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\ttenantInsertCacheMut.RLock()\n\tcache, cached := tenantInsertCache[key]\n\ttenantInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\ttenantAllColumns,\n\t\t\ttenantColumnsWithDefault,\n\t\t\ttenantColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(tenantType, tenantMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(tenantType, tenantMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `tenants` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `tenants` () VALUES ()%s%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `tenants` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, tenantPrimaryKeyColumns))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to insert into tenants\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, identifierCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to populate default values for tenants\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttenantInsertCacheMut.Lock()\n\t\ttenantInsertCache[key] = cache\n\t\ttenantInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (c *BigQueryClient) Insert(ctx context.Context, issues []Issue) error {\n\tinserter := c.Table.Inserter()\n\tinserter.IgnoreUnknownValues = true\n\n\tif err := inserter.Put(ctx, issues); err != nil {\n\t\tputErr := err.(bigquery.PutMultiError)\n\t\tfor _, rowErr := range putErr {\n\t\t\tlog.From(ctx).Error(\"inserting row\", zap.Error(rowErr.Errors))\n\t\t}\n\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Channel) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no channels provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(channelColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tchannelInsertCacheMut.RLock()\n\tcache, cached := channelInsertCache[key]\n\tchannelInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tchannelAllColumns,\n\t\t\tchannelColumnsWithDefault,\n\t\t\tchannelColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(channelType, channelMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(channelType, channelMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"channels\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"channels\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into channels\")\n\t}\n\n\tif !cached {\n\t\tchannelInsertCacheMut.Lock()\n\t\tchannelInsertCache[key] = cache\n\t\tchannelInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func (o *Skin) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no skin provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(skinColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tskinInsertCacheMut.RLock()\n\tcache, cached := skinInsertCache[key]\n\tskinInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tskinColumns,\n\t\t\tskinColumnsWithDefault,\n\t\t\tskinColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(skinType, skinMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(skinType, skinMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO `skin` (`%s`) %%sVALUES (%s)%%s\", strings.Join(wl, \"`,`\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO `skin` () VALUES ()\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.retQuery = fmt.Sprintf(\"SELECT `%s` FROM `skin` WHERE %s\", strings.Join(returnColumns, \"`,`\"), strmangle.WhereClause(\"`\", \"`\", 0, skinPrimaryKeyColumns))\n\t\t}\n\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into skin\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.Index,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for skin\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tskinInsertCacheMut.Lock()\n\t\tskinInsertCache[key] = cache\n\t\tskinInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *RecordMeasure) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Latency) Insert(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no latencies provided for insertion\")\n\t}\n\n\tvar err error\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.UpdatedAt.IsZero() {\n\t\t\to.UpdatedAt = currTime\n\t\t}\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeInsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(latencyColumnsWithDefault, o)\n\n\tkey := makeCacheKey(columns, nzDefaults)\n\tlatencyInsertCacheMut.RLock()\n\tcache, cached := latencyInsertCache[key]\n\tlatencyInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := columns.InsertColumnSet(\n\t\t\tlatencyAllColumns,\n\t\t\tlatencyColumnsWithDefault,\n\t\t\tlatencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(latencyType, latencyMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(latencyType, latencyMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"latencies\\\" (\\\"%s\\\") %%sVALUES (%s)%%s\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.UseIndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"latencies\\\" %sDEFAULT VALUES%s\"\n\t\t}\n\n\t\tvar queryOutput, queryReturning string\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tqueryReturning = fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(cache.query, queryOutput, queryReturning)\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to insert into latencies\")\n\t}\n\n\tif !cached {\n\t\tlatencyInsertCacheMut.Lock()\n\t\tlatencyInsertCache[key] = cache\n\t\tlatencyInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(ctx, exec)\n}", "func Insert(tid int, database *sql.DB, entries <-chan MarksStudent, results chan<- MarksStudent) {\n\n\tstatement, _ := database.Prepare(\"INSERT INTO student_marks (uid, name, maths, physics, chemistry) VALUES (?, ?, ?, ?, ?)\")\n\tfor entry := range entries {\n\t\t//fmt.Println(tid, entry)\n\t\tstatement.Exec(entry.UID, entry.Name, entry.Maths,\n\t\t\tentry.Physics, entry.Chem)\n\t\tresults <- entry\n\t}\n\n}", "func (ds *MySQL) Create(data interface{}) (interface{}, error) {\n\tif data == nil {\n\t\treturn nil, errors.New(\"create_data_nil\")\n\t}\n\t// Checking for auto generated uuid. If found — generating\n\tuuidx := ds.generateUUID()\n\tdataMap := data.(map[string]interface{})\n\tif len(uuidx) > 0 {\n\t\tfor key, v := range uuidx {\n\t\t\tdataMap[key] = v\n\t\t}\n\t\tdata = dataMap\n\t}\n\t// Starting to build INSERT query\n\tbuilder := ds.adapter.Builder()\n\tbuilder.Insert(ds.source).Values(data)\n\tSQL := builder.Build()\n\n\tif ds.debug {\n\t\tfmt.Println(\"Create SQL: \", SQL)\n\t}\n\tfmt.Println(\"SQL is: \", SQL)\n\tresult, err := ds.adapter.Exec(SQL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// We have auto increment id that is returned\n\tif id, err := result.LastInsertId(); err == nil {\n\t\treturn ds.FindByID(id)\n\t}\n\t// We have primary key\n\tif ds.key != \"\" && dataMap[ds.key] != nil {\n\t\treturn ds.FindByID(dataMap[ds.key])\n\t}\n\t// We have nothing, just returning payload back\n\treturn data, nil\n}", "func (o *Transaction) InsertP(exec boil.Executor, whitelist ...string) {\n\tif err := o.Insert(exec, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.58229846", "0.5762543", "0.57507", "0.5735412", "0.573035", "0.55946255", "0.5570468", "0.5566235", "0.55581963", "0.5526916", "0.55217713", "0.55116796", "0.5486399", "0.5478079", "0.54741436", "0.5463372", "0.54504305", "0.5438887", "0.54303813", "0.5425824", "0.54163724", "0.5413742", "0.5386225", "0.538115", "0.5354728", "0.5353232", "0.5344402", "0.53392583", "0.53335136", "0.5282264", "0.5278563", "0.52588725", "0.52563345", "0.525595", "0.52492905", "0.5246899", "0.52445096", "0.5244272", "0.524253", "0.52333254", "0.52330273", "0.52279955", "0.5225176", "0.5215029", "0.5213214", "0.5209906", "0.5193231", "0.51915497", "0.51774883", "0.5166813", "0.51452523", "0.5136032", "0.5135487", "0.5133972", "0.51329416", "0.5125651", "0.5109947", "0.5107513", "0.5105031", "0.5087854", "0.5087744", "0.508636", "0.5082162", "0.50819254", "0.50817776", "0.50817776", "0.50817776", "0.50817776", "0.50817776", "0.50817776", "0.50653625", "0.50628465", "0.50596106", "0.50563663", "0.50525486", "0.50469005", "0.5045558", "0.50374043", "0.50323427", "0.50320095", "0.5031887", "0.5025966", "0.5024997", "0.5023493", "0.5022746", "0.50189304", "0.5018002", "0.50099224", "0.4997575", "0.49939698", "0.4992173", "0.4992086", "0.49819857", "0.4980739", "0.49778813", "0.4973911", "0.4970708", "0.49696982", "0.4968572", "0.49682286", "0.49614668" ]
0.0
-1
UpdateG a single AssetRevision record. See Update for whitelist behavior description.
func (o *AssetRevision) UpdateG(whitelist ...string) error { return o.Update(boil.GetDB(), whitelist...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *AssetRevision) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tassetRevisionUpdateCacheMut.RLock()\n\tcache, cached := assetRevisionUpdateCache[key]\n\tassetRevisionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(assetRevisionColumns, assetRevisionPrimaryKeyColumns, whitelist)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"public: unable to update asset_revision, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"asset_revision\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, assetRevisionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, append(wl, assetRevisionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to update asset_revision row\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionUpdateCacheMut.Lock()\n\t\tassetRevisionUpdateCache[key] = cache\n\t\tassetRevisionUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *AssetRevision) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AssetRevision) SetAssetG(insert bool, related *Asset) error {\n\treturn o.SetAsset(boil.GetDB(), insert, related)\n}", "func (o *Author) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Rental) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Inventory) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *BraceletPhoto) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o AssetRevisionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o *AssetRevision) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *StockCvterm) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func FindAssetRevisionG(orig int, selectCols ...string) (*AssetRevision, error) {\n\treturn FindAssetRevision(boil.GetDB(), orig, selectCols...)\n}", "func (o *Transaction) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Transaction) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Stock) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Vote) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (hg *Hg) Update() error {\n\treturn nil\n}", "func (o *Source) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *RSSAnnouncement) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (s *SmartContract) UpdateAsset(ctx contractapi.TransactionContextInterface, id string, newColor string, newSize int, newValue int) error {\n\n\tasset, err := s.ReadAsset(ctx, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclientID, err := s.GetSubmittingClientIdentity(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif clientID != asset.Owner {\n\t\treturn fmt.Errorf(\"submitting client not authorized to update asset, does not own asset\")\n\t}\n\n\tasset.Color = newColor\n\tasset.Size = newSize\n\tasset.AppraisedValue = newValue\n\n\tassetJSON, err := json.Marshal(asset)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ctx.GetStub().PutState(id, assetJSON)\n}", "func AssetRevisionExistsG(orig int) (bool, error) {\n\treturn AssetRevisionExists(boil.GetDB(), orig)\n}", "func (o *AssetRevision) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tassetRevisionUpsertCacheMut.RLock()\n\tcache, cached := assetRevisionUpsertCache[key]\n\tassetRevisionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"public: unable to upsert asset_revision, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(assetRevisionPrimaryKeyColumns))\n\t\t\tcopy(conflict, assetRevisionPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"asset_revision\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to upsert asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionUpsertCacheMut.Lock()\n\t\tassetRevisionUpsertCache[key] = cache\n\t\tassetRevisionUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *InstrumentClass) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Offer) UpdateG(columns boil.Columns) (int64, error) {\n\treturn o.Update(boil.GetDB(), columns)\n}", "func (o *Cvtermsynonym) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Comment) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func FindAssetRevisionGP(orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(boil.GetDB(), orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (o *FeatureCvtermDbxref) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func AssetRevisionsG(mods ...qm.QueryMod) assetRevisionQuery {\n\treturn AssetRevisions(boil.GetDB(), mods...)\n}", "func (o *Organism) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (a Author) Update(cfg *config.Config, authorID string, data map[string]string) {\n\n}", "func (g *Group) Update(r *Revocation) {\n\ttmp := new(big.Int).Neg(r.x)\n\ttmp.Add(tmp, bn256.Order)\n\tt := new(bn256.G2).ScalarMult(r.aStar, tmp)\n\tg.w.Add(g.g2, t)\n\n\tg.g1 = r.a\n\tg.g2 = r.aStar\n\n\tg.precompute()\n}", "func (o *AssetRevision) SetAssetGP(insert bool, related *Asset) {\n\tif err := o.SetAsset(boil.GetDB(), insert, related); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) AssetG(mods ...qm.QueryMod) assetQuery {\n\treturn o.AssetByFk(boil.GetDB(), mods...)\n}", "func (o *Item) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *RentalRower) UpdateG(columns boil.Columns) (int64, error) {\n\treturn o.Update(boil.GetDB(), columns)\n}", "func (o *AssetRevision) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PremiumCode) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *FeatureRelationship) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func Update(c *gin.Context) {\n\tvar r UpdateRequest\n\tc.BindJSON(&r)\n\tid, err := book.IDFromString(c.Param(\"id\"))\n\tif err != nil {\n\t\tlog.Println(\"UpdateBook IDFromString\", err)\n\t\tc.JSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\"error\": fmt.Sprintf(\"not found book to update by id %q\", id)},\n\t\t)\n\t\treturn\n\t}\n\tb := db.Find(id)\n\tb = updatedBook(r, b)\n\tdb.Save(b)\n\tc.JSON(http.StatusOK, gin.H{\"data\": b})\n}", "func (o *Jet) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Shelf) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func Update(c *gin.Context) {\r\n\toldpost := getById(c)\r\n\tvar newpost Post\r\n\tif err := c.ShouldBindJSON(&newpost); err != nil {\r\n\t\tc.JSON(http.StatusBadRequest, gin.H{\r\n\t\t\t\"messege\": err.Error(),\r\n\t\t\t\"data\": \"\",\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\toldpost.Title = newpost.Title\r\n\toldpost.Des = newpost.Des\r\n\tif newpost.Status != \"\" {\r\n\t\toldpost.Status = newpost.Status\r\n\t}\r\n\r\n\tdb.Save(&oldpost)\r\n\r\n\tc.JSON(http.StatusOK, gin.H{\r\n\t\t\"messege\": \"Post has been updated\",\r\n\t\t\"data\": oldpost,\r\n\t})\r\n}", "func (s *Store) Update(c *gin.Context) {\n\n}", "func (svc *AdminBuildService) Update(b *library.Build) (*library.Build, *Response, error) {\n\t// set the API endpoint path we send the request to\n\tu := \"/api/v1/admin/build\"\n\n\t// library Build type we want to return\n\tv := new(library.Build)\n\n\t// send request using client\n\tresp, err := svc.client.Call(\"PUT\", u, b, v)\n\n\treturn v, resp, err\n}", "func (o *ClaimInList) UpdateG(columns boil.Columns) error {\n\treturn o.Update(boil.GetDB(), columns)\n}", "func (r *pgRepository) Update(ctx context.Context, tenant string, item *model.APIDefinition) error {\n\tif item == nil {\n\t\treturn apperrors.NewInternalError(\"item cannot be nil\")\n\t}\n\n\tentity := r.conv.ToEntity(item)\n\n\treturn r.updater.UpdateSingle(ctx, resource.API, tenant, entity)\n}", "func (o *Tenant) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *AutomodRuleDatum) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Address) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (r repository) Update(ctx context.Context, link entity.Link) error {\n\treturn r.db.With(ctx).Model(&link).Update()\n}", "func (o *PremiumSlot) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Skin) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Picture) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (client *LicenseStatusClient) Update(obj *models.LicenseStatus, options ...session.ApiOptionsParams) (*models.LicenseStatus, error) {\n\tvar robj *models.LicenseStatus\n\tpath := client.getAPIPath(*obj.UUID)\n\terr := client.aviSession.Put(path, obj, &robj, options...)\n\treturn robj, err\n}", "func (reg *defaultRegistry) UpdateRevision(revision *engine.Revision) error {\n\t_, err := reg.store.Save(revision, store.WithReplaceOrForceGen())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error while updating revision: %s\", err)\n\t}\n\n\treturn nil\n}", "func (o *Failure) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) SetAsset(exec boil.Executor, insert bool, related *Asset) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(exec); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"asset_revision\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"asset\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, assetRevisionPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.Orig}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, updateQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tif _, err = exec.Exec(updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.Asset = related.ID\n\n\tif o.R == nil {\n\t\to.R = &assetRevisionR{\n\t\t\tAsset: related,\n\t\t}\n\t} else {\n\t\to.R.Asset = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &assetR{\n\t\t\tAssetRevisions: AssetRevisionSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.AssetRevisions = append(related.R.AssetRevisions, o)\n\t}\n\n\treturn nil\n}", "func AssetRevisionExistsGP(orig int) bool {\n\te, err := AssetRevisionExists(boil.GetDB(), orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (o *Ticket) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (m *MovieModel) Update(movie *Movie) error {\n\n\t/* potential to use uuid here\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = uuid_generate_v4()\n\tWHERE id = $5 AND\n\t**/\n\t// Add version = $6, so we can stop race conditions\n\tquery := `\n\tUPDATE movies\n\tSET title = $1, year = $2, runtime = $3, genres = $4, version = version + 1\n\tWHERE id = $5 AND version = $6\n\tRETURNING version \n\t`\n\n\t// create the arg slice contaninig the values for the placeholder params.\n\targs := []interface{}{\n\t\tmovie.Title,\n\t\tmovie.Year,\n\t\tmovie.Runtime,\n\t\tpq.Array(movie.Genres),\n\t\tmovie.ID,\n\t\tmovie.Version, // Add the expected movie version\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\t// If no matching row could be found (version has been changed)\n\terr := m.DB.QueryRowContext(ctx, query, args...).Scan(&movie.Version)\n\tif err != nil {\n\t\tswitch {\n\t\tcase errors.Is(err, sql.ErrNoRows):\n\t\t\treturn ErrEditConflict\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (r *BillingRepository) EditTransactiongByID(tx *models.Transaction) error {\n\terr := r.C.Update(bson.M{\"_id\": tx.ID},\n\t\tbson.M{\"$set\": bson.M{\n\t\t\t\"statement\": \t\ttx.Statement,\n\t\t\t\"paymentref\": \ttx.PaymentRef,\n\t\t\t\"status\": \ttx.Status,\n\t\t\t\"updatedat\":\t\ttime.Now(),\n\t\t}})\n\treturn err\n}", "func (o *DMessageEmbed) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func EditRelease(ctx *context.APIContext) {\n\t// swagger:operation PATCH /repos/{owner}/{repo}/releases/{id} repository repoEditRelease\n\t// ---\n\t// summary: Update a release\n\t// consumes:\n\t// - application/json\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: id\n\t// in: path\n\t// description: id of the release to edit\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// - name: body\n\t// in: body\n\t// schema:\n\t// \"$ref\": \"#/definitions/EditReleaseOption\"\n\t// responses:\n\t// \"200\":\n\t// \"$ref\": \"#/responses/Release\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/notFound\"\n\n\tform := web.GetForm(ctx).(*api.EditReleaseOption)\n\tid := ctx.ParamsInt64(\":id\")\n\trel, err := repo_model.GetReleaseByID(ctx, id)\n\tif err != nil && !repo_model.IsErrReleaseNotExist(err) {\n\t\tctx.Error(http.StatusInternalServerError, \"GetReleaseByID\", err)\n\t\treturn\n\t}\n\tif err != nil && repo_model.IsErrReleaseNotExist(err) ||\n\t\trel.IsTag || rel.RepoID != ctx.Repo.Repository.ID {\n\t\tctx.NotFound()\n\t\treturn\n\t}\n\n\tif len(form.TagName) > 0 {\n\t\trel.TagName = form.TagName\n\t}\n\tif len(form.Target) > 0 {\n\t\trel.Target = form.Target\n\t}\n\tif len(form.Title) > 0 {\n\t\trel.Title = form.Title\n\t}\n\tif len(form.Note) > 0 {\n\t\trel.Note = form.Note\n\t}\n\tif form.IsDraft != nil {\n\t\trel.IsDraft = *form.IsDraft\n\t}\n\tif form.IsPrerelease != nil {\n\t\trel.IsPrerelease = *form.IsPrerelease\n\t}\n\tif err := release_service.UpdateRelease(ctx.Doer, ctx.Repo.GitRepo, rel, nil, nil, nil); err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"UpdateRelease\", err)\n\t\treturn\n\t}\n\n\t// reload data from database\n\trel, err = repo_model.GetReleaseByID(ctx, id)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"GetReleaseByID\", err)\n\t\treturn\n\t}\n\tif err := rel.LoadAttributes(ctx); err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"LoadAttributes\", err)\n\t\treturn\n\t}\n\tctx.JSON(http.StatusOK, convert.ToAPIRelease(ctx, ctx.Repo.Repository, rel))\n}", "func (o *RowerGroup) UpdateG(columns boil.Columns) (int64, error) {\n\treturn o.Update(boil.GetDB(), columns)\n}", "func (m ComicsModel) Update(comics *Comics) error {\n\tquery := `UPDATE comics\n\t\t\tSET title = $1, year = $2, pages = $3, version = version + 1\n\t\t\tWHERE id = $4 AND version = $5\n\t\t\tRETURNING version`\n\targs := []interface{}{\n\t\tcomics.Title,\n\t\tcomics.Year,\n\t\tcomics.Pages,\n\t\tcomics.ID,\n\t\tcomics.Version,\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\terr := m.DB.QueryRowContext(ctx, query, args...).Scan(&comics.Version)\n\tif err != nil {\n\t\tswitch {\n\t\tcase errors.Is(err, sql.ErrNoRows):\n\t\t\treturn ErrEditConflict\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (r UpdateRevisionRequest) Send(ctx context.Context) (*UpdateRevisionResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &UpdateRevisionResponse{\n\t\tUpdateRevisionOutput: r.Request.Data.(*UpdateRevisionOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (pg *PGStorage) Update(a *Address) error {\n\tvar err error\n\t_, err = pg.con.Exec(`\n\t\t\tUPDATE address set ballance = $1, income = $2, outcome = $3\n\t\t\tWHERE id=$4`,\n\t\ta.Ballance,\n\t\ta.Income,\n\t\ta.Outcome,\n\t\ta.ID,\n\t)\n\treturn err\n}", "func (o AssetRevisionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (b *ProjectModels) UpdateGallery(id int, data ProjectGallery) error {\n\terr := configs.GetDB.Model(&data).Where(\"id = ?\", id).Updates(&data).Error\n\treturn err\n}", "func (o *RecordMeasure) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (m BooksModel) Update(books *Books) error {\n\tquery := `UPDATE books\n\t\t\tSET title = $1, year = $2, pages = $3, version = version + 1\n\t\t\tWHERE id = $4 AND version = $5\n\t\t\tRETURNING version`\n\targs := []interface{}{\n\t\tbooks.Title,\n\t\tbooks.Year,\n\t\tbooks.Pages,\n\t\tbooks.ID,\n\t\tbooks.Version,\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)\n\tdefer cancel()\n\n\terr := m.DB.QueryRowContext(ctx, query, args...).Scan(&books.Version)\n\tif err != nil {\n\t\tswitch {\n\t\tcase errors.Is(err, sql.ErrNoRows):\n\t\t\treturn ErrEditConflict\n\t\tdefault:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (o *Auth) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (p Database) Edit(d interface{}) (string, error) {\n\tjsonBuf, err := json.Marshal(d)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tidRev := idAndRev{}\n\tmust(json.Unmarshal(jsonBuf, &idRev))\n\tif idRev.ID == \"\" {\n\t\treturn \"\", errNoID\n\t}\n\tif idRev.Rev == \"\" {\n\t\treturn \"\", errNoRev\n\t}\n\tu := fmt.Sprintf(\"%s/%s\", p.DBURL(), url.QueryEscape(idRev.ID))\n\tir := Response{}\n\tif _, err = interact(\"PUT\", u, p.defaultHdrs, jsonBuf, &ir); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn ir.Rev, nil\n}", "func (r *resourceFrameworkShare) Update(ctx context.Context, req resource.UpdateRequest, resp *resource.UpdateResponse) {\n}", "func (o *BookCategory) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Rental) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (t *badgerTableVersion) Update(id []byte, obj ObjectVersion) ([]byte, error) {\n\tkey := t.getRefKey(id)\n\n\tval, err := obj.Marshal()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\th := t.hf()\n\tobj.Hash(h)\n\tobjID := h.Sum(nil)\n\n\terr = t.db.Update(func(txn *badger.Txn) error {\n\t\t_, err := txn.Get(key)\n\t\tif err != nil {\n\t\t\treturn translateError(err)\n\t\t}\n\n\t\tobjKey := t.getObjKey(id, objID)\n\t\terr = txn.Set(objKey, val)\n\t\tif err == nil {\n\t\t\terr = txn.Set(key, objID)\n\t\t}\n\n\t\treturn err\n\t})\n\n\tt.log.Printf(\"Update version key=%q\\n\", key)\n\n\treturn objID, err\n}", "func (o *Inventory) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (client *GalleryImageVersionsClient) update(ctx context.Context, resourceGroupName string, galleryName string, galleryImageName string, galleryImageVersionName string, galleryImageVersion GalleryImageVersionUpdate, options *GalleryImageVersionsBeginUpdateOptions) (*azcore.Response, error) {\n\treq, err := client.updateCreateRequest(ctx, resourceGroupName, galleryName, galleryImageName, galleryImageVersionName, galleryImageVersion, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.con.Pipeline().Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !resp.HasStatusCode(http.StatusOK) {\n\t\treturn nil, client.updateHandleError(resp)\n\t}\n\treturn resp, nil\n}", "func (operator *AccessOperator) UpdateRelease(cxt context.Context, option *ReleaseOption) error {\n\t//business first\n\tbusiness, _, err := getBusinessAndApp(operator, operator.Business, option.AppName)\n\tif err != nil {\n\t\treturn err\n\t}\n\trequest := &accessserver.UpdateReleaseReq{\n\t\tSeq: pkgcommon.Sequence(),\n\t\tBid: business.Bid,\n\t\tReleaseid: option.ReleaseID,\n\t\tName: option.Name,\n\t\tOperator: operator.User,\n\t}\n\tgrpcOptions := []grpc.CallOption{\n\t\tgrpc.WaitForReady(true),\n\t}\n\tresponse, err := operator.Client.UpdateRelease(cxt, request, grpcOptions...)\n\tif err != nil {\n\t\tlogger.V(3).Infof(\"UpdateRelease %s failed, %s\", option.Name, err.Error())\n\t\treturn err\n\t}\n\tif response.ErrCode != common.ErrCode_E_OK {\n\t\tlogger.V(3).Infof(\"UpdateRelease %s successfully, but response Err, %s\", option.ReleaseID, response.ErrMsg)\n\t\treturn fmt.Errorf(\"%s\", response.ErrMsg)\n\t}\n\treturn nil\n\n}", "func (o *BookCategoryAssign) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Store) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (m *Manager) Update(ctx context.Context, r *hub.ChartRepository) error {\n\tquery := \"select update_chart_repository($1::uuid, $2::jsonb)\"\n\tuserID := ctx.Value(hub.UserIDKey).(string)\n\trJSON, _ := json.Marshal(r)\n\t_, err := m.db.Exec(ctx, query, userID, rJSON)\n\treturn err\n}", "func (o *ItemSide) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *AssetRevision) Reload(exec boil.Executor) error {\n\tret, err := FindAssetRevision(exec, o.Orig)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (this *RouterTable) UpdateRevision() (previous, current int64) {\n\tprev := this.Revision\n\tthis.Revision = time.Now().Unix()\n\t//reset the dynmap\n\tthis.DynMap = dynmap.NewDynMap()\n\treturn prev, this.Revision\n}", "func (repo *Repository) Update(ctx context.Context, claims auth.Claims, req UpdateRequest, now time.Time) error {\n\tspan, ctx := tracer.StartSpanFromContext(ctx, \"internal.expenditure.Update\")\n\tdefer span.Finish()\n\n\tif claims.Audience == \"\" {\n\t\treturn errors.WithStack(ErrForbidden)\n\t}\n\t// Admin users can update branches they have access to.\n\tif !claims.HasRole(auth.RoleAdmin) {\n\t\treturn errors.WithStack(ErrForbidden)\n\t}\n\n\t// Validate the request.\n\tv := webcontext.Validator()\n\terr := v.Struct(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcols := models.M{}\n\tif req.Amount != nil {\n\t\tcols[models.RepsExpenseColumns.Amount] = *req.Amount\n\t}\n\tif req.Reason != nil {\n\t\tcols[models.RepsExpenseColumns.Reason] = *req.Reason\n\t}\n\tif len(cols) == 0 {\n\t\treturn nil\n\t}\n\n\t// If now empty set it to the current time.\n\tif now.IsZero() {\n\t\tnow = time.Now()\n\t}\n\n\t// Always store the time as UTC.\n\tnow = now.UTC()\n\t// Postgres truncates times to milliseconds when storing. We and do the same\n\t// here so the value we return is consistent with what we store.\n\tnow = now.Truncate(time.Millisecond)\n\n\tcols[models.BranchColumns.UpdatedAt] = now\n\n\t_, err = models.RepsExpenses(models.RepsExpenseWhere.ID.EQ(req.ID)).UpdateAll(ctx, repo.DbConn, cols)\n\n\treturn nil\n}", "func (a *Advert) Update(e *models.Advert) error {\n\t_, err := Storage.Run(\"UPDATE advert SET locality = ?, link = ?, hash_id = ?, price = ?, name = ?, description = ?, status = ?, created = ? WHERE id = ?\",\n\t\te.Locality, e.Link, e.HashID, e.Price, e.Name, e.Description, e.Status, e.GetCreated().Unix(), e.ID)\n\treturn err\n}", "func Update(ctx context.Context, vo ProjectVO.Instance) {\n\tds.Update(ctx, vo)\n\tcache.InvalidateCache(ctx)\n}", "func (o *AuthUserUserPermission) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (g *GistFile) Update(interface{}) (*http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}" ]
[ "0.62834185", "0.6270282", "0.5811454", "0.57556576", "0.57354647", "0.5684118", "0.56083536", "0.5535632", "0.55213845", "0.55205476", "0.551485", "0.5511936", "0.5471882", "0.5471882", "0.54289025", "0.54180944", "0.54090554", "0.5369409", "0.5346684", "0.53351533", "0.5314897", "0.5283723", "0.5240856", "0.52363527", "0.52345026", "0.5222833", "0.52172995", "0.5211249", "0.520213", "0.5199405", "0.5187508", "0.5170703", "0.5168798", "0.51682454", "0.51643264", "0.5161232", "0.5154989", "0.51528454", "0.5143824", "0.5141408", "0.5119707", "0.5098518", "0.5096106", "0.5086248", "0.5086139", "0.50788397", "0.5076114", "0.507431", "0.5071555", "0.5067869", "0.5067703", "0.50430316", "0.50336766", "0.50245804", "0.502451", "0.501705", "0.50134254", "0.5010276", "0.5007777", "0.5004328", "0.49864644", "0.49828252", "0.49808824", "0.49783397", "0.49752575", "0.49709734", "0.4955562", "0.494411", "0.4933748", "0.49276075", "0.4920212", "0.49126622", "0.4903443", "0.490117", "0.48999402", "0.4880249", "0.48787692", "0.4873722", "0.4863464", "0.48565516", "0.4850468", "0.48468274", "0.48457965", "0.48444128", "0.48320845", "0.48320374", "0.48236403", "0.48188064", "0.48095572", "0.47982243", "0.4791176", "0.47906095", "0.47879902", "0.47872168", "0.4785916", "0.47702762", "0.47670403", "0.47661397", "0.47590142", "0.47430986" ]
0.7077478
0
UpdateGP a single AssetRevision record. UpdateGP takes a whitelist of column names that should be updated. Panics on error. See Update for whitelist behavior description.
func (o *AssetRevision) UpdateGP(whitelist ...string) { if err := o.Update(boil.GetDB(), whitelist...); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Author) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Address) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) UpdateGP(columns boil.Columns) {\n\terr := o.Update(boil.GetDB(), columns)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Phenotypeprop) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasure) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) UpdateGP(ctx context.Context, columns boil.Columns) int64 {\n\trowsAff, err := o.Update(ctx, boil.GetContextDB(), columns)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *AuthMessage) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) UpdateGP(whitelist ...string) {\n\tif err := o.Update(boil.GetDB(), whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *StockCvterm) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Cvtermsynonym) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PremiumCode) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *StockCvterm) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Jet) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Author) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) SetAssetGP(insert bool, related *Asset) {\n\tif err := o.SetAsset(boil.GetDB(), insert, related); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Transaction) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Author) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Phenotypeprop) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AssetRevision) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Cvtermsynonym) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Inventory) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *BraceletPhoto) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Rental) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *APIKey) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AutomodRuleDatum) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *FeatureRelationship) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Jet) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *PremiumSlot) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Stock) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Offer) UpdateG(columns boil.Columns) (int64, error) {\n\treturn o.Update(boil.GetDB(), columns)\n}", "func (o *InstrumentClass) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Inventory) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o AssetRevisionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o *Item) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Address) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *AuthUserUserPermission) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *BraceletPhoto) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Picture) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *Comment) UpdateG(whitelist ...string) error {\n\treturn o.Update(boil.GetDB(), whitelist...)\n}", "func (o *BlackCard) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RSSAnnouncement) UpdateG(ctx context.Context, columns boil.Columns) (int64, error) {\n\treturn o.Update(ctx, boil.GetContextDB(), columns)\n}", "func (o *Comment) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o StockCvtermSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}" ]
[ "0.70530504", "0.7043095", "0.7043095", "0.7030146", "0.7028213", "0.6974584", "0.6943412", "0.691446", "0.6848551", "0.68293357", "0.67873293", "0.67292076", "0.67214954", "0.6708014", "0.6691071", "0.66745", "0.6667388", "0.6662034", "0.6617402", "0.6608754", "0.6581439", "0.65617317", "0.65389544", "0.64818996", "0.64740217", "0.64593184", "0.6427415", "0.64264053", "0.64123416", "0.63977474", "0.6360182", "0.6328576", "0.63175887", "0.6299114", "0.62791", "0.6269023", "0.6236092", "0.6221699", "0.6218057", "0.62179226", "0.62059885", "0.62059885", "0.6200466", "0.61535674", "0.61442894", "0.6137942", "0.6137279", "0.61304283", "0.61186165", "0.60296035", "0.60208577", "0.5979721", "0.5897243", "0.58716744", "0.5870841", "0.5857118", "0.58551675", "0.5837737", "0.5803901", "0.57997185", "0.57866937", "0.57805467", "0.5769964", "0.57689506", "0.57625526", "0.57577443", "0.572941", "0.5706733", "0.5682884", "0.5682884", "0.56774133", "0.56748354", "0.5668244", "0.56678045", "0.56498814", "0.5621432", "0.5620647", "0.558878", "0.55871356", "0.55726224", "0.55702156", "0.5552773", "0.55436337", "0.55128634", "0.5512812", "0.5511202", "0.5503818", "0.5497425", "0.54856426", "0.54657763", "0.54656494", "0.5464921", "0.5463595", "0.54581106", "0.54560286", "0.54550564", "0.5452954", "0.54489845", "0.5446011", "0.5445676" ]
0.79560786
0
UpdateP uses an executor to update the AssetRevision, and panics on error. See Update for whitelist behavior description.
func (o *AssetRevision) UpdateP(exec boil.Executor, whitelist ...string) { err := o.Update(exec, whitelist...) if err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Transaction) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) UpdateP(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) int64 {\n\trowsAff, err := o.Update(ctx, exec, columns)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *AuthUserUserPermission) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tassetRevisionUpdateCacheMut.RLock()\n\tcache, cached := assetRevisionUpdateCache[key]\n\tassetRevisionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(assetRevisionColumns, assetRevisionPrimaryKeyColumns, whitelist)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"public: unable to update asset_revision, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"asset_revision\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, assetRevisionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, append(wl, assetRevisionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to update asset_revision row\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionUpdateCacheMut.Lock()\n\t\tassetRevisionUpdateCache[key] = cache\n\t\tassetRevisionUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *DMessageEmbed) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) UpdateP(exec boil.Executor, columns boil.Columns) {\n\terr := o.Update(exec, columns)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (ru *RevisionUpdate) Exec(ctx context.Context) error {\n\t_, err := ru.Save(ctx)\n\treturn err\n}", "func (pu *PostUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (pu *PostUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (pu *PharmacistUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (pu *PetUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (pu *PetUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (o *FeatureRelationship) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (pfmu *ParticipantFlowModuleUpdate) Exec(ctx context.Context) error {\n\t_, err := pfmu.Save(ctx)\n\treturn err\n}", "func (q assetRevisionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (ruo *RevisionUpdateOne) Exec(ctx context.Context) error {\n\t_, err := ruo.Save(ctx)\n\treturn err\n}", "func (o *Failure) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (puo *PostUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (puo *PostUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (pu *ProductUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (pu *ProductUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (o *Jet) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Phenotypeprop) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (pu *PendingloanbindingUpdate) Exec(ctx context.Context) error {\n\t_, err := pu.Save(ctx)\n\treturn err\n}", "func (o *Stock) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (_m *Repository) Update(p *entity.Person, commitChan <-chan bool, doneChan chan<- bool) {\n\t_m.Called(p, commitChan, doneChan)\n}", "func (piu *ProviderIDUpdate) Exec(ctx context.Context) error {\n\t_, err := piu.Save(ctx)\n\treturn err\n}", "func (puo *PharmacistUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (o *RecordMeasure) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (upu *UnsavedPostUpdate) Exec(ctx context.Context) error {\n\t_, err := upu.Save(ctx)\n\treturn err\n}", "func (o *AssetRevision) SetAssetP(exec boil.Executor, insert bool, related *Asset) {\n\tif err := o.SetAsset(exec, insert, related); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (_Smartchef *SmartchefTransactor) UpdatePool(opts *bind.TransactOpts, _pid *big.Int) (*types.Transaction, error) {\n\treturn _Smartchef.contract.Transact(opts, \"updatePool\", _pid)\n}", "func (o *AssetRevision) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tassetRevisionUpsertCacheMut.RLock()\n\tcache, cached := assetRevisionUpsertCache[key]\n\tassetRevisionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"public: unable to upsert asset_revision, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(assetRevisionPrimaryKeyColumns))\n\t\t\tcopy(conflict, assetRevisionPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"asset_revision\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to upsert asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionUpsertCacheMut.Lock()\n\t\tassetRevisionUpsertCache[key] = cache\n\t\tassetRevisionUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (p *MockDeployPlugin) Update(cluster *lang.Cluster, deployName string, params util.NestedParameterMap, eventLog *event.Log) error {\n\ttime.Sleep(p.SleepTime)\n\treturn nil\n}", "func (puo *PetUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (puo *PetUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (o *BraceletPhoto) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (au *AdminUpdate) Exec(ctx context.Context) error {\n\t_, err := au.Save(ctx)\n\treturn err\n}", "func (t *Table) ExecuteUpdate(u *UpdateCommand) {\n\tu.real.execute(t, u.allocated[:u.allocSplit])\n\tu.virt.execute(t, u.allocated[u.allocSplit:])\n\tu.clear()\n}", "func AssetRevisionExistsP(exec boil.Executor, orig int) bool {\n\te, err := AssetRevisionExists(exec, orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (gau *GithubAssetUpdate) Exec(ctx context.Context) error {\n\t_, err := gau.Save(ctx)\n\treturn err\n}", "func (up *Updater) UpdateTo(rel *Release, cmdPath string) error {\n\tvar client http.Client\n\tsrc, redirectURL, err := up.api.Repositories.DownloadReleaseAsset(up.apiCtx, rel.RepoOwner, rel.RepoName, rel.AssetID, &client)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to call GitHub Releases API for getting an asset(ID: %d) for repository '%s/%s': %s\", rel.AssetID, rel.RepoOwner, rel.RepoName, err)\n\t}\n\tif redirectURL != \"\" {\n\t\tlog.Println(\"Redirect URL was returned while trying to download a release asset from GitHub API. Falling back to downloading from asset URL directly:\", redirectURL)\n\t\tsrc, err = up.downloadDirectlyFromURL(redirectURL)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tdefer src.Close()\n\n\tdata, err := ioutil.ReadAll(src)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed reading asset body: %v\", err)\n\t}\n\n\tif up.validator == nil {\n\t\treturn uncompressAndUpdate(bytes.NewReader(data), rel.AssetURL, cmdPath)\n\t}\n\n\tvalidationSrc, validationRedirectURL, err := up.api.Repositories.DownloadReleaseAsset(up.apiCtx, rel.RepoOwner, rel.RepoName, rel.ValidationAssetID, &client)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to call GitHub Releases API for getting an validation asset(ID: %d) for repository '%s/%s': %s\", rel.ValidationAssetID, rel.RepoOwner, rel.RepoName, err)\n\t}\n\tif validationRedirectURL != \"\" {\n\t\tlog.Println(\"Redirect URL was returned while trying to download a release validation asset from GitHub API. Falling back to downloading from asset URL directly:\", redirectURL)\n\t\tvalidationSrc, err = up.downloadDirectlyFromURL(validationRedirectURL)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tdefer validationSrc.Close()\n\n\tvalidationData, err := ioutil.ReadAll(validationSrc)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed reading validation asset body: %v\", err)\n\t}\n\n\tif err := up.validator.Validate(data, validationData); err != nil {\n\t\treturn fmt.Errorf(\"Failed validating asset content: %v\", err)\n\t}\n\n\treturn uncompressAndUpdate(bytes.NewReader(data), rel.AssetURL, cmdPath)\n}", "func (c *updateCmd) Run(_ *buildChild) error {\n\treturn nil\n}", "func (o *Comment) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (pfmuo *ParticipantFlowModuleUpdateOne) Exec(ctx context.Context) error {\n\t_, err := pfmuo.Save(ctx)\n\treturn err\n}", "func (plugin *ipamPlugin) Update(args *cniSkel.CmdArgs) error {\n\treturn nil\n}", "func (o *AuthUserUserPermission) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (puo *ProductUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (puo *ProductUpdateOne) Exec(ctx context.Context) error {\n\t_, err := puo.Save(ctx)\n\treturn err\n}", "func (pm *basePackageManager) Update() error {\n\t_, _, err := RunCommandWithRetry(pm.cmder.UpdateCmd(), nil)\n\treturn err\n}", "func (upd *Update) Execute(vcursor VCursor, bindVars map[string]*querypb.BindVariable, wantfields bool) (*sqltypes.Result, error) {\n\tif upd.QueryTimeout != 0 {\n\t\tcancel := vcursor.SetContextTimeout(time.Duration(upd.QueryTimeout) * time.Millisecond)\n\t\tdefer cancel()\n\t}\n\n\tswitch upd.Opcode {\n\tcase Unsharded:\n\t\treturn upd.execUpdateUnsharded(vcursor, bindVars)\n\tcase Equal:\n\t\treturn upd.execUpdateEqual(vcursor, bindVars)\n\tcase In:\n\t\treturn upd.execUpdateIn(vcursor, bindVars)\n\tcase Scatter:\n\t\treturn upd.execUpdateByDestination(vcursor, bindVars, key.DestinationAllShards{})\n\tcase ByDestination:\n\t\treturn upd.execUpdateByDestination(vcursor, bindVars, upd.TargetDestination)\n\tdefault:\n\t\t// Unreachable.\n\t\treturn nil, fmt.Errorf(\"unsupported opcode: %v\", upd)\n\t}\n}", "func (o *Source) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (iu *ItemUpdate) Exec(ctx context.Context) error {\n\t_, err := iu.Save(ctx)\n\treturn err\n}", "func (o VoteSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *VSP) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tvspUpdateCacheMut.RLock()\n\tcache, cached := vspUpdateCache[key]\n\tvspUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update vsp, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"vsp\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, vspPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(vspType, vspMapping, append(wl, vspPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update vsp row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for vsp\")\n\t}\n\n\tif !cached {\n\t\tvspUpdateCacheMut.Lock()\n\t\tvspUpdateCache[key] = cache\n\t\tvspUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Author) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (r *pgRepository) Update(ctx context.Context, tenant string, item *model.APIDefinition) error {\n\tif item == nil {\n\t\treturn apperrors.NewInternalError(\"item cannot be nil\")\n\t}\n\n\tentity := r.conv.ToEntity(item)\n\n\treturn r.updater.UpdateSingle(ctx, resource.API, tenant, entity)\n}", "func (a *paymentUsecase) Update(c context.Context, ar *models.Payment) (*models.Payment, error) {\n\tctx, cancel := context.WithTimeout(c, a.contextTimeout)\n\tdefer cancel()\n\n\tar.UpdatedAt = time.Now()\n\treturn a.repo.Update(ctx, ar)\n}", "func (o *Picture) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (up *updater) Update() UpdateProgress {\n\tupdatePlan := up.updatePlan\n\ttoCreateJobs := updatePlan.GetToCreateJobs()\n\tjobs := make([]*batchv1.Job, len(toCreateJobs))\n\tfor index, job := range toCreateJobs {\n\t\tjobs[index] = &job\n\t}\n\ttoApplyDeployments := updatePlan.GetToApplyDeployments()\n\tdeployments := make([]*v1.Deployment, len(toApplyDeployments))\n\tfor index, deployment := range toApplyDeployments {\n\t\tdeployments[index] = &deployment\n\t}\n\n\tupdateProgress := &updateProgressConfiguration{\n\t\tjobs: jobs,\n\t\tdeployments: deployments,\n\t\tfailed: false,\n\t}\n\tup.updateProgress = updateProgress\n\tgo up.runUpdate()\n\treturn updateProgress\n}", "func (lu *LibraryUpdate) Exec(ctx context.Context) error {\n\t_, err := lu.Save(ctx)\n\treturn err\n}", "func (o *Rental) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (s *permisoService) Update(id string, p *model.Permiso) error {\n\n\t// p.UpdatedAt = time.Now()\n\n\t// return s.service.Update(id, p)\n\n\treturn nil\n}" ]
[ "0.6953268", "0.6953268", "0.6867607", "0.6802554", "0.67360014", "0.67205167", "0.66652924", "0.6634668", "0.6610693", "0.659195", "0.65255624", "0.6499854", "0.63843244", "0.63363504", "0.6268577", "0.625415", "0.62433475", "0.6228706", "0.6219512", "0.6211925", "0.6168595", "0.6135913", "0.61087614", "0.6079751", "0.60609114", "0.60178417", "0.6000936", "0.59923214", "0.59861386", "0.594356", "0.5896243", "0.5879821", "0.5794369", "0.5781947", "0.5695911", "0.56887984", "0.55883116", "0.55883116", "0.5535773", "0.5478805", "0.5478805", "0.546499", "0.54563105", "0.54378617", "0.54335546", "0.5401946", "0.5384659", "0.53617394", "0.533662", "0.53298736", "0.5320338", "0.5320338", "0.5316089", "0.5316089", "0.5312221", "0.52984905", "0.52918106", "0.52464586", "0.5239035", "0.52231205", "0.5220522", "0.5220497", "0.5207614", "0.5207614", "0.5195045", "0.5183914", "0.51779866", "0.51767373", "0.51714236", "0.51623243", "0.51623243", "0.51616794", "0.5141356", "0.51334435", "0.51184076", "0.51165617", "0.511483", "0.5109652", "0.51085556", "0.51030886", "0.5091599", "0.5083243", "0.50824803", "0.5077973", "0.50767046", "0.50767046", "0.50725603", "0.5070337", "0.5049867", "0.5041313", "0.50349474", "0.5030064", "0.50271434", "0.5024545", "0.50110734", "0.50096345", "0.5009549", "0.50061816", "0.5005693", "0.49924955" ]
0.7700634
0
Update uses an executor to update the AssetRevision. Whitelist behavior: If a whitelist is provided, only the columns given are updated. No whitelist behavior: Without a whitelist, columns are inferred by the following rules: All columns are inferred to start with All primary keys are subtracted from this set Update does not automatically update the record in case of default values. Use .Reload() to refresh the records.
func (o *AssetRevision) Update(exec boil.Executor, whitelist ...string) error { var err error if err = o.doBeforeUpdateHooks(exec); err != nil { return err } key := makeCacheKey(whitelist, nil) assetRevisionUpdateCacheMut.RLock() cache, cached := assetRevisionUpdateCache[key] assetRevisionUpdateCacheMut.RUnlock() if !cached { wl := strmangle.UpdateColumnSet(assetRevisionColumns, assetRevisionPrimaryKeyColumns, whitelist) if len(whitelist) == 0 { wl = strmangle.SetComplement(wl, []string{"created_at"}) } if len(wl) == 0 { return errors.New("public: unable to update asset_revision, could not build whitelist") } cache.query = fmt.Sprintf("UPDATE \"asset_revision\" SET %s WHERE %s", strmangle.SetParamNames("\"", "\"", 1, wl), strmangle.WhereClause("\"", "\"", len(wl)+1, assetRevisionPrimaryKeyColumns), ) cache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, append(wl, assetRevisionPrimaryKeyColumns...)) if err != nil { return err } } values := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping) if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, cache.query) fmt.Fprintln(boil.DebugWriter, values) } _, err = exec.Exec(cache.query, values...) if err != nil { return errors.Wrap(err, "public: unable to update asset_revision row") } if !cached { assetRevisionUpdateCacheMut.Lock() assetRevisionUpdateCache[key] = cache assetRevisionUpdateCacheMut.Unlock() } return o.doAfterUpdateHooks(exec) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Author) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tauthorUpdateCacheMut.RLock()\n\tcache, cached := authorUpdateCache[key]\n\tauthorUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to update authors, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"authors\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, authorPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(authorType, authorMapping, append(wl, authorPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update authors row\")\n\t}\n\n\tif !cached {\n\t\tauthorUpdateCacheMut.Lock()\n\t\tauthorUpdateCache[key] = cache\n\t\tauthorUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\ttransactionUpdateCacheMut.RLock()\n\tcache, cached := transactionUpdateCache[key]\n\ttransactionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"model: unable to update transaction, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `transaction` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, transactionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, append(wl, transactionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to update transaction row\")\n\t}\n\n\tif !cached {\n\t\ttransactionUpdateCacheMut.Lock()\n\t\ttransactionUpdateCache[key] = cache\n\t\ttransactionUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\ttransactionUpdateCacheMut.RLock()\n\tcache, cached := transactionUpdateCache[key]\n\ttransactionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update transactions, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `transactions` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, transactionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, append(wl, transactionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update transactions row\")\n\t}\n\n\tif !cached {\n\t\ttransactionUpdateCacheMut.Lock()\n\t\ttransactionUpdateCache[key] = cache\n\t\ttransactionUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Inventory) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tinventoryUpdateCacheMut.RLock()\n\tcache, cached := inventoryUpdateCache[key]\n\tinventoryUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update inventory, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `inventory` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, inventoryPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(inventoryType, inventoryMapping, append(wl, inventoryPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update inventory row\")\n\t}\n\n\tif !cached {\n\t\tinventoryUpdateCacheMut.Lock()\n\t\tinventoryUpdateCache[key] = cache\n\t\tinventoryUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Rental) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\trentalUpdateCacheMut.RLock()\n\tcache, cached := rentalUpdateCache[key]\n\trentalUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\trentalColumns,\n\t\t\trentalPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"sqlboiler: unable to update rental, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `rental` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, rentalPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(rentalType, rentalMapping, append(wl, rentalPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update rental row\")\n\t}\n\n\tif !cached {\n\t\trentalUpdateCacheMut.Lock()\n\t\trentalUpdateCache[key] = cache\n\t\trentalUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Stock) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tstockUpdateCacheMut.RLock()\n\tcache, cached := stockUpdateCache[key]\n\tstockUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tstockColumns,\n\t\t\tstockPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update stock, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `stock` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, stockPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(stockType, stockMapping, append(wl, stockPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update stock row\")\n\t}\n\n\tif !cached {\n\t\tstockUpdateCacheMut.Lock()\n\t\tstockUpdateCache[key] = cache\n\t\tstockUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Source) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tsourceUpdateCacheMut.RLock()\n\tcache, cached := sourceUpdateCache[key]\n\tsourceUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to update sources, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"sources\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, sourcePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, append(wl, sourcePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update sources row\")\n\t}\n\n\tif !cached {\n\t\tsourceUpdateCacheMut.Lock()\n\t\tsourceUpdateCache[key] = cache\n\t\tsourceUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *AssetRevision) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tassetRevisionUpsertCacheMut.RLock()\n\tcache, cached := assetRevisionUpsertCache[key]\n\tassetRevisionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"public: unable to upsert asset_revision, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(assetRevisionPrimaryKeyColumns))\n\t\t\tcopy(conflict, assetRevisionPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"asset_revision\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to upsert asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionUpsertCacheMut.Lock()\n\t\tassetRevisionUpsertCache[key] = cache\n\t\tassetRevisionUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o AssetRevisionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"public: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), assetRevisionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tquery := fmt.Sprintf(\n\t\t\"UPDATE \\\"asset_revision\\\" SET %s WHERE (\\\"orig\\\") IN (%s)\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(assetRevisionPrimaryKeyColumns), len(colNames)+1, len(assetRevisionPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to update all in assetRevision slice\")\n\t}\n\n\treturn nil\n}", "func (o *Jet) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tjetUpdateCacheMut.RLock()\n\tcache, cached := jetUpdateCache[key]\n\tjetUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update jets, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `jets` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, jetPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(jetType, jetMapping, append(wl, jetPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update jets row\")\n\t}\n\n\tif !cached {\n\t\tjetUpdateCacheMut.Lock()\n\t\tjetUpdateCache[key] = cache\n\t\tjetUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Vote) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tvoteUpdateCacheMut.RLock()\n\tcache, cached := voteUpdateCache[key]\n\tvoteUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update vote, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, votePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, append(wl, votePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update vote row\")\n\t}\n\n\tif !cached {\n\t\tvoteUpdateCacheMut.Lock()\n\t\tvoteUpdateCache[key] = cache\n\t\tvoteUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *RawVisit) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\trawVisitUpdateCacheMut.RLock()\n\tcache, cached := rawVisitUpdateCache[key]\n\trawVisitUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update raw_visits, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"raw_visits\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, rawVisitPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, append(wl, rawVisitPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update raw_visits row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitUpdateCacheMut.Lock()\n\t\trawVisitUpdateCache[key] = cache\n\t\trawVisitUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *AssetRevision) UpdateP(exec boil.Executor, whitelist ...string) {\n\terr := o.Update(exec, whitelist...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *HoldenAt) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tholdenAtUpdateCacheMut.RLock()\n\tcache, cached := holdenAtUpdateCache[key]\n\tholdenAtUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update HoldenAt, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"HoldenAt\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, holdenAtPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, append(wl, holdenAtPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update HoldenAt row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for HoldenAt\")\n\t}\n\n\tif !cached {\n\t\tholdenAtUpdateCacheMut.Lock()\n\t\tholdenAtUpdateCache[key] = cache\n\t\tholdenAtUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Offer) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tofferUpdateCacheMut.RLock()\n\tcache, cached := offerUpdateCache[key]\n\tofferUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"stellarcore: unable to update offers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"offers\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, offerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(offerType, offerMapping, append(wl, offerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update offers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by update for offers\")\n\t}\n\n\tif !cached {\n\t\tofferUpdateCacheMut.Lock()\n\t\tofferUpdateCache[key] = cache\n\t\tofferUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *Block) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tblockUpdateCacheMut.RLock()\n\tcache, cached := blockUpdateCache[key]\n\tblockUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update block, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"block\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, blockPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(blockType, blockMapping, append(wl, blockPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update block row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for block\")\n\t}\n\n\tif !cached {\n\t\tblockUpdateCacheMut.Lock()\n\t\tblockUpdateCache[key] = cache\n\t\tblockUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Vendor) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tvendorUpdateCacheMut.RLock()\n\tcache, cached := vendorUpdateCache[key]\n\tvendorUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tvendorColumns,\n\t\t\tvendorPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update vendors, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `vendors` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, vendorPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(vendorType, vendorMapping, append(wl, vendorPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update vendors row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for vendors\")\n\t}\n\n\tif !cached {\n\t\tvendorUpdateCacheMut.Lock()\n\t\tvendorUpdateCache[key] = cache\n\t\tvendorUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *BraceletPhoto) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tbraceletPhotoUpdateCacheMut.RLock()\n\tcache, cached := braceletPhotoUpdateCache[key]\n\tbraceletPhotoUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(braceletPhotoColumns, braceletPhotoPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update bracelet_photo, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `bracelet_photo` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, braceletPhotoPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(braceletPhotoType, braceletPhotoMapping, append(wl, braceletPhotoPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update bracelet_photo row\")\n\t}\n\n\tif !cached {\n\t\tbraceletPhotoUpdateCacheMut.Lock()\n\t\tbraceletPhotoUpdateCache[key] = cache\n\t\tbraceletPhotoUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Shelf) Update(exec boil.Executor, whitelist ...string) error {\n\to.whitelist = whitelist\n\twhitelist = o.Whitelist()\n\n\to.operation = \"UPDATE\"\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tshelfUpdateCacheMut.RLock()\n\tcache, cached := shelfUpdateCache[key]\n\tshelfUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(shelfColumns, shelfPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update shelf, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `shelf` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, shelfPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(shelfType, shelfMapping, append(wl, shelfPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update shelf row\")\n\t}\n\n\tif !cached {\n\t\tshelfUpdateCacheMut.Lock()\n\t\tshelfUpdateCache[key] = cache\n\t\tshelfUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *RSSAnnouncement) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\trssAnnouncementUpdateCacheMut.RLock()\n\tcache, cached := rssAnnouncementUpdateCache[key]\n\trssAnnouncementUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trssAnnouncementAllColumns,\n\t\t\trssAnnouncementPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update rss_announcements, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"rss_announcements\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, rssAnnouncementPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(rssAnnouncementType, rssAnnouncementMapping, append(wl, rssAnnouncementPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update rss_announcements row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for rss_announcements\")\n\t}\n\n\tif !cached {\n\t\trssAnnouncementUpdateCacheMut.Lock()\n\t\trssAnnouncementUpdateCache[key] = cache\n\t\trssAnnouncementUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Comment) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tcommentUpdateCacheMut.RLock()\n\tcache, cached := commentUpdateCache[key]\n\tcommentUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tcommentColumns,\n\t\t\tcommentPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update comment, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `comment` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, commentPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(commentType, commentMapping, append(wl, commentPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update comment row\")\n\t}\n\n\tif !cached {\n\t\tcommentUpdateCacheMut.Lock()\n\t\tcommentUpdateCache[key] = cache\n\t\tcommentUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Tenant) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\ttenantUpdateCacheMut.RLock()\n\tcache, cached := tenantUpdateCache[key]\n\ttenantUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\ttenantAllColumns,\n\t\t\ttenantPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"dbmodel: unable to update tenants, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `tenants` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, tenantPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(tenantType, tenantMapping, append(wl, tenantPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update tenants row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by update for tenants\")\n\t}\n\n\tif !cached {\n\t\ttenantUpdateCacheMut.Lock()\n\t\ttenantUpdateCache[key] = cache\n\t\ttenantUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *PremiumSlot) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tpremiumSlotUpdateCacheMut.RLock()\n\tcache, cached := premiumSlotUpdateCache[key]\n\tpremiumSlotUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update premium_slots, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"premium_slots\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, premiumSlotPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, append(wl, premiumSlotPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update premium_slots row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for premium_slots\")\n\t}\n\n\tif !cached {\n\t\tpremiumSlotUpdateCacheMut.Lock()\n\t\tpremiumSlotUpdateCache[key] = cache\n\t\tpremiumSlotUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Repository) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\trepositoryUpdateCacheMut.RLock()\n\tcache, cached := repositoryUpdateCache[key]\n\trepositoryUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trepositoryAllColumns,\n\t\t\trepositoryPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update repositories, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `repositories` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, repositoryPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(repositoryType, repositoryMapping, append(wl, repositoryPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update repositories row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for repositories\")\n\t}\n\n\tif !cached {\n\t\trepositoryUpdateCacheMut.Lock()\n\t\trepositoryUpdateCache[key] = cache\n\t\trepositoryUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Address) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\taddressUpdateCacheMut.RLock()\n\tcache, cached := addressUpdateCache[key]\n\taddressUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\taddressColumns,\n\t\t\taddressPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"sqlboiler: unable to update address, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `address` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, addressPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(addressType, addressMapping, append(wl, addressPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update address row\")\n\t}\n\n\tif !cached {\n\t\taddressUpdateCacheMut.Lock()\n\t\taddressUpdateCache[key] = cache\n\t\taddressUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Skin) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tskinUpdateCacheMut.RLock()\n\tcache, cached := skinUpdateCache[key]\n\tskinUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tskinColumns,\n\t\t\tskinPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update skin, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `skin` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, skinPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(skinType, skinMapping, append(wl, skinPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update skin row\")\n\t}\n\n\tif !cached {\n\t\tskinUpdateCacheMut.Lock()\n\t\tskinUpdateCache[key] = cache\n\t\tskinUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *AssetRevision) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tassetRevisionInsertCacheMut.RLock()\n\tcache, cached := assetRevisionInsertCache[key]\n\tassetRevisionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"asset_revision\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"asset_revision\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to insert into asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionInsertCacheMut.Lock()\n\t\tassetRevisionInsertCache[key] = cache\n\t\tassetRevisionInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func (o *Currency) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tcurrencyUpdateCacheMut.RLock()\n\tcache, cached := currencyUpdateCache[key]\n\tcurrencyUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update currency, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"currency\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, currencyPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(currencyType, currencyMapping, append(wl, currencyPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update currency row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for currency\")\n\t}\n\n\tif !cached {\n\t\tcurrencyUpdateCacheMut.Lock()\n\t\tcurrencyUpdateCache[key] = cache\n\t\tcurrencyUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *Weather) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tweatherUpdateCacheMut.RLock()\n\tcache, cached := weatherUpdateCache[key]\n\tweatherUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"db: unable to update weather, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"prh\\\".\\\"weather\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, weatherPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(weatherType, weatherMapping, append(wl, weatherPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update weather row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by update for weather\")\n\t}\n\n\tif !cached {\n\t\tweatherUpdateCacheMut.Lock()\n\t\tweatherUpdateCache[key] = cache\n\t\tweatherUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *RestaurantRank) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\trestaurantRankUpdateCacheMut.RLock()\n\tcache, cached := restaurantRankUpdateCache[key]\n\trestaurantRankUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trestaurantRankAllColumns,\n\t\t\trestaurantRankPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update restaurant_rank, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"restaurant_rank\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, restaurantRankPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(restaurantRankType, restaurantRankMapping, append(wl, restaurantRankPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update restaurant_rank row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for restaurant_rank\")\n\t}\n\n\tif !cached {\n\t\trestaurantRankUpdateCacheMut.Lock()\n\t\trestaurantRankUpdateCache[key] = cache\n\t\trestaurantRankUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *InstrumentClass) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tinstrumentClassUpdateCacheMut.RLock()\n\tcache, cached := instrumentClassUpdateCache[key]\n\tinstrumentClassUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update instrument_class, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"instruments\\\".\\\"instrument_class\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, instrumentClassPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, append(wl, instrumentClassPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update instrument_class row\")\n\t}\n\n\tif !cached {\n\t\tinstrumentClassUpdateCacheMut.Lock()\n\t\tinstrumentClassUpdateCache[key] = cache\n\t\tinstrumentClassUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Task) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\ttaskUpdateCacheMut.RLock()\n\tcache, cached := taskUpdateCache[key]\n\ttaskUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update tasks, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"tasks\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, taskPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, append(wl, taskPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update tasks row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskUpdateCacheMut.Lock()\n\t\ttaskUpdateCache[key] = cache\n\t\ttaskUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Description) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tdescriptionUpdateCacheMut.RLock()\n\tcache, cached := descriptionUpdateCache[key]\n\tdescriptionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tdescriptionColumns,\n\t\t\tdescriptionPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update descriptions, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `descriptions` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, descriptionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(descriptionType, descriptionMapping, append(wl, descriptionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update descriptions row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for descriptions\")\n\t}\n\n\tif !cached {\n\t\tdescriptionUpdateCacheMut.Lock()\n\t\tdescriptionUpdateCache[key] = cache\n\t\tdescriptionUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *PremiumCode) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tpremiumCodeUpdateCacheMut.RLock()\n\tcache, cached := premiumCodeUpdateCache[key]\n\tpremiumCodeUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpremiumCodeAllColumns,\n\t\t\tpremiumCodePrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update premium_codes, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"premium_codes\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, premiumCodePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(premiumCodeType, premiumCodeMapping, append(wl, premiumCodePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update premium_codes row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for premium_codes\")\n\t}\n\n\tif !cached {\n\t\tpremiumCodeUpdateCacheMut.Lock()\n\t\tpremiumCodeUpdateCache[key] = cache\n\t\tpremiumCodeUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Origin) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\toriginUpdateCacheMut.RLock()\n\tcache, cached := originUpdateCache[key]\n\toriginUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\toriginColumns,\n\t\t\toriginPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update origins, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"origins\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, originPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(originType, originMapping, append(wl, originPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update origins row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for origins\")\n\t}\n\n\tif !cached {\n\t\toriginUpdateCacheMut.Lock()\n\t\toriginUpdateCache[key] = cache\n\t\toriginUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Post) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpostUpdateCacheMut.RLock()\n\tcache, cached := postUpdateCache[key]\n\tpostUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpostColumns,\n\t\t\tpostPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"orm: unable to update posts, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"posts\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, postPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(postType, postMapping, append(wl, postPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to update posts row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: failed to get rows affected by update for posts\")\n\t}\n\n\tif !cached {\n\t\tpostUpdateCacheMut.Lock()\n\t\tpostUpdateCache[key] = cache\n\t\tpostUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *ForeignLegalResource) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tforeignLegalResourceUpdateCacheMut.RLock()\n\tcache, cached := foreignLegalResourceUpdateCache[key]\n\tforeignLegalResourceUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tforeignLegalResourceAllColumns,\n\t\t\tforeignLegalResourcePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update ForeignLegalResources, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"ForeignLegalResources\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, foreignLegalResourcePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(foreignLegalResourceType, foreignLegalResourceMapping, append(wl, foreignLegalResourcePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update ForeignLegalResources row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for ForeignLegalResources\")\n\t}\n\n\tif !cached {\n\t\tforeignLegalResourceUpdateCacheMut.Lock()\n\t\tforeignLegalResourceUpdateCache[key] = cache\n\t\tforeignLegalResourceUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\treturn Model.Data(r).Where(gdb.GetWhereConditionOfStruct(r)).Update()\n}", "func (o *ClaimInList) Update(exec boil.Executor, columns boil.Columns) error {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tclaimInListUpdateCacheMut.RLock()\n\tcache, cached := claimInListUpdateCache[key]\n\tclaimInListUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tclaimInListAllColumns,\n\t\t\tclaimInListPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"model: unable to update claim_in_list, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `claim_in_list` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, claimInListPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(claimInListType, claimInListMapping, append(wl, claimInListPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to update claim_in_list row\")\n\t}\n\n\tif !cached {\n\t\tclaimInListUpdateCacheMut.Lock()\n\t\tclaimInListUpdateCache[key] = cache\n\t\tclaimInListUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *VSP) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tvspUpdateCacheMut.RLock()\n\tcache, cached := vspUpdateCache[key]\n\tvspUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update vsp, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"vsp\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, vspPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(vspType, vspMapping, append(wl, vspPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update vsp row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for vsp\")\n\t}\n\n\tif !cached {\n\t\tvspUpdateCacheMut.Lock()\n\t\tvspUpdateCache[key] = cache\n\t\tvspUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *PaymentObject) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpaymentObjectUpdateCacheMut.RLock()\n\tcache, cached := paymentObjectUpdateCache[key]\n\tpaymentObjectUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update payment_objects, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `payment_objects` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, paymentObjectPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, append(wl, paymentObjectPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update payment_objects row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for payment_objects\")\n\t}\n\n\tif !cached {\n\t\tpaymentObjectUpdateCacheMut.Lock()\n\t\tpaymentObjectUpdateCache[key] = cache\n\t\tpaymentObjectUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *RecipeAdditive) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\trecipeAdditiveUpdateCacheMut.RLock()\n\tcache, cached := recipeAdditiveUpdateCache[key]\n\trecipeAdditiveUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trecipeAdditiveAllColumns,\n\t\t\trecipeAdditivePrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update recipe_additive, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, recipeAdditivePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(recipeAdditiveType, recipeAdditiveMapping, append(wl, recipeAdditivePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update recipe_additive row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for recipe_additive\")\n\t}\n\n\tif !cached {\n\t\trecipeAdditiveUpdateCacheMut.Lock()\n\t\trecipeAdditiveUpdateCache[key] = cache\n\t\trecipeAdditiveUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Failure) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tfailureUpdateCacheMut.RLock()\n\tcache, cached := failureUpdateCache[key]\n\tfailureUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailurePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update failure, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `failure` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, failurePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(failureType, failureMapping, append(wl, failurePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update failure row\")\n\t}\n\n\tif !cached {\n\t\tfailureUpdateCacheMut.Lock()\n\t\tfailureUpdateCache[key] = cache\n\t\tfailureUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *APIKey) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tapiKeyUpdateCacheMut.RLock()\n\tcache, cached := apiKeyUpdateCache[key]\n\tapiKeyUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update api_keys, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"api_keys\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, apiKeyPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, append(wl, apiKeyPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update api_keys row\")\n\t}\n\n\tif !cached {\n\t\tapiKeyUpdateCacheMut.Lock()\n\t\tapiKeyUpdateCache[key] = cache\n\t\tapiKeyUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Organism) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\torganismUpdateCacheMut.RLock()\n\tcache, cached := organismUpdateCache[key]\n\torganismUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(organismColumns, organismPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"chado: unable to update organism, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"organism\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, organismPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(organismType, organismMapping, append(wl, organismPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update organism row\")\n\t}\n\n\tif !cached {\n\t\torganismUpdateCacheMut.Lock()\n\t\torganismUpdateCache[key] = cache\n\t\torganismUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Source) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tsourceUpdateCacheMut.RLock()\n\tcache, cached := sourceUpdateCache[key]\n\tsourceUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tsourceAllColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t)\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"mdbmodels: unable to update sources, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"sources\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, sourcePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, append(wl, sourcePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to update sources row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by update for sources\")\n\t}\n\n\tif !cached {\n\t\tsourceUpdateCacheMut.Lock()\n\t\tsourceUpdateCache[key] = cache\n\t\tsourceUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *ExchangeCurrency) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\texchangeCurrencyUpdateCacheMut.RLock()\n\tcache, cached := exchangeCurrencyUpdateCache[key]\n\texchangeCurrencyUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\texchangeCurrencyColumns,\n\t\t\texchangeCurrencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update exchange_currency, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"exchange_currency\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, exchangeCurrencyPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(exchangeCurrencyType, exchangeCurrencyMapping, append(wl, exchangeCurrencyPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update exchange_currency row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for exchange_currency\")\n\t}\n\n\tif !cached {\n\t\texchangeCurrencyUpdateCacheMut.Lock()\n\t\texchangeCurrencyUpdateCache[key] = cache\n\t\texchangeCurrencyUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *BlackCard) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tblackCardUpdateCacheMut.RLock()\n\tcache, cached := blackCardUpdateCache[key]\n\tblackCardUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tblackCardColumns,\n\t\t\tblackCardPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update black_cards, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"black_cards\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, blackCardPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(blackCardType, blackCardMapping, append(wl, blackCardPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update black_cards row\")\n\t}\n\n\tif !cached {\n\t\tblackCardUpdateCacheMut.Lock()\n\t\tblackCardUpdateCache[key] = cache\n\t\tblackCardUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Vote) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tvoteUpdateCacheMut.RLock()\n\tcache, cached := voteUpdateCache[key]\n\tvoteUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update vote, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"vote\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, votePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, append(wl, votePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update vote row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for vote\")\n\t}\n\n\tif !cached {\n\t\tvoteUpdateCacheMut.Lock()\n\t\tvoteUpdateCache[key] = cache\n\t\tvoteUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *MempoolBin) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tmempoolBinUpdateCacheMut.RLock()\n\tcache, cached := mempoolBinUpdateCache[key]\n\tmempoolBinUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tmempoolBinAllColumns,\n\t\t\tmempoolBinPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update mempool_bin, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"mempool_bin\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, mempoolBinPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(mempoolBinType, mempoolBinMapping, append(wl, mempoolBinPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update mempool_bin row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for mempool_bin\")\n\t}\n\n\tif !cached {\n\t\tmempoolBinUpdateCacheMut.Lock()\n\t\tmempoolBinUpdateCache[key] = cache\n\t\tmempoolBinUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Cvtermsynonym) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tcvtermsynonymUpdateCacheMut.RLock()\n\tcache, cached := cvtermsynonymUpdateCache[key]\n\tcvtermsynonymUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(cvtermsynonymColumns, cvtermsynonymPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"chado: unable to update cvtermsynonym, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"cvtermsynonym\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, cvtermsynonymPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(cvtermsynonymType, cvtermsynonymMapping, append(wl, cvtermsynonymPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update cvtermsynonym row\")\n\t}\n\n\tif !cached {\n\t\tcvtermsynonymUpdateCacheMut.Lock()\n\t\tcvtermsynonymUpdateCache[key] = cache\n\t\tcvtermsynonymUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (q assetRevisionQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to update all for asset_revision\")\n\t}\n\n\treturn nil\n}", "func (o *Item) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\titemUpdateCacheMut.RLock()\n\tcache, cached := itemUpdateCache[key]\n\titemUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\titemAllColumns,\n\t\t\titemPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update items, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"items\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 0, itemPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(itemType, itemMapping, append(wl, itemPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update items row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for items\")\n\t}\n\n\tif !cached {\n\t\titemUpdateCacheMut.Lock()\n\t\titemUpdateCache[key] = cache\n\t\titemUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Project) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tprojectUpdateCacheMut.RLock()\n\tcache, cached := projectUpdateCache[key]\n\tprojectUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tprojectAllColumns,\n\t\t\tprojectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update project, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `project` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, projectPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(projectType, projectMapping, append(wl, projectPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update project row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for project\")\n\t}\n\n\tif !cached {\n\t\tprojectUpdateCacheMut.Lock()\n\t\tprojectUpdateCache[key] = cache\n\t\tprojectUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *StockCvterm) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tstockCvtermUpdateCacheMut.RLock()\n\tcache, cached := stockCvtermUpdateCache[key]\n\tstockCvtermUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(stockCvtermColumns, stockCvtermPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"chado: unable to update stock_cvterm, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"stock_cvterm\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, stockCvtermPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(stockCvtermType, stockCvtermMapping, append(wl, stockCvtermPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update stock_cvterm row\")\n\t}\n\n\tif !cached {\n\t\tstockCvtermUpdateCacheMut.Lock()\n\t\tstockCvtermUpdateCache[key] = cache\n\t\tstockCvtermUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Auth) Update(exec boil.Executor, whitelist ...string) error {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tauthUpdateCacheMut.RLock()\n\tcache, cached := authUpdateCache[key]\n\tauthUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthPrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update auths, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"auths\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, authPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(authType, authMapping, append(wl, authPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update auths row\")\n\t}\n\n\tif !cached {\n\t\tauthUpdateCacheMut.Lock()\n\t\tauthUpdateCache[key] = cache\n\t\tauthUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *Employee) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\temployeeUpdateCacheMut.RLock()\n\tcache, cached := employeeUpdateCache[key]\n\temployeeUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\temployeeAllColumns,\n\t\t\temployeePrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update employee, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"employee\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, employeePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(employeeType, employeeMapping, append(wl, employeePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update employee row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for employee\")\n\t}\n\n\tif !cached {\n\t\temployeeUpdateCacheMut.Lock()\n\t\temployeeUpdateCache[key] = cache\n\t\temployeeUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *RentalRower) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\trentalRowerUpdateCacheMut.RLock()\n\tcache, cached := rentalRowerUpdateCache[key]\n\trentalRowerUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\trentalRowerColumns,\n\t\t\trentalRowerPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update rental_rowers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"rental_rowers\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, rentalRowerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(rentalRowerType, rentalRowerMapping, append(wl, rentalRowerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update rental_rowers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for rental_rowers\")\n\t}\n\n\tif !cached {\n\t\trentalRowerUpdateCacheMut.Lock()\n\t\trentalRowerUpdateCache[key] = cache\n\t\trentalRowerUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *Building) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tbuildingUpdateCacheMut.RLock()\n\tcache, cached := buildingUpdateCache[key]\n\tbuildingUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tbuildingColumns,\n\t\t\tbuildingPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"record: unable to update buildings, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"buildings\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, buildingPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(buildingType, buildingMapping, append(wl, buildingPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to update buildings row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: failed to get rows affected by update for buildings\")\n\t}\n\n\tif !cached {\n\t\tbuildingUpdateCacheMut.Lock()\n\t\tbuildingUpdateCache[key] = cache\n\t\tbuildingUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Job) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tjobUpdateCacheMut.RLock()\n\tcache, cached := jobUpdateCache[key]\n\tjobUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update jobs, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"jobs\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, jobPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, append(wl, jobPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update jobs row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpdateCacheMut.Lock()\n\t\tjobUpdateCache[key] = cache\n\t\tjobUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *CMFBalanceChargeAdmin) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tcmfBalanceChargeAdminUpdateCacheMut.RLock()\n\tcache, cached := cmfBalanceChargeAdminUpdateCache[key]\n\tcmfBalanceChargeAdminUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tcmfBalanceChargeAdminAllColumns,\n\t\t\tcmfBalanceChargeAdminPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update cmf_balance_charge_admin, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `cmf_balance_charge_admin` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, cmfBalanceChargeAdminPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(cmfBalanceChargeAdminType, cmfBalanceChargeAdminMapping, append(wl, cmfBalanceChargeAdminPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update cmf_balance_charge_admin row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for cmf_balance_charge_admin\")\n\t}\n\n\tif !cached {\n\t\tcmfBalanceChargeAdminUpdateCacheMut.Lock()\n\t\tcmfBalanceChargeAdminUpdateCache[key] = cache\n\t\tcmfBalanceChargeAdminUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (r *Entity) Update() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Data(r).Where(where, args).Update()\n}", "func (o AssetRevisionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Project) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tprojectUpdateCacheMut.RLock()\n\tcache, cached := projectUpdateCache[key]\n\tprojectUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tprojectAllColumns,\n\t\t\tprojectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update projects, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `projects` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, projectPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(projectType, projectMapping, append(wl, projectPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update projects row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for projects\")\n\t}\n\n\tif !cached {\n\t\tprojectUpdateCacheMut.Lock()\n\t\tprojectUpdateCache[key] = cache\n\t\tprojectUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *ActivityLog) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tactivityLogUpdateCacheMut.RLock()\n\tcache, cached := activityLogUpdateCache[key]\n\tactivityLogUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tactivityLogAllColumns,\n\t\t\tactivityLogPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"dbmodel: unable to update activity_logs, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"activity_logs\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 0, activityLogPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(activityLogType, activityLogMapping, append(wl, activityLogPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update activity_logs row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by update for activity_logs\")\n\t}\n\n\tif !cached {\n\t\tactivityLogUpdateCacheMut.Lock()\n\t\tactivityLogUpdateCache[key] = cache\n\t\tactivityLogUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *DestinationRank) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tdestinationRankUpdateCacheMut.RLock()\n\tcache, cached := destinationRankUpdateCache[key]\n\tdestinationRankUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tdestinationRankAllColumns,\n\t\t\tdestinationRankPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update destination_rank, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"destination_rank\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, destinationRankPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(destinationRankType, destinationRankMapping, append(wl, destinationRankPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update destination_rank row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for destination_rank\")\n\t}\n\n\tif !cached {\n\t\tdestinationRankUpdateCacheMut.Lock()\n\t\tdestinationRankUpdateCache[key] = cache\n\t\tdestinationRankUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *AutomodRuleDatum) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tautomodRuleDatumUpdateCacheMut.RLock()\n\tcache, cached := automodRuleDatumUpdateCache[key]\n\tautomodRuleDatumUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tautomodRuleDatumAllColumns,\n\t\t\tautomodRuleDatumPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update automod_rule_data, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"automod_rule_data\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, automodRuleDatumPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(automodRuleDatumType, automodRuleDatumMapping, append(wl, automodRuleDatumPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update automod_rule_data row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for automod_rule_data\")\n\t}\n\n\tif !cached {\n\t\tautomodRuleDatumUpdateCacheMut.Lock()\n\t\tautomodRuleDatumUpdateCache[key] = cache\n\t\tautomodRuleDatumUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *AuthUserUserPermission) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tauthUserUserPermissionUpdateCacheMut.RLock()\n\tcache, cached := authUserUserPermissionUpdateCache[key]\n\tauthUserUserPermissionUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(authUserUserPermissionColumns, authUserUserPermissionPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update auth_user_user_permissions, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `auth_user_user_permissions` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, authUserUserPermissionPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(authUserUserPermissionType, authUserUserPermissionMapping, append(wl, authUserUserPermissionPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update auth_user_user_permissions row\")\n\t}\n\n\tif !cached {\n\t\tauthUserUserPermissionUpdateCacheMut.Lock()\n\t\tauthUserUserPermissionUpdateCache[key] = cache\n\t\tauthUserUserPermissionUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (ru *RevisionUpdate) Exec(ctx context.Context) error {\n\t_, err := ru.Save(ctx)\n\treturn err\n}", "func (o *WithdrawalCrypto) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\twithdrawalCryptoUpdateCacheMut.RLock()\n\tcache, cached := withdrawalCryptoUpdateCache[key]\n\twithdrawalCryptoUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\twithdrawalCryptoAllColumns,\n\t\t\twithdrawalCryptoPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"sqlite3: unable to update withdrawal_crypto, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"withdrawal_crypto\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 0, withdrawalCryptoPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(withdrawalCryptoType, withdrawalCryptoMapping, append(wl, withdrawalCryptoPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to update withdrawal_crypto row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: failed to get rows affected by update for withdrawal_crypto\")\n\t}\n\n\tif !cached {\n\t\twithdrawalCryptoUpdateCacheMut.Lock()\n\t\twithdrawalCryptoUpdateCache[key] = cache\n\t\twithdrawalCryptoUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *CMFTurntable) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tcmfTurntableUpdateCacheMut.RLock()\n\tcache, cached := cmfTurntableUpdateCache[key]\n\tcmfTurntableUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tcmfTurntableAllColumns,\n\t\t\tcmfTurntablePrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update cmf_turntable, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `cmf_turntable` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, cmfTurntablePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(cmfTurntableType, cmfTurntableMapping, append(wl, cmfTurntablePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update cmf_turntable row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for cmf_turntable\")\n\t}\n\n\tif !cached {\n\t\tcmfTurntableUpdateCacheMut.Lock()\n\t\tcmfTurntableUpdateCache[key] = cache\n\t\tcmfTurntableUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Channel) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tchannelUpdateCacheMut.RLock()\n\tcache, cached := channelUpdateCache[key]\n\tchannelUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tchannelAllColumns,\n\t\t\tchannelPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update channels, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"channels\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, channelPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(channelType, channelMapping, append(wl, channelPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update channels row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for channels\")\n\t}\n\n\tif !cached {\n\t\tchannelUpdateCacheMut.Lock()\n\t\tchannelUpdateCache[key] = cache\n\t\tchannelUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (repo *Repository) Update(ctx context.Context, claims auth.Claims, req UpdateRequest, now time.Time) error {\n\tspan, ctx := tracer.StartSpanFromContext(ctx, \"internal.expenditure.Update\")\n\tdefer span.Finish()\n\n\tif claims.Audience == \"\" {\n\t\treturn errors.WithStack(ErrForbidden)\n\t}\n\t// Admin users can update branches they have access to.\n\tif !claims.HasRole(auth.RoleAdmin) {\n\t\treturn errors.WithStack(ErrForbidden)\n\t}\n\n\t// Validate the request.\n\tv := webcontext.Validator()\n\terr := v.Struct(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcols := models.M{}\n\tif req.Amount != nil {\n\t\tcols[models.RepsExpenseColumns.Amount] = *req.Amount\n\t}\n\tif req.Reason != nil {\n\t\tcols[models.RepsExpenseColumns.Reason] = *req.Reason\n\t}\n\tif len(cols) == 0 {\n\t\treturn nil\n\t}\n\n\t// If now empty set it to the current time.\n\tif now.IsZero() {\n\t\tnow = time.Now()\n\t}\n\n\t// Always store the time as UTC.\n\tnow = now.UTC()\n\t// Postgres truncates times to milliseconds when storing. We and do the same\n\t// here so the value we return is consistent with what we store.\n\tnow = now.Truncate(time.Millisecond)\n\n\tcols[models.BranchColumns.UpdatedAt] = now\n\n\t_, err = models.RepsExpenses(models.RepsExpenseWhere.ID.EQ(req.ID)).UpdateAll(ctx, repo.DbConn, cols)\n\n\treturn nil\n}", "func (o BlockSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"block\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, blockPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in block slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all block\")\n\t}\n\treturn rowsAff, nil\n}", "func (o *Ticket) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tticketUpdateCacheMut.RLock()\n\tcache, cached := ticketUpdateCache[key]\n\tticketUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update tickets, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"tickets\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, ticketPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(ticketType, ticketMapping, append(wl, ticketPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update tickets row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for tickets\")\n\t}\n\n\tif !cached {\n\t\tticketUpdateCacheMut.Lock()\n\t\tticketUpdateCache[key] = cache\n\t\tticketUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Kvstore) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tkvstoreUpdateCacheMut.RLock()\n\tcache, cached := kvstoreUpdateCache[key]\n\tkvstoreUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update kvstore, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"kvstore\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, kvstorePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, append(wl, kvstorePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update kvstore row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for kvstore\")\n\t}\n\n\tif !cached {\n\t\tkvstoreUpdateCacheMut.Lock()\n\t\tkvstoreUpdateCache[key] = cache\n\t\tkvstoreUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (m *File) Update(attr string, value interface{}) error {\n\treturn UnscopedDb().Model(m).UpdateColumn(attr, value).Error\n}", "func (o AuthorSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"mdbmdbmodels: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"authors\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, authorPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update all in author slice\")\n\t}\n\n\treturn nil\n}", "func (o *ORM) Update(model interface{}) error {\n\tt := time.Now()\n\tid, clauses := o.clauses(model, false)\n\n\tstmt, args, err := o.builder.Update(\"articles\").\n\t\tSetMap(clauses).\n\t\tWhere(squirrel.Eq{\"id\": id}).\n\t\tSuffix(\"RETURNING *\").\n\t\tToSql()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer o.log(t, stmt, args)\n\n\treturn o.conn.QueryRowx(stmt, args...).StructScan(model)\n}", "func (o *Customer) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tcustomerUpdateCacheMut.RLock()\n\tcache, cached := customerUpdateCache[key]\n\tcustomerUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tcustomerColumns,\n\t\t\tcustomerPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update customers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"customers\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, customerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(customerType, customerMapping, append(wl, customerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update customers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for customers\")\n\t}\n\n\tif !cached {\n\t\tcustomerUpdateCacheMut.Lock()\n\t\tcustomerUpdateCache[key] = cache\n\t\tcustomerUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *AuthUser) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tkey := makeCacheKey(whitelist, nil)\n\tauthUserUpdateCacheMut.RLock()\n\tcache, cached := authUserUpdateCache[key]\n\tauthUserUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(authUserColumns, authUserPrimaryKeyColumns, whitelist)\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update auth_user, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `auth_user` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, authUserPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(authUserType, authUserMapping, append(wl, authUserPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update auth_user row\")\n\t}\n\n\tif !cached {\n\t\tauthUserUpdateCacheMut.Lock()\n\t\tauthUserUpdateCache[key] = cache\n\t\tauthUserUpdateCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (s *Session) Update(dest interface{}) (int64, error) {\n\ts.initStatemnt()\n\ts.statement.Update()\n\tscanner, err := NewScanner(dest)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer scanner.Close()\n\tif s.statement.table == \"\" {\n\t\ts.statement.From(scanner.GetTableName())\n\t}\n\tupdateFields := make([]string, 0)\n\tpks := make([]interface{}, 0)\n\tprimaryKey := \"\"\n\tfor n, f := range scanner.Model.Fields {\n\t\tif !f.IsReadOnly && !f.IsPrimaryKey {\n\t\t\tupdateFields = append(updateFields, n)\n\t\t}\n\t\tif f.IsPrimaryKey {\n\t\t\tprimaryKey = n\n\t\t}\n\t}\n\tif primaryKey == \"\" {\n\t\treturn 0, ModelMustHavePrimaryKey\n\t}\n\ts.Columns(updateFields...)\n\tif scanner.entityPointer.Kind() == reflect.Slice {\n\t\tfor i := 0; i < scanner.entityPointer.Len(); i++ {\n\t\t\tval := make([]interface{}, 0)\n\t\t\tsub := scanner.entityPointer.Index(i)\n\t\t\tif sub.Kind() == reflect.Ptr {\n\t\t\t\tsubElem := sub.Elem()\n\t\t\t\tfor _, fn := range updateFields {\n\t\t\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tfv := subElem.Field(f.idx)\n\t\t\t\t\tval = append(val, fv.Interface())\n\t\t\t\t}\n\t\t\t\tprimaryF, _ := scanner.Model.Fields[primaryKey]\n\t\t\t\tfv := subElem.Field(primaryF.idx)\n\t\t\t\tpks = append(pks, fv.Interface())\n\t\t\t} else {\n\t\t\t\tfor _, fn := range updateFields {\n\t\t\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tfv := sub.Field(f.idx)\n\t\t\t\t\tval = append(val, fv.Interface())\n\t\t\t\t}\n\t\t\t\tprimaryF, _ := scanner.Model.Fields[primaryKey]\n\t\t\t\tfv := sub.Field(primaryF.idx)\n\t\t\t\tpks = append(pks, fv.Interface())\n\t\t\t}\n\t\t\ts.statement.Values(val)\n\t\t}\n\n\t} else if scanner.entityPointer.Kind() == reflect.Struct {\n\t\tval := make([]interface{}, 0)\n\t\tfor _, fn := range updateFields {\n\t\t\tf, ok := scanner.Model.Fields[fn]\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfv := scanner.entityPointer.Field(f.idx)\n\t\t\tval = append(val, fv.Interface())\n\t\t}\n\t\tprimaryF, _ := scanner.Model.Fields[primaryKey]\n\t\tfv := scanner.entityPointer.Field(primaryF.idx)\n\t\tpks = append(pks, fv.Interface())\n\t\ts.statement.Values(val)\n\t} else {\n\t\treturn 0, UpdateExpectSliceOrStruct\n\t}\n\ts.Where(Eq{scanner.Model.PkName: pks})\n\tsql, args, err := s.statement.ToSQL()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\ts.logger.Debugf(\"[Session Update] sql: %s, args: %v\", sql, args)\n\ts.initCtx()\n\tsResult, err := s.ExecContext(s.ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn sResult.RowsAffected()\n}", "func (o *Segment) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tsegmentUpdateCacheMut.RLock()\n\tcache, cached := segmentUpdateCache[key]\n\tsegmentUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tsegmentAllColumns,\n\t\t\tsegmentPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"boiler: unable to update segment, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"segment\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, segmentPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(segmentType, segmentMapping, append(wl, segmentPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to update segment row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: failed to get rows affected by update for segment\")\n\t}\n\n\tif !cached {\n\t\tsegmentUpdateCacheMut.Lock()\n\t\tsegmentUpdateCache[key] = cache\n\t\tsegmentUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *BookCategoryAssign) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tbookCategoryAssignUpdateCacheMut.RLock()\n\tcache, cached := bookCategoryAssignUpdateCache[key]\n\tbookCategoryAssignUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tbookCategoryAssignAllColumns,\n\t\t\tbookCategoryAssignPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update book_category_assign, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `book_category_assign` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, bookCategoryAssignPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(bookCategoryAssignType, bookCategoryAssignMapping, append(wl, bookCategoryAssignPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update book_category_assign row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for book_category_assign\")\n\t}\n\n\tif !cached {\n\t\tbookCategoryAssignUpdateCacheMut.Lock()\n\t\tbookCategoryAssignUpdateCache[key] = cache\n\t\tbookCategoryAssignUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Board) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tboardUpdateCacheMut.RLock()\n\tcache, cached := boardUpdateCache[key]\n\tboardUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"rdb: unable to update boards, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `boards` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, boardPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(boardType, boardMapping, append(wl, boardPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to update boards row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: failed to get rows affected by update for boards\")\n\t}\n\n\tif !cached {\n\t\tboardUpdateCacheMut.Lock()\n\t\tboardUpdateCache[key] = cache\n\t\tboardUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (blt Bolt) Update(execute dbtx.Execute) error {\n\treturn blt.db.Update(func(tx *b.Tx) error {\n\t\treturn execute(tx.Bucket(blt.Bucket))\n\t})\n}", "func (o PremiumSlotSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"premium_slots\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, premiumSlotPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all premiumSlot\")\n\t}\n\treturn rowsAff, nil\n}", "func (o *Doc) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tdocUpdateCacheMut.RLock()\n\tcache, cached := docUpdateCache[key]\n\tdocUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tdocAllColumns,\n\t\t\tdocPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update doc, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `doc` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, docPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(docType, docMapping, append(wl, docPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update doc row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for doc\")\n\t}\n\n\tif !cached {\n\t\tdocUpdateCacheMut.Lock()\n\t\tdocUpdateCache[key] = cache\n\t\tdocUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Picture) Update(exec boil.Executor, whitelist ...string) error {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn err\n\t}\n\tkey := makeCacheKey(whitelist, nil)\n\tpictureUpdateCacheMut.RLock()\n\tcache, cached := pictureUpdateCache[key]\n\tpictureUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := strmangle.UpdateColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpicturePrimaryKeyColumns,\n\t\t\twhitelist,\n\t\t)\n\n\t\tif len(whitelist) == 0 {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn errors.New(\"models: unable to update pictures, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `pictures` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, picturePrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(pictureType, pictureMapping, append(wl, picturePrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\t_, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update pictures row\")\n\t}\n\n\tif !cached {\n\t\tpictureUpdateCacheMut.Lock()\n\t\tpictureUpdateCache[key] = cache\n\t\tpictureUpdateCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpdateHooks(exec)\n}", "func (o *StockKeepingUnitContent) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tstockKeepingUnitContentUpdateCacheMut.RLock()\n\tcache, cached := stockKeepingUnitContentUpdateCache[key]\n\tstockKeepingUnitContentUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tstockKeepingUnitContentAllColumns,\n\t\t\tstockKeepingUnitContentPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"db: unable to update stock_keeping_unit_content, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"stock_keeping_unit_content\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, stockKeepingUnitContentPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(stockKeepingUnitContentType, stockKeepingUnitContentMapping, append(wl, stockKeepingUnitContentPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update stock_keeping_unit_content row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by update for stock_keeping_unit_content\")\n\t}\n\n\tif !cached {\n\t\tstockKeepingUnitContentUpdateCacheMut.Lock()\n\t\tstockKeepingUnitContentUpdateCache[key] = cache\n\t\tstockKeepingUnitContentUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *OauthClient) Update(exec boil.Executor, columns boil.Columns) (int64, error) {\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\toauthClientUpdateCacheMut.RLock()\n\tcache, cached := oauthClientUpdateCache[key]\n\toauthClientUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update oauth_clients, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `oauth_clients` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, oauthClientPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, append(wl, oauthClientPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.Exec(cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update oauth_clients row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for oauth_clients\")\n\t}\n\n\tif !cached {\n\t\toauthClientUpdateCacheMut.Lock()\n\t\toauthClientUpdateCache[key] = cache\n\t\toauthClientUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(exec)\n}", "func (o *TrainingCost) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\ttrainingCostUpdateCacheMut.RLock()\n\tcache, cached := trainingCostUpdateCache[key]\n\ttrainingCostUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update training_costs, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"training_costs\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, trainingCostPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, append(wl, trainingCostPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update training_costs row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostUpdateCacheMut.Lock()\n\t\ttrainingCostUpdateCache[key] = cache\n\t\ttrainingCostUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *UsernameListing) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tkey := makeCacheKey(columns, nil)\n\tusernameListingUpdateCacheMut.RLock()\n\tcache, cached := usernameListingUpdateCache[key]\n\tusernameListingUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tusernameListingAllColumns,\n\t\t\tusernameListingPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update username_listings, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"username_listings\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, usernameListingPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(usernameListingType, usernameListingMapping, append(wl, usernameListingPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update username_listings row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for username_listings\")\n\t}\n\n\tif !cached {\n\t\tusernameListingUpdateCacheMut.Lock()\n\t\tusernameListingUpdateCache[key] = cache\n\t\tusernameListingUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VendorSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vendors` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vendor slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vendor\")\n\t}\n\treturn rowsAff, nil\n}", "func (o *Illness) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tillnessUpdateCacheMut.RLock()\n\tcache, cached := illnessUpdateCache[key]\n\tillnessUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tillnessAllColumns,\n\t\t\tillnessPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"models: unable to update illness, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE \\\"illness\\\" SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, wl),\n\t\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", len(wl)+1, illnessPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(illnessType, illnessMapping, append(wl, illnessPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update illness row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by update for illness\")\n\t}\n\n\tif !cached {\n\t\tillnessUpdateCacheMut.Lock()\n\t\tillnessUpdateCache[key] = cache\n\t\tillnessUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}", "func (o *Peer) Update(ctx context.Context, exec boil.ContextExecutor, columns boil.Columns) (int64, error) {\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tvar err error\n\tif err = o.doBeforeUpdateHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\tkey := makeCacheKey(columns, nil)\n\tpeerUpdateCacheMut.RLock()\n\tcache, cached := peerUpdateCache[key]\n\tpeerUpdateCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl := columns.UpdateColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerPrimaryKeyColumns,\n\t\t)\n\n\t\tif !columns.IsWhitelist() {\n\t\t\twl = strmangle.SetComplement(wl, []string{\"created_at\"})\n\t\t}\n\t\tif len(wl) == 0 {\n\t\t\treturn 0, errors.New(\"model: unable to update peers, could not build whitelist\")\n\t\t}\n\n\t\tcache.query = fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, wl),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, peerPrimaryKeyColumns),\n\t\t)\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, append(wl, peerPrimaryKeyColumns...))\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tvalues := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cache.valueMapping)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tvar result sql.Result\n\tresult, err = exec.ExecContext(ctx, cache.query, values...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update peers row\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: failed to get rows affected by update for peers\")\n\t}\n\n\tif !cached {\n\t\tpeerUpdateCacheMut.Lock()\n\t\tpeerUpdateCache[key] = cache\n\t\tpeerUpdateCacheMut.Unlock()\n\t}\n\n\treturn rowsAff, o.doAfterUpdateHooks(ctx, exec)\n}" ]
[ "0.67680526", "0.65728456", "0.656442", "0.6516539", "0.6376052", "0.63365155", "0.6237881", "0.6209677", "0.6199092", "0.6088644", "0.6077436", "0.60091984", "0.59750617", "0.5958057", "0.5934554", "0.59304446", "0.5921469", "0.58936304", "0.5891069", "0.5888901", "0.5868908", "0.58337367", "0.5787405", "0.5783977", "0.5770641", "0.57495445", "0.5741291", "0.5737329", "0.57073885", "0.56925863", "0.5687053", "0.56829065", "0.56727225", "0.5668402", "0.5666729", "0.5647444", "0.56449616", "0.56249374", "0.56249374", "0.56249374", "0.56249374", "0.56249374", "0.56228083", "0.56021845", "0.5593849", "0.5569857", "0.5553852", "0.55538243", "0.554403", "0.5541271", "0.55407816", "0.55346435", "0.5527814", "0.5517472", "0.55130553", "0.5512793", "0.55111915", "0.5508118", "0.55036634", "0.55017924", "0.5501679", "0.5498429", "0.5493703", "0.5476905", "0.5445214", "0.5434761", "0.54299045", "0.5428679", "0.54253626", "0.5400462", "0.53950745", "0.53904605", "0.5386887", "0.53851885", "0.5363767", "0.53626174", "0.5361696", "0.53580785", "0.5348438", "0.5346064", "0.534232", "0.5340865", "0.53316545", "0.53258073", "0.5306922", "0.5296861", "0.5295216", "0.5283398", "0.5277769", "0.52758485", "0.5274774", "0.52706087", "0.5265454", "0.52627414", "0.5261844", "0.5254869", "0.524763", "0.5236829", "0.52253056", "0.52195203" ]
0.79449517
0
UpdateAllP updates all rows with matching column names, and panics on error.
func (q assetRevisionQuery) UpdateAllP(cols M) { if err := q.UpdateAll(cols); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q failureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q apiKeyQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q sourceQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q shelfQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q inventoryQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q skinQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockCvtermQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q instrumentClassQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authorQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q organismQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q braceletPhotoQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cvtermsynonymQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q addressQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q voteQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q blackCardQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authMessageQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q pictureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserUserPermissionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q commentQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q dMessageEmbedQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authTokenQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) UpdateAllP(ctx context.Context, exec boil.ContextExecutor, cols M) int64 {\n\trowsAff, err := q.UpdateAll(ctx, exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o VoteSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q claimInListQuery) UpdateAllP(exec boil.Executor, cols M) {\n\terr := q.UpdateAll(exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureCvtermDbxrefQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SkinSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q recordMeasureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureRelationshipQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ClaimInListSlice) UpdateAllP(exec boil.Executor, cols M) {\n\terr := o.UpdateAll(exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all for phenotypeprop\")\n\t}\n\n\treturn nil\n}", "func (o PeerPropertySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"peer_properties\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, peerPropertyPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in peerProperty slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all peerProperty\")\n\t}\n\treturn rowsAff, nil\n}", "func (q vspQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o WeatherSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"prh\\\".\\\"weather\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, weatherPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all weather\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TenantSlice) UpdateAllP(ctx context.Context, exec boil.ContextExecutor, cols M) int64 {\n\trowsAff, err := o.UpdateAll(ctx, exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o PictureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VSPSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vsp\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, vspPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vsp slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vsp\")\n\t}\n\treturn rowsAff, nil\n}", "func (o PhenotypepropSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"chado: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"UPDATE \\\"phenotypeprop\\\" SET %s WHERE (\\\"phenotypeprop_id\\\") IN (%s)\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(phenotypepropPrimaryKeyColumns), len(colNames)+1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all in phenotypeprop slice\")\n\t}\n\n\treturn nil\n}", "func (o PeerSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"model: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update all in peer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to retrieve rows affected all in update all peer\")\n\t}\n\treturn rowsAff, nil\n}", "func (o JetSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `jets` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in jet slice\")\n\t}\n\n\treturn nil\n}", "func (q jetQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for jets\")\n\t}\n\n\treturn nil\n}", "func (o FailureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q failureQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for failure\")\n\t}\n\n\treturn nil\n}", "func (o VoteSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\treturn nil\n}", "func (q taskQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q peerPropertyQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for peer_properties\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for peer_properties\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FailureSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), failurePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `failure` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, failurePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in failure slice\")\n\t}\n\n\treturn nil\n}", "func (q peerQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update all for peers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to retrieve rows affected for peers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q offerQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all for offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OfferSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"offers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, offerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all offer\")\n\t}\n\treturn rowsAff, nil\n}", "func (q sourceQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to update all for sources\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to retrieve rows affected for sources\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o JetSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vote\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, votePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vote\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currency\")\n\t}\n\treturn rowsAff, nil\n}", "func (q kvstoreQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q weatherQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all for weather\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected for weather\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ProjectSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `project` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in project slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all project\")\n\t}\n\treturn rowsAff, nil\n}", "func (o PremiumSlotSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"premium_slots\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, premiumSlotPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all premiumSlot\")\n\t}\n\treturn rowsAff, nil\n}", "func (o HoldenAtSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"HoldenAt\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, holdenAtPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in holdenAt slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all holdenAt\")\n\t}\n\treturn rowsAff, nil\n}", "func (q projectQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for project\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for project\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q jobQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q illnessQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for illness\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for illness\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for currency\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.8506606", "0.8503917", "0.8446253", "0.84416735", "0.84044635", "0.83883965", "0.8385268", "0.83843184", "0.83843184", "0.8367998", "0.83427215", "0.83184147", "0.8316005", "0.83028984", "0.8291803", "0.824599", "0.8233303", "0.822652", "0.8225935", "0.82193744", "0.82168454", "0.82106304", "0.8203605", "0.81981945", "0.81957304", "0.8177243", "0.8177243", "0.81722116", "0.8171742", "0.8168828", "0.81581074", "0.81525284", "0.81474614", "0.8131442", "0.8118917", "0.8118389", "0.8110256", "0.8099511", "0.80807936", "0.8074887", "0.8071829", "0.8059451", "0.80500776", "0.8025218", "0.8014976", "0.8010819", "0.8004283", "0.7979328", "0.79576045", "0.7946249", "0.7928875", "0.7894847", "0.7892139", "0.7889921", "0.7879699", "0.78758615", "0.7834231", "0.78199893", "0.780099", "0.77839315", "0.77813536", "0.77361774", "0.77288777", "0.76997715", "0.76902276", "0.7672655", "0.7626026", "0.7617204", "0.7592486", "0.75732976", "0.7558458", "0.75514036", "0.75404257", "0.75019807", "0.74698526", "0.74612135", "0.7448351", "0.74483216", "0.74073166", "0.7400068", "0.73807716", "0.73722714", "0.7365531", "0.7351957", "0.7346847", "0.73448855", "0.7335343", "0.7325256", "0.73146355", "0.730903", "0.7304962", "0.73033464", "0.7299254", "0.72983956", "0.7297866", "0.729751", "0.7297194", "0.72923994", "0.72922647", "0.7289267" ]
0.80665153
41
UpdateAll updates all rows with the specified column values.
func (q assetRevisionQuery) UpdateAll(cols M) error { queries.SetUpdate(q.Query, cols) _, err := q.Query.Exec() if err != nil { return errors.Wrap(err, "public: unable to update all for asset_revision") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o WeatherSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"prh\\\".\\\"weather\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, weatherPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all weather\")\n\t}\n\treturn rowsAff, nil\n}", "func (q holdenAtQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q weatherQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all for weather\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected for weather\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\treturn nil\n}", "func (q cmfTurntableQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_turntable\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_turntable\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RentalSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"sqlboiler: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `rental` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update all in rental slice\")\n\t}\n\n\treturn nil\n}", "func (q currencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q jetQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for jets\")\n\t}\n\n\treturn nil\n}", "func (q buildingQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to update all for buildings\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to retrieve rows affected for buildings\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currency\")\n\t}\n\treturn rowsAff, nil\n}", "func (o JetSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `jets` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in jet slice\")\n\t}\n\n\treturn nil\n}", "func (q kvstoreQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o HoldenAtSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"HoldenAt\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, holdenAtPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in holdenAt slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all holdenAt\")\n\t}\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vote\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, votePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vote\")\n\t}\n\treturn rowsAff, nil\n}", "func (q voteQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StockSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `stock` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in stock slice\")\n\t}\n\n\treturn nil\n}", "func (q automodRuleDatumQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for automod_rule_data\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for automod_rule_data\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"model: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `transaction` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to update all in transaction slice\")\n\t}\n\n\treturn nil\n}", "func (o FriendshipSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `friendship` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in friendship slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all friendship\")\n\t}\n\treturn rowsAff, nil\n}", "func (q friendshipQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for friendship\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for friendship\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFTurntableSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_turntable` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfTurntable slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfTurntable\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `transactions` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in transaction slice\")\n\t}\n\n\treturn nil\n}", "func (q illnessQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for illness\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for illness\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q rentalQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update all for rental\")\n\t}\n\n\treturn nil\n}", "func (q itemQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for items\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for items\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currentChartDataMinutelyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for current_chart_data_minutely\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for current_chart_data_minutely\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q employeeQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for employee\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for employee\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q transactionQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to update all for transaction\")\n\t}\n\n\treturn nil\n}", "func (q taskQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ExchangeCurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), exchangeCurrencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"exchange_currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, exchangeCurrencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in exchangeCurrency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all exchangeCurrency\")\n\t}\n\treturn rowsAff, nil\n}", "func (o IllnessSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"illness\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, illnessPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in illness slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all illness\")\n\t}\n\treturn rowsAff, nil\n}", "func (q jobQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q rentalRowerQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for rental_rowers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for rental_rowers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for vote\")\n\t}\n\n\treturn nil\n}", "func (q shelfQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for shelf\")\n\t}\n\n\treturn nil\n}", "func (q cmfBalanceChargeAdminQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_balance_charge_admin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_balance_charge_admin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o InventorySlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), inventoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `inventory` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, inventoryPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in inventory slice\")\n\t}\n\n\treturn nil\n}", "func (o UtxoSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"utxo\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, utxoPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in utxo slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all utxo\")\n\t}\n\treturn rowsAff, nil\n}", "func (q storeQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for stores\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for stores\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o EmployeeSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), employeePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"employee\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, employeePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in employee slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all employee\")\n\t}\n\treturn rowsAff, nil\n}", "func (q stockQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for stock\")\n\t}\n\n\treturn nil\n}", "func (o AutomodRuleDatumSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), automodRuleDatumPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"automod_rule_data\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, automodRuleDatumPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in automodRuleDatum slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all automodRuleDatum\")\n\t}\n\treturn rowsAff, nil\n}", "func (q utxoQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for utxo\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for utxo\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q repositoryQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for repositories\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for repositories\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q withdrawalCryptoQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to update all for withdrawal_crypto\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to retrieve rows affected for withdrawal_crypto\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q blockQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for block\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q exchangeCurrencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for exchange_currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for exchange_currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TicketSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"tickets\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, ticketPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in ticket slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all ticket\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RawVisitSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"raw_visits\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rawVisitPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rawVisit slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rawVisit\")\n\t}\n\treturn rowsAff, nil\n}", "func (o AddressSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"sqlboiler: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), addressPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `address` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, addressPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update all in address slice\")\n\t}\n\n\treturn nil\n}", "func (q boardQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to update all for boards\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to retrieve rows affected for boards\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o BuildingSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"record: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"buildings\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, buildingPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to update all in building slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to retrieve rows affected all in update all building\")\n\t}\n\treturn rowsAff, nil\n}", "func (q offerQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all for offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFBalanceChargeAdminSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_balance_charge_admin` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfBalanceChargeAdmin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfBalanceChargeAdmin\")\n\t}\n\treturn rowsAff, nil\n}", "func (q latencyQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for latencies\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for latencies\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o NotificationSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"notification\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, notificationPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in notification slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all notification\")\n\t}\n\treturn rowsAff, nil\n}", "func (o BoardSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"rdb: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `boards` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to update all in board slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to retrieve rows affected all in update all board\")\n\t}\n\treturn rowsAff, nil\n}", "func (q ticketQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tickets\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tickets\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q activityLogQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update all for activity_logs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to retrieve rows affected for activity_logs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthorSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"mdbmdbmodels: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"authors\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, authorPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update all in author slice\")\n\t}\n\n\treturn nil\n}", "func (o TrainingCostSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"training_costs\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, trainingCostPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in trainingCost slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all trainingCost\")\n\t}\n\treturn rowsAff, nil\n}", "func (q transactionQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for transactions\")\n\t}\n\n\treturn nil\n}", "func (o ItemSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"items\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in item slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all item\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RestaurantRankSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), restaurantRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"restaurant_rank\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, restaurantRankPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in restaurantRank slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all restaurantRank\")\n\t}\n\treturn rowsAff, nil\n}", "func (q sourceQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to update all for sources\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to retrieve rows affected for sources\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q chatQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to update all for chat\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to retrieve rows affected for chat\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CommentSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), commentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `comment` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, commentPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in comment slice\")\n\t}\n\n\treturn nil\n}", "func (q rawVisitQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for raw_visits\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for raw_visits\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q organismQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all for organism\")\n\t}\n\n\treturn nil\n}", "func (q notificationQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for notification\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for notification\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q recipeLipidQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for recipe_lipid\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for recipe_lipid\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CurrentChartDataMinutelySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"current_chart_data_minutely\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currentChartDataMinutelyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currentChartDataMinutely slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currentChartDataMinutely\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RentalRowerSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"rental_rowers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rentalRowerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rentalRower slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rentalRower\")\n\t}\n\treturn rowsAff, nil\n}", "func (q instrumentClassQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for instrument_class\")\n\t}\n\n\treturn nil\n}", "func (o RSSAnnouncementSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rssAnnouncementPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"rss_announcements\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rssAnnouncementPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rssAnnouncement slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rssAnnouncement\")\n\t}\n\treturn rowsAff, nil\n}", "func (o UsernameListingSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), usernameListingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"username_listings\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, usernameListingPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in usernameListing slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all usernameListing\")\n\t}\n\treturn rowsAff, nil\n}", "func (o SourceSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"mdbmdbmodels: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"sources\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, sourcePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update all in source slice\")\n\t}\n\n\treturn nil\n}", "func (o ActivityLogSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"dbmodel: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), activityLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"activity_logs\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, activityLogPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update all in activityLog slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to retrieve rows affected all in update all activityLog\")\n\t}\n\treturn rowsAff, nil\n}", "func (q mempoolBinQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for mempool_bin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for mempool_bin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PeerPropertySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"peer_properties\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, peerPropertyPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in peerProperty slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all peerProperty\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RecipeAdditiveSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeAdditivePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"recipe_additive\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, recipeAdditivePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in recipeAdditive slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all recipeAdditive\")\n\t}\n\treturn rowsAff, nil\n}", "func (o WithdrawalCryptoSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"sqlite3: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"withdrawal_crypto\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to update all in withdrawalCrypto slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to retrieve rows affected all in update all withdrawalCrypto\")\n\t}\n\treturn rowsAff, nil\n}", "func (q btcTXOutputQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for btc_tx_output\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q nodeQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for node\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for node\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q vspQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q recipeAdditiveQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for recipe_additive\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for recipe_additive\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StoreSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"stores\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, storePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in store slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all store\")\n\t}\n\treturn rowsAff, nil\n}", "func (o KvstoreSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"kvstore\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, kvstorePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in kvstore slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all kvstore\")\n\t}\n\treturn rowsAff, nil\n}", "func (q phenotypepropQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all for phenotypeprop\")\n\t}\n\n\treturn nil\n}", "func (o StockKeepingUnitContentSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockKeepingUnitContentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"stock_keeping_unit_content\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, stockKeepingUnitContentPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in stockKeepingUnitContent slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all stockKeepingUnitContent\")\n\t}\n\treturn rowsAff, nil\n}", "func (o OfferSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"offers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, offerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all offer\")\n\t}\n\treturn rowsAff, nil\n}", "func (q paymentObjectQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for payment_objects\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for payment_objects\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RepositorySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `repositories` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in repository slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all repository\")\n\t}\n\treturn rowsAff, nil\n}", "func (q sourceQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to update all for sources\")\n\t}\n\n\treturn nil\n}", "func (q cmfUserSuperQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_user_super\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_user_super\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StorestateSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storestatePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"storestate\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, storestatePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in storestate slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all storestate\")\n\t}\n\treturn rowsAff, nil\n}", "func (o EmailSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"mysql: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `email` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to update all in email slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to retrieve rows affected all in update all email\")\n\t}\n\treturn rowsAff, nil\n}", "func (q restaurantRankQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for restaurant_rank\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for restaurant_rank\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q channelQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for channels\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for channels\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q itemSideQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for item_sides\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for item_sides\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q docQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for doc\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for doc\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.796935", "0.7907914", "0.78601724", "0.7845937", "0.78452486", "0.783391", "0.7828506", "0.78112113", "0.7809873", "0.7807201", "0.77594817", "0.775496", "0.7748255", "0.77446365", "0.774176", "0.77298295", "0.7727604", "0.7727594", "0.7718456", "0.7712316", "0.7712261", "0.76888955", "0.7684777", "0.76846427", "0.7682325", "0.7665387", "0.7657949", "0.7649154", "0.7646723", "0.76438224", "0.76413614", "0.76385176", "0.76335645", "0.76272875", "0.7624369", "0.76207817", "0.7610351", "0.7608142", "0.76041776", "0.76006484", "0.75962263", "0.75923306", "0.758551", "0.7584796", "0.75819564", "0.75786465", "0.757115", "0.75643307", "0.7560392", "0.7556611", "0.7555309", "0.7552172", "0.7550516", "0.75448257", "0.7541764", "0.753147", "0.75313705", "0.7527766", "0.75261515", "0.75201786", "0.7518788", "0.7516934", "0.7514436", "0.7509356", "0.7506713", "0.7497875", "0.7494303", "0.7493662", "0.7492334", "0.74919057", "0.7491449", "0.749079", "0.7487326", "0.7482481", "0.74808663", "0.74808276", "0.7477945", "0.7477454", "0.74763024", "0.7475494", "0.747485", "0.74688554", "0.74683845", "0.7468319", "0.7467841", "0.74677646", "0.7467017", "0.74668735", "0.7463987", "0.74639547", "0.7462768", "0.74625003", "0.7460804", "0.74594957", "0.7457434", "0.7454676", "0.7452063", "0.74499416", "0.7449087", "0.74456555", "0.74450713" ]
0.0
-1
UpdateAllG updates all rows with the specified column values.
func (o AssetRevisionSlice) UpdateAllG(cols M) error { return o.UpdateAll(boil.GetDB(), cols) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q rentalRowerQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (q storeQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q itemQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q automodRuleDatumQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q itemSideQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q rowerGroupQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (q exchangeCurrencyQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (q notificationQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (o JetSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q tenantQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q premiumCodeQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q ticketQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o RentalRowerSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q premiumSlotQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q usernameListingQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o RentalSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o TransactionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o TransactionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q rssAnnouncementQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o StockCvtermSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o StorestateSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o StoreSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o CvtermsynonymSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o StockSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o ExchangeCurrencySlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o SourceSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o OfferSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o PremiumCodeSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o PhenotypepropSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o OrganismSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o ItemSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o InstrumentClassSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o ShelfSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AutomodRuleDatumSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o VoteSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o UsernameListingSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o APIKeySlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o PremiumSlotSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q claimInListQuery) UpdateAllG(cols M) error {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (o AuthUserSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q bookCategoryQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o TicketSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o InventorySlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q bookCategoryAssignQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o ItemSideSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o RentalSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AuthorSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o BraceletPhotoSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o NotificationSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o JetSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RowerGroupSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AddressSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o SkinSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o ShelfSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o TenantSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q tenantQuery) UpdateAllGP(ctx context.Context, cols M) int64 {\n\trowsAff, err := q.UpdateAll(ctx, boil.GetContextDB(), cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o StockSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o CvtermsynonymSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RSSAnnouncementSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o TransactionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AuthTokenSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o BlackCardSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o StockCvtermSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o OrganismSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AuthMessageSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o BookCategoryAssignSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o ClaimInListSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o InstrumentClassSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o SkinSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BookCategorySlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q userGoogleQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to update all for user_google\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to retrieve rows affected for user_google\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FeatureCvtermDbxrefSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o AddressSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o PhenotypepropSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) UpdateAllGP(ctx context.Context, cols M) int64 {\n\trowsAff, err := o.UpdateAll(ctx, boil.GetContextDB(), cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o CommentSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.83586043", "0.82939416", "0.8253447", "0.8198915", "0.81822294", "0.81634235", "0.81623614", "0.80981153", "0.80860525", "0.8081597", "0.8071907", "0.805967", "0.8036535", "0.80231774", "0.801802", "0.80056363", "0.798706", "0.798706", "0.79787683", "0.79612833", "0.7960956", "0.79545605", "0.79510385", "0.7937114", "0.7916863", "0.7915056", "0.7903883", "0.7897257", "0.78668594", "0.7860494", "0.7857197", "0.78559595", "0.7847742", "0.7846182", "0.7831482", "0.7823623", "0.7812915", "0.7790786", "0.7789533", "0.778793", "0.77861834", "0.77742434", "0.77620673", "0.77598727", "0.77560914", "0.7751727", "0.774709", "0.77419144", "0.77417403", "0.7740431", "0.77393246", "0.77339727", "0.77311504", "0.7731049", "0.7711347", "0.769092", "0.76802355", "0.76683325", "0.7665991", "0.76659155", "0.76429355", "0.76363355", "0.7628139", "0.7628139", "0.7626582", "0.76215374", "0.76212955", "0.7593552", "0.7588841", "0.7550371", "0.75488776", "0.75488746", "0.7537991", "0.7519415", "0.7517477", "0.7511742", "0.7504801", "0.7501962", "0.7491655", "0.7478013", "0.74647725", "0.7464039", "0.7461011", "0.7444668", "0.74225605", "0.7421819", "0.74054134", "0.7387722", "0.7362812", "0.73431516", "0.7333147", "0.7327689", "0.7314093", "0.72969633", "0.7279011", "0.7266431", "0.7262244", "0.72582823", "0.72483903", "0.72303814" ]
0.7668796
57
UpdateAllGP updates all rows with the specified column values, and panics on error.
func (o AssetRevisionSlice) UpdateAllGP(cols M) { if err := o.UpdateAll(boil.GetDB(), cols); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o FailureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) UpdateAllGP(ctx context.Context, cols M) int64 {\n\trowsAff, err := q.UpdateAll(ctx, boil.GetContextDB(), cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o RentalSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) UpdateAllGP(ctx context.Context, cols M) int64 {\n\trowsAff, err := o.UpdateAll(ctx, boil.GetContextDB(), cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o SkinSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalRowerQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (o AuthorSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q itemQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q automodRuleDatumQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q storeQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o RecordMeasureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q premiumCodeQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q userGoogleQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to update all for user_google\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to retrieve rows affected for user_google\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q stockQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rowerGroupQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (q premiumSlotQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q failureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q notificationQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (o FeatureCvtermDbxrefSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q itemSideQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q ticketQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q apiKeyQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ClaimInListSlice) UpdateAllGP(cols M) {\n\terr := o.UpdateAll(boil.GetDB(), cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockCvtermQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q sourceQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q exchangeCurrencyQuery) UpdateAllG(cols M) (int64, error) {\n\treturn q.UpdateAll(boil.GetDB(), cols)\n}", "func (q rentalQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q instrumentClassQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q organismQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PremiumCodeSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q authUserUserPermissionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o UserGoogleSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"model2: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `user_google` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to update all in userGoogle slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to retrieve rows affected all in update all userGoogle\")\n\t}\n\treturn rowsAff, nil\n}", "func (o FailureSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q usernameListingQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q addressQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o AuthUserUserPermissionSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o RentalRowerSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q inventoryQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cvtermsynonymQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q pictureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authTokenQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PremiumSlotSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o JetSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o APIKeySlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q trainingCostQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for training_costs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for training_costs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OfferSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q featureRelationshipQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q shelfQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q commentQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q blackCardQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q braceletPhotoQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rssAnnouncementQuery) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn q.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (o AutomodRuleDatumSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q buildingQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to update all for buildings\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to retrieve rows affected for buildings\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q taskQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthUserSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (q phenotypepropQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all for phenotypeprop\")\n\t}\n\n\treturn nil\n}", "func (q skinQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rowerGroupQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q authorQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o UsernameListingSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}", "func (q jobQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o SourceSlice) UpdateAllG(cols M) error {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o StorestateSlice) UpdateAllG(cols M) (int64, error) {\n\treturn o.UpdateAll(boil.GetDB(), cols)\n}", "func (o ItemSlice) UpdateAllG(ctx context.Context, cols M) (int64, error) {\n\treturn o.UpdateAll(ctx, boil.GetContextDB(), cols)\n}" ]
[ "0.8076004", "0.80503494", "0.7996403", "0.7996403", "0.7995395", "0.79852635", "0.796148", "0.7938542", "0.79333127", "0.792682", "0.7914768", "0.7896731", "0.78956175", "0.78546804", "0.78536206", "0.78491974", "0.7828256", "0.77774894", "0.7771724", "0.7770692", "0.77571183", "0.7742444", "0.7741817", "0.77056456", "0.76410425", "0.7568315", "0.7561251", "0.75601876", "0.754756", "0.75367075", "0.7530217", "0.7525052", "0.7519933", "0.748896", "0.74858075", "0.7463491", "0.74621123", "0.7461098", "0.74464226", "0.74437475", "0.74247384", "0.7408114", "0.7403065", "0.73975354", "0.7393246", "0.7366806", "0.7364688", "0.7362535", "0.7354402", "0.73510015", "0.7333695", "0.73325044", "0.7326225", "0.7326225", "0.73231745", "0.7321074", "0.7317045", "0.73111886", "0.7307748", "0.73067516", "0.730254", "0.73014325", "0.7296001", "0.72844905", "0.72760516", "0.72725034", "0.72520316", "0.7234799", "0.72087455", "0.7202449", "0.7200648", "0.7190264", "0.7190019", "0.7188777", "0.7188763", "0.7173887", "0.71724766", "0.71581584", "0.7155819", "0.7140816", "0.7139127", "0.71378106", "0.71372676", "0.7135035", "0.7130739", "0.713035", "0.7122604", "0.7122343", "0.71110564", "0.7096023", "0.7095698", "0.70917195", "0.708564", "0.70800126", "0.70624304", "0.70585394", "0.7058162", "0.705565", "0.70494014", "0.7048695" ]
0.75096726
33
UpdateAllP updates all rows with the specified column values, and panics on error.
func (o AssetRevisionSlice) UpdateAllP(exec boil.Executor, cols M) { if err := o.UpdateAll(exec, cols); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q failureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q apiKeyQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q shelfQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q sourceQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q inventoryQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q instrumentClassQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockCvtermQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q voteQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserUserPermissionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q skinQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q organismQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authorQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q blackCardQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) UpdateAllP(ctx context.Context, exec boil.ContextExecutor, cols M) int64 {\n\trowsAff, err := q.UpdateAll(ctx, exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o StockSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q commentQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q braceletPhotoQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q addressQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q assetRevisionQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cvtermsynonymQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q pictureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authMessageQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q dMessageEmbedQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authTokenQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q claimInListQuery) UpdateAllP(exec boil.Executor, cols M) {\n\terr := q.UpdateAll(exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SkinSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q recordMeasureQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureCvtermDbxrefQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) UpdateAllP(ctx context.Context, exec boil.ContextExecutor, cols M) int64 {\n\trowsAff, err := o.UpdateAll(ctx, exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o BlackCardSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureRelationshipQuery) UpdateAllP(cols M) {\n\tif err := q.UpdateAll(cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PeerPropertySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"peer_properties\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, peerPropertyPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in peerProperty slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all peerProperty\")\n\t}\n\treturn rowsAff, nil\n}", "func (o ClaimInListSlice) UpdateAllP(exec boil.Executor, cols M) {\n\terr := o.UpdateAll(exec, cols)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q vspQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FeatureRelationshipSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all for phenotypeprop\")\n\t}\n\n\treturn nil\n}", "func (o VSPSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vsp\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, vspPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vsp slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vsp\")\n\t}\n\treturn rowsAff, nil\n}", "func (o WeatherSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"prh\\\".\\\"weather\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, weatherPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all weather\")\n\t}\n\treturn rowsAff, nil\n}", "func (o PictureSlice) UpdateAllP(exec boil.Executor, cols M) {\n\tif err := o.UpdateAll(exec, cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"chado: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"UPDATE \\\"phenotypeprop\\\" SET %s WHERE (\\\"phenotypeprop_id\\\") IN (%s)\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(phenotypepropPrimaryKeyColumns), len(colNames)+1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all in phenotypeprop slice\")\n\t}\n\n\treturn nil\n}", "func (o PeerSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"model: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update all in peer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to retrieve rows affected all in update all peer\")\n\t}\n\treturn rowsAff, nil\n}", "func (q taskQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vote\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, votePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vote\")\n\t}\n\treturn rowsAff, nil\n}", "func (q peerPropertyQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for peer_properties\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for peer_properties\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\treturn nil\n}", "func (o CMFBalanceChargeAdminSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_balance_charge_admin` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfBalanceChargeAdmin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfBalanceChargeAdmin\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currency\")\n\t}\n\treturn rowsAff, nil\n}", "func (q offerQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all for offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PremiumSlotSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"premium_slots\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, premiumSlotPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all premiumSlot\")\n\t}\n\treturn rowsAff, nil\n}", "func (q currencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q illnessQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for illness\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for illness\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TaskSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"tasks\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, taskPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in task slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all task\")\n\t}\n\treturn rowsAff, nil\n}", "func (q peerQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update all for peers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to retrieve rows affected for peers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FailureSlice) UpdateAllGP(cols M) {\n\tif err := o.UpdateAll(boil.GetDB(), cols); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `jets` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in jet slice\")\n\t}\n\n\treturn nil\n}", "func (o OfferSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"offers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, offerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all offer\")\n\t}\n\treturn rowsAff, nil\n}", "func (q currentChartDataMinutelyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for current_chart_data_minutely\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for current_chart_data_minutely\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PostSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"orm: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"posts\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, postPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to update all in post slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to retrieve rows affected all in update all post\")\n\t}\n\treturn rowsAff, nil\n}", "func (q jobQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q kvstoreQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o EmailSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"mysql: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `email` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to update all in email slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to retrieve rows affected all in update all email\")\n\t}\n\treturn rowsAff, nil\n}", "func (q premiumSlotQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for premium_slots\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for premium_slots\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q cmfBalanceChargeAdminQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_balance_charge_admin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_balance_charge_admin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q holdenAtQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q failureQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for failure\")\n\t}\n\n\treturn nil\n}" ]
[ "0.8451245", "0.8365498", "0.8360608", "0.83537006", "0.83537006", "0.83125263", "0.83121735", "0.82942456", "0.82700384", "0.8266638", "0.8262217", "0.82367194", "0.8218402", "0.82120603", "0.8206835", "0.8193837", "0.81888837", "0.8184062", "0.81806445", "0.8173364", "0.8162036", "0.8162036", "0.81540024", "0.8150164", "0.8121507", "0.812106", "0.81171906", "0.81170374", "0.8109586", "0.8107272", "0.8074578", "0.8066455", "0.80655104", "0.80611056", "0.8057067", "0.80488175", "0.8038233", "0.80351305", "0.802825", "0.8007613", "0.7992862", "0.79921734", "0.7986602", "0.7986513", "0.79738665", "0.79528445", "0.7946062", "0.794383", "0.79359335", "0.79260033", "0.79249334", "0.7868768", "0.78601795", "0.7828214", "0.7815232", "0.77793854", "0.7765878", "0.77455616", "0.7745081", "0.77412254", "0.7727731", "0.7719704", "0.7717632", "0.7706361", "0.76728", "0.764232", "0.7623729", "0.76058775", "0.7584506", "0.75808114", "0.75547516", "0.7523267", "0.7503107", "0.7493312", "0.7474518", "0.74600184", "0.7422612", "0.7415305", "0.7376499", "0.7369581", "0.73427224", "0.7338348", "0.7333676", "0.7326115", "0.7324031", "0.7315027", "0.7309779", "0.7303017", "0.730298", "0.73021203", "0.73018456", "0.72967744", "0.72955376", "0.72844696", "0.72810096", "0.72749454", "0.7273412", "0.7272844", "0.7264533", "0.72624975" ]
0.769914
64
UpdateAll updates all rows with the specified column values, using an executor.
func (o AssetRevisionSlice) UpdateAll(exec boil.Executor, cols M) error { ln := int64(len(o)) if ln == 0 { return nil } if len(cols) == 0 { return errors.New("public: update all requires at least one column argument") } colNames := make([]string, len(cols)) args := make([]interface{}, len(cols)) i := 0 for name, value := range cols { colNames[i] = name args[i] = value i++ } // Append all of the primary key values for each column for _, obj := range o { pkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), assetRevisionPrimaryKeyMapping) args = append(args, pkeyArgs...) } query := fmt.Sprintf( "UPDATE \"asset_revision\" SET %s WHERE (\"orig\") IN (%s)", strmangle.SetParamNames("\"", "\"", 1, colNames), strmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(assetRevisionPrimaryKeyColumns), len(colNames)+1, len(assetRevisionPrimaryKeyColumns)), ) if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, query) fmt.Fprintln(boil.DebugWriter, args...) } _, err := exec.Exec(query, args...) if err != nil { return errors.Wrap(err, "public: unable to update all in assetRevision slice") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o WeatherSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"prh\\\".\\\"weather\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, weatherPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all weather\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TaskSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"tasks\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, taskPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in task slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all task\")\n\t}\n\treturn rowsAff, nil\n}", "func (q taskQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vote\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, votePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vote\")\n\t}\n\treturn rowsAff, nil\n}", "func (o OfferSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"offers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, offerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all offer\")\n\t}\n\treturn rowsAff, nil\n}", "func (o BoardSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"rdb: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `boards` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to update all in board slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to retrieve rows affected all in update all board\")\n\t}\n\treturn rowsAff, nil\n}", "func (o ChannelSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"channels\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, channelPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in channel slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all channel\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currency\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CMFBalanceChargeAdminSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_balance_charge_admin` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfBalanceChargeAdmin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfBalanceChargeAdmin\")\n\t}\n\treturn rowsAff, nil\n}", "func (o VoteSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vote` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in vote slice\")\n\t}\n\n\treturn nil\n}", "func (o NodeSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), nodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"node\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, nodePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in node slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all node\")\n\t}\n\treturn rowsAff, nil\n}", "func (o ExchangeCurrencySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), exchangeCurrencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"exchange_currency\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, exchangeCurrencyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in exchangeCurrency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all exchangeCurrency\")\n\t}\n\treturn rowsAff, nil\n}", "func (o UtxoSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"utxo\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, utxoPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in utxo slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all utxo\")\n\t}\n\treturn rowsAff, nil\n}", "func (q offerQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all for offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o MempoolBinSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), mempoolBinPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"mempool_bin\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, mempoolBinPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in mempoolBin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all mempoolBin\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TicketSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"tickets\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, ticketPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in ticket slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all ticket\")\n\t}\n\treturn rowsAff, nil\n}", "func (o BlockSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"block\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, blockPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in block slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all block\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TrainingCostSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"training_costs\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, trainingCostPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in trainingCost slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all trainingCost\")\n\t}\n\treturn rowsAff, nil\n}", "func (o JetSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `jets` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in jet slice\")\n\t}\n\n\treturn nil\n}", "func (o RawVisitSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"raw_visits\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rawVisitPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rawVisit slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rawVisit\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RestaurantRankSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), restaurantRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"restaurant_rank\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, restaurantRankPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in restaurantRank slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all restaurantRank\")\n\t}\n\treturn rowsAff, nil\n}", "func (o EmailSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"mysql: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `email` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to update all in email slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to retrieve rows affected all in update all email\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RepositorySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `repositories` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in repository slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all repository\")\n\t}\n\treturn rowsAff, nil\n}", "func (q weatherQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all for weather\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected for weather\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o WithdrawalCryptoSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"sqlite3: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"withdrawal_crypto\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to update all in withdrawalCrypto slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to retrieve rows affected all in update all withdrawalCrypto\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"model: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `transaction` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to update all in transaction slice\")\n\t}\n\n\treturn nil\n}", "func (o HoldenAtSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"HoldenAt\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, holdenAtPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in holdenAt slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all holdenAt\")\n\t}\n\treturn rowsAff, nil\n}", "func (o StockKeepingUnitContentSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"db: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockKeepingUnitContentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"stock_keeping_unit_content\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, stockKeepingUnitContentPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to update all in stockKeepingUnitContent slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to retrieve rows affected all in update all stockKeepingUnitContent\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `transactions` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in transaction slice\")\n\t}\n\n\treturn nil\n}", "func (o JobSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"jobs\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, jobPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in job slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all job\")\n\t}\n\treturn rowsAff, nil\n}", "func (o StockSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `stock` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in stock slice\")\n\t}\n\n\treturn nil\n}", "func (q channelQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for channels\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for channels\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFTurntableSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_turntable` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfTurntable slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfTurntable\")\n\t}\n\treturn rowsAff, nil\n}", "func (q blockQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for block\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o EmployeeSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), employeePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"employee\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, employeePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in employee slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all employee\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CurrentChartDataMinutelySlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"current_chart_data_minutely\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, currentChartDataMinutelyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in currentChartDataMinutely slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all currentChartDataMinutely\")\n\t}\n\treturn rowsAff, nil\n}", "func (o ActivityLogSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"dbmodel: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), activityLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"activity_logs\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, activityLogPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update all in activityLog slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to retrieve rows affected all in update all activityLog\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TenantSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"dbmodel: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `tenants` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to update all in tenant slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to retrieve rows affected all in update all tenant\")\n\t}\n\treturn rowsAff, nil\n}", "func (q jobQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RentalSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"sqlboiler: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `rental` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update all in rental slice\")\n\t}\n\n\treturn nil\n}", "func (o PeerSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"model: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `peers` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to update all in peer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to retrieve rows affected all in update all peer\")\n\t}\n\treturn rowsAff, nil\n}", "func (o FriendshipSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `friendship` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in friendship slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all friendship\")\n\t}\n\treturn rowsAff, nil\n}", "func (o TreeSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), treePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"trees\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, treePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in tree slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all tree\")\n\t}\n\treturn rowsAff, nil\n}", "func (o UsernameListingSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), usernameListingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"username_listings\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, usernameListingPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in usernameListing slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all usernameListing\")\n\t}\n\treturn rowsAff, nil\n}", "func (o PremiumSlotSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"premium_slots\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, premiumSlotPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all premiumSlot\")\n\t}\n\treturn rowsAff, nil\n}", "func (q withdrawalCryptoQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to update all for withdrawal_crypto\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to retrieve rows affected for withdrawal_crypto\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q boardQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to update all for boards\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to retrieve rows affected for boards\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q nodeQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for node\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for node\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ItemSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"items\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in item slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all item\")\n\t}\n\treturn rowsAff, nil\n}", "func (q repositoryQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for repositories\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for repositories\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFFamilyUserPoliciesTakeSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_family_user_policies_take` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfFamilyUserPoliciesTake slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfFamilyUserPoliciesTake\")\n\t}\n\treturn rowsAff, nil\n}", "func (o DestinationRankSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), destinationRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"destination_rank\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, destinationRankPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in destinationRank slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all destinationRank\")\n\t}\n\treturn rowsAff, nil\n}", "func (q voteQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o BTCTXOutputSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), btcTXOutputPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `btc_tx_output` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, btcTXOutputPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in btcTXOutput slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all btcTXOutput\")\n\t}\n\treturn rowsAff, nil\n}", "func (o StockCvtermSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"chado: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockCvtermPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"UPDATE \\\"stock_cvterm\\\" SET %s WHERE (\\\"stock_cvterm_id\\\") IN (%s)\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(stockCvtermPrimaryKeyColumns), len(colNames)+1, len(stockCvtermPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to update all in stockCvterm slice\")\n\t}\n\n\treturn nil\n}", "func (o ChatSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"dal: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"chat\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, chatPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to update all in chat slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to retrieve rows affected all in update all chat\")\n\t}\n\treturn rowsAff, nil\n}", "func (o SubscriberSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), subscriberPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `subscribers` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, subscriberPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in subscriber slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all subscriber\")\n\t}\n\treturn rowsAff, nil\n}", "func (o RentalRowerSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"rental_rowers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rentalRowerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rentalRower slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rentalRower\")\n\t}\n\treturn rowsAff, nil\n}", "func (q exchangeCurrencyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for exchange_currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for exchange_currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o IllnessSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"illness\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, illnessPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in illness slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all illness\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CommentSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), commentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `comment` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, commentPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in comment slice\")\n\t}\n\n\treturn nil\n}", "func (o AddressSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"sqlboiler: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), addressPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `address` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, addressPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to update all in address slice\")\n\t}\n\n\treturn nil\n}", "func (o StorestateSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"stellarcore: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storestatePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"storestate\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, storestatePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to update all in storestate slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to retrieve rows affected all in update all storestate\")\n\t}\n\treturn rowsAff, nil\n}", "func (o SegmentSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"boiler: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"segment\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, segmentPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to update all in segment slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to retrieve rows affected all in update all segment\")\n\t}\n\treturn rowsAff, nil\n}", "func (o InstrumentClassSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), instrumentClassPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"instruments\\\".\\\"instrument_class\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, instrumentClassPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in instrumentClass slice\")\n\t}\n\n\treturn nil\n}", "func (o KvstoreSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"kvstore\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, kvstorePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in kvstore slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all kvstore\")\n\t}\n\treturn rowsAff, nil\n}", "func (q utxoQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for utxo\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for utxo\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFUserExperienceLogSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserExperienceLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `cmf_user_experience_log` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserExperienceLogPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in cmfUserExperienceLog slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all cmfUserExperienceLog\")\n\t}\n\treturn rowsAff, nil\n}", "func (q cmfBalanceChargeAdminQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_balance_charge_admin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_balance_charge_admin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PaymentObjectSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `payment_objects` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in paymentObject slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all paymentObject\")\n\t}\n\treturn rowsAff, nil\n}", "func (q holdenAtQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthTokenSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authTokenPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"auth_tokens\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, authTokenPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in authToken slice\")\n\t}\n\n\treturn nil\n}", "func (o RSSAnnouncementSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rssAnnouncementPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"rss_announcements\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, rssAnnouncementPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in rssAnnouncement slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all rssAnnouncement\")\n\t}\n\treturn rowsAff, nil\n}", "func (q ticketQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for tickets\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for tickets\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currentChartDataMinutelyQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for current_chart_data_minutely\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for current_chart_data_minutely\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o NotificationSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"notification\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, notificationPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in notification slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all notification\")\n\t}\n\treturn rowsAff, nil\n}", "func (q jetQuery) UpdateAll(cols M) error {\n\tqueries.SetUpdate(q.Query, cols)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all for jets\")\n\t}\n\n\treturn nil\n}", "func (o OauthClientSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), oauthClientPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `oauth_clients` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, oauthClientPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in oauthClient slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all oauthClient\")\n\t}\n\treturn rowsAff, nil\n}", "func (o PostSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"orm: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"posts\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, postPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to update all in post slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to retrieve rows affected all in update all post\")\n\t}\n\treturn rowsAff, nil\n}", "func (q itemQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for items\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for items\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o LatencySlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), latencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"latencies\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, latencyPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in latency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all latency\")\n\t}\n\treturn rowsAff, nil\n}", "func (o OriginSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), originPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"origins\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, originPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in origin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all origin\")\n\t}\n\treturn rowsAff, nil\n}", "func (o BuildingSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"record: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"buildings\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, buildingPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to update all in building slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to retrieve rows affected all in update all building\")\n\t}\n\treturn rowsAff, nil\n}", "func (q cmfTurntableQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for cmf_turntable\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for cmf_turntable\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FilesStorageSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), filesStoragePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `files_storages` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, filesStoragePrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in filesStorage slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all filesStorage\")\n\t}\n\treturn rowsAff, nil\n}", "func (o DescriptionSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), descriptionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `descriptions` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, descriptionPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in description slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all description\")\n\t}\n\treturn rowsAff, nil\n}", "func (o InventorySlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), inventoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `inventory` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, inventoryPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in inventory slice\")\n\t}\n\n\treturn nil\n}", "func (q kvstoreQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthUserUserPermissionSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserUserPermissionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"UPDATE `auth_user_user_permissions` SET %s WHERE (`id`) IN (%s)\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(authUserUserPermissionPrimaryKeyColumns), len(colNames)+1, len(authUserUserPermissionPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in authUserUserPermission slice\")\n\t}\n\n\treturn nil\n}", "func (o VSPSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"vsp\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, vspPrimaryKeyColumns, len(o)))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vsp slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vsp\")\n\t}\n\treturn rowsAff, nil\n}", "func (o CustomerSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"customers\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, customerPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in customer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all customer\")\n\t}\n\treturn rowsAff, nil\n}", "func (o BookCategoryAssignSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), bookCategoryAssignPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `book_category_assign` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, bookCategoryAssignPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in bookCategoryAssign slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all bookCategoryAssign\")\n\t}\n\treturn rowsAff, nil\n}", "func (q chatQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to update all for chat\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to retrieve rows affected for chat\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q rowerGroupQuery) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthUserSlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"UPDATE `auth_user` SET %s WHERE (`id`) IN (%s)\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(authUserPrimaryKeyColumns), len(colNames)+1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in authUser slice\")\n\t}\n\n\treturn nil\n}", "func (q rawVisitQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for raw_visits\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for raw_visits\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q btcTXOutputQuery) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tqueries.SetUpdate(q.Query, cols)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all for btc_tx_output\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o APIKeySlice) UpdateAll(exec boil.Executor, cols M) error {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"api_keys\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, apiKeyPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to update all in apiKey slice\")\n\t}\n\n\treturn nil\n}", "func (o ContentUnitDerivationSlice) UpdateAll(exec boil.Executor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"mdbmodels: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE \\\"content_unit_derivations\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), len(colNames)+1, contentUnitDerivationPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to update all in contentUnitDerivation slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to retrieve rows affected all in update all contentUnitDerivation\")\n\t}\n\treturn rowsAff, nil\n}", "func (o VendorSlice) UpdateAll(ctx context.Context, exec boil.ContextExecutor, cols M) (int64, error) {\n\tln := int64(len(o))\n\tif ln == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cols) == 0 {\n\t\treturn 0, errors.New(\"models: update all requires at least one column argument\")\n\t}\n\n\tcolNames := make([]string, len(cols))\n\targs := make([]interface{}, len(cols))\n\n\ti := 0\n\tfor name, value := range cols {\n\t\tcolNames[i] = name\n\t\targs[i] = value\n\t\ti++\n\t}\n\n\t// Append all of the primary key values for each column\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\"UPDATE `vendors` SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, colNames),\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(o)))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to update all in vendor slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to retrieve rows affected all in update all vendor\")\n\t}\n\treturn rowsAff, nil\n}" ]
[ "0.7353619", "0.7285703", "0.72350585", "0.7130394", "0.7117237", "0.7108043", "0.7105222", "0.7088896", "0.7074127", "0.70737535", "0.70358866", "0.70282245", "0.7018634", "0.69852245", "0.69732857", "0.6970341", "0.69643295", "0.69515", "0.6948007", "0.69398", "0.69355875", "0.6930986", "0.6926865", "0.69177556", "0.6917679", "0.69148034", "0.6887811", "0.68858343", "0.6884799", "0.68814653", "0.68755686", "0.6867739", "0.6866365", "0.6855836", "0.6847635", "0.6845703", "0.68289685", "0.68288857", "0.68280226", "0.6812352", "0.6810345", "0.68044853", "0.680062", "0.67977643", "0.6796646", "0.67955667", "0.67850304", "0.6784755", "0.67751443", "0.6771545", "0.67640555", "0.6746981", "0.67440206", "0.6742376", "0.67422837", "0.6738283", "0.67359537", "0.67356026", "0.67300636", "0.6721772", "0.6719795", "0.6714271", "0.6713641", "0.67111516", "0.6708665", "0.670802", "0.67077464", "0.6705267", "0.670253", "0.6702079", "0.6692309", "0.66922975", "0.6692191", "0.6690346", "0.6688378", "0.6685846", "0.6679148", "0.6652131", "0.665065", "0.66489285", "0.6647793", "0.6635045", "0.66332984", "0.66324073", "0.6619597", "0.661603", "0.6615817", "0.66149974", "0.6612279", "0.6603562", "0.66014695", "0.6600848", "0.6600813", "0.6599926", "0.65994287", "0.6598781", "0.659756", "0.6594381", "0.6593819", "0.65927184", "0.6591095" ]
0.0
-1
UpsertG attempts an insert, and does an update or ignore on conflict.
func (o *AssetRevision) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error { return o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Store) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Storestate) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *PremiumCode) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *AutomodRuleDatum) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Offer) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Source) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *RentalRower) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Jet) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *PremiumSlot) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Ticket) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *UsernameListing) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *BraceletPhoto) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Auth) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Transaction) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Transaction) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *StockCvterm) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Cvtermsynonym) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *InstrumentClass) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *ExchangeCurrency) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Stock) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *APIKey) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *RowerGroup) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *RSSAnnouncement) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Author) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Failure) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Organism) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *BlackCard) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Skin) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Inventory) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Shelf) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Picture) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Rental) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *FeatureCvtermDbxref) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Vote) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Notification) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Auth) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *AuthToken) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *AuthUser) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Source) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Phenotypeprop) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *InstrumentClass) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *AuthMessage) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *APIKey) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BookCategory) UpsertG(ctx context.Context, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateColumns, insertColumns)\n}", "func (o *Cvtermsynonym) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasure) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Comment) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *BookCategoryAssign) UpsertG(ctx context.Context, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateColumns, insertColumns)\n}", "func (o *BraceletPhoto) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) UpsertG(updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, insertColumns)\n}", "func (o *Jet) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (m *MySQL)UpSert(i interface{})(sql.Result, error){\n\tif err := m.Exists(i); err != nil {\n\t\treturn m.Insert(i)\n\t}\n\treturn m.Update(i)\n}", "func (o *Rental) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *AuthToken) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *UserGoogle) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"model2: no user_google provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(userGoogleColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLUserGoogleUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tuserGoogleUpsertCacheMut.RLock()\n\tcache, cached := userGoogleUpsertCache[key]\n\tuserGoogleUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tuserGoogleAllColumns,\n\t\t\tuserGoogleColumnsWithDefault,\n\t\t\tuserGoogleColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tuserGoogleAllColumns,\n\t\t\tuserGooglePrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"model2: unable to upsert user_google, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`user_google`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `user_google` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(userGoogleType, userGoogleMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(userGoogleType, userGoogleMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to upsert for user_google\")\n\t}\n\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(userGoogleType, userGoogleMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to retrieve unique values for user_google\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to populate default values for user_google\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tuserGoogleUpsertCacheMut.Lock()\n\t\tuserGoogleUpsertCache[key] = cache\n\t\tuserGoogleUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Skin) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (repository *GormRepository) Upsert(uow *UnitOfWork, entity interface{}, queryProcessors []QueryProcessor) microappError.DatabaseError {\n\tdb := uow.DB\n\tif queryProcessors != nil {\n\t\tvar err error\n\t\tfor _, queryProcessor := range queryProcessors {\n\t\t\tdb, err = queryProcessor(db, entity)\n\t\t\tif err != nil {\n\t\t\t\treturn microappError.NewDatabaseError(err)\n\t\t\t}\n\t\t}\n\t}\n\tresult := db.Model(entity).Updates(entity)\n\tif result.Error != nil {\n\t\treturn microappError.NewDatabaseError(result.Error)\n\t}\n\n\tif result.RowsAffected == 0 {\n\t\tif err := uow.DB.Create(entity).Error; err != nil {\n\t\t\treturn microappError.NewDatabaseError(err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *AuthMessage) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasure) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Phenotypeprop) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Kvstore) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no kvstore provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(kvstoreColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tkvstoreUpsertCacheMut.RLock()\n\tcache, cached := kvstoreUpsertCache[key]\n\tkvstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstoreColumnsWithDefault,\n\t\t\tkvstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert kvstore, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(kvstorePrimaryKeyColumns))\n\t\t\tcopy(conflict, kvstorePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"kvstore\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert kvstore\")\n\t}\n\n\tif !cached {\n\t\tkvstoreUpsertCacheMut.Lock()\n\t\tkvstoreUpsertCache[key] = cache\n\t\tkvstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Store) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stores provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstoreUpsertCacheMut.RLock()\n\tcache, cached := storeUpsertCache[key]\n\tstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstoreColumnsWithDefault,\n\t\t\tstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert stores, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storePrimaryKeyColumns))\n\t\t\tcopy(conflict, storePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"stores\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storeType, storeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storeType, storeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert stores\")\n\t}\n\n\tif !cached {\n\t\tstoreUpsertCacheMut.Lock()\n\t\tstoreUpsertCache[key] = cache\n\t\tstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Address) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Job) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jobs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jobColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjobUpsertCacheMut.RLock()\n\tcache, cached := jobUpsertCache[key]\n\tjobUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobColumnsWithDefault,\n\t\t\tjobColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jobs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(jobPrimaryKeyColumns))\n\t\t\tcopy(conflict, jobPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"jobs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jobType, jobMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpsertCacheMut.Lock()\n\t\tjobUpsertCache[key] = cache\n\t\tjobUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Vote) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(votePrimaryKeyColumns))\n\t\t\tcopy(conflict, votePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vote\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vote\")\n\t}\n\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *OauthClient) Upsert(exec boil.Executor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no oauth_clients provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t}\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(oauthClientColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLOauthClientUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\toauthClientUpsertCacheMut.RLock()\n\tcache, cached := oauthClientUpsertCache[key]\n\toauthClientUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientColumnsWithDefault,\n\t\t\toauthClientColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert oauth_clients, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"oauth_clients\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `oauth_clients` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for oauth_clients\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == oauthClientMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(oauthClientType, oauthClientMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for oauth_clients\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\terr = exec.QueryRow(cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for oauth_clients\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\toauthClientUpsertCacheMut.Lock()\n\t\toauthClientUpsertCache[key] = cache\n\t\toauthClientUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Auth) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auths provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.Time.IsZero() {\n\t\to.CreatedAt.Time = currTime\n\t\to.CreatedAt.Valid = true\n\t}\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthUpsertCacheMut.RLock()\n\tcache, cached := authUpsertCache[key]\n\tauthUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthColumnsWithDefault,\n\t\t\tauthColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auths, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(authPrimaryKeyColumns))\n\t\t\tcopy(conflict, authPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"auths\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authType, authMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authType, authMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert auths\")\n\t}\n\n\tif !cached {\n\t\tauthUpsertCacheMut.Lock()\n\t\tauthUpsertCache[key] = cache\n\t\tauthUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (db *DBService) Upsert(jobs ...*DataHistoryJobResult) error {\n\tif len(jobs) == 0 {\n\t\treturn nil\n\t}\n\tctx := context.TODO()\n\n\ttx, err := db.sql.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"beginTx %w\", err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terrRB := tx.Rollback()\n\t\t\tif errRB != nil {\n\t\t\t\tlog.Errorf(log.DatabaseMgr, \"Insert tx.Rollback %v\", errRB)\n\t\t\t}\n\t\t}\n\t}()\n\n\tswitch db.driver {\n\tcase database.DBSQLite3, database.DBSQLite:\n\t\terr = upsertSqlite(ctx, tx, jobs...)\n\tcase database.DBPostgreSQL:\n\t\terr = upsertPostgres(ctx, tx, jobs...)\n\tdefault:\n\t\treturn database.ErrNoDatabaseProvided\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (o *TrainingCost) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no training_costs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(trainingCostColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttrainingCostUpsertCacheMut.RLock()\n\tcache, cached := trainingCostUpsertCache[key]\n\ttrainingCostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostColumnsWithDefault,\n\t\t\ttrainingCostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert training_costs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(trainingCostPrimaryKeyColumns))\n\t\t\tcopy(conflict, trainingCostPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"training_costs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostUpsertCacheMut.Lock()\n\t\ttrainingCostUpsertCache[key] = cache\n\t\ttrainingCostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Task) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tasks provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(taskColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttaskUpsertCacheMut.RLock()\n\tcache, cached := taskUpsertCache[key]\n\ttaskUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskColumnsWithDefault,\n\t\t\ttaskColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tasks, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(taskPrimaryKeyColumns))\n\t\t\tcopy(conflict, taskPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tasks\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(taskType, taskMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskUpsertCacheMut.Lock()\n\t\ttaskUpsertCache[key] = cache\n\t\ttaskUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *AuthItemGroup) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_item_groups provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authItemGroupColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLAuthItemGroupUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthItemGroupUpsertCacheMut.RLock()\n\tcache, cached := authItemGroupUpsertCache[key]\n\tauthItemGroupUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupColumnsWithDefault,\n\t\t\tauthItemGroupColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auth_item_groups, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"auth_item_groups\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `auth_item_groups` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for auth_item_groups\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authItemGroupMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for auth_item_groups\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_item_groups\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthItemGroupUpsertCacheMut.Lock()\n\t\tauthItemGroupUpsertCache[key] = cache\n\t\tauthItemGroupUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (db *DB) Upsert(key interface{}, value interface{}) error {\n\treturn db.bolt.Update(func(tx *bolt.Tx) error {\n\t\treturn db.UpsertTx(tx, key, value)\n\t})\n}", "func (p *PointsSgMutator) DoUpsert() bool { //nolint:dupl false positive\n\t_, err := p.Adapter.Replace(p.SpaceName(), p.ToArray())\n\treturn !L.IsError(err, `PointsSg.DoUpsert failed: `+p.SpaceName())\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no transaction provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"model: unable to upsert transaction, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transaction\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transaction` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to upsert for transaction\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for transaction\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Building) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"record: no buildings provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(buildingColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbuildingUpsertCacheMut.RLock()\n\tcache, cached := buildingUpsertCache[key]\n\tbuildingUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tbuildingColumns,\n\t\t\tbuildingColumnsWithDefault,\n\t\t\tbuildingColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tbuildingColumns,\n\t\t\tbuildingPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"record: unable to upsert buildings, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(buildingPrimaryKeyColumns))\n\t\t\tcopy(conflict, buildingPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"buildings\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(buildingType, buildingMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(buildingType, buildingMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to upsert buildings\")\n\t}\n\n\tif !cached {\n\t\tbuildingUpsertCacheMut.Lock()\n\t\tbuildingUpsertCache[key] = cache\n\t\tbuildingUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Storestate) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no storestate provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storestateColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstorestateUpsertCacheMut.RLock()\n\tcache, cached := storestateUpsertCache[key]\n\tstorestateUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestateColumnsWithDefault,\n\t\t\tstorestateColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestatePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"stellarcore: unable to upsert storestate, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storestatePrimaryKeyColumns))\n\t\t\tcopy(conflict, storestatePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"storestate\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storestateType, storestateMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storestateType, storestateMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to upsert storestate\")\n\t}\n\n\tif !cached {\n\t\tstorestateUpsertCacheMut.Lock()\n\t\tstorestateUpsertCache[key] = cache\n\t\tstorestateUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func Upsert(s Session, dbname string, collection string, selector map[string]interface{}, updator map[string]interface{}) error {\n\t_, err := s.DB(dbname).C(collection).Upsert(selector, updator)\n\treturn err\n}", "func (o *RawVisit) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no raw_visits provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rawVisitColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trawVisitUpsertCacheMut.RLock()\n\tcache, cached := rawVisitUpsertCache[key]\n\trawVisitUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitColumnsWithDefault,\n\t\t\trawVisitColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert raw_visits, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rawVisitPrimaryKeyColumns))\n\t\t\tcopy(conflict, rawVisitPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"raw_visits\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitUpsertCacheMut.Lock()\n\t\trawVisitUpsertCache[key] = cache\n\t\trawVisitUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no transactions provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert transactions, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transactions\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transactions` WHERE `transaction_id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for transactions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.TransactionID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"TransactionID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.TransactionID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for transactions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}" ]
[ "0.77679014", "0.76335657", "0.7589262", "0.7540506", "0.75095516", "0.7490963", "0.74864453", "0.7466031", "0.7457427", "0.7432935", "0.74103045", "0.7405938", "0.7384538", "0.73754615", "0.73754615", "0.73720944", "0.7368038", "0.7342968", "0.7330346", "0.7322401", "0.7322164", "0.73121226", "0.7309201", "0.73089075", "0.7289928", "0.7262276", "0.7255437", "0.7253122", "0.7247316", "0.7228726", "0.72135305", "0.7188866", "0.71741205", "0.71624666", "0.7159617", "0.71080065", "0.70942897", "0.7085319", "0.7076301", "0.7051621", "0.7048401", "0.7047714", "0.7034955", "0.7008088", "0.6950653", "0.6914357", "0.6890965", "0.6883914", "0.6872496", "0.6866803", "0.6864405", "0.68448555", "0.68389577", "0.6828873", "0.6828633", "0.6828633", "0.68270546", "0.6825547", "0.67983085", "0.679643", "0.6746633", "0.67436504", "0.67398214", "0.670185", "0.6686429", "0.6681724", "0.6675501", "0.66740614", "0.6660241", "0.66499114", "0.66383237", "0.6627643", "0.6627514", "0.6568353", "0.65310645", "0.6531011", "0.6523112", "0.6509043", "0.6497557", "0.6470513", "0.64269817", "0.64227194", "0.6393415", "0.63783735", "0.6372924", "0.63696605", "0.63423705", "0.63213086", "0.6283714", "0.62554777", "0.62487286", "0.62288475", "0.62196255", "0.62182885", "0.6213454", "0.62039053", "0.6184752", "0.61775124", "0.6169372", "0.6147309" ]
0.74116534
10
UpsertGP attempts an insert, and does an update or ignore on conflict. Panics on error.
func (o *AssetRevision) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) { if err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Auth) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Phenotypeprop) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PremiumCode) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Shelf) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Store) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *PremiumSlot) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *AutomodRuleDatum) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Jet) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Storestate) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *BraceletPhoto) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Source) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *FeatureRelationship) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *APIKey) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Ticket) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Cvtermsynonym) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Offer) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *RowerGroup) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *RecordMeasure) UpsertGP(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Comment) UpsertGP(updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Transaction) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Transaction) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *UsernameListing) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *ExchangeCurrency) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *UserGoogle) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"model2: no user_google provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(userGoogleColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLUserGoogleUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tuserGoogleUpsertCacheMut.RLock()\n\tcache, cached := userGoogleUpsertCache[key]\n\tuserGoogleUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tuserGoogleAllColumns,\n\t\t\tuserGoogleColumnsWithDefault,\n\t\t\tuserGoogleColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tuserGoogleAllColumns,\n\t\t\tuserGooglePrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"model2: unable to upsert user_google, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`user_google`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `user_google` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(userGoogleType, userGoogleMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(userGoogleType, userGoogleMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to upsert for user_google\")\n\t}\n\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(userGoogleType, userGoogleMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to retrieve unique values for user_google\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to populate default values for user_google\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tuserGoogleUpsertCacheMut.Lock()\n\t\tuserGoogleUpsertCache[key] = cache\n\t\tuserGoogleUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *StockCvterm) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *RSSAnnouncement) UpsertG(ctx context.Context, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *FeatureCvtermDbxref) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *RentalRower) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Phenotypeprop) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *AssetRevision) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Stock) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Failure) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Organism) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Address) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *AuthToken) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *BlackCard) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Notification) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, insertColumns)\n}", "func (o *Inventory) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *AuthUser) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Author) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Vote) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *DMessageEmbed) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *AuthMessage) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Skin) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *AuthItemGroup) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_item_groups provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authItemGroupColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLAuthItemGroupUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthItemGroupUpsertCacheMut.RLock()\n\tcache, cached := authItemGroupUpsertCache[key]\n\tauthItemGroupUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupColumnsWithDefault,\n\t\t\tauthItemGroupColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auth_item_groups, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"auth_item_groups\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `auth_item_groups` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for auth_item_groups\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authItemGroupMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for auth_item_groups\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_item_groups\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthItemGroupUpsertCacheMut.Lock()\n\t\tauthItemGroupUpsertCache[key] = cache\n\t\tauthItemGroupUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (p *PointsSgMutator) DoUpsert() bool { //nolint:dupl false positive\n\t_, err := p.Adapter.Replace(p.SpaceName(), p.ToArray())\n\treturn !L.IsError(err, `PointsSg.DoUpsert failed: `+p.SpaceName())\n}", "func (o *Rental) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *ClaimInList) UpsertGP(updateColumns, insertColumns boil.Columns) {\n\tif err := o.Upsert(boil.GetDB(), updateColumns, insertColumns); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *Kvstore) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no kvstore provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(kvstoreColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tkvstoreUpsertCacheMut.RLock()\n\tcache, cached := kvstoreUpsertCache[key]\n\tkvstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstoreColumnsWithDefault,\n\t\t\tkvstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert kvstore, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(kvstorePrimaryKeyColumns))\n\t\t\tcopy(conflict, kvstorePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"kvstore\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert kvstore\")\n\t}\n\n\tif !cached {\n\t\tkvstoreUpsertCacheMut.Lock()\n\t\tkvstoreUpsertCache[key] = cache\n\t\tkvstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *APIKey) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *Job) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jobs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jobColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjobUpsertCacheMut.RLock()\n\tcache, cached := jobUpsertCache[key]\n\tjobUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobColumnsWithDefault,\n\t\t\tjobColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jobs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(jobPrimaryKeyColumns))\n\t\t\tcopy(conflict, jobPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"jobs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jobType, jobMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpsertCacheMut.Lock()\n\t\tjobUpsertCache[key] = cache\n\t\tjobUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Comment) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (db *DB) Upsert(key interface{}, value interface{}) error {\n\treturn db.bolt.Update(func(tx *bolt.Tx) error {\n\t\treturn db.UpsertTx(tx, key, value)\n\t})\n}", "func (o *AuthUserUserPermission) UpsertG(updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateColumns, whitelist...)\n}", "func (o *InstrumentClass) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (oee *OtxEpubEpub) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif oee._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.otx_epub_epub (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6` +\n\t\t`) ON CONFLICT (document_ptr_id) DO UPDATE SET (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) = (` +\n\t\t`EXCLUDED.document_ptr_id, EXCLUDED.publisher, EXCLUDED.source, EXCLUDED.oebps_folder, EXCLUDED.manifest, EXCLUDED.contents` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\t_, err = db.Exec(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\toee._exists = true\n\n\treturn nil\n}", "func (o *Source) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (repository *GormRepository) Upsert(uow *UnitOfWork, entity interface{}, queryProcessors []QueryProcessor) microappError.DatabaseError {\n\tdb := uow.DB\n\tif queryProcessors != nil {\n\t\tvar err error\n\t\tfor _, queryProcessor := range queryProcessors {\n\t\t\tdb, err = queryProcessor(db, entity)\n\t\t\tif err != nil {\n\t\t\t\treturn microappError.NewDatabaseError(err)\n\t\t\t}\n\t\t}\n\t}\n\tresult := db.Model(entity).Updates(entity)\n\tif result.Error != nil {\n\t\treturn microappError.NewDatabaseError(result.Error)\n\t}\n\n\tif result.RowsAffected == 0 {\n\t\tif err := uow.DB.Create(entity).Error; err != nil {\n\t\t\treturn microappError.NewDatabaseError(err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (db *DBService) Upsert(jobs ...*DataHistoryJobResult) error {\n\tif len(jobs) == 0 {\n\t\treturn nil\n\t}\n\tctx := context.TODO()\n\n\ttx, err := db.sql.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"beginTx %w\", err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terrRB := tx.Rollback()\n\t\t\tif errRB != nil {\n\t\t\t\tlog.Errorf(log.DatabaseMgr, \"Insert tx.Rollback %v\", errRB)\n\t\t\t}\n\t\t}\n\t}()\n\n\tswitch db.driver {\n\tcase database.DBSQLite3, database.DBSQLite:\n\t\terr = upsertSqlite(ctx, tx, jobs...)\n\tcase database.DBPostgreSQL:\n\t\terr = upsertPostgres(ctx, tx, jobs...)\n\tdefault:\n\t\treturn database.ErrNoDatabaseProvided\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (o *Store) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stores provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstoreUpsertCacheMut.RLock()\n\tcache, cached := storeUpsertCache[key]\n\tstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstoreColumnsWithDefault,\n\t\t\tstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert stores, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storePrimaryKeyColumns))\n\t\t\tcopy(conflict, storePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"stores\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storeType, storeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storeType, storeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert stores\")\n\t}\n\n\tif !cached {\n\t\tstoreUpsertCacheMut.Lock()\n\t\tstoreUpsertCache[key] = cache\n\t\tstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Node) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no node provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(nodeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tnodeUpsertCacheMut.RLock()\n\tcache, cached := nodeUpsertCache[key]\n\tnodeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodeColumnsWithDefault,\n\t\t\tnodeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert node, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(nodePrimaryKeyColumns))\n\t\t\tcopy(conflict, nodePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"node\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(nodeType, nodeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(nodeType, nodeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert node\")\n\t}\n\n\tif !cached {\n\t\tnodeUpsertCacheMut.Lock()\n\t\tnodeUpsertCache[key] = cache\n\t\tnodeUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Task) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tasks provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(taskColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttaskUpsertCacheMut.RLock()\n\tcache, cached := taskUpsertCache[key]\n\ttaskUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskColumnsWithDefault,\n\t\t\ttaskColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tasks, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(taskPrimaryKeyColumns))\n\t\t\tcopy(conflict, taskPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tasks\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(taskType, taskMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskUpsertCacheMut.Lock()\n\t\ttaskUpsertCache[key] = cache\n\t\ttaskUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Vote) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(votePrimaryKeyColumns))\n\t\t\tcopy(conflict, votePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vote\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vote\")\n\t}\n\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *BookCategoryAssign) UpsertG(ctx context.Context, updateColumns, insertColumns boil.Columns) error {\n\treturn o.Upsert(ctx, boil.GetContextDB(), updateColumns, insertColumns)\n}", "func (o *Picture) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no pictures provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(pictureColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpictureUpsertCacheMut.RLock()\n\tcache, cached := pictureUpsertCache[key]\n\tpictureUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpictureColumnsWithDefault,\n\t\t\tpictureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpicturePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert pictures, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"pictures\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `pictures` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(pictureType, pictureMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(pictureType, pictureMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for pictures\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for pictures\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpictureUpsertCacheMut.Lock()\n\t\tpictureUpsertCache[key] = cache\n\t\tpictureUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *TrainingCost) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no training_costs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(trainingCostColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttrainingCostUpsertCacheMut.RLock()\n\tcache, cached := trainingCostUpsertCache[key]\n\ttrainingCostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostColumnsWithDefault,\n\t\t\ttrainingCostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert training_costs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(trainingCostPrimaryKeyColumns))\n\t\t\tcopy(conflict, trainingCostPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"training_costs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostUpsertCacheMut.Lock()\n\t\ttrainingCostUpsertCache[key] = cache\n\t\ttrainingCostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *RecordMeasure) UpsertG(updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\treturn o.Upsert(boil.GetDB(), updateOnConflict, conflictColumns, updateColumns, whitelist...)\n}", "func (o *RawVisit) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no raw_visits provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rawVisitColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trawVisitUpsertCacheMut.RLock()\n\tcache, cached := rawVisitUpsertCache[key]\n\trawVisitUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitColumnsWithDefault,\n\t\t\trawVisitColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert raw_visits, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rawVisitPrimaryKeyColumns))\n\t\t\tcopy(conflict, rawVisitPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"raw_visits\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitUpsertCacheMut.Lock()\n\t\trawVisitUpsertCache[key] = cache\n\t\trawVisitUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}" ]
[ "0.7733437", "0.7640093", "0.75672746", "0.7556934", "0.7555976", "0.7523594", "0.7523594", "0.7461622", "0.7455481", "0.74375397", "0.74352455", "0.742481", "0.7375996", "0.734567", "0.732444", "0.732366", "0.73226404", "0.7308071", "0.73027486", "0.72790086", "0.72602683", "0.725878", "0.7241007", "0.7221743", "0.718071", "0.7178772", "0.7074968", "0.70601684", "0.7055122", "0.70409644", "0.69344443", "0.6896031", "0.6890317", "0.68802977", "0.68719137", "0.6867128", "0.68597186", "0.68558323", "0.68557715", "0.68492687", "0.6847846", "0.6841111", "0.6814789", "0.681098", "0.68053734", "0.67940867", "0.6790792", "0.6781078", "0.67522633", "0.67522633", "0.67494696", "0.67433494", "0.6721107", "0.6706449", "0.6704375", "0.67039603", "0.6680255", "0.6677797", "0.66758925", "0.666667", "0.66635513", "0.6662717", "0.6647788", "0.66326094", "0.6624193", "0.66166157", "0.66107035", "0.6576848", "0.6557735", "0.6537507", "0.6500007", "0.6469041", "0.64680374", "0.64573455", "0.64373034", "0.6435111", "0.6361326", "0.6351779", "0.6343012", "0.6320986", "0.63193345", "0.6285653", "0.6270629", "0.6266012", "0.6249909", "0.623631", "0.62276137", "0.62092763", "0.62038666", "0.6201447", "0.6194681", "0.6185531", "0.6178721", "0.61754036", "0.61722344", "0.61535734", "0.61376274", "0.6126189", "0.6123651", "0.6122007" ]
0.7462726
7
UpsertP attempts an insert using an executor, and does an update or ignore on conflict. UpsertP panics on error.
func (o *AssetRevision) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) { if err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(votePrimaryKeyColumns))\n\t\t\tcopy(conflict, votePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vote\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vote\")\n\t}\n\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Offer) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no offers provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(offerColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tofferUpsertCacheMut.RLock()\n\tcache, cached := offerUpsertCache[key]\n\tofferUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferColumnsWithDefault,\n\t\t\tofferColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"stellarcore: unable to upsert offers, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(offerPrimaryKeyColumns))\n\t\t\tcopy(conflict, offerPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"offers\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(offerType, offerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(offerType, offerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to upsert offers\")\n\t}\n\n\tif !cached {\n\t\tofferUpsertCacheMut.Lock()\n\t\tofferUpsertCache[key] = cache\n\t\tofferUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Task) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tasks provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(taskColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttaskUpsertCacheMut.RLock()\n\tcache, cached := taskUpsertCache[key]\n\ttaskUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskColumnsWithDefault,\n\t\t\ttaskColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tasks, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(taskPrimaryKeyColumns))\n\t\t\tcopy(conflict, taskPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tasks\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(taskType, taskMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskUpsertCacheMut.Lock()\n\t\ttaskUpsertCache[key] = cache\n\t\ttaskUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no transaction provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"model: unable to upsert transaction, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transaction\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transaction` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to upsert for transaction\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for transaction\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *AuthUserUserPermission) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jets provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jetColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjetUpsertCacheMut.RLock()\n\tcache, cached := jetUpsertCache[key]\n\tjetUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetColumnsWithDefault,\n\t\t\tjetColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jets, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"jets\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `jets` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jetType, jetMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jetType, jetMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for jets\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for jets\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tjetUpsertCacheMut.Lock()\n\t\tjetUpsertCache[key] = cache\n\t\tjetUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Vote) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"vote\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `vote` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for vote\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == voteMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for vote\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Phenotypeprop) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Kvstore) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no kvstore provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(kvstoreColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tkvstoreUpsertCacheMut.RLock()\n\tcache, cached := kvstoreUpsertCache[key]\n\tkvstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstoreColumnsWithDefault,\n\t\t\tkvstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert kvstore, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(kvstorePrimaryKeyColumns))\n\t\t\tcopy(conflict, kvstorePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"kvstore\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert kvstore\")\n\t}\n\n\tif !cached {\n\t\tkvstoreUpsertCacheMut.Lock()\n\t\tkvstoreUpsertCache[key] = cache\n\t\tkvstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *VSP) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vsp provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(vspColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvspUpsertCacheMut.RLock()\n\tcache, cached := vspUpsertCache[key]\n\tvspUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspColumnsWithDefault,\n\t\t\tvspColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vsp, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(vspPrimaryKeyColumns))\n\t\t\tcopy(conflict, vspPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vsp\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(vspType, vspMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(vspType, vspMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vsp\")\n\t}\n\n\tif !cached {\n\t\tvspUpsertCacheMut.Lock()\n\t\tvspUpsertCache[key] = cache\n\t\tvspUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *FeatureRelationship) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Post) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"orm: no posts provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(postColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpostUpsertCacheMut.RLock()\n\tcache, cached := postUpsertCache[key]\n\tpostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpostColumns,\n\t\t\tpostColumnsWithDefault,\n\t\t\tpostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpostColumns,\n\t\t\tpostPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"orm: unable to upsert posts, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(postPrimaryKeyColumns))\n\t\t\tcopy(conflict, postPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"posts\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(postType, postMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(postType, postMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to upsert posts\")\n\t}\n\n\tif !cached {\n\t\tpostUpsertCacheMut.Lock()\n\t\tpostUpsertCache[key] = cache\n\t\tpostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no transactions provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert transactions, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transactions\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transactions` WHERE `transaction_id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for transactions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.TransactionID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"TransactionID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.TransactionID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for transactions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Failure) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no failure provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(failureColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tfailureUpsertCacheMut.RLock()\n\tcache, cached := failureUpsertCache[key]\n\tfailureUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailureColumnsWithDefault,\n\t\t\tfailureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailurePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert failure, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"failure\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `failure` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(failureType, failureMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(failureType, failureMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for failure\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == failureMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for failure\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfailureUpsertCacheMut.Lock()\n\t\tfailureUpsertCache[key] = cache\n\t\tfailureUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Store) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stores provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstoreUpsertCacheMut.RLock()\n\tcache, cached := storeUpsertCache[key]\n\tstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstoreColumnsWithDefault,\n\t\t\tstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert stores, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storePrimaryKeyColumns))\n\t\t\tcopy(conflict, storePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"stores\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storeType, storeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storeType, storeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert stores\")\n\t}\n\n\tif !cached {\n\t\tstoreUpsertCacheMut.Lock()\n\t\tstoreUpsertCache[key] = cache\n\t\tstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *PremiumSlot) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no premium_slots provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(premiumSlotColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpremiumSlotUpsertCacheMut.RLock()\n\tcache, cached := premiumSlotUpsertCache[key]\n\tpremiumSlotUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotColumnsWithDefault,\n\t\t\tpremiumSlotColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert premium_slots, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(premiumSlotPrimaryKeyColumns))\n\t\t\tcopy(conflict, premiumSlotPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"premium_slots\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert premium_slots\")\n\t}\n\n\tif !cached {\n\t\tpremiumSlotUpsertCacheMut.Lock()\n\t\tpremiumSlotUpsertCache[key] = cache\n\t\tpremiumSlotUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *PremiumCode) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no premium_codes provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(premiumCodeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpremiumCodeUpsertCacheMut.RLock()\n\tcache, cached := premiumCodeUpsertCache[key]\n\tpremiumCodeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpremiumCodeAllColumns,\n\t\t\tpremiumCodeColumnsWithDefault,\n\t\t\tpremiumCodeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpremiumCodeAllColumns,\n\t\t\tpremiumCodePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert premium_codes, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(premiumCodePrimaryKeyColumns))\n\t\t\tcopy(conflict, premiumCodePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"premium_codes\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(premiumCodeType, premiumCodeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(premiumCodeType, premiumCodeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert premium_codes\")\n\t}\n\n\tif !cached {\n\t\tpremiumCodeUpsertCacheMut.Lock()\n\t\tpremiumCodeUpsertCache[key] = cache\n\t\tpremiumCodeUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (oee *OtxEpubEpub) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif oee._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.otx_epub_epub (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6` +\n\t\t`) ON CONFLICT (document_ptr_id) DO UPDATE SET (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) = (` +\n\t\t`EXCLUDED.document_ptr_id, EXCLUDED.publisher, EXCLUDED.source, EXCLUDED.oebps_folder, EXCLUDED.manifest, EXCLUDED.contents` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\t_, err = db.Exec(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\toee._exists = true\n\n\treturn nil\n}", "func (o *Peer) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no peers provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(peerColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLPeerUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpeerUpsertCacheMut.RLock()\n\tcache, cached := peerUpsertCache[key]\n\tpeerUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerColumnsWithDefault,\n\t\t\tpeerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"model: unable to upsert peers, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"peers\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `peers` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(peerType, peerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to upsert for peers\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == peerMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(peerType, peerMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to retrieve unique values for peers\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for peers\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpeerUpsertCacheMut.Lock()\n\t\tpeerUpsertCache[key] = cache\n\t\tpeerUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *RecordMeasure) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Repository) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no repositories provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(repositoryColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLRepositoryUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trepositoryUpsertCacheMut.RLock()\n\tcache, cached := repositoryUpsertCache[key]\n\trepositoryUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trepositoryAllColumns,\n\t\t\trepositoryColumnsWithDefault,\n\t\t\trepositoryColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trepositoryAllColumns,\n\t\t\trepositoryPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert repositories, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"repositories\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `repositories` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(repositoryType, repositoryMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(repositoryType, repositoryMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for repositories\")\n\t}\n\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(repositoryType, repositoryMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for repositories\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for repositories\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\trepositoryUpsertCacheMut.Lock()\n\t\trepositoryUpsertCache[key] = cache\n\t\trepositoryUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Storestate) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no storestate provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storestateColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstorestateUpsertCacheMut.RLock()\n\tcache, cached := storestateUpsertCache[key]\n\tstorestateUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestateColumnsWithDefault,\n\t\t\tstorestateColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestatePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"stellarcore: unable to upsert storestate, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storestatePrimaryKeyColumns))\n\t\t\tcopy(conflict, storestatePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"storestate\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storestateType, storestateMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storestateType, storestateMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to upsert storestate\")\n\t}\n\n\tif !cached {\n\t\tstorestateUpsertCacheMut.Lock()\n\t\tstorestateUpsertCache[key] = cache\n\t\tstorestateUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Weather) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"db: no weather provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(weatherColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tweatherUpsertCacheMut.RLock()\n\tcache, cached := weatherUpsertCache[key]\n\tweatherUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherColumnsWithDefault,\n\t\t\tweatherColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"db: unable to upsert weather, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(weatherPrimaryKeyColumns))\n\t\t\tcopy(conflict, weatherPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"prh\\\".\\\"weather\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(weatherType, weatherMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(weatherType, weatherMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to upsert weather\")\n\t}\n\n\tif !cached {\n\t\tweatherUpsertCacheMut.Lock()\n\t\tweatherUpsertCache[key] = cache\n\t\tweatherUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *RawVisit) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no raw_visits provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rawVisitColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trawVisitUpsertCacheMut.RLock()\n\tcache, cached := rawVisitUpsertCache[key]\n\trawVisitUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitColumnsWithDefault,\n\t\t\trawVisitColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert raw_visits, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rawVisitPrimaryKeyColumns))\n\t\t\tcopy(conflict, rawVisitPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"raw_visits\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitUpsertCacheMut.Lock()\n\t\trawVisitUpsertCache[key] = cache\n\t\trawVisitUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Node) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no node provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(nodeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tnodeUpsertCacheMut.RLock()\n\tcache, cached := nodeUpsertCache[key]\n\tnodeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodeColumnsWithDefault,\n\t\t\tnodeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert node, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(nodePrimaryKeyColumns))\n\t\t\tcopy(conflict, nodePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"node\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(nodeType, nodeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(nodeType, nodeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert node\")\n\t}\n\n\tif !cached {\n\t\tnodeUpsertCacheMut.Lock()\n\t\tnodeUpsertCache[key] = cache\n\t\tnodeUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Auth) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auths provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.Time.IsZero() {\n\t\to.CreatedAt.Time = currTime\n\t\to.CreatedAt.Valid = true\n\t}\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthUpsertCacheMut.RLock()\n\tcache, cached := authUpsertCache[key]\n\tauthUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthColumnsWithDefault,\n\t\t\tauthColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auths, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(authPrimaryKeyColumns))\n\t\t\tcopy(conflict, authPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"auths\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authType, authMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authType, authMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert auths\")\n\t}\n\n\tif !cached {\n\t\tauthUpsertCacheMut.Lock()\n\t\tauthUpsertCache[key] = cache\n\t\tauthUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *CurrentChartDataMinutely) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no current_chart_data_minutely provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currentChartDataMinutelyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcurrentChartDataMinutelyUpsertCacheMut.RLock()\n\tcache, cached := currentChartDataMinutelyUpsertCache[key]\n\tcurrentChartDataMinutelyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tcurrentChartDataMinutelyColumns,\n\t\t\tcurrentChartDataMinutelyColumnsWithDefault,\n\t\t\tcurrentChartDataMinutelyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tcurrentChartDataMinutelyColumns,\n\t\t\tcurrentChartDataMinutelyPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert current_chart_data_minutely, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(currentChartDataMinutelyPrimaryKeyColumns))\n\t\t\tcopy(conflict, currentChartDataMinutelyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"current_chart_data_minutely\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert current_chart_data_minutely\")\n\t}\n\n\tif !cached {\n\t\tcurrentChartDataMinutelyUpsertCacheMut.Lock()\n\t\tcurrentChartDataMinutelyUpsertCache[key] = cache\n\t\tcurrentChartDataMinutelyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Utxo) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no utxo provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(utxoColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tutxoUpsertCacheMut.RLock()\n\tcache, cached := utxoUpsertCache[key]\n\tutxoUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tutxoAllColumns,\n\t\t\tutxoColumnsWithDefault,\n\t\t\tutxoColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tutxoAllColumns,\n\t\t\tutxoPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert utxo, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(utxoPrimaryKeyColumns))\n\t\t\tcopy(conflict, utxoPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"utxo\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(utxoType, utxoMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(utxoType, utxoMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert utxo\")\n\t}\n\n\tif !cached {\n\t\tutxoUpsertCacheMut.Lock()\n\t\tutxoUpsertCache[key] = cache\n\t\tutxoUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Picture) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no pictures provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(pictureColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpictureUpsertCacheMut.RLock()\n\tcache, cached := pictureUpsertCache[key]\n\tpictureUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpictureColumnsWithDefault,\n\t\t\tpictureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpicturePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert pictures, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"pictures\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `pictures` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(pictureType, pictureMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(pictureType, pictureMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for pictures\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for pictures\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpictureUpsertCacheMut.Lock()\n\t\tpictureUpsertCache[key] = cache\n\t\tpictureUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Origin) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no origins provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(originColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\toriginUpsertCacheMut.RLock()\n\tcache, cached := originUpsertCache[key]\n\toriginUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\toriginColumns,\n\t\t\toriginColumnsWithDefault,\n\t\t\toriginColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\toriginColumns,\n\t\t\toriginPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert origins, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(originPrimaryKeyColumns))\n\t\t\tcopy(conflict, originPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"origins\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(originType, originMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(originType, originMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert origins\")\n\t}\n\n\tif !cached {\n\t\toriginUpsertCacheMut.Lock()\n\t\toriginUpsertCache[key] = cache\n\t\toriginUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *OauthClient) Upsert(exec boil.Executor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no oauth_clients provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t}\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(oauthClientColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLOauthClientUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\toauthClientUpsertCacheMut.RLock()\n\tcache, cached := oauthClientUpsertCache[key]\n\toauthClientUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientColumnsWithDefault,\n\t\t\toauthClientColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert oauth_clients, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"oauth_clients\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `oauth_clients` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for oauth_clients\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == oauthClientMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(oauthClientType, oauthClientMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for oauth_clients\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\terr = exec.QueryRow(cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for oauth_clients\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\toauthClientUpsertCacheMut.Lock()\n\t\toauthClientUpsertCache[key] = cache\n\t\toauthClientUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *ClaimInList) UpsertP(exec boil.Executor, updateColumns, insertColumns boil.Columns) {\n\tif err := o.Upsert(exec, updateColumns, insertColumns); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (m *MySQL)UpSert(i interface{})(sql.Result, error){\n\tif err := m.Exists(i); err != nil {\n\t\treturn m.Insert(i)\n\t}\n\treturn m.Update(i)\n}", "func (o *Inventory) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no inventory provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(inventoryColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tinventoryUpsertCacheMut.RLock()\n\tcache, cached := inventoryUpsertCache[key]\n\tinventoryUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryColumnsWithDefault,\n\t\t\tinventoryColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert inventory, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"inventory\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `inventory` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(inventoryType, inventoryMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(inventoryType, inventoryMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for inventory\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == inventoryMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for inventory\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tinventoryUpsertCacheMut.Lock()\n\t\tinventoryUpsertCache[key] = cache\n\t\tinventoryUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Job) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jobs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jobColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjobUpsertCacheMut.RLock()\n\tcache, cached := jobUpsertCache[key]\n\tjobUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobColumnsWithDefault,\n\t\t\tjobColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jobs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(jobPrimaryKeyColumns))\n\t\t\tcopy(conflict, jobPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"jobs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jobType, jobMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpsertCacheMut.Lock()\n\t\tjobUpsertCache[key] = cache\n\t\tjobUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *APIKey) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no api_keys provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(apiKeyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tapiKeyUpsertCacheMut.RLock()\n\tcache, cached := apiKeyUpsertCache[key]\n\tapiKeyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyColumnsWithDefault,\n\t\t\tapiKeyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert api_keys, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(apiKeyPrimaryKeyColumns))\n\t\t\tcopy(conflict, apiKeyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"api_keys\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert api_keys\")\n\t}\n\n\tif !cached {\n\t\tapiKeyUpsertCacheMut.Lock()\n\t\tapiKeyUpsertCache[key] = cache\n\t\tapiKeyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *HoldenAt) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no HoldenAt provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(holdenAtColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tholdenAtUpsertCacheMut.RLock()\n\tcache, cached := holdenAtUpsertCache[key]\n\tholdenAtUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtColumnsWithDefault,\n\t\t\tholdenAtColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert HoldenAt, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(holdenAtPrimaryKeyColumns))\n\t\t\tcopy(conflict, holdenAtPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"HoldenAt\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert HoldenAt\")\n\t}\n\n\tif !cached {\n\t\tholdenAtUpsertCacheMut.Lock()\n\t\tholdenAtUpsertCache[key] = cache\n\t\tholdenAtUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *PublisherSearchIdx) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no publisher_search_idx provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(publisherSearchIdxColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpublisherSearchIdxUpsertCacheMut.RLock()\n\tcache, cached := publisherSearchIdxUpsertCache[key]\n\tpublisherSearchIdxUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpublisherSearchIdxAllColumns,\n\t\t\tpublisherSearchIdxColumnsWithDefault,\n\t\t\tpublisherSearchIdxColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpublisherSearchIdxAllColumns,\n\t\t\tpublisherSearchIdxPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert publisher_search_idx, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(publisherSearchIdxPrimaryKeyColumns))\n\t\t\tcopy(conflict, publisherSearchIdxPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQuerySQLite(dialect, \"\\\"publisher_search_idx\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(publisherSearchIdxType, publisherSearchIdxMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(publisherSearchIdxType, publisherSearchIdxMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert publisher_search_idx\")\n\t}\n\n\tif !cached {\n\t\tpublisherSearchIdxUpsertCacheMut.Lock()\n\t\tpublisherSearchIdxUpsertCache[key] = cache\n\t\tpublisherSearchIdxUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *BTCTXOutput) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no btc_tx_output provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(btcTXOutputColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLBTCTXOutputUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbtcTXOutputUpsertCacheMut.RLock()\n\tcache, cached := btcTXOutputUpsertCache[key]\n\tbtcTXOutputUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tbtcTXOutputAllColumns,\n\t\t\tbtcTXOutputColumnsWithDefault,\n\t\t\tbtcTXOutputColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tbtcTXOutputAllColumns,\n\t\t\tbtcTXOutputPrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert btc_tx_output, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`btc_tx_output`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `btc_tx_output` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for btc_tx_output\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == btcTXOutputMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for btc_tx_output\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for btc_tx_output\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tbtcTXOutputUpsertCacheMut.Lock()\n\t\tbtcTXOutputUpsertCache[key] = cache\n\t\tbtcTXOutputUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Block) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no block provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(blockColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tblockUpsertCacheMut.RLock()\n\tcache, cached := blockUpsertCache[key]\n\tblockUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockColumnsWithDefault,\n\t\t\tblockColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert block, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(blockPrimaryKeyColumns))\n\t\t\tcopy(conflict, blockPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"block\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(blockType, blockMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(blockType, blockMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert block\")\n\t}\n\n\tif !cached {\n\t\tblockUpsertCacheMut.Lock()\n\t\tblockUpsertCache[key] = cache\n\t\tblockUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Subscriber) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no subscribers provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(subscriberColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLSubscriberUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsubscriberUpsertCacheMut.RLock()\n\tcache, cached := subscriberUpsertCache[key]\n\tsubscriberUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsubscriberColumns,\n\t\t\tsubscriberColumnsWithDefault,\n\t\t\tsubscriberColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsubscriberColumns,\n\t\t\tsubscriberPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert subscribers, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"subscribers\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `subscribers` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(subscriberType, subscriberMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(subscriberType, subscriberMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for subscribers\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == subscriberMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(subscriberType, subscriberMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for subscribers\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for subscribers\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tsubscriberUpsertCacheMut.Lock()\n\t\tsubscriberUpsertCache[key] = cache\n\t\tsubscriberUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *PeerProperty) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no peer_properties provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(peerPropertyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpeerPropertyUpsertCacheMut.RLock()\n\tcache, cached := peerPropertyUpsertCache[key]\n\tpeerPropertyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpeerPropertyAllColumns,\n\t\t\tpeerPropertyColumnsWithDefault,\n\t\t\tpeerPropertyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpeerPropertyAllColumns,\n\t\t\tpeerPropertyPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert peer_properties, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(peerPropertyPrimaryKeyColumns))\n\t\t\tcopy(conflict, peerPropertyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"peer_properties\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(peerPropertyType, peerPropertyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(peerPropertyType, peerPropertyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert peer_properties\")\n\t}\n\n\tif !cached {\n\t\tpeerPropertyUpsertCacheMut.Lock()\n\t\tpeerPropertyUpsertCache[key] = cache\n\t\tpeerPropertyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (db *DBService) Upsert(jobs ...*DataHistoryJobResult) error {\n\tif len(jobs) == 0 {\n\t\treturn nil\n\t}\n\tctx := context.TODO()\n\n\ttx, err := db.sql.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"beginTx %w\", err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terrRB := tx.Rollback()\n\t\t\tif errRB != nil {\n\t\t\t\tlog.Errorf(log.DatabaseMgr, \"Insert tx.Rollback %v\", errRB)\n\t\t\t}\n\t\t}\n\t}()\n\n\tswitch db.driver {\n\tcase database.DBSQLite3, database.DBSQLite:\n\t\terr = upsertSqlite(ctx, tx, jobs...)\n\tcase database.DBPostgreSQL:\n\t\terr = upsertPostgres(ctx, tx, jobs...)\n\tdefault:\n\t\treturn database.ErrNoDatabaseProvided\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (o *Source) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmodels: no sources provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(sourceColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsourceUpsertCacheMut.RLock()\n\tcache, cached := sourceUpsertCache[key]\n\tsourceUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsourceAllColumns,\n\t\t\tsourceColumnsWithDefault,\n\t\t\tsourceColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsourceAllColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmodels: unable to upsert sources, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(sourcePrimaryKeyColumns))\n\t\t\tcopy(conflict, sourcePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"sources\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(sourceType, sourceMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to upsert sources\")\n\t}\n\n\tif !cached {\n\t\tsourceUpsertCacheMut.Lock()\n\t\tsourceUpsertCache[key] = cache\n\t\tsourceUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Source) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no sources provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(sourceColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsourceUpsertCacheMut.RLock()\n\tcache, cached := sourceUpsertCache[key]\n\tsourceUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourceColumnsWithDefault,\n\t\t\tsourceColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to upsert sources, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(sourcePrimaryKeyColumns))\n\t\t\tcopy(conflict, sourcePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"sources\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(sourceType, sourceMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to upsert sources\")\n\t}\n\n\tif !cached {\n\t\tsourceUpsertCacheMut.Lock()\n\t\tsourceUpsertCache[key] = cache\n\t\tsourceUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (repository *GormRepository) Upsert(uow *UnitOfWork, entity interface{}, queryProcessors []QueryProcessor) microappError.DatabaseError {\n\tdb := uow.DB\n\tif queryProcessors != nil {\n\t\tvar err error\n\t\tfor _, queryProcessor := range queryProcessors {\n\t\t\tdb, err = queryProcessor(db, entity)\n\t\t\tif err != nil {\n\t\t\t\treturn microappError.NewDatabaseError(err)\n\t\t\t}\n\t\t}\n\t}\n\tresult := db.Model(entity).Updates(entity)\n\tif result.Error != nil {\n\t\treturn microappError.NewDatabaseError(result.Error)\n\t}\n\n\tif result.RowsAffected == 0 {\n\t\tif err := uow.DB.Create(entity).Error; err != nil {\n\t\t\treturn microappError.NewDatabaseError(err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *RecipeLipid) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no recipe_lipid provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(recipeLipidColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trecipeLipidUpsertCacheMut.RLock()\n\tcache, cached := recipeLipidUpsertCache[key]\n\trecipeLipidUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trecipeLipidAllColumns,\n\t\t\trecipeLipidColumnsWithDefault,\n\t\t\trecipeLipidColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trecipeLipidAllColumns,\n\t\t\trecipeLipidPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert recipe_lipid, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(recipeLipidPrimaryKeyColumns))\n\t\t\tcopy(conflict, recipeLipidPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"recipe_lipid\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(recipeLipidType, recipeLipidMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(recipeLipidType, recipeLipidMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert recipe_lipid\")\n\t}\n\n\tif !cached {\n\t\trecipeLipidUpsertCacheMut.Lock()\n\t\trecipeLipidUpsertCache[key] = cache\n\t\trecipeLipidUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Phenotypeprop) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no phenotypeprop provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(phenotypepropColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tphenotypepropUpsertCacheMut.RLock()\n\tcache, cached := phenotypepropUpsertCache[key]\n\tphenotypepropUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tphenotypepropColumns,\n\t\t\tphenotypepropColumnsWithDefault,\n\t\t\tphenotypepropColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tphenotypepropColumns,\n\t\t\tphenotypepropPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"chado: unable to upsert phenotypeprop, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(phenotypepropPrimaryKeyColumns))\n\t\t\tcopy(conflict, phenotypepropPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"phenotypeprop\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(phenotypepropType, phenotypepropMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(phenotypepropType, phenotypepropMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to upsert for phenotypeprop\")\n\t}\n\n\tif !cached {\n\t\tphenotypepropUpsertCacheMut.Lock()\n\t\tphenotypepropUpsertCache[key] = cache\n\t\tphenotypepropUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (s *State) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif s._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.state (` +\n\t\t`\"id\", \"namespace\", \"context_id\", \"created_at\", \"state\", \"data\", \"event\", \"processing_error\"` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6, $7, $8` +\n\t\t`) ON CONFLICT (\"id\") DO UPDATE SET (` +\n\t\t`\"id\", \"namespace\", \"context_id\", \"created_at\", \"state\", \"data\", \"event\", \"processing_error\"` +\n\t\t`) = (` +\n\t\t`EXCLUDED.\"id\", EXCLUDED.\"namespace\", EXCLUDED.\"context_id\", EXCLUDED.\"created_at\", EXCLUDED.\"state\", EXCLUDED.\"data\", EXCLUDED.\"event\", EXCLUDED.\"processing_error\"` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, s.ID, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError)\n\t_, err = db.Exec(sqlstr, s.ID, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ts._exists = true\n\n\treturn nil\n}", "func (o *RentalRower) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no rental_rowers provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rentalRowerColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trentalRowerUpsertCacheMut.RLock()\n\tcache, cached := rentalRowerUpsertCache[key]\n\trentalRowerUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trentalRowerColumns,\n\t\t\trentalRowerColumnsWithDefault,\n\t\t\trentalRowerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trentalRowerColumns,\n\t\t\trentalRowerPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert rental_rowers, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rentalRowerPrimaryKeyColumns))\n\t\t\tcopy(conflict, rentalRowerPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"rental_rowers\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rentalRowerType, rentalRowerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rentalRowerType, rentalRowerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert rental_rowers\")\n\t}\n\n\tif !cached {\n\t\trentalRowerUpsertCacheMut.Lock()\n\t\trentalRowerUpsertCache[key] = cache\n\t\trentalRowerUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Ticket) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tickets provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(ticketColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tticketUpsertCacheMut.RLock()\n\tcache, cached := ticketUpsertCache[key]\n\tticketUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketColumnsWithDefault,\n\t\t\tticketColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tickets, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(ticketPrimaryKeyColumns))\n\t\t\tcopy(conflict, ticketPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tickets\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(ticketType, ticketMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(ticketType, ticketMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tickets\")\n\t}\n\n\tif !cached {\n\t\tticketUpsertCacheMut.Lock()\n\t\tticketUpsertCache[key] = cache\n\t\tticketUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *AuthItemGroup) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_item_groups provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authItemGroupColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLAuthItemGroupUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthItemGroupUpsertCacheMut.RLock()\n\tcache, cached := authItemGroupUpsertCache[key]\n\tauthItemGroupUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupColumnsWithDefault,\n\t\t\tauthItemGroupColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auth_item_groups, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"auth_item_groups\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `auth_item_groups` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for auth_item_groups\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authItemGroupMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for auth_item_groups\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_item_groups\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthItemGroupUpsertCacheMut.Lock()\n\t\tauthItemGroupUpsertCache[key] = cache\n\t\tauthItemGroupUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *BraceletPhoto) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no bracelet_photo provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(braceletPhotoColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbraceletPhotoUpsertCacheMut.RLock()\n\tcache, cached := braceletPhotoUpsertCache[key]\n\tbraceletPhotoUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tbraceletPhotoColumns,\n\t\t\tbraceletPhotoColumnsWithDefault,\n\t\t\tbraceletPhotoColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tbraceletPhotoColumns,\n\t\t\tbraceletPhotoPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert bracelet_photo, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"bracelet_photo\", update, whitelist)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `bracelet_photo` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(braceletPhotoType, braceletPhotoMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(braceletPhotoType, braceletPhotoMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for bracelet_photo\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == braceletPhotoMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for bracelet_photo\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tbraceletPhotoUpsertCacheMut.Lock()\n\t\tbraceletPhotoUpsertCache[key] = cache\n\t\tbraceletPhotoUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Currency) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no currency provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currencyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcurrencyUpsertCacheMut.RLock()\n\tcache, cached := currencyUpsertCache[key]\n\tcurrencyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyColumnsWithDefault,\n\t\t\tcurrencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert currency, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(currencyPrimaryKeyColumns))\n\t\t\tcopy(conflict, currencyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"currency\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currencyType, currencyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(currencyType, currencyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert currency\")\n\t}\n\n\tif !cached {\n\t\tcurrencyUpsertCacheMut.Lock()\n\t\tcurrencyUpsertCache[key] = cache\n\t\tcurrencyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Author) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no authors provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authorColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthorUpsertCacheMut.RLock()\n\tcache, cached := authorUpsertCache[key]\n\tauthorUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorColumnsWithDefault,\n\t\t\tauthorColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to upsert authors, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(authorPrimaryKeyColumns))\n\t\t\tcopy(conflict, authorPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"authors\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authorType, authorMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authorType, authorMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to upsert authors\")\n\t}\n\n\tif !cached {\n\t\tauthorUpsertCacheMut.Lock()\n\t\tauthorUpsertCache[key] = cache\n\t\tauthorUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (a *Actor) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif a._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.actor (` +\n\t\t`actor_id, first_name, last_name, last_update` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4` +\n\t\t`) ON CONFLICT (actor_id) DO UPDATE SET (` +\n\t\t`actor_id, first_name, last_name, last_update` +\n\t\t`) = (` +\n\t\t`EXCLUDED.actor_id, EXCLUDED.first_name, EXCLUDED.last_name, EXCLUDED.last_update` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate)\n\t_, err = db.Exec(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ta._exists = true\n\n\treturn nil\n}", "func (o *Email) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"mysql: no email provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(emailColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLEmailUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\temailUpsertCacheMut.RLock()\n\tcache, cached := emailUpsertCache[key]\n\temailUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\temailColumns,\n\t\t\temailColumnsWithDefault,\n\t\t\temailColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\temailColumns,\n\t\t\temailPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mysql: unable to upsert email, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"email\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `email` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(emailType, emailMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(emailType, emailMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to upsert for email\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == emailMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(emailType, emailMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to retrieve unique values for email\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to populate default values for email\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\temailUpsertCacheMut.Lock()\n\t\temailUpsertCache[key] = cache\n\t\temailUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Board) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"rdb: no boards provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(boardColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLBoardUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tboardUpsertCacheMut.RLock()\n\tcache, cached := boardUpsertCache[key]\n\tboardUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardColumnsWithDefault,\n\t\t\tboardColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"rdb: unable to upsert boards, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"boards\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `boards` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(boardType, boardMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(boardType, boardMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to upsert for boards\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == boardMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(boardType, boardMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to retrieve unique values for boards\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to populate default values for boards\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tboardUpsertCacheMut.Lock()\n\t\tboardUpsertCache[key] = cache\n\t\tboardUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Latency) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no latencies provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(latencyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tlatencyUpsertCacheMut.RLock()\n\tcache, cached := latencyUpsertCache[key]\n\tlatencyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tlatencyAllColumns,\n\t\t\tlatencyColumnsWithDefault,\n\t\t\tlatencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tlatencyAllColumns,\n\t\t\tlatencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert latencies, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(latencyPrimaryKeyColumns))\n\t\t\tcopy(conflict, latencyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"latencies\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(latencyType, latencyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(latencyType, latencyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert latencies\")\n\t}\n\n\tif !cached {\n\t\tlatencyUpsertCacheMut.Lock()\n\t\tlatencyUpsertCache[key] = cache\n\t\tlatencyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *DestinationRank) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no destination_rank provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(destinationRankColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tdestinationRankUpsertCacheMut.RLock()\n\tcache, cached := destinationRankUpsertCache[key]\n\tdestinationRankUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tdestinationRankAllColumns,\n\t\t\tdestinationRankColumnsWithDefault,\n\t\t\tdestinationRankColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tdestinationRankAllColumns,\n\t\t\tdestinationRankPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert destination_rank, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(destinationRankPrimaryKeyColumns))\n\t\t\tcopy(conflict, destinationRankPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"destination_rank\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(destinationRankType, destinationRankMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(destinationRankType, destinationRankMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert destination_rank\")\n\t}\n\n\tif !cached {\n\t\tdestinationRankUpsertCacheMut.Lock()\n\t\tdestinationRankUpsertCache[key] = cache\n\t\tdestinationRankUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *TrainingCost) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no training_costs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(trainingCostColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttrainingCostUpsertCacheMut.RLock()\n\tcache, cached := trainingCostUpsertCache[key]\n\ttrainingCostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostColumnsWithDefault,\n\t\t\ttrainingCostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert training_costs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(trainingCostPrimaryKeyColumns))\n\t\t\tcopy(conflict, trainingCostPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"training_costs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostUpsertCacheMut.Lock()\n\t\ttrainingCostUpsertCache[key] = cache\n\t\ttrainingCostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *InstrumentClass) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no instrument_class provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(instrumentClassColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tinstrumentClassUpsertCacheMut.RLock()\n\tcache, cached := instrumentClassUpsertCache[key]\n\tinstrumentClassUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassColumnsWithDefault,\n\t\t\tinstrumentClassColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert instrument_class, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(instrumentClassPrimaryKeyColumns))\n\t\t\tcopy(conflict, instrumentClassPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"instruments\\\".\\\"instrument_class\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert instrument_class\")\n\t}\n\n\tif !cached {\n\t\tinstrumentClassUpsertCacheMut.Lock()\n\t\tinstrumentClassUpsertCache[key] = cache\n\t\tinstrumentClassUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *RSSAnnouncement) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no rss_announcements provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rssAnnouncementColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trssAnnouncementUpsertCacheMut.RLock()\n\tcache, cached := rssAnnouncementUpsertCache[key]\n\trssAnnouncementUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trssAnnouncementAllColumns,\n\t\t\trssAnnouncementColumnsWithDefault,\n\t\t\trssAnnouncementColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trssAnnouncementAllColumns,\n\t\t\trssAnnouncementPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert rss_announcements, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rssAnnouncementPrimaryKeyColumns))\n\t\t\tcopy(conflict, rssAnnouncementPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"rss_announcements\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rssAnnouncementType, rssAnnouncementMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rssAnnouncementType, rssAnnouncementMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert rss_announcements\")\n\t}\n\n\tif !cached {\n\t\trssAnnouncementUpsertCacheMut.Lock()\n\t\trssAnnouncementUpsertCache[key] = cache\n\t\trssAnnouncementUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (db *DB) Upsert(key interface{}, value interface{}) error {\n\treturn db.bolt.Update(func(tx *bolt.Tx) error {\n\t\treturn db.UpsertTx(tx, key, value)\n\t})\n}", "func (o *MempoolBin) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no mempool_bin provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(mempoolBinColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tmempoolBinUpsertCacheMut.RLock()\n\tcache, cached := mempoolBinUpsertCache[key]\n\tmempoolBinUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tmempoolBinAllColumns,\n\t\t\tmempoolBinColumnsWithDefault,\n\t\t\tmempoolBinColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tmempoolBinAllColumns,\n\t\t\tmempoolBinPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert mempool_bin, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(mempoolBinPrimaryKeyColumns))\n\t\t\tcopy(conflict, mempoolBinPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"mempool_bin\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(mempoolBinType, mempoolBinMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(mempoolBinType, mempoolBinMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert mempool_bin\")\n\t}\n\n\tif !cached {\n\t\tmempoolBinUpsertCacheMut.Lock()\n\t\tmempoolBinUpsertCache[key] = cache\n\t\tmempoolBinUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Tree) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no trees provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(treeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttreeUpsertCacheMut.RLock()\n\tcache, cached := treeUpsertCache[key]\n\ttreeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttreeAllColumns,\n\t\t\ttreeColumnsWithDefault,\n\t\t\ttreeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttreeAllColumns,\n\t\t\ttreePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert trees, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(treePrimaryKeyColumns))\n\t\t\tcopy(conflict, treePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"trees\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(treeType, treeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(treeType, treeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert trees\")\n\t}\n\n\tif !cached {\n\t\ttreeUpsertCacheMut.Lock()\n\t\ttreeUpsertCache[key] = cache\n\t\ttreeUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Doc) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no doc provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(docColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLDocUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tdocUpsertCacheMut.RLock()\n\tcache, cached := docUpsertCache[key]\n\tdocUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tdocAllColumns,\n\t\t\tdocColumnsWithDefault,\n\t\t\tdocColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tdocAllColumns,\n\t\t\tdocPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert doc, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"doc\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `doc` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(docType, docMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(docType, docMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for doc\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == docMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(docType, docMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for doc\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for doc\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tdocUpsertCacheMut.Lock()\n\t\tdocUpsertCache[key] = cache\n\t\tdocUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *PaymentObject) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no payment_objects provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(paymentObjectColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLPaymentObjectUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpaymentObjectUpsertCacheMut.RLock()\n\tcache, cached := paymentObjectUpsertCache[key]\n\tpaymentObjectUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectColumnsWithDefault,\n\t\t\tpaymentObjectColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert payment_objects, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`payment_objects`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `payment_objects` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for payment_objects\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == paymentObjectMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for payment_objects\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for payment_objects\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpaymentObjectUpsertCacheMut.Lock()\n\t\tpaymentObjectUpsertCache[key] = cache\n\t\tpaymentObjectUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *FilesStorage) Upsert(exec boil.Executor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no files_storages provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(filesStorageColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLFilesStorageUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tfilesStorageUpsertCacheMut.RLock()\n\tcache, cached := filesStorageUpsertCache[key]\n\tfilesStorageUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tfilesStorageColumns,\n\t\t\tfilesStorageColumnsWithDefault,\n\t\t\tfilesStorageColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tfilesStorageColumns,\n\t\t\tfilesStoragePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert files_storages, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"files_storages\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `files_storages` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(filesStorageType, filesStorageMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(filesStorageType, filesStorageMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for files_storages\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == filesStorageMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(filesStorageType, filesStorageMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for files_storages\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for files_storages\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfilesStorageUpsertCacheMut.Lock()\n\t\tfilesStorageUpsertCache[key] = cache\n\t\tfilesStorageUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (ds *DjangoSession) Upsert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase ds._deleted: // deleted\n\t\treturn logerror(&ErrUpsertFailed{ErrMarkedForDeletion})\n\t}\n\t// upsert\n\tconst sqlstr = `MERGE django.django_sessiont ` +\n\t\t`USING (` +\n\t\t`SELECT :1 session_key, :2 session_data, :3 expire_date ` +\n\t\t`FROM DUAL ) s ` +\n\t\t`ON s.session_key = t.session_key ` +\n\t\t`WHEN MATCHED THEN ` +\n\t\t`UPDATE SET ` +\n\t\t`t.session_data = s.session_data, t.expire_date = s.expire_date ` +\n\t\t`WHEN NOT MATCHED THEN ` +\n\t\t`INSERT (` +\n\t\t`session_key, session_data, expire_date` +\n\t\t`) VALUES (` +\n\t\t`s.session_key, s.session_data, s.expire_date` +\n\t\t`);`\n\t// run\n\tlogf(sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate)\n\tif _, err := db.ExecContext(ctx, sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate); err != nil {\n\t\treturn logerror(err)\n\t}\n\t// set exists\n\tds._exists = true\n\treturn nil\n}", "func (o *Project) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no projects provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(projectColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLProjectUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tprojectUpsertCacheMut.RLock()\n\tcache, cached := projectUpsertCache[key]\n\tprojectUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tprojectAllColumns,\n\t\t\tprojectColumnsWithDefault,\n\t\t\tprojectColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tprojectAllColumns,\n\t\t\tprojectPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert projects, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"projects\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `projects` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(projectType, projectMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(projectType, projectMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for projects\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == projectMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(projectType, projectMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for projects\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for projects\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tprojectUpsertCacheMut.Lock()\n\t\tprojectUpsertCache[key] = cache\n\t\tprojectUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Segment) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"boiler: no segment provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(segmentColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsegmentUpsertCacheMut.RLock()\n\tcache, cached := segmentUpsertCache[key]\n\tsegmentUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsegmentAllColumns,\n\t\t\tsegmentColumnsWithDefault,\n\t\t\tsegmentColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsegmentAllColumns,\n\t\t\tsegmentPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"boiler: unable to upsert segment, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(segmentPrimaryKeyColumns))\n\t\t\tcopy(conflict, segmentPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"segment\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(segmentType, segmentMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(segmentType, segmentMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to upsert segment\")\n\t}\n\n\tif !cached {\n\t\tsegmentUpsertCacheMut.Lock()\n\t\tsegmentUpsertCache[key] = cache\n\t\tsegmentUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}" ]
[ "0.77784735", "0.77784735", "0.77512676", "0.7726924", "0.76303315", "0.7577626", "0.7575267", "0.7478117", "0.7474516", "0.7442074", "0.7431867", "0.74048245", "0.7382313", "0.73467124", "0.7344757", "0.7247732", "0.72265697", "0.72006553", "0.71860015", "0.71807057", "0.71773845", "0.71540743", "0.71434146", "0.7121397", "0.7115753", "0.7102843", "0.7083053", "0.7072151", "0.69327784", "0.6852024", "0.68133914", "0.67739636", "0.6771646", "0.6740653", "0.6712392", "0.6704586", "0.6703198", "0.6702578", "0.6690043", "0.66868484", "0.6680748", "0.66786397", "0.66747105", "0.6666932", "0.6658183", "0.6646749", "0.6620963", "0.6619499", "0.6599067", "0.65851533", "0.65401155", "0.65378547", "0.6534207", "0.6487538", "0.6487266", "0.64867246", "0.6476725", "0.6455898", "0.6434498", "0.6419876", "0.64116997", "0.6402408", "0.6402114", "0.64001745", "0.63895506", "0.6380112", "0.637426", "0.6374256", "0.63622326", "0.63264763", "0.63216764", "0.6296512", "0.6274215", "0.6273176", "0.62665915", "0.62489116", "0.6236984", "0.6231978", "0.6229136", "0.6228264", "0.6225895", "0.6224675", "0.62163943", "0.62109387", "0.6206338", "0.6189859", "0.6188925", "0.6186011", "0.61843824", "0.6180214", "0.6168418", "0.61654466", "0.61608523", "0.6158614", "0.6157961", "0.61381036", "0.6136549", "0.6135764", "0.6135004", "0.61312395" ]
0.7367751
13
Upsert attempts an insert using an executor, and does an update or ignore on conflict.
func (o *AssetRevision) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error { if o == nil { return errors.New("public: no asset_revision provided for upsert") } if err := o.doBeforeUpsertHooks(exec); err != nil { return err } nzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o) // Build cache key in-line uglily - mysql vs postgres problems buf := strmangle.GetBuffer() if updateOnConflict { buf.WriteByte('t') } else { buf.WriteByte('f') } buf.WriteByte('.') for _, c := range conflictColumns { buf.WriteString(c) } buf.WriteByte('.') for _, c := range updateColumns { buf.WriteString(c) } buf.WriteByte('.') for _, c := range whitelist { buf.WriteString(c) } buf.WriteByte('.') for _, c := range nzDefaults { buf.WriteString(c) } key := buf.String() strmangle.PutBuffer(buf) assetRevisionUpsertCacheMut.RLock() cache, cached := assetRevisionUpsertCache[key] assetRevisionUpsertCacheMut.RUnlock() var err error if !cached { var ret []string whitelist, ret = strmangle.InsertColumnSet( assetRevisionColumns, assetRevisionColumnsWithDefault, assetRevisionColumnsWithoutDefault, nzDefaults, whitelist, ) update := strmangle.UpdateColumnSet( assetRevisionColumns, assetRevisionPrimaryKeyColumns, updateColumns, ) if len(update) == 0 { return errors.New("public: unable to upsert asset_revision, could not build update column list") } conflict := conflictColumns if len(conflict) == 0 { conflict = make([]string, len(assetRevisionPrimaryKeyColumns)) copy(conflict, assetRevisionPrimaryKeyColumns) } cache.query = queries.BuildUpsertQueryPostgres(dialect, "\"asset_revision\"", updateOnConflict, ret, update, conflict, whitelist) cache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, whitelist) if err != nil { return err } if len(ret) != 0 { cache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, ret) if err != nil { return err } } } value := reflect.Indirect(reflect.ValueOf(o)) vals := queries.ValuesFromMapping(value, cache.valueMapping) var returns []interface{} if len(cache.retMapping) != 0 { returns = queries.PtrsFromMapping(value, cache.retMapping) } if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, cache.query) fmt.Fprintln(boil.DebugWriter, vals) } if len(cache.retMapping) != 0 { err = exec.QueryRow(cache.query, vals...).Scan(returns...) if err == sql.ErrNoRows { err = nil // Postgres doesn't return anything when there's no update } } else { _, err = exec.Exec(cache.query, vals...) } if err != nil { return errors.Wrap(err, "public: unable to upsert asset_revision") } if !cached { assetRevisionUpsertCacheMut.Lock() assetRevisionUpsertCache[key] = cache assetRevisionUpsertCacheMut.Unlock() } return o.doAfterUpsertHooks(exec) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Vote) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvoteAllColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(votePrimaryKeyColumns))\n\t\t\tcopy(conflict, votePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vote\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vote\")\n\t}\n\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no transactions provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert transactions, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transactions\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transactions` WHERE `transaction_id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for transactions\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.TransactionID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"TransactionID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.TransactionID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for transactions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no transaction provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(transactionColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttransactionUpsertCacheMut.RLock()\n\tcache, cached := transactionUpsertCache[key]\n\ttransactionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionColumnsWithDefault,\n\t\t\ttransactionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\ttransactionColumns,\n\t\t\ttransactionPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"model: unable to upsert transaction, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"transaction\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `transaction` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(transactionType, transactionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(transactionType, transactionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to upsert for transaction\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == transactionMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for transaction\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\ttransactionUpsertCacheMut.Lock()\n\t\ttransactionUpsertCache[key] = cache\n\t\ttransactionUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Task) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tasks provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(taskColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttaskUpsertCacheMut.RLock()\n\tcache, cached := taskUpsertCache[key]\n\ttaskUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskColumnsWithDefault,\n\t\t\ttaskColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttaskAllColumns,\n\t\t\ttaskPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tasks, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(taskPrimaryKeyColumns))\n\t\t\tcopy(conflict, taskPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tasks\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(taskType, taskMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(taskType, taskMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tasks\")\n\t}\n\n\tif !cached {\n\t\ttaskUpsertCacheMut.Lock()\n\t\ttaskUpsertCache[key] = cache\n\t\ttaskUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Weather) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"db: no weather provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(weatherColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tweatherUpsertCacheMut.RLock()\n\tcache, cached := weatherUpsertCache[key]\n\tweatherUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherColumnsWithDefault,\n\t\t\tweatherColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tweatherColumns,\n\t\t\tweatherPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"db: unable to upsert weather, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(weatherPrimaryKeyColumns))\n\t\t\tcopy(conflict, weatherPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"prh\\\".\\\"weather\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(weatherType, weatherMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(weatherType, weatherMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to upsert weather\")\n\t}\n\n\tif !cached {\n\t\tweatherUpsertCacheMut.Lock()\n\t\tweatherUpsertCache[key] = cache\n\t\tweatherUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Store) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stores provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstoreUpsertCacheMut.RLock()\n\tcache, cached := storeUpsertCache[key]\n\tstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstoreColumnsWithDefault,\n\t\t\tstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstoreAllColumns,\n\t\t\tstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert stores, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storePrimaryKeyColumns))\n\t\t\tcopy(conflict, storePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"stores\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storeType, storeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storeType, storeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert stores\")\n\t}\n\n\tif !cached {\n\t\tstoreUpsertCacheMut.Lock()\n\t\tstoreUpsertCache[key] = cache\n\t\tstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Vote) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vote provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(voteColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvoteUpsertCacheMut.RLock()\n\tcache, cached := voteUpsertCache[key]\n\tvoteUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvoteColumnsWithDefault,\n\t\t\tvoteColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tvoteColumns,\n\t\t\tvotePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vote, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"vote\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `vote` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(voteType, voteMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(voteType, voteMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for vote\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == voteMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for vote\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tvoteUpsertCacheMut.Lock()\n\t\tvoteUpsertCache[key] = cache\n\t\tvoteUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Repository) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no repositories provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(repositoryColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLRepositoryUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trepositoryUpsertCacheMut.RLock()\n\tcache, cached := repositoryUpsertCache[key]\n\trepositoryUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trepositoryAllColumns,\n\t\t\trepositoryColumnsWithDefault,\n\t\t\trepositoryColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trepositoryAllColumns,\n\t\t\trepositoryPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert repositories, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"repositories\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `repositories` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(repositoryType, repositoryMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(repositoryType, repositoryMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for repositories\")\n\t}\n\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(repositoryType, repositoryMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for repositories\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for repositories\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\trepositoryUpsertCacheMut.Lock()\n\t\trepositoryUpsertCache[key] = cache\n\t\trepositoryUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *RawVisit) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no raw_visits provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rawVisitColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trawVisitUpsertCacheMut.RLock()\n\tcache, cached := rawVisitUpsertCache[key]\n\trawVisitUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitColumnsWithDefault,\n\t\t\trawVisitColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trawVisitAllColumns,\n\t\t\trawVisitPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert raw_visits, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rawVisitPrimaryKeyColumns))\n\t\t\tcopy(conflict, rawVisitPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"raw_visits\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rawVisitType, rawVisitMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert raw_visits\")\n\t}\n\n\tif !cached {\n\t\trawVisitUpsertCacheMut.Lock()\n\t\trawVisitUpsertCache[key] = cache\n\t\trawVisitUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *OauthClient) Upsert(exec boil.Executor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no oauth_clients provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t}\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(oauthClientColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLOauthClientUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\toauthClientUpsertCacheMut.RLock()\n\tcache, cached := oauthClientUpsertCache[key]\n\toauthClientUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientColumnsWithDefault,\n\t\t\toauthClientColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\toauthClientAllColumns,\n\t\t\toauthClientPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert oauth_clients, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"oauth_clients\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `oauth_clients` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(oauthClientType, oauthClientMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for oauth_clients\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == oauthClientMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(oauthClientType, oauthClientMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for oauth_clients\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\terr = exec.QueryRow(cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for oauth_clients\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\toauthClientUpsertCacheMut.Lock()\n\t\toauthClientUpsertCache[key] = cache\n\t\toauthClientUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Jet) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jets provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jetColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjetUpsertCacheMut.RLock()\n\tcache, cached := jetUpsertCache[key]\n\tjetUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetColumnsWithDefault,\n\t\t\tjetColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tjetColumns,\n\t\t\tjetPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jets, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"jets\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `jets` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jetType, jetMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jetType, jetMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for jets\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for jets\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tjetUpsertCacheMut.Lock()\n\t\tjetUpsertCache[key] = cache\n\t\tjetUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Failure) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no failure provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(failureColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tfailureUpsertCacheMut.RLock()\n\tcache, cached := failureUpsertCache[key]\n\tfailureUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailureColumnsWithDefault,\n\t\t\tfailureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tfailureColumns,\n\t\t\tfailurePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert failure, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"failure\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `failure` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(failureType, failureMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(failureType, failureMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for failure\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == failureMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for failure\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfailureUpsertCacheMut.Lock()\n\t\tfailureUpsertCache[key] = cache\n\t\tfailureUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Kvstore) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no kvstore provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(kvstoreColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tkvstoreUpsertCacheMut.RLock()\n\tcache, cached := kvstoreUpsertCache[key]\n\tkvstoreUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstoreColumnsWithDefault,\n\t\t\tkvstoreColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tkvstoreAllColumns,\n\t\t\tkvstorePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert kvstore, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(kvstorePrimaryKeyColumns))\n\t\t\tcopy(conflict, kvstorePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"kvstore\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(kvstoreType, kvstoreMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert kvstore\")\n\t}\n\n\tif !cached {\n\t\tkvstoreUpsertCacheMut.Lock()\n\t\tkvstoreUpsertCache[key] = cache\n\t\tkvstoreUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Inventory) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no inventory provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(inventoryColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tinventoryUpsertCacheMut.RLock()\n\tcache, cached := inventoryUpsertCache[key]\n\tinventoryUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryColumnsWithDefault,\n\t\t\tinventoryColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tinventoryColumns,\n\t\t\tinventoryPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert inventory, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"inventory\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `inventory` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(inventoryType, inventoryMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(inventoryType, inventoryMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for inventory\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == inventoryMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for inventory\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tinventoryUpsertCacheMut.Lock()\n\t\tinventoryUpsertCache[key] = cache\n\t\tinventoryUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (m *MySQL)UpSert(i interface{})(sql.Result, error){\n\tif err := m.Exists(i); err != nil {\n\t\treturn m.Insert(i)\n\t}\n\treturn m.Update(i)\n}", "func (o *Origin) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no origins provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(originColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\toriginUpsertCacheMut.RLock()\n\tcache, cached := originUpsertCache[key]\n\toriginUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\toriginColumns,\n\t\t\toriginColumnsWithDefault,\n\t\t\toriginColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\toriginColumns,\n\t\t\toriginPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert origins, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(originPrimaryKeyColumns))\n\t\t\tcopy(conflict, originPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"origins\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(originType, originMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(originType, originMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert origins\")\n\t}\n\n\tif !cached {\n\t\toriginUpsertCacheMut.Lock()\n\t\toriginUpsertCache[key] = cache\n\t\toriginUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Offer) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no offers provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(offerColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tofferUpsertCacheMut.RLock()\n\tcache, cached := offerUpsertCache[key]\n\tofferUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferColumnsWithDefault,\n\t\t\tofferColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tofferColumns,\n\t\t\tofferPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"stellarcore: unable to upsert offers, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(offerPrimaryKeyColumns))\n\t\t\tcopy(conflict, offerPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"offers\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(offerType, offerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(offerType, offerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to upsert offers\")\n\t}\n\n\tif !cached {\n\t\tofferUpsertCacheMut.Lock()\n\t\tofferUpsertCache[key] = cache\n\t\tofferUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (a *Actor) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif a._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.actor (` +\n\t\t`actor_id, first_name, last_name, last_update` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4` +\n\t\t`) ON CONFLICT (actor_id) DO UPDATE SET (` +\n\t\t`actor_id, first_name, last_name, last_update` +\n\t\t`) = (` +\n\t\t`EXCLUDED.actor_id, EXCLUDED.first_name, EXCLUDED.last_name, EXCLUDED.last_update` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate)\n\t_, err = db.Exec(sqlstr, a.ActorID, a.FirstName, a.LastName, a.LastUpdate)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ta._exists = true\n\n\treturn nil\n}", "func (o *Utxo) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no utxo provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(utxoColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tutxoUpsertCacheMut.RLock()\n\tcache, cached := utxoUpsertCache[key]\n\tutxoUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tutxoAllColumns,\n\t\t\tutxoColumnsWithDefault,\n\t\t\tutxoColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tutxoAllColumns,\n\t\t\tutxoPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert utxo, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(utxoPrimaryKeyColumns))\n\t\t\tcopy(conflict, utxoPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"utxo\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(utxoType, utxoMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(utxoType, utxoMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert utxo\")\n\t}\n\n\tif !cached {\n\t\tutxoUpsertCacheMut.Lock()\n\t\tutxoUpsertCache[key] = cache\n\t\tutxoUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *HoldenAt) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no HoldenAt provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(holdenAtColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tholdenAtUpsertCacheMut.RLock()\n\tcache, cached := holdenAtUpsertCache[key]\n\tholdenAtUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtColumnsWithDefault,\n\t\t\tholdenAtColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tholdenAtAllColumns,\n\t\t\tholdenAtPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert HoldenAt, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(holdenAtPrimaryKeyColumns))\n\t\t\tcopy(conflict, holdenAtPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"HoldenAt\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(holdenAtType, holdenAtMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert HoldenAt\")\n\t}\n\n\tif !cached {\n\t\tholdenAtUpsertCacheMut.Lock()\n\t\tholdenAtUpsertCache[key] = cache\n\t\tholdenAtUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Currency) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no currency provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currencyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcurrencyUpsertCacheMut.RLock()\n\tcache, cached := currencyUpsertCache[key]\n\tcurrencyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyColumnsWithDefault,\n\t\t\tcurrencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tcurrencyColumns,\n\t\t\tcurrencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert currency, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(currencyPrimaryKeyColumns))\n\t\t\tcopy(conflict, currencyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"currency\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currencyType, currencyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(currencyType, currencyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert currency\")\n\t}\n\n\tif !cached {\n\t\tcurrencyUpsertCacheMut.Lock()\n\t\tcurrencyUpsertCache[key] = cache\n\t\tcurrencyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Ticket) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no tickets provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(ticketColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tticketUpsertCacheMut.RLock()\n\tcache, cached := ticketUpsertCache[key]\n\tticketUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketColumnsWithDefault,\n\t\t\tticketColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tticketAllColumns,\n\t\t\tticketPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert tickets, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(ticketPrimaryKeyColumns))\n\t\t\tcopy(conflict, ticketPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"tickets\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(ticketType, ticketMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(ticketType, ticketMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert tickets\")\n\t}\n\n\tif !cached {\n\t\tticketUpsertCacheMut.Lock()\n\t\tticketUpsertCache[key] = cache\n\t\tticketUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Job) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no jobs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(jobColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tjobUpsertCacheMut.RLock()\n\tcache, cached := jobUpsertCache[key]\n\tjobUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobColumnsWithDefault,\n\t\t\tjobColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tjobColumns,\n\t\t\tjobPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert jobs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(jobPrimaryKeyColumns))\n\t\t\tcopy(conflict, jobPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"jobs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(jobType, jobMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(jobType, jobMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert jobs\")\n\t}\n\n\tif !cached {\n\t\tjobUpsertCacheMut.Lock()\n\t\tjobUpsertCache[key] = cache\n\t\tjobUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Block) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no block provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(blockColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tblockUpsertCacheMut.RLock()\n\tcache, cached := blockUpsertCache[key]\n\tblockUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockColumnsWithDefault,\n\t\t\tblockColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tblockAllColumns,\n\t\t\tblockPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert block, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(blockPrimaryKeyColumns))\n\t\t\tcopy(conflict, blockPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"block\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(blockType, blockMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(blockType, blockMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert block\")\n\t}\n\n\tif !cached {\n\t\tblockUpsertCacheMut.Lock()\n\t\tblockUpsertCache[key] = cache\n\t\tblockUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Subscriber) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no subscribers provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(subscriberColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLSubscriberUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsubscriberUpsertCacheMut.RLock()\n\tcache, cached := subscriberUpsertCache[key]\n\tsubscriberUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsubscriberColumns,\n\t\t\tsubscriberColumnsWithDefault,\n\t\t\tsubscriberColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsubscriberColumns,\n\t\t\tsubscriberPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert subscribers, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"subscribers\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `subscribers` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(subscriberType, subscriberMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(subscriberType, subscriberMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for subscribers\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == subscriberMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(subscriberType, subscriberMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for subscribers\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for subscribers\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tsubscriberUpsertCacheMut.Lock()\n\t\tsubscriberUpsertCache[key] = cache\n\t\tsubscriberUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (repository *GormRepository) Upsert(uow *UnitOfWork, entity interface{}, queryProcessors []QueryProcessor) microappError.DatabaseError {\n\tdb := uow.DB\n\tif queryProcessors != nil {\n\t\tvar err error\n\t\tfor _, queryProcessor := range queryProcessors {\n\t\t\tdb, err = queryProcessor(db, entity)\n\t\t\tif err != nil {\n\t\t\t\treturn microappError.NewDatabaseError(err)\n\t\t\t}\n\t\t}\n\t}\n\tresult := db.Model(entity).Updates(entity)\n\tif result.Error != nil {\n\t\treturn microappError.NewDatabaseError(result.Error)\n\t}\n\n\tif result.RowsAffected == 0 {\n\t\tif err := uow.DB.Create(entity).Error; err != nil {\n\t\t\treturn microappError.NewDatabaseError(err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *APIKey) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no api_keys provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(apiKeyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tapiKeyUpsertCacheMut.RLock()\n\tcache, cached := apiKeyUpsertCache[key]\n\tapiKeyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyColumnsWithDefault,\n\t\t\tapiKeyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tapiKeyColumns,\n\t\t\tapiKeyPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert api_keys, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(apiKeyPrimaryKeyColumns))\n\t\t\tcopy(conflict, apiKeyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"api_keys\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(apiKeyType, apiKeyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert api_keys\")\n\t}\n\n\tif !cached {\n\t\tapiKeyUpsertCacheMut.Lock()\n\t\tapiKeyUpsertCache[key] = cache\n\t\tapiKeyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Node) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no node provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(nodeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tnodeUpsertCacheMut.RLock()\n\tcache, cached := nodeUpsertCache[key]\n\tnodeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodeColumnsWithDefault,\n\t\t\tnodeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tnodeAllColumns,\n\t\t\tnodePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert node, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(nodePrimaryKeyColumns))\n\t\t\tcopy(conflict, nodePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"node\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(nodeType, nodeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(nodeType, nodeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert node\")\n\t}\n\n\tif !cached {\n\t\tnodeUpsertCacheMut.Lock()\n\t\tnodeUpsertCache[key] = cache\n\t\tnodeUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Board) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"rdb: no boards provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(boardColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLBoardUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tboardUpsertCacheMut.RLock()\n\tcache, cached := boardUpsertCache[key]\n\tboardUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardColumnsWithDefault,\n\t\t\tboardColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tboardAllColumns,\n\t\t\tboardPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"rdb: unable to upsert boards, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"boards\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `boards` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(boardType, boardMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(boardType, boardMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to upsert for boards\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == boardMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(boardType, boardMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to retrieve unique values for boards\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to populate default values for boards\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tboardUpsertCacheMut.Lock()\n\t\tboardUpsertCache[key] = cache\n\t\tboardUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (blt Bolt) Upsert(key, value []byte) error {\n\treturn blt.db.Update(func(tx *b.Tx) error {\n\t\treturn tx.Bucket(blt.Bucket).Put(key, value)\n\t})\n}", "func (o *PremiumSlot) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no premium_slots provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(premiumSlotColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpremiumSlotUpsertCacheMut.RLock()\n\tcache, cached := premiumSlotUpsertCache[key]\n\tpremiumSlotUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotColumnsWithDefault,\n\t\t\tpremiumSlotColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpremiumSlotAllColumns,\n\t\t\tpremiumSlotPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert premium_slots, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(premiumSlotPrimaryKeyColumns))\n\t\t\tcopy(conflict, premiumSlotPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"premium_slots\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(premiumSlotType, premiumSlotMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert premium_slots\")\n\t}\n\n\tif !cached {\n\t\tpremiumSlotUpsertCacheMut.Lock()\n\t\tpremiumSlotUpsertCache[key] = cache\n\t\tpremiumSlotUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *InstrumentClass) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no instrument_class provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(instrumentClassColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tinstrumentClassUpsertCacheMut.RLock()\n\tcache, cached := instrumentClassUpsertCache[key]\n\tinstrumentClassUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassColumnsWithDefault,\n\t\t\tinstrumentClassColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tinstrumentClassColumns,\n\t\t\tinstrumentClassPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert instrument_class, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(instrumentClassPrimaryKeyColumns))\n\t\t\tcopy(conflict, instrumentClassPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"instruments\\\".\\\"instrument_class\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(instrumentClassType, instrumentClassMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert instrument_class\")\n\t}\n\n\tif !cached {\n\t\tinstrumentClassUpsertCacheMut.Lock()\n\t\tinstrumentClassUpsertCache[key] = cache\n\t\tinstrumentClassUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *CurrentChartDataMinutely) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no current_chart_data_minutely provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(currentChartDataMinutelyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcurrentChartDataMinutelyUpsertCacheMut.RLock()\n\tcache, cached := currentChartDataMinutelyUpsertCache[key]\n\tcurrentChartDataMinutelyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tcurrentChartDataMinutelyColumns,\n\t\t\tcurrentChartDataMinutelyColumnsWithDefault,\n\t\t\tcurrentChartDataMinutelyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tcurrentChartDataMinutelyColumns,\n\t\t\tcurrentChartDataMinutelyPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert current_chart_data_minutely, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(currentChartDataMinutelyPrimaryKeyColumns))\n\t\t\tcopy(conflict, currentChartDataMinutelyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"current_chart_data_minutely\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(currentChartDataMinutelyType, currentChartDataMinutelyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert current_chart_data_minutely\")\n\t}\n\n\tif !cached {\n\t\tcurrentChartDataMinutelyUpsertCacheMut.Lock()\n\t\tcurrentChartDataMinutelyUpsertCache[key] = cache\n\t\tcurrentChartDataMinutelyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (db *DB) Upsert(key interface{}, value interface{}) error {\n\treturn db.bolt.Update(func(tx *bolt.Tx) error {\n\t\treturn db.UpsertTx(tx, key, value)\n\t})\n}", "func (o *BTCTXOutput) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no btc_tx_output provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(btcTXOutputColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLBTCTXOutputUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbtcTXOutputUpsertCacheMut.RLock()\n\tcache, cached := btcTXOutputUpsertCache[key]\n\tbtcTXOutputUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tbtcTXOutputAllColumns,\n\t\t\tbtcTXOutputColumnsWithDefault,\n\t\t\tbtcTXOutputColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tbtcTXOutputAllColumns,\n\t\t\tbtcTXOutputPrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert btc_tx_output, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`btc_tx_output`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `btc_tx_output` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for btc_tx_output\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == btcTXOutputMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(btcTXOutputType, btcTXOutputMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for btc_tx_output\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for btc_tx_output\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tbtcTXOutputUpsertCacheMut.Lock()\n\t\tbtcTXOutputUpsertCache[key] = cache\n\t\tbtcTXOutputUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *PremiumCode) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no premium_codes provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(premiumCodeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpremiumCodeUpsertCacheMut.RLock()\n\tcache, cached := premiumCodeUpsertCache[key]\n\tpremiumCodeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpremiumCodeAllColumns,\n\t\t\tpremiumCodeColumnsWithDefault,\n\t\t\tpremiumCodeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpremiumCodeAllColumns,\n\t\t\tpremiumCodePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert premium_codes, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(premiumCodePrimaryKeyColumns))\n\t\t\tcopy(conflict, premiumCodePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"premium_codes\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(premiumCodeType, premiumCodeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(premiumCodeType, premiumCodeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert premium_codes\")\n\t}\n\n\tif !cached {\n\t\tpremiumCodeUpsertCacheMut.Lock()\n\t\tpremiumCodeUpsertCache[key] = cache\n\t\tpremiumCodeUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *ExchangeCurrency) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no exchange_currency provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(exchangeCurrencyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\texchangeCurrencyUpsertCacheMut.RLock()\n\tcache, cached := exchangeCurrencyUpsertCache[key]\n\texchangeCurrencyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\texchangeCurrencyColumns,\n\t\t\texchangeCurrencyColumnsWithDefault,\n\t\t\texchangeCurrencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\texchangeCurrencyColumns,\n\t\t\texchangeCurrencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert exchange_currency, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(exchangeCurrencyPrimaryKeyColumns))\n\t\t\tcopy(conflict, exchangeCurrencyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"exchange_currency\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(exchangeCurrencyType, exchangeCurrencyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(exchangeCurrencyType, exchangeCurrencyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert exchange_currency\")\n\t}\n\n\tif !cached {\n\t\texchangeCurrencyUpsertCacheMut.Lock()\n\t\texchangeCurrencyUpsertCache[key] = cache\n\t\texchangeCurrencyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (d *DB) Upsert(table string, record interface{}) error {\n\tif d.Conn == nil {\n\t\tpanic(\"database is not initialized\")\n\t}\n\n\ttxn := d.Conn.Txn(true)\n\tdefer txn.Abort()\n\n\terr := txn.Insert(table, record)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttxn.Commit()\n\n\treturn nil\n}", "func (o *Storestate) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no storestate provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(storestateColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstorestateUpsertCacheMut.RLock()\n\tcache, cached := storestateUpsertCache[key]\n\tstorestateUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestateColumnsWithDefault,\n\t\t\tstorestateColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstorestateColumns,\n\t\t\tstorestatePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"stellarcore: unable to upsert storestate, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(storestatePrimaryKeyColumns))\n\t\t\tcopy(conflict, storestatePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"storestate\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(storestateType, storestateMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(storestateType, storestateMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to upsert storestate\")\n\t}\n\n\tif !cached {\n\t\tstorestateUpsertCacheMut.Lock()\n\t\tstorestateUpsertCache[key] = cache\n\t\tstorestateUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (r *Connectors) Upsert(col string, value interface{}, opts *gocb.UpsertOptions) (*gocb.MutationResult, error) {\n\tcollection := r.Bucket.DefaultCollection()\n\treturn collection.Upsert(col, value, opts)\n}", "func (s *State) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif s._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.state (` +\n\t\t`\"id\", \"namespace\", \"context_id\", \"created_at\", \"state\", \"data\", \"event\", \"processing_error\"` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6, $7, $8` +\n\t\t`) ON CONFLICT (\"id\") DO UPDATE SET (` +\n\t\t`\"id\", \"namespace\", \"context_id\", \"created_at\", \"state\", \"data\", \"event\", \"processing_error\"` +\n\t\t`) = (` +\n\t\t`EXCLUDED.\"id\", EXCLUDED.\"namespace\", EXCLUDED.\"context_id\", EXCLUDED.\"created_at\", EXCLUDED.\"state\", EXCLUDED.\"data\", EXCLUDED.\"event\", EXCLUDED.\"processing_error\"` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, s.ID, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError)\n\t_, err = db.Exec(sqlstr, s.ID, s.Namespace, s.ContextID, s.CreatedAt, s.State, s.Data, s.Event, s.ProcessingError)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\ts._exists = true\n\n\treturn nil\n}", "func (o *TrainingCost) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no training_costs provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(trainingCostColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttrainingCostUpsertCacheMut.RLock()\n\tcache, cached := trainingCostUpsertCache[key]\n\ttrainingCostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostColumnsWithDefault,\n\t\t\ttrainingCostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttrainingCostAllColumns,\n\t\t\ttrainingCostPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert training_costs, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(trainingCostPrimaryKeyColumns))\n\t\t\tcopy(conflict, trainingCostPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"training_costs\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(trainingCostType, trainingCostMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert training_costs\")\n\t}\n\n\tif !cached {\n\t\ttrainingCostUpsertCacheMut.Lock()\n\t\ttrainingCostUpsertCache[key] = cache\n\t\ttrainingCostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Auth) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auths provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.Time.IsZero() {\n\t\to.CreatedAt.Time = currTime\n\t\to.CreatedAt.Valid = true\n\t}\n\to.UpdatedAt.Time = currTime\n\to.UpdatedAt.Valid = true\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthUpsertCacheMut.RLock()\n\tcache, cached := authUpsertCache[key]\n\tauthUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthColumnsWithDefault,\n\t\t\tauthColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tauthColumns,\n\t\t\tauthPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auths, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(authPrimaryKeyColumns))\n\t\t\tcopy(conflict, authPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"auths\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authType, authMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authType, authMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert auths\")\n\t}\n\n\tif !cached {\n\t\tauthUpsertCacheMut.Lock()\n\t\tauthUpsertCache[key] = cache\n\t\tauthUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Peer) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no peers provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(peerColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLPeerUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpeerUpsertCacheMut.RLock()\n\tcache, cached := peerUpsertCache[key]\n\tpeerUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerColumnsWithDefault,\n\t\t\tpeerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpeerAllColumns,\n\t\t\tpeerPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"model: unable to upsert peers, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"peers\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `peers` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(peerType, peerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(peerType, peerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to upsert for peers\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == peerMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(peerType, peerMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to retrieve unique values for peers\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to populate default values for peers\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpeerUpsertCacheMut.Lock()\n\t\tpeerUpsertCache[key] = cache\n\t\tpeerUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *PublisherSearchIdx) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no publisher_search_idx provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(publisherSearchIdxColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpublisherSearchIdxUpsertCacheMut.RLock()\n\tcache, cached := publisherSearchIdxUpsertCache[key]\n\tpublisherSearchIdxUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpublisherSearchIdxAllColumns,\n\t\t\tpublisherSearchIdxColumnsWithDefault,\n\t\t\tpublisherSearchIdxColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpublisherSearchIdxAllColumns,\n\t\t\tpublisherSearchIdxPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert publisher_search_idx, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(publisherSearchIdxPrimaryKeyColumns))\n\t\t\tcopy(conflict, publisherSearchIdxPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQuerySQLite(dialect, \"\\\"publisher_search_idx\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(publisherSearchIdxType, publisherSearchIdxMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(publisherSearchIdxType, publisherSearchIdxMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert publisher_search_idx\")\n\t}\n\n\tif !cached {\n\t\tpublisherSearchIdxUpsertCacheMut.Lock()\n\t\tpublisherSearchIdxUpsertCache[key] = cache\n\t\tpublisherSearchIdxUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (m ConcurrentMap[T]) Upsert(key string, val T, cb UpsertFunc[T]) T {\n\tshard := m.getShard(key)\n\tshard.Lock()\n\told, ok := shard.items[key]\n\tres := cb(ok, old, val)\n\tshard.items[key] = res\n\tshard.Unlock()\n\treturn res\n}", "func (o *Stock) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no stock provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(stockColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstockUpsertCacheMut.RLock()\n\tcache, cached := stockUpsertCache[key]\n\tstockUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tstockColumns,\n\t\t\tstockColumnsWithDefault,\n\t\t\tstockColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tstockColumns,\n\t\t\tstockPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert stock, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"stock\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `stock` WHERE `stock_id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(stockType, stockMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(stockType, stockMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for stock\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.StockID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == stockMapping[\"StockID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.StockID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for stock\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tstockUpsertCacheMut.Lock()\n\t\tstockUpsertCache[key] = cache\n\t\tstockUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Post) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"orm: no posts provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(postColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpostUpsertCacheMut.RLock()\n\tcache, cached := postUpsertCache[key]\n\tpostUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpostColumns,\n\t\t\tpostColumnsWithDefault,\n\t\t\tpostColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpostColumns,\n\t\t\tpostPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"orm: unable to upsert posts, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(postPrimaryKeyColumns))\n\t\t\tcopy(conflict, postPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"posts\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(postType, postMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(postType, postMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to upsert posts\")\n\t}\n\n\tif !cached {\n\t\tpostUpsertCacheMut.Lock()\n\t\tpostUpsertCache[key] = cache\n\t\tpostUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *StockKeepingUnitContent) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"db: no stock_keeping_unit_content provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\to.UpdatedAt = currTime\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(stockKeepingUnitContentColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstockKeepingUnitContentUpsertCacheMut.RLock()\n\tcache, cached := stockKeepingUnitContentUpsertCache[key]\n\tstockKeepingUnitContentUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tstockKeepingUnitContentAllColumns,\n\t\t\tstockKeepingUnitContentColumnsWithDefault,\n\t\t\tstockKeepingUnitContentColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tstockKeepingUnitContentAllColumns,\n\t\t\tstockKeepingUnitContentPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"db: unable to upsert stock_keeping_unit_content, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(stockKeepingUnitContentPrimaryKeyColumns))\n\t\t\tcopy(conflict, stockKeepingUnitContentPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"stock_keeping_unit_content\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(stockKeepingUnitContentType, stockKeepingUnitContentMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(stockKeepingUnitContentType, stockKeepingUnitContentMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to upsert stock_keeping_unit_content\")\n\t}\n\n\tif !cached {\n\t\tstockKeepingUnitContentUpsertCacheMut.Lock()\n\t\tstockKeepingUnitContentUpsertCache[key] = cache\n\t\tstockKeepingUnitContentUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (db *DBService) Upsert(jobs ...*DataHistoryJobResult) error {\n\tif len(jobs) == 0 {\n\t\treturn nil\n\t}\n\tctx := context.TODO()\n\n\ttx, err := db.sql.BeginTx(ctx, nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"beginTx %w\", err)\n\t}\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terrRB := tx.Rollback()\n\t\t\tif errRB != nil {\n\t\t\t\tlog.Errorf(log.DatabaseMgr, \"Insert tx.Rollback %v\", errRB)\n\t\t\t}\n\t\t}\n\t}()\n\n\tswitch db.driver {\n\tcase database.DBSQLite3, database.DBSQLite:\n\t\terr = upsertSqlite(ctx, tx, jobs...)\n\tcase database.DBPostgreSQL:\n\t\terr = upsertPostgres(ctx, tx, jobs...)\n\tdefault:\n\t\treturn database.ErrNoDatabaseProvided\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (o *Vendor) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vendors provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(vendorColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLVendorUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvendorUpsertCacheMut.RLock()\n\tcache, cached := vendorUpsertCache[key]\n\tvendorUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvendorColumns,\n\t\t\tvendorColumnsWithDefault,\n\t\t\tvendorColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvendorColumns,\n\t\t\tvendorPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vendors, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"vendors\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `vendors` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(vendorType, vendorMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(vendorType, vendorMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for vendors\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == vendorMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(vendorType, vendorMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for vendors\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for vendors\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tvendorUpsertCacheMut.Lock()\n\t\tvendorUpsertCache[key] = cache\n\t\tvendorUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (ds *DjangoSession) Upsert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase ds._deleted: // deleted\n\t\treturn logerror(&ErrUpsertFailed{ErrMarkedForDeletion})\n\t}\n\t// upsert\n\tconst sqlstr = `MERGE django.django_sessiont ` +\n\t\t`USING (` +\n\t\t`SELECT :1 session_key, :2 session_data, :3 expire_date ` +\n\t\t`FROM DUAL ) s ` +\n\t\t`ON s.session_key = t.session_key ` +\n\t\t`WHEN MATCHED THEN ` +\n\t\t`UPDATE SET ` +\n\t\t`t.session_data = s.session_data, t.expire_date = s.expire_date ` +\n\t\t`WHEN NOT MATCHED THEN ` +\n\t\t`INSERT (` +\n\t\t`session_key, session_data, expire_date` +\n\t\t`) VALUES (` +\n\t\t`s.session_key, s.session_data, s.expire_date` +\n\t\t`);`\n\t// run\n\tlogf(sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate)\n\tif _, err := db.ExecContext(ctx, sqlstr, ds.SessionKey, ds.SessionData, ds.ExpireDate); err != nil {\n\t\treturn logerror(err)\n\t}\n\t// set exists\n\tds._exists = true\n\treturn nil\n}", "func (o *Doc) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no doc provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(docColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLDocUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tdocUpsertCacheMut.RLock()\n\tcache, cached := docUpsertCache[key]\n\tdocUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tdocAllColumns,\n\t\t\tdocColumnsWithDefault,\n\t\t\tdocColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tdocAllColumns,\n\t\t\tdocPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert doc, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"doc\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `doc` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(docType, docMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(docType, docMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for doc\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == docMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(docType, docMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for doc\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for doc\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tdocUpsertCacheMut.Lock()\n\t\tdocUpsertCache[key] = cache\n\t\tdocUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (r ExecutionResultRepoCassandra) Upsert(ctx context.Context, partnerID, taskName string, results ...ExecutionResult) (errTotal error) {\n\tfor _, result := range results {\n\t\tbatch := cassandra.Session.NewBatch(gocql.UnloggedBatch)\n\n\t\tresultFields := []interface{}{\n\t\t\tresult.ManagedEndpointID,\n\t\t\tresult.TaskInstanceID,\n\t\t\tresult.UpdatedAt,\n\t\t\tresult.ExecutionStatus,\n\t\t\tresult.StdOut,\n\t\t\tresult.StdErr,\n\t\t\tconfig.Config.DataRetentionIntervalDay * secondsInDay,\n\t\t}\n\n\t\tquery := `INSERT INTO script_execution_results (\n\t\t\t\tmanaged_endpoint_id, \n\t\t\t\ttask_instance_id, \n\t\t\t\tupdated_at, \n\t\t\t\texecution_status, \n\t\t\t\tstd_out, \n\t\t\t\tstd_err\n\t\t\t) VALUES (?, ?, ?, ?, ?, ?) USING TTL ?`\n\n\t\tbatch.Query(query, resultFields...)\n\n\t\tqueryMatView := `INSERT INTO script_execution_results_by_task_instance_id_mv (\n\t\t\t\tmanaged_endpoint_id, \n\t\t\t\ttask_instance_id, \n\t\t\t\tupdated_at, \n\t\t\t\texecution_status, \n\t\t\t\tstd_out, \n\t\t\t\tstd_err\n\t\t\t) VALUES (?, ?, ?, ?, ?, ?) USING TTL ?`\n\n\t\tbatch.Query(queryMatView, resultFields...)\n\n\t\tqueryLastExecutionTable := `INSERT INTO last_task_executions (\n\t\t\t\tpartner_id, \n\t\t\t\tendpoint_id, \n\t\t\t\trun_time, \n\t\t\t\tname, \n\t\t\t\tstatus)\n\t\t\t VALUES (?, ?, ?, ?, ?)`\n\n\t\tresultFieldsLastExecution := []interface{}{\n\t\t\tpartnerID,\n\t\t\tresult.ManagedEndpointID,\n\t\t\tresult.UpdatedAt,\n\t\t\ttaskName,\n\t\t\tresult.ExecutionStatus,\n\t\t}\n\n\t\tbatch.Query(queryLastExecutionTable, resultFieldsLastExecution...)\n\n\t\tif err := cassandra.Session.ExecuteBatch(batch); err != nil {\n\t\t\t//IF batch too large we have to insert one by one\n\t\t\terr = cassandra.QueryCassandra(ctx, query, resultFields...).Exec()\n\t\t\tif err != nil {\n\t\t\t\terrTotal = fmt.Errorf(errTotalFmt, errTotal, err)\n\t\t\t}\n\n\t\t\terr = cassandra.QueryCassandra(ctx, queryMatView, resultFields...).Exec()\n\t\t\tif err != nil {\n\t\t\t\terrTotal = fmt.Errorf(errTotalFmt, errTotal, err)\n\t\t\t}\n\n\t\t\terr = cassandra.QueryCassandra(ctx, queryLastExecutionTable, resultFieldsLastExecution...).Exec()\n\t\t\tif err != nil {\n\t\t\t\terrTotal = fmt.Errorf(errTotalFmt, errTotal, err)\n\t\t\t}\n\t\t}\n\n\t}\n\treturn\n}", "func (o *Author) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no authors provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authorColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthorUpsertCacheMut.RLock()\n\tcache, cached := authorUpsertCache[key]\n\tauthorUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorColumnsWithDefault,\n\t\t\tauthorColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tauthorColumns,\n\t\t\tauthorPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to upsert authors, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(authorPrimaryKeyColumns))\n\t\t\tcopy(conflict, authorPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"authors\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authorType, authorMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authorType, authorMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to upsert authors\")\n\t}\n\n\tif !cached {\n\t\tauthorUpsertCacheMut.Lock()\n\t\tauthorUpsertCache[key] = cache\n\t\tauthorUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Channel) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no channels provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(channelColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tchannelUpsertCacheMut.RLock()\n\tcache, cached := channelUpsertCache[key]\n\tchannelUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tchannelAllColumns,\n\t\t\tchannelColumnsWithDefault,\n\t\t\tchannelColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tchannelAllColumns,\n\t\t\tchannelPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert channels, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(channelPrimaryKeyColumns))\n\t\t\tcopy(conflict, channelPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"channels\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(channelType, channelMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(channelType, channelMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert channels\")\n\t}\n\n\tif !cached {\n\t\tchannelUpsertCacheMut.Lock()\n\t\tchannelUpsertCache[key] = cache\n\t\tchannelUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *AuthItemGroup) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no auth_item_groups provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(authItemGroupColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLAuthItemGroupUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tauthItemGroupUpsertCacheMut.RLock()\n\tcache, cached := authItemGroupUpsertCache[key]\n\tauthItemGroupUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupColumnsWithDefault,\n\t\t\tauthItemGroupColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tauthItemGroupAllColumns,\n\t\t\tauthItemGroupPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert auth_item_groups, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"auth_item_groups\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `auth_item_groups` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for auth_item_groups\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == authItemGroupMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(authItemGroupType, authItemGroupMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for auth_item_groups\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for auth_item_groups\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tauthItemGroupUpsertCacheMut.Lock()\n\t\tauthItemGroupUpsertCache[key] = cache\n\t\tauthItemGroupUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Segment) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"boiler: no segment provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(segmentColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsegmentUpsertCacheMut.RLock()\n\tcache, cached := segmentUpsertCache[key]\n\tsegmentUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsegmentAllColumns,\n\t\t\tsegmentColumnsWithDefault,\n\t\t\tsegmentColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsegmentAllColumns,\n\t\t\tsegmentPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"boiler: unable to upsert segment, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(segmentPrimaryKeyColumns))\n\t\t\tcopy(conflict, segmentPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"segment\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(segmentType, segmentMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(segmentType, segmentMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to upsert segment\")\n\t}\n\n\tif !cached {\n\t\tsegmentUpsertCacheMut.Lock()\n\t\tsegmentUpsertCache[key] = cache\n\t\tsegmentUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (t *Territory) Upsert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase t._deleted: // deleted\n\t\treturn logerror(&ErrUpsertFailed{ErrMarkedForDeletion})\n\t}\n\t// upsert\n\tconst sqlstr = `INSERT INTO territories (` +\n\t\t`territory_id, territory_description, region_id` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3` +\n\t\t`)` +\n\t\t` ON CONFLICT (territory_id) DO ` +\n\t\t`UPDATE SET ` +\n\t\t`territory_description = EXCLUDED.territory_description, region_id = EXCLUDED.region_id `\n\t// run\n\tlogf(sqlstr, t.TerritoryID, t.TerritoryDescription, t.RegionID)\n\tif _, err := db.ExecContext(ctx, sqlstr, t.TerritoryID, t.TerritoryDescription, t.RegionID); err != nil {\n\t\treturn logerror(err)\n\t}\n\t// set exists\n\tt._exists = true\n\treturn nil\n}", "func (fc *FakeCollection) Upsert(col string, value interface{}, opts *gocb.UpsertOptions) (*gocb.MutationResult, error) {\n\tif fc.Force == \"error\" {\n\t\treturn &gocb.MutationResult{}, errors.New(\"Forced collection upsert error\")\n\t}\n\treturn &gocb.MutationResult{}, nil\n}", "func (tbl AssociationTable) Upsert(v *Association, wh where.Expression) error {\n\tcol := tbl.Dialect().Quoter().Quote(tbl.pk)\n\tqName := tbl.quotedName()\n\twhs, args := where.Where(wh, tbl.Dialect().Quoter())\n\n\tquery := fmt.Sprintf(\"SELECT %s FROM %s %s\", col, qName, whs)\n\trows, err := support.Query(tbl, query, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer rows.Close()\n\n\tif !rows.Next() {\n\t\treturn tbl.Insert(require.One, v)\n\t}\n\n\tvar id int64\n\terr = rows.Scan(&id)\n\tif err != nil {\n\t\treturn tbl.Logger().LogIfError(err)\n\t}\n\n\tif rows.Next() {\n\t\treturn require.ErrWrongSize(2, \"expected to find no more than 1 but got at least 2 using %q\", wh)\n\t}\n\n\tv.Id = id\n\t_, err = tbl.Update(require.One, v)\n\treturn err\n}", "func (o *BraceletPhoto) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no bracelet_photo provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(braceletPhotoColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbraceletPhotoUpsertCacheMut.RLock()\n\tcache, cached := braceletPhotoUpsertCache[key]\n\tbraceletPhotoUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tbraceletPhotoColumns,\n\t\t\tbraceletPhotoColumnsWithDefault,\n\t\t\tbraceletPhotoColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tbraceletPhotoColumns,\n\t\t\tbraceletPhotoPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert bracelet_photo, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"bracelet_photo\", update, whitelist)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `bracelet_photo` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(braceletPhotoType, braceletPhotoMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(braceletPhotoType, braceletPhotoMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for bracelet_photo\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == braceletPhotoMapping[\"ID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for bracelet_photo\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tbraceletPhotoUpsertCacheMut.Lock()\n\t\tbraceletPhotoUpsertCache[key] = cache\n\t\tbraceletPhotoUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (oee *OtxEpubEpub) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif oee._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.otx_epub_epub (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6` +\n\t\t`) ON CONFLICT (document_ptr_id) DO UPDATE SET (` +\n\t\t`document_ptr_id, publisher, source, oebps_folder, manifest, contents` +\n\t\t`) = (` +\n\t\t`EXCLUDED.document_ptr_id, EXCLUDED.publisher, EXCLUDED.source, EXCLUDED.oebps_folder, EXCLUDED.manifest, EXCLUDED.contents` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\t_, err = db.Exec(sqlstr, oee.DocumentPtrID, oee.Publisher, oee.Source, oee.OebpsFolder, oee.Manifest, oee.Contents)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\toee._exists = true\n\n\treturn nil\n}", "func (o *Email) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"mysql: no email provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(emailColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLEmailUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\temailUpsertCacheMut.RLock()\n\tcache, cached := emailUpsertCache[key]\n\temailUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\temailColumns,\n\t\t\temailColumnsWithDefault,\n\t\t\temailColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\temailColumns,\n\t\t\temailPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mysql: unable to upsert email, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"email\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `email` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(emailType, emailMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(emailType, emailMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to upsert for email\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int64(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == emailMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(emailType, emailMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to retrieve unique values for email\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to populate default values for email\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\temailUpsertCacheMut.Lock()\n\t\temailUpsertCache[key] = cache\n\t\temailUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Latency) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no latencies provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\to.UpdatedAt = currTime\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(latencyColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tlatencyUpsertCacheMut.RLock()\n\tcache, cached := latencyUpsertCache[key]\n\tlatencyUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tlatencyAllColumns,\n\t\t\tlatencyColumnsWithDefault,\n\t\t\tlatencyColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tlatencyAllColumns,\n\t\t\tlatencyPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert latencies, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(latencyPrimaryKeyColumns))\n\t\t\tcopy(conflict, latencyPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"latencies\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(latencyType, latencyMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(latencyType, latencyMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert latencies\")\n\t}\n\n\tif !cached {\n\t\tlatencyUpsertCacheMut.Lock()\n\t\tlatencyUpsertCache[key] = cache\n\t\tlatencyUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Picture) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no pictures provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(pictureColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpictureUpsertCacheMut.RLock()\n\tcache, cached := pictureUpsertCache[key]\n\tpictureUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpictureColumnsWithDefault,\n\t\t\tpictureColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tpictureColumns,\n\t\t\tpicturePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert pictures, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"pictures\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `pictures` WHERE `id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(pictureType, pictureMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(pictureType, pictureMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\t_, err = exec.Exec(cache.query, vals...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for pictures\")\n\t}\n\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.ID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for pictures\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpictureUpsertCacheMut.Lock()\n\t\tpictureUpsertCache[key] = cache\n\t\tpictureUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (bb *BooktestBook) Upsert(ctx context.Context, db DB) error {\n\tswitch {\n\tcase bb._deleted: // deleted\n\t\treturn logerror(&ErrUpsertFailed{ErrMarkedForDeletion})\n\t}\n\t// upsert\n\tconst sqlstr = `INSERT INTO public.booktest_book (` +\n\t\t`book_id, isbn, book_type, title, year, available, author_id` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3, $4, $5, $6, $7` +\n\t\t`)` +\n\t\t` ON CONFLICT (book_id) DO ` +\n\t\t`UPDATE SET ` +\n\t\t`isbn = EXCLUDED.isbn, book_type = EXCLUDED.book_type, title = EXCLUDED.title, year = EXCLUDED.year, available = EXCLUDED.available, author_id = EXCLUDED.author_id `\n\t// run\n\tlogf(sqlstr, bb.BookID, bb.Isbn, bb.BookType, bb.Title, bb.Year, bb.Available, bb.AuthorID)\n\tif _, err := db.ExecContext(ctx, sqlstr, bb.BookID, bb.Isbn, bb.BookType, bb.Title, bb.Year, bb.Available, bb.AuthorID); err != nil {\n\t\treturn err\n\t}\n\t// set exists\n\tbb._exists = true\n\treturn nil\n}", "func (o *Illness) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no illness provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(illnessColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tillnessUpsertCacheMut.RLock()\n\tcache, cached := illnessUpsertCache[key]\n\tillnessUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tillnessAllColumns,\n\t\t\tillnessColumnsWithDefault,\n\t\t\tillnessColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tillnessAllColumns,\n\t\t\tillnessPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert illness, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(illnessPrimaryKeyColumns))\n\t\t\tcopy(conflict, illnessPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"illness\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(illnessType, illnessMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(illnessType, illnessMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert illness\")\n\t}\n\n\tif !cached {\n\t\tillnessUpsertCacheMut.Lock()\n\t\tillnessUpsertCache[key] = cache\n\t\tillnessUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *StockCvterm) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no stock_cvterm provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(stockCvtermColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tstockCvtermUpsertCacheMut.RLock()\n\tcache, cached := stockCvtermUpsertCache[key]\n\tstockCvtermUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tstockCvtermColumns,\n\t\t\tstockCvtermColumnsWithDefault,\n\t\t\tstockCvtermColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tstockCvtermColumns,\n\t\t\tstockCvtermPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"chado: unable to upsert stock_cvterm, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(stockCvtermPrimaryKeyColumns))\n\t\t\tcopy(conflict, stockCvtermPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"stock_cvterm\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(stockCvtermType, stockCvtermMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(stockCvtermType, stockCvtermMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to upsert for stock_cvterm\")\n\t}\n\n\tif !cached {\n\t\tstockCvtermUpsertCacheMut.Lock()\n\t\tstockCvtermUpsertCache[key] = cache\n\t\tstockCvtermUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *Source) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmodels: no sources provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(sourceColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsourceUpsertCacheMut.RLock()\n\tcache, cached := sourceUpsertCache[key]\n\tsourceUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tsourceAllColumns,\n\t\t\tsourceColumnsWithDefault,\n\t\t\tsourceColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tsourceAllColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmodels: unable to upsert sources, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(sourcePrimaryKeyColumns))\n\t\t\tcopy(conflict, sourcePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"sources\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(sourceType, sourceMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to upsert sources\")\n\t}\n\n\tif !cached {\n\t\tsourceUpsertCacheMut.Lock()\n\t\tsourceUpsertCache[key] = cache\n\t\tsourceUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *MempoolBin) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no mempool_bin provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(mempoolBinColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tmempoolBinUpsertCacheMut.RLock()\n\tcache, cached := mempoolBinUpsertCache[key]\n\tmempoolBinUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tmempoolBinAllColumns,\n\t\t\tmempoolBinColumnsWithDefault,\n\t\t\tmempoolBinColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tmempoolBinAllColumns,\n\t\t\tmempoolBinPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert mempool_bin, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(mempoolBinPrimaryKeyColumns))\n\t\t\tcopy(conflict, mempoolBinPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"mempool_bin\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(mempoolBinType, mempoolBinMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(mempoolBinType, mempoolBinMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert mempool_bin\")\n\t}\n\n\tif !cached {\n\t\tmempoolBinUpsertCacheMut.Lock()\n\t\tmempoolBinUpsertCache[key] = cache\n\t\tmempoolBinUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Jet) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (at *AuthtokenToken) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif at._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.authtoken_token (` +\n\t\t`key, created, user_id` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3` +\n\t\t`) ON CONFLICT (key) DO UPDATE SET (` +\n\t\t`key, created, user_id` +\n\t\t`) = (` +\n\t\t`EXCLUDED.key, EXCLUDED.created, EXCLUDED.user_id` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, at.Key, at.Created, at.UserID)\n\t_, err = db.Exec(sqlstr, at.Key, at.Created, at.UserID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\tat._exists = true\n\n\treturn nil\n}", "func (o *RSSAnnouncement) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no rss_announcements provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rssAnnouncementColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trssAnnouncementUpsertCacheMut.RLock()\n\tcache, cached := rssAnnouncementUpsertCache[key]\n\trssAnnouncementUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trssAnnouncementAllColumns,\n\t\t\trssAnnouncementColumnsWithDefault,\n\t\t\trssAnnouncementColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trssAnnouncementAllColumns,\n\t\t\trssAnnouncementPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert rss_announcements, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rssAnnouncementPrimaryKeyColumns))\n\t\t\tcopy(conflict, rssAnnouncementPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"rss_announcements\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rssAnnouncementType, rssAnnouncementMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rssAnnouncementType, rssAnnouncementMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert rss_announcements\")\n\t}\n\n\tif !cached {\n\t\trssAnnouncementUpsertCacheMut.Lock()\n\t\trssAnnouncementUpsertCache[key] = cache\n\t\trssAnnouncementUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *DestinationRank) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no destination_rank provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(destinationRankColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tdestinationRankUpsertCacheMut.RLock()\n\tcache, cached := destinationRankUpsertCache[key]\n\tdestinationRankUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tdestinationRankAllColumns,\n\t\t\tdestinationRankColumnsWithDefault,\n\t\t\tdestinationRankColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tdestinationRankAllColumns,\n\t\t\tdestinationRankPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert destination_rank, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(destinationRankPrimaryKeyColumns))\n\t\t\tcopy(conflict, destinationRankPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"destination_rank\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(destinationRankType, destinationRankMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(destinationRankType, destinationRankMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert destination_rank\")\n\t}\n\n\tif !cached {\n\t\tdestinationRankUpsertCacheMut.Lock()\n\t\tdestinationRankUpsertCache[key] = cache\n\t\tdestinationRankUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Source) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no sources provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(sourceColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tsourceUpsertCacheMut.RLock()\n\tcache, cached := sourceUpsertCache[key]\n\tsourceUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourceColumnsWithDefault,\n\t\t\tsourceColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tsourceColumns,\n\t\t\tsourcePrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"mdbmdbmodels: unable to upsert sources, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(sourcePrimaryKeyColumns))\n\t\t\tcopy(conflict, sourcePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"sources\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(sourceType, sourceMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(sourceType, sourceMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to upsert sources\")\n\t}\n\n\tif !cached {\n\t\tsourceUpsertCacheMut.Lock()\n\t\tsourceUpsertCache[key] = cache\n\t\tsourceUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (node *IntermediateNode) SafeUpsert(manager TransactionManager, updateKey, value string) InsertionResult {\n\tidx := node.indexContaining(updateKey)\n\tchild := node.Children[idx]\n\tmanager.Add(child.GetMux())\n\tresult := child.SafeUpsert(manager, updateKey, value)\n\tif !result.DidSplit() {\n\t\treturn result\n\t}\n\tnode.Keys = insert(node.Keys, idx, result.SplitKey)\n\tnode.Children[idx] = result.Left\n\tnode.Children = insertNode(node.Children, idx+1, result.Right)\n\tif len(node.Keys) > node.MaxKeys {\n\t\tleft, right, splitKey := node.Split()\n\t\treturn InsertionResult{left, right, splitKey, result.Created}\n\t}\n\treturn InsertionResult{Created: result.Created}\n}", "func (node *LeafNode) SafeUpsert(manager TransactionManager, updateKey, value string) InsertionResult {\n\tidx := 0\n\tfor idx < len(node.Keys) && updateKey > node.Keys[idx] {\n\t\tidx++\n\t}\n\tif idx != len(node.Keys) && updateKey == node.Keys[idx] {\n\t\tnode.Values[idx] = value\n\t\treturn InsertionResult{Created: false}\n\t}\n\tnode.Keys = insert(node.Keys, idx, updateKey)\n\tnode.Values = insert(node.Values, idx, value)\n\tif len(node.Keys) > node.MaxKeys {\n\t\tleft, right, splitKey := node.Split()\n\t\treturn InsertionResult{left, right, splitKey, true}\n\t}\n\treturn InsertionResult{Created: true}\n}", "func (o *Failure) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AutomodRuleDatum) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no automod_rule_data provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(automodRuleDatumColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tautomodRuleDatumUpsertCacheMut.RLock()\n\tcache, cached := automodRuleDatumUpsertCache[key]\n\tautomodRuleDatumUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tautomodRuleDatumAllColumns,\n\t\t\tautomodRuleDatumColumnsWithDefault,\n\t\t\tautomodRuleDatumColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tautomodRuleDatumAllColumns,\n\t\t\tautomodRuleDatumPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert automod_rule_data, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(automodRuleDatumPrimaryKeyColumns))\n\t\t\tcopy(conflict, automodRuleDatumPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"automod_rule_data\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(automodRuleDatumType, automodRuleDatumMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(automodRuleDatumType, automodRuleDatumMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert automod_rule_data\")\n\t}\n\n\tif !cached {\n\t\tautomodRuleDatumUpsertCacheMut.Lock()\n\t\tautomodRuleDatumUpsertCache[key] = cache\n\t\tautomodRuleDatumUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Organism) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no organism provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(organismColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\torganismUpsertCacheMut.RLock()\n\tcache, cached := organismUpsertCache[key]\n\torganismUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\torganismColumns,\n\t\t\torganismColumnsWithDefault,\n\t\t\torganismColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\torganismColumns,\n\t\t\torganismPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"chado: unable to upsert organism, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(organismPrimaryKeyColumns))\n\t\t\tcopy(conflict, organismPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"organism\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(organismType, organismMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(organismType, organismMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to upsert for organism\")\n\t}\n\n\tif !cached {\n\t\torganismUpsertCacheMut.Lock()\n\t\torganismUpsertCache[key] = cache\n\t\torganismUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *PaymentObject) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no payment_objects provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(paymentObjectColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLPaymentObjectUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tpaymentObjectUpsertCacheMut.RLock()\n\tcache, cached := paymentObjectUpsertCache[key]\n\tpaymentObjectUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectColumnsWithDefault,\n\t\t\tpaymentObjectColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tpaymentObjectAllColumns,\n\t\t\tpaymentObjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert payment_objects, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`payment_objects`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `payment_objects` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for payment_objects\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == paymentObjectMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(paymentObjectType, paymentObjectMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for payment_objects\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for payment_objects\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tpaymentObjectUpsertCacheMut.Lock()\n\t\tpaymentObjectUpsertCache[key] = cache\n\t\tpaymentObjectUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Description) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no descriptions provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t}\n\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(descriptionColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLDescriptionUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tdescriptionUpsertCacheMut.RLock()\n\tcache, cached := descriptionUpsertCache[key]\n\tdescriptionUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tdescriptionColumns,\n\t\t\tdescriptionColumnsWithDefault,\n\t\t\tdescriptionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tdescriptionColumns,\n\t\t\tdescriptionPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert descriptions, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"descriptions\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `descriptions` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(descriptionType, descriptionMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(descriptionType, descriptionMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for descriptions\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = uint(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == descriptionMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(descriptionType, descriptionMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for descriptions\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for descriptions\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tdescriptionUpsertCacheMut.Lock()\n\t\tdescriptionUpsertCache[key] = cache\n\t\tdescriptionUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (r *Redis) Upsert(kite *protocol.Kite, value *kontrolprotocol.RegisterValue) error {\n\tr.log.Debug(\"UPSERT\")\n\n\treturn r.Add(kite, value)\n}", "func (o *Building) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"record: no buildings provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(buildingColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tbuildingUpsertCacheMut.RLock()\n\tcache, cached := buildingUpsertCache[key]\n\tbuildingUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tbuildingColumns,\n\t\t\tbuildingColumnsWithDefault,\n\t\t\tbuildingColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tbuildingColumns,\n\t\t\tbuildingPrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"record: unable to upsert buildings, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(buildingPrimaryKeyColumns))\n\t\t\tcopy(conflict, buildingPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"buildings\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(buildingType, buildingMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(buildingType, buildingMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to upsert buildings\")\n\t}\n\n\tif !cached {\n\t\tbuildingUpsertCacheMut.Lock()\n\t\tbuildingUpsertCache[key] = cache\n\t\tbuildingUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *CMFFamilyUserPoliciesTake) Upsert(ctx context.Context, exec boil.ContextExecutor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no cmf_family_user_policies_take provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t\to.UpdatedAt = currTime\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(cmfFamilyUserPoliciesTakeColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLCMFFamilyUserPoliciesTakeUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcmfFamilyUserPoliciesTakeUpsertCacheMut.RLock()\n\tcache, cached := cmfFamilyUserPoliciesTakeUpsertCache[key]\n\tcmfFamilyUserPoliciesTakeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tcmfFamilyUserPoliciesTakeAllColumns,\n\t\t\tcmfFamilyUserPoliciesTakeColumnsWithDefault,\n\t\t\tcmfFamilyUserPoliciesTakeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tcmfFamilyUserPoliciesTakeAllColumns,\n\t\t\tcmfFamilyUserPoliciesTakePrimaryKeyColumns,\n\t\t)\n\n\t\tif !updateColumns.IsNone() && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert cmf_family_user_policies_take, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"`cmf_family_user_policies_take`\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `cmf_family_user_policies_take` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(cmfFamilyUserPoliciesTakeType, cmfFamilyUserPoliciesTakeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(cmfFamilyUserPoliciesTakeType, cmfFamilyUserPoliciesTakeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tresult, err := exec.ExecContext(ctx, cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for cmf_family_user_policies_take\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == cmfFamilyUserPoliciesTakeMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(cmfFamilyUserPoliciesTakeType, cmfFamilyUserPoliciesTakeMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for cmf_family_user_policies_take\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.retQuery)\n\t\tfmt.Fprintln(writer, nzUniqueCols...)\n\t}\n\terr = exec.QueryRowContext(ctx, cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for cmf_family_user_policies_take\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tcmfFamilyUserPoliciesTakeUpsertCacheMut.Lock()\n\t\tcmfFamilyUserPoliciesTakeUpsertCache[key] = cache\n\t\tcmfFamilyUserPoliciesTakeUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *InstrumentClass) UpsertP(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateOnConflict, conflictColumns, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *UsernameListing) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no username_listings provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif queries.MustTime(o.CreatedAt).IsZero() {\n\t\t\tqueries.SetScanner(&o.CreatedAt, currTime)\n\t\t}\n\t\tqueries.SetScanner(&o.UpdatedAt, currTime)\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(usernameListingColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tusernameListingUpsertCacheMut.RLock()\n\tcache, cached := usernameListingUpsertCache[key]\n\tusernameListingUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tusernameListingAllColumns,\n\t\t\tusernameListingColumnsWithDefault,\n\t\t\tusernameListingColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tusernameListingAllColumns,\n\t\t\tusernameListingPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert username_listings, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(usernameListingPrimaryKeyColumns))\n\t\t\tcopy(conflict, usernameListingPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"username_listings\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(usernameListingType, usernameListingMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(usernameListingType, usernameListingMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert username_listings\")\n\t}\n\n\tif !cached {\n\t\tusernameListingUpsertCacheMut.Lock()\n\t\tusernameListingUpsertCache[key] = cache\n\t\tusernameListingUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *ScheduleSubject) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no schedule_subject provided for upsert\")\n\t}\n\tif !boil.TimestampsAreSkipped(ctx) {\n\t\tcurrTime := time.Now().In(boil.GetLocation())\n\n\t\tif o.CreatedAt.IsZero() {\n\t\t\to.CreatedAt = currTime\n\t\t}\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(scheduleSubjectColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tscheduleSubjectUpsertCacheMut.RLock()\n\tcache, cached := scheduleSubjectUpsertCache[key]\n\tscheduleSubjectUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tscheduleSubjectAllColumns,\n\t\t\tscheduleSubjectColumnsWithDefault,\n\t\t\tscheduleSubjectColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tscheduleSubjectAllColumns,\n\t\t\tscheduleSubjectPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert schedule_subject, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(scheduleSubjectPrimaryKeyColumns))\n\t\t\tcopy(conflict, scheduleSubjectPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"schedule_subject\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(scheduleSubjectType, scheduleSubjectMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(scheduleSubjectType, scheduleSubjectMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert schedule_subject\")\n\t}\n\n\tif !cached {\n\t\tscheduleSubjectUpsertCacheMut.Lock()\n\t\tscheduleSubjectUpsertCache[key] = cache\n\t\tscheduleSubjectUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *Tree) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no trees provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(ctx, exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(treeColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\ttreeUpsertCacheMut.RLock()\n\tcache, cached := treeUpsertCache[key]\n\ttreeUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\ttreeAllColumns,\n\t\t\ttreeColumnsWithDefault,\n\t\t\ttreeColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\ttreeAllColumns,\n\t\t\ttreePrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert trees, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(treePrimaryKeyColumns))\n\t\t\tcopy(conflict, treePrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"trees\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(treeType, treeMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(treeType, treeMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert trees\")\n\t}\n\n\tif !cached {\n\t\ttreeUpsertCacheMut.Lock()\n\t\ttreeUpsertCache[key] = cache\n\t\ttreeUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(ctx, exec)\n}", "func (o *VSP) Upsert(ctx context.Context, exec boil.ContextExecutor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no vsp provided for upsert\")\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(vspColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tvspUpsertCacheMut.RLock()\n\tcache, cached := vspUpsertCache[key]\n\tvspUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspColumnsWithDefault,\n\t\t\tvspColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tvspAllColumns,\n\t\t\tvspPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert vsp, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(vspPrimaryKeyColumns))\n\t\t\tcopy(conflict, vspPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"vsp\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(vspType, vspMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(vspType, vspMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, cache.query)\n\t\tfmt.Fprintln(writer, vals)\n\t}\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRowContext(ctx, cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.ExecContext(ctx, cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert vsp\")\n\t}\n\n\tif !cached {\n\t\tvspUpsertCacheMut.Lock()\n\t\tvspUpsertCache[key] = cache\n\t\tvspUpsertCacheMut.Unlock()\n\t}\n\n\treturn nil\n}", "func (o *Cvtermsynonym) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no cvtermsynonym provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(cvtermsynonymColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tcvtermsynonymUpsertCacheMut.RLock()\n\tcache, cached := cvtermsynonymUpsertCache[key]\n\tcvtermsynonymUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tvar ret []string\n\t\twhitelist, ret = strmangle.InsertColumnSet(\n\t\t\tcvtermsynonymColumns,\n\t\t\tcvtermsynonymColumnsWithDefault,\n\t\t\tcvtermsynonymColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\tcvtermsynonymColumns,\n\t\t\tcvtermsynonymPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"chado: unable to upsert cvtermsynonym, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(cvtermsynonymPrimaryKeyColumns))\n\t\t\tcopy(conflict, cvtermsynonymPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = queries.BuildUpsertQueryPostgres(dialect, \"\\\"cvtermsynonym\\\"\", updateOnConflict, ret, update, conflict, whitelist)\n\n\t\tcache.valueMapping, err = queries.BindMapping(cvtermsynonymType, cvtermsynonymMapping, whitelist)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(cvtermsynonymType, cvtermsynonymMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to upsert for cvtermsynonym\")\n\t}\n\n\tif !cached {\n\t\tcvtermsynonymUpsertCacheMut.Lock()\n\t\tcvtermsynonymUpsertCache[key] = cache\n\t\tcvtermsynonymUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func Upsert(s Session, dbname string, collection string, selector map[string]interface{}, updator map[string]interface{}) error {\n\t_, err := s.DB(dbname).C(collection).Upsert(selector, updator)\n\treturn err\n}", "func (o *RentalRower) Upsert(exec boil.Executor, updateOnConflict bool, conflictColumns []string, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no rental_rowers provided for upsert\")\n\t}\n\tcurrTime := time.Now().In(boil.GetLocation())\n\n\tif o.CreatedAt.IsZero() {\n\t\to.CreatedAt = currTime\n\t}\n\to.UpdatedAt = currTime\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(rentalRowerColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tif updateOnConflict {\n\t\tbuf.WriteByte('t')\n\t} else {\n\t\tbuf.WriteByte('f')\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range conflictColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\trentalRowerUpsertCacheMut.RLock()\n\tcache, cached := rentalRowerUpsertCache[key]\n\trentalRowerUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\trentalRowerColumns,\n\t\t\trentalRowerColumnsWithDefault,\n\t\t\trentalRowerColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\trentalRowerColumns,\n\t\t\trentalRowerPrimaryKeyColumns,\n\t\t)\n\n\t\tif updateOnConflict && len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert rental_rowers, could not build update column list\")\n\t\t}\n\n\t\tconflict := conflictColumns\n\t\tif len(conflict) == 0 {\n\t\t\tconflict = make([]string, len(rentalRowerPrimaryKeyColumns))\n\t\t\tcopy(conflict, rentalRowerPrimaryKeyColumns)\n\t\t}\n\t\tcache.query = buildUpsertQueryPostgres(dialect, \"\\\"rental_rowers\\\"\", updateOnConflict, ret, update, conflict, insert)\n\n\t\tcache.valueMapping, err = queries.BindMapping(rentalRowerType, rentalRowerMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(rentalRowerType, rentalRowerMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(returns...)\n\t\tif err == sql.ErrNoRows {\n\t\t\terr = nil // Postgres doesn't return anything when there's no update\n\t\t}\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert rental_rowers\")\n\t}\n\n\tif !cached {\n\t\trentalRowerUpsertCacheMut.Lock()\n\t\trentalRowerUpsertCache[key] = cache\n\t\trentalRowerUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (o *FilesStorage) Upsert(exec boil.Executor, updateColumns, insertColumns boil.Columns) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no files_storages provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(filesStorageColumnsWithDefault, o)\n\tnzUniques := queries.NonZeroDefaultSet(mySQLFilesStorageUniqueColumns, o)\n\n\tif len(nzUniques) == 0 {\n\t\treturn errors.New(\"cannot upsert with a table that cannot conflict on a unique column\")\n\t}\n\n\t// Build cache key in-line uglily - mysql vs psql problems\n\tbuf := strmangle.GetBuffer()\n\tbuf.WriteString(strconv.Itoa(updateColumns.Kind))\n\tfor _, c := range updateColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tbuf.WriteString(strconv.Itoa(insertColumns.Kind))\n\tfor _, c := range insertColumns.Cols {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzUniques {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\tfilesStorageUpsertCacheMut.RLock()\n\tcache, cached := filesStorageUpsertCache[key]\n\tfilesStorageUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := insertColumns.InsertColumnSet(\n\t\t\tfilesStorageColumns,\n\t\t\tfilesStorageColumnsWithDefault,\n\t\t\tfilesStorageColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t)\n\t\tupdate := updateColumns.UpdateColumnSet(\n\t\t\tfilesStorageColumns,\n\t\t\tfilesStoragePrimaryKeyColumns,\n\t\t)\n\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"models: unable to upsert files_storages, could not build update column list\")\n\t\t}\n\n\t\tret = strmangle.SetComplement(ret, nzUniques)\n\t\tcache.query = buildUpsertQueryMySQL(dialect, \"files_storages\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `files_storages` WHERE %s\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, nzUniques),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(filesStorageType, filesStorageMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(filesStorageType, filesStorageMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to upsert for files_storages\")\n\t}\n\n\tvar lastID int64\n\tvar uniqueMap []uint64\n\tvar nzUniqueCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.ID = int(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == filesStorageMapping[\"id\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tuniqueMap, err = queries.BindMapping(filesStorageType, filesStorageMapping, nzUniques)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to retrieve unique values for files_storages\")\n\t}\n\tnzUniqueCols = queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), uniqueMap)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, nzUniqueCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, nzUniqueCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to populate default values for files_storages\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\tfilesStorageUpsertCacheMut.Lock()\n\t\tfilesStorageUpsertCache[key] = cache\n\t\tfilesStorageUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (mm *Model) Upsert(selector, update interface{}, keys ...string) (*mgo.ChangeInfo, error) {\n\treturn mm.change(func(c CachedCollection) (*mgo.ChangeInfo, error) {\n\t\treturn c.Upsert(selector, update, keys...)\n\t})\n}", "func (o *Stock) UpsertP(exec boil.Executor, updateColumns []string, whitelist ...string) {\n\tif err := o.Upsert(exec, updateColumns, whitelist...); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) Upsert(exec boil.Executor, updateColumns []string, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no address provided for upsert\")\n\t}\n\n\tif err := o.doBeforeUpsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(addressColumnsWithDefault, o)\n\n\t// Build cache key in-line uglily - mysql vs postgres problems\n\tbuf := strmangle.GetBuffer()\n\tfor _, c := range updateColumns {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range whitelist {\n\t\tbuf.WriteString(c)\n\t}\n\tbuf.WriteByte('.')\n\tfor _, c := range nzDefaults {\n\t\tbuf.WriteString(c)\n\t}\n\tkey := buf.String()\n\tstrmangle.PutBuffer(buf)\n\n\taddressUpsertCacheMut.RLock()\n\tcache, cached := addressUpsertCache[key]\n\taddressUpsertCacheMut.RUnlock()\n\n\tvar err error\n\n\tif !cached {\n\t\tinsert, ret := strmangle.InsertColumnSet(\n\t\t\taddressColumns,\n\t\t\taddressColumnsWithDefault,\n\t\t\taddressColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tupdate := strmangle.UpdateColumnSet(\n\t\t\taddressColumns,\n\t\t\taddressPrimaryKeyColumns,\n\t\t\tupdateColumns,\n\t\t)\n\t\tif len(update) == 0 {\n\t\t\treturn errors.New(\"sqlboiler: unable to upsert address, could not build update column list\")\n\t\t}\n\n\t\tcache.query = queries.BuildUpsertQueryMySQL(dialect, \"address\", update, insert)\n\t\tcache.retQuery = fmt.Sprintf(\n\t\t\t\"SELECT %s FROM `address` WHERE `address_id`=?\",\n\t\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, ret), \",\"),\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(addressType, addressMapping, insert)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(ret) != 0 {\n\t\t\tcache.retMapping, err = queries.BindMapping(addressType, addressMapping, ret)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\tvar returns []interface{}\n\tif len(cache.retMapping) != 0 {\n\t\treturns = queries.PtrsFromMapping(value, cache.retMapping)\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tresult, err := exec.Exec(cache.query, vals...)\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to upsert for address\")\n\t}\n\n\tvar lastID int64\n\tvar identifierCols []interface{}\n\n\tif len(cache.retMapping) == 0 {\n\t\tgoto CacheNoHooks\n\t}\n\n\tlastID, err = result.LastInsertId()\n\tif err != nil {\n\t\treturn ErrSyncFail\n\t}\n\n\to.AddressID = uint16(lastID)\n\tif lastID != 0 && len(cache.retMapping) == 1 && cache.retMapping[0] == addressMapping[\"AddressID\"] {\n\t\tgoto CacheNoHooks\n\t}\n\n\tidentifierCols = []interface{}{\n\t\to.AddressID,\n\t}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.retQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, identifierCols...)\n\t}\n\n\terr = exec.QueryRow(cache.retQuery, identifierCols...).Scan(returns...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to populate default values for address\")\n\t}\n\nCacheNoHooks:\n\tif !cached {\n\t\taddressUpsertCacheMut.Lock()\n\t\taddressUpsertCache[key] = cache\n\t\taddressUpsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterUpsertHooks(exec)\n}", "func (owwb *OtxWeblinkWeblinkBasket) Upsert(db XODB) error {\n\tvar err error\n\n\t// if already exist, bail\n\tif owwb._exists {\n\t\treturn errors.New(\"insert failed: already exists\")\n\t}\n\n\t// sql query\n\tconst sqlstr = `INSERT INTO public.otx_weblink_weblink_baskets (` +\n\t\t`id, weblink_id, basket_id` +\n\t\t`) VALUES (` +\n\t\t`$1, $2, $3` +\n\t\t`) ON CONFLICT (id) DO UPDATE SET (` +\n\t\t`id, weblink_id, basket_id` +\n\t\t`) = (` +\n\t\t`EXCLUDED.id, EXCLUDED.weblink_id, EXCLUDED.basket_id` +\n\t\t`)`\n\n\t// run query\n\tXOLog(sqlstr, owwb.ID, owwb.WeblinkID, owwb.BasketID)\n\t_, err = db.Exec(sqlstr, owwb.ID, owwb.WeblinkID, owwb.BasketID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set existence\n\towwb._exists = true\n\n\treturn nil\n}" ]
[ "0.68628186", "0.68343914", "0.6825077", "0.6821852", "0.6818972", "0.67183787", "0.6684278", "0.6676654", "0.6652459", "0.6646507", "0.664303", "0.6634662", "0.6628523", "0.6597861", "0.65875816", "0.65823644", "0.6564014", "0.65455353", "0.6542315", "0.65241", "0.65157294", "0.65138656", "0.650647", "0.6485777", "0.6470378", "0.64657044", "0.64382076", "0.6437108", "0.6431643", "0.6419845", "0.64089596", "0.64061207", "0.64052", "0.640207", "0.6390413", "0.63679624", "0.6364532", "0.63622797", "0.63594604", "0.6338381", "0.63381284", "0.6318268", "0.63125175", "0.6281451", "0.6271804", "0.6267557", "0.62669724", "0.6259036", "0.62507623", "0.62233", "0.6218213", "0.6216297", "0.6206129", "0.62057406", "0.62044656", "0.6192901", "0.61883324", "0.6180873", "0.6168208", "0.61676836", "0.6154734", "0.61427486", "0.61367905", "0.61367905", "0.6133178", "0.61271906", "0.61257046", "0.6125193", "0.61210686", "0.6118935", "0.61045784", "0.6092245", "0.6091544", "0.6086343", "0.6085684", "0.60816526", "0.60804576", "0.607789", "0.6077513", "0.60749537", "0.60613596", "0.60581183", "0.60577947", "0.60527456", "0.6049223", "0.6046077", "0.6043317", "0.6030511", "0.6027827", "0.6015263", "0.6015122", "0.6013268", "0.60127795", "0.6008334", "0.60054064", "0.60037655", "0.6001631", "0.59937763", "0.5992555", "0.59862494", "0.5982436" ]
0.0
-1
DeleteP deletes a single AssetRevision record with an executor. DeleteP will match against the primary key column to find the record to delete. Panics on error.
func (o *AssetRevision) DeleteP(exec boil.Executor) { if err := o.Delete(exec); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Transaction) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKey) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) DeleteP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := o.Delete(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *Jet) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthToken) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) DeleteP(exec boil.Executor) {\n\terr := o.Delete(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasure) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *FeatureRelationship) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q assetRevisionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Phenotypeprop) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *VSP) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no VSP provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), vspPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"vsp\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (cr APIContractRepository) Delete(ctx context.Context, projectID uint, clusterID uint, revisionID uuid.UUID) error {\n\tconf := models.APIContractRevision{\n\t\tID: revisionID,\n\t\tProjectID: int(projectID),\n\t}\n\n\tif clusterID != 0 {\n\t\tconf.ClusterID = int(clusterID)\n\t}\n\n\ttx := cr.db.Delete(&conf)\n\tif tx.Error != nil {\n\t\treturn tx.Error\n\t}\n\treturn nil\n}", "func (param *TOParameter) Delete() (error, tc.ApiErrorType) {\n\tlog.Debugf(\"about to run exec query: %s with parameter: %++v\", deleteQuery(), param)\n\tresult, err := param.ReqInfo.Tx.NamedExec(deleteQuery(), param)\n\tif err != nil {\n\t\tlog.Errorf(\"received error: %++v from delete execution\", err)\n\t\treturn tc.DBError, tc.SystemError\n\t}\n\trowsAffected, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn tc.DBError, tc.SystemError\n\t}\n\tif rowsAffected < 1 {\n\t\treturn errors.New(\"no parameter with that id found\"), tc.DataMissingError\n\t}\n\tif rowsAffected > 1 {\n\t\treturn fmt.Errorf(\"this create affected too many rows: %d\", rowsAffected), tc.SystemError\n\t}\n\n\treturn nil, tc.NoError\n}", "func (process *Process) Delete() {\n\tsql := \"DELETE FROM `process` WHERE id=?\"\n\tquery, err := database.Connection.Prepare(sql)\n\tif err != nil {\n\t\tfmt.Println(\"Delete #1 error for process:\")\n\t\tfmt.Println(err)\n\t}\n\tquery.Exec(process.Action.ID)\n}", "func (cop *ComponentOrderProcess) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif cop._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetComponentOrderProcessTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE id = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, cop.ID)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, cop.ID)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, cop.ID)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tcop._deleted = true\n\n\treturn nil\n}", "func (r *PoolNAPTRResource) Delete(id string) error {\n\tif err := r.c.ModQuery(\"DELETE\", BasePath+PoolNAPTREndpoint+\"/\"+id, nil); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o AuthorSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (p *Poet) Delete(db *sql.DB) error {\n\tvar (\n\t\tpoems int\n\t\terr error\n\t)\n\n\t// assume that we delete the poet directory within the caller scope\n\n\tif countPoetPoemsStmt == nil {\n\t\tstmt := `SELECT COUNT(*) FROM poems p WHERE p.author = $1`\n\t\tcountPoetPoemsStmt, err = db.Prepare(stmt)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// how many poems has this poet published?\n\terr = countPoetPoemsStmt.QueryRow(p.Id).Scan(&poems)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// if the poet has published at least 1 poem, we will keep its meta-data up here.\n\t// otherwise, we will delete the entire record.\n\tif poems == 0 {\n\t\t// delete this poet and its files\n\t\tif deletePoetStmt == nil {\n\t\t\tstmt := `DELETE FROM poets WHERE id = $1`\n\t\t\tdeletePoetStmt, err = db.Prepare(stmt)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\t_, err = deletePoetStmt.Exec(p.Id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t} else {\n\t\t// delete the files and update the deathdate\n\t\tif softDeletePoetStmt == nil {\n\t\t\tstmt := `UPDATE poets SET deleted = true, deathDate = $2 WHERE id = $1`\n\t\t\tsoftDeletePoetStmt, err = db.Prepare(stmt)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\t_, err = softDeletePoetStmt.Exec(p.Id, time.Now())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func (o *Vote) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Vote provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), votePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"vote\\\" WHERE \\\"hash\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (p Database) Delete(id, rev string) error {\n\theaders := map[string][]string{\n\t\t\"If-Match\": []string{rev},\n\t}\n\tu := fmt.Sprintf(\"%s/%s\", p.DBURL(), id)\n\tir := Response{}\n\tif _, err := interact(\"DELETE\", u, headers, nil, &ir); err != nil {\n\t\treturn err\n\t}\n\tif !ir.Ok {\n\t\treturn fmt.Errorf(\"%s: %s\", ir.Error, ir.Reason)\n\t}\n\treturn nil\n}", "func (o *Offer) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"stellarcore: no Offer provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), offerPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"offers\\\" WHERE \\\"offerid\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete from offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by delete for offers\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (b *Executor) Delete() (err error) {\n\tif b.builder != nil {\n\t\terr = b.builder.Delete()\n\t\tb.builder = nil\n\t}\n\treturn err\n}", "func (repository Posts) Delete(postID uint64) error {\n\tdbStatement, err := repository.db.Prepare(\n\t\t\"DELETE FROM posts WHERE id = ?\",\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer dbStatement.Close()\n\n\tif _, err = dbStatement.Exec(postID); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (q tenantQuery) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := q.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *Vote) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), votePrimaryKeyMapping)\n\tsql := \"DELETE FROM `vote` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from vote\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c VaultClient) Delete(p string) error {\n\tif c.client == nil {\n\t\treturn nil\n\t}\n\n\tkeyPath := c.prefix(p)\n\n\t_, err := c.client.Logical().Delete(keyPath)\n\tif err != nil {\n\t\tlog.WithField(\"vault.path\", keyPath).WithError(err).Trace(\"delete\")\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *Payment) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !p._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif p._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM public.payment WHERE payment_id = $1`\n\n\t// run query\n\tXOLog(sqlstr, p.PaymentID)\n\t_, err = db.Exec(sqlstr, p.PaymentID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tp._deleted = true\n\n\treturn nil\n}", "func (resolver *ResolverTODO) Delete(params graphql.ResolveParams) (interface{}, error) {\n\t_, err := resolver.Db.Delete(params.Args)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn nil, nil\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (r *RpcPool) Delete(hostport string) {\n\tr.l.Lock()\n\tdefer r.l.Unlock()\n\tdelete(r.m, hostport)\n}", "func (k *xyzProvider) Delete(ctx context.Context, req *pulumirpc.DeleteRequest) (*pbempty.Empty, error) {\n\turn := resource.URN(req.GetUrn())\n\tty := urn.Type()\n\tif ty != \"xyz:index:Random\" {\n\t\treturn nil, fmt.Errorf(\"Unknown resource type '%s'\", ty)\n\t}\n\n\t// Note that for our Random resource, we don't have to do anything on Delete.\n\treturn &pbempty.Empty{}, nil\n}", "func (o *AssetRevision) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (p *planner) Delete(n *parser.Delete, autoCommit bool) (planNode, *roachpb.Error) {\n\ttableDesc, pErr := p.getAliasedTableLease(n.Table)\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\tif err := p.checkPrivilege(tableDesc, privilege.DELETE); err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\t// TODO(tamird,pmattis): avoid going through Select to avoid encoding\n\t// and decoding keys.\n\trows, pErr := p.SelectClause(&parser.SelectClause{\n\t\tExprs: tableDesc.allColumnsSelector(),\n\t\tFrom: []parser.TableExpr{n.Table},\n\t\tWhere: n.Where,\n\t})\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\tsel := rows.(*selectNode)\n\n\trh, err := makeReturningHelper(p, n.Returning, tableDesc.Name, tableDesc.Columns)\n\tif err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\tif p.evalCtx.PrepareOnly {\n\t\t// Return the result column types.\n\t\treturn rh.getResults()\n\t}\n\n\t// Construct a map from column ID to the index the value appears at within a\n\t// row.\n\tcolIDtoRowIndex, err := makeColIDtoRowIndex(rows, tableDesc)\n\tif err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\tprimaryIndex := tableDesc.PrimaryIndex\n\tprimaryIndexKeyPrefix := MakeIndexKeyPrefix(tableDesc.ID, primaryIndex.ID)\n\n\t// Determine the secondary indexes that need to be updated as well.\n\tindexes := tableDesc.Indexes\n\t// Also include all the indexes under mutation; mutation state is\n\t// irrelevant for deletions.\n\tfor _, m := range tableDesc.Mutations {\n\t\tif index := m.GetIndex(); index != nil {\n\t\t\tindexes = append(indexes, *index)\n\t\t}\n\t}\n\n\tif isSystemConfigID(tableDesc.GetID()) {\n\t\t// Mark transaction as operating on the system DB.\n\t\tp.txn.SetSystemConfigTrigger()\n\t}\n\n\t// Check if we can avoid doing a round-trip to read the values and just\n\t// \"fast-path\" skip to deleting the key ranges without reading them first.\n\t// TODO(dt): We could probably be smarter when presented with an index-join,\n\t// but this goes away anyway once we push-down more of SQL.\n\tif scan, ok := sel.table.node.(*scanNode); ok && canDeleteWithoutScan(n, scan, len(indexes)) {\n\t\tcols, err := rh.getResults()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn p.fastDelete(scan, cols, autoCommit)\n\t}\n\n\tb := p.txn.NewBatch()\n\n\tfor rows.Next() {\n\t\trowVals := rows.Values()\n\n\t\tprimaryIndexKey, _, err := encodeIndexKey(\n\t\t\t&primaryIndex, colIDtoRowIndex, rowVals, primaryIndexKeyPrefix)\n\t\tif err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\n\t\tsecondaryIndexEntries, err := encodeSecondaryIndexes(\n\t\t\ttableDesc.ID, indexes, colIDtoRowIndex, rowVals)\n\t\tif err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\n\t\tfor _, secondaryIndexEntry := range secondaryIndexEntries {\n\t\t\tif log.V(2) {\n\t\t\t\tlog.Infof(\"Del %s\", secondaryIndexEntry.key)\n\t\t\t}\n\t\t\tb.Del(secondaryIndexEntry.key)\n\t\t}\n\n\t\t// Delete the row.\n\t\trowStartKey := roachpb.Key(primaryIndexKey)\n\t\trowEndKey := rowStartKey.PrefixEnd()\n\t\tif log.V(2) {\n\t\t\tlog.Infof(\"DelRange %s - %s\", rowStartKey, rowEndKey)\n\t\t}\n\t\tb.DelRange(rowStartKey, rowEndKey, false)\n\n\t\tif err := rh.append(rowVals); err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\t}\n\n\tif pErr := rows.PErr(); pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\tif autoCommit {\n\t\t// An auto-txn can commit the transaction with the batch. This is an\n\t\t// optimization to avoid an extra round-trip to the transaction\n\t\t// coordinator.\n\t\tpErr = p.txn.CommitInBatch(b)\n\t} else {\n\t\tpErr = p.txn.Run(b)\n\t}\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\treturn rh.getResults()\n}", "func (p *cinderProvisioner) Delete(pv *v1.PersistentVolume) error {\n\tann, ok := pv.Annotations[provisionerIDAnn]\n\tif !ok {\n\t\treturn errors.New(\"identity annotation not found on PV\")\n\t}\n\tif ann != p.identity {\n\t\treturn &controller.IgnoredError{\n\t\t\tReason: \"identity annotation on PV does not match ours\",\n\t\t}\n\t}\n\t// TODO when beta is removed, have to check kube version and pick v1/beta\n\t// accordingly: maybe the controller lib should offer a function for that\n\n\tvolumeID, ok := pv.Annotations[cinderVolumeID]\n\tif !ok {\n\t\treturn errors.New(cinderVolumeID + \" annotation not found on PV\")\n\t}\n\n\tctx := deleteCtx{p, pv}\n\tmapper, err := newVolumeMapperFromPV(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tmapper.AuthTeardown(ctx)\n\n\terr = disconnectCinderVolume(p, volumeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = unreserveCinderVolume(p, volumeID)\n\tif err != nil {\n\t\t// TODO: Create placeholder PV?\n\t\tglog.Errorf(\"Failed to unreserve volume: %v\", err)\n\t\treturn err\n\t}\n\n\terr = deleteCinderVolume(p, volumeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tglog.V(2).Infof(\"Successfully deleted cinder volume %s\", volumeID)\n\treturn nil\n}", "func (o RentalSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := o.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (p *glusterBlockProvisioner) Delete(volume *v1.PersistentVolume) error {\n\tann, ok := volume.Annotations[\"glusterBlockProvisionerIdentity\"]\n\tif !ok {\n\t\treturn errors.New(\"identity annotation not found on PV\")\n\t}\n\tif ann != string(p.identity) {\n\t\treturn &controller.IgnoredError{\"identity annotation on PV does not match ours\"}\n\t}\n\n\treturn nil\n}", "func (psh *psHandle) RecordDelete(dnsserver, domain string, rec *models.RecordConfig) error {\n\n\tvar c string\n\tif rec.Type == \"NAPTR\" {\n\t\tc = generatePSDeleteNaptr(dnsserver, domain, rec)\n\t\t//printer.Printf(\"DEBUG: deleteNAPTR: %s\\n\", c)\n\t} else {\n\t\tc = generatePSDelete(dnsserver, domain, rec)\n\t}\n\n\t//eLog(c)\n\t_, stderr, err := psh.shell.Execute(c)\n\tif err != nil {\n\t\tprinter.Printf(\"PowerShell code was:\\nSTART\\n%s\\nEND\\n\", c)\n\t\treturn err\n\t}\n\tif stderr != \"\" {\n\t\tprinter.Printf(\"STDERROR = %q\\n\", stderr)\n\t\tprinter.Printf(\"PowerShell code was:\\nSTART\\n%s\\nEND\\n\", c)\n\t\treturn fmt.Errorf(\"unexpected stderr from PSDelete: %q\", stderr)\n\t}\n\treturn nil\n}", "func (tbl DbCompoundTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := tbl.deleteRows(wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (s *Session) Delete(dest interface{}) (int64, error) {\n\ts.initStatemnt()\n\ts.statement.Delete()\n\tscanner, err := NewScanner(dest)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer scanner.Close()\n\tif s.statement.table == \"\" {\n\t\ts.statement.From(scanner.GetTableName())\n\t}\n\tpks := make([]interface{}, 0)\n\tif scanner.Model.PkName == \"\" {\n\t\treturn 0, ModelMissingPrimaryKey\n\t}\n\tif scanner.entityPointer.Kind() == reflect.Slice {\n\t\tfor i := 0; i < scanner.entityPointer.Len(); i++ {\n\t\t\tsub := scanner.entityPointer.Index(i)\n\t\t\tif sub.Kind() == reflect.Ptr {\n\t\t\t\tpks = append(pks, sub.Elem().Field(scanner.Model.PkIdx).Interface())\n\t\t\t} else {\n\t\t\t\tpks = append(pks, sub.Field(scanner.Model.PkIdx).Interface())\n\t\t\t}\n\t\t}\n\t} else if scanner.entityPointer.Kind() == reflect.Struct {\n\t\tpks = append(pks, scanner.entityPointer.Field(scanner.Model.PkIdx).Interface())\n\t} else {\n\t\treturn 0, DeleteExpectSliceOrStruct\n\t}\n\ts.Where(Eq{scanner.Model.PkName: pks})\n\tsql, args, err := s.statement.ToSQL()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\ts.logger.Debugf(\"[Session Delete] sql: %s, args: %v\", sql, args)\n\ts.initCtx()\n\tsResult, err := s.ExecContext(s.ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn sResult.RowsAffected()\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (pmt *Payment) Delete(db *bolt.DB) error {\n\tid := paymentID(pmt.Height, pmt.CreatedOn, pmt.Account)\n\treturn deleteEntry(db, paymentBkt, id)\n}", "func (o *Author) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), authorPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"authors\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete from authors\")\n\t}\n\n\treturn nil\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Where(where, args).Delete()\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (d *DB) Delete(a *Article) {\n\tst, err := d.db.Prepare(\"update articles set deleted = true where id = ?\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tdefer st.Close()\n\n\tif _, err := st.Exec(a.id); err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (o *PremiumCode) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no PremiumCode provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), premiumCodePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"premium_codes\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from premium_codes\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for premium_codes\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (t *Transaction) Delete(db *pg.DB) error {\n\tt.Status = \"canceled\"\n\t_, err := db.Model(t).WherePK().Update()\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn nil\n}", "func (i *ProjectIPServiceOp) Delete(ipReservationID string) (*Response, error) {\n\tif validateErr := ValidateUUID(ipReservationID); validateErr != nil {\n\t\treturn nil, validateErr\n\t}\n\treturn deleteFromIP(i.client, ipReservationID)\n}", "func (pubManager PublicationManager) Delete(id int64) error {\n\n\tvar title string\n\trow := pubManager.dbGetMasterFile.QueryRow(id)\n\terr := row.Scan(&title)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// delete all purchases relative to this publication\n\t_, err = pubManager.db.Exec(`DELETE FROM purchase WHERE publication_id=?`, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// delete the publication\n\t_, err = pubManager.db.Exec(\"DELETE FROM publication WHERE id = ?\", id)\n\treturn err\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFPaidprogramComment) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no CMFPaidprogramComment provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cmfPaidprogramCommentPrimaryKeyMapping)\n\tsql := \"DELETE FROM `cmf_paidprogram_comment` WHERE `id`=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from cmf_paidprogram_comment\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for cmf_paidprogram_comment\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (k Keeper) DeleteCDP(ctx sdk.Context, cdp types.CDP) {\n\tstore := prefix.NewStore(ctx.KVStore(k.key), types.CdpKeyPrefix)\n\tdb, _ := k.GetDenomPrefix(ctx, cdp.Collateral[0].Denom)\n\tstore.Delete(types.CdpKey(db, cdp.ID))\n\n}", "func (s *permStore) Delete(ctx context.Context, perm *core.Perm) error {\n\treturn s.db.Lock(func(execer db.Execer, binder db.Binder) error {\n\t\tparams := toParams(perm)\n\t\tstmt, args, err := binder.BindNamed(stmtDelete, params)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = execer.Exec(stmt, args...)\n\t\treturn err\n\t})\n}", "func (o FailureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (a *ResourcepoolApiService) DeleteResourcepoolLeaseExecute(r ApiDeleteResourcepoolLeaseRequest) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = http.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tformFiles []formFile\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"ResourcepoolApiService.DeleteResourcepoolLease\")\n\tif err != nil {\n\t\treturn nil, &GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v1/resourcepool/Leases/{Moid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"Moid\"+\"}\", url.PathEscape(parameterToString(r.moid, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, formFiles)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := &GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 401 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 403 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v Error\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tvar v Error\n\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\tif err != nil {\n\t\t\tnewErr.error = err.Error()\n\t\t\treturn localVarHTTPResponse, newErr\n\t\t}\n\t\tnewErr.model = v\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (e *engine) DeletePipeline(ctx context.Context, p *library.Pipeline) error {\n\te.logger.WithFields(logrus.Fields{\n\t\t\"pipeline\": p.GetCommit(),\n\t}).Tracef(\"deleting pipeline %s from the database\", p.GetCommit())\n\n\t// cast the library type to database type\n\t//\n\t// https://pkg.go.dev/github.com/go-vela/types/database#PipelineFromLibrary\n\tpipeline := database.PipelineFromLibrary(p)\n\n\t// send query to the database\n\treturn e.client.\n\t\tTable(constants.TablePipeline).\n\t\tDelete(pipeline).\n\t\tError\n}", "func (c *PhatClient) Delete(subpath string) error {\n\targs := &phatdb.DBCommand{\"DELETE\", subpath, \"\"}\n\t_, err := c.processCallWithRetry(args)\n\treturn err\n}", "func (dao AccountPUCDAOPsql) Delete(obj *models.AccountPUC) error {\n\tquery := \"DELETE FROM accounts_puc WHERE id = $1\"\n\tdb := get()\n\tdefer db.Close()\n\n\tstmt, err := db.Prepare(query)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tdefer stmt.Close()\n\n\tresult, err := stmt.Exec(obj.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif rowsAffected, _ := result.RowsAffected(); rowsAffected == 0 {\n\t\t\treturn errors.New(\"No se eliminó ningún registro\")\n\t}\n\tobj = new(models.AccountPUC)\n\treturn nil\n}", "func (o *Post) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"orm: no Post provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), postPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"posts\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to delete from posts\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: failed to get rows affected by delete for posts\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func FindAssetRevisionP(exec boil.Executor, orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(exec, orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func Delete(t *testing.T, tp *Tapestry) {\n\ttp.Lock()\n\tnode, _, err := tp.RandNode()\n\tif err == nil {\n\t\t// obtain a random key\n\t\tvar key string\n\t\tfor key = range node.blobstore.blobs {\n\t\t\tbreak\n\t\t}\n\n\t\t// if there is some key stored here\n\t\tif key != \"\" {\n\t\t\tif node.Remove(key) {\n\t\t\t\t// successful deletion\n\t\t\t\tdelete(tp.Blobs, key)\n\t\t\t\ttp.RemoveKey(t, key)\n\t\t\t\tOut.Printf(\"Successfully deleted %v from %v\", key, node)\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"Failed to delete key %v from %v (wasn't stored)\", key, node)\n\t\t\t}\n\t\t}\n\t}\n\ttp.Unlock()\n}", "func (q claimInListQuery) DeleteAllP(exec boil.Executor) {\n\terr := q.DeleteAll(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.7497784", "0.7497784", "0.716189", "0.7144478", "0.7141534", "0.71057224", "0.7053826", "0.70534617", "0.70474064", "0.7032722", "0.69220173", "0.6898116", "0.68363637", "0.6799", "0.6758734", "0.67540485", "0.6729994", "0.6713826", "0.6712764", "0.66883254", "0.66620076", "0.65564597", "0.6519134", "0.6515932", "0.6495175", "0.64936393", "0.6470855", "0.63994694", "0.6383703", "0.6346581", "0.6330952", "0.6298532", "0.6261497", "0.6244354", "0.60417926", "0.5767005", "0.5689529", "0.5594974", "0.558982", "0.55360013", "0.55005974", "0.5436602", "0.53775847", "0.53657866", "0.53490925", "0.53490925", "0.5318679", "0.5313865", "0.5268907", "0.5264874", "0.52558917", "0.5251896", "0.5250442", "0.5235613", "0.5233988", "0.5209127", "0.5183562", "0.5179501", "0.517677", "0.51719767", "0.5170135", "0.5155484", "0.5154907", "0.51479495", "0.51353437", "0.5122702", "0.51217395", "0.51215327", "0.51169205", "0.5109482", "0.51083773", "0.51080817", "0.5106817", "0.51017195", "0.51003903", "0.50941676", "0.5082097", "0.5082097", "0.5082097", "0.5082097", "0.5082097", "0.50732845", "0.5070391", "0.5066258", "0.5065474", "0.5065403", "0.50639224", "0.50637513", "0.5056472", "0.50546354", "0.50519997", "0.5048576", "0.5045276", "0.5034162", "0.5031796", "0.5020046", "0.50157577", "0.5010111", "0.50059426", "0.4997509" ]
0.8252762
0
DeleteG deletes a single AssetRevision record. DeleteG will match against the primary key column to find the record to delete.
func (o *AssetRevision) DeleteG() error { if o == nil { return errors.New("public: no AssetRevision provided for deletion") } return o.Delete(boil.GetDB()) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *AssetRevision) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Author) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Transaction) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *StockCvterm) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Rental) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RentalRower) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *BraceletPhoto) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Cvtermsynonym) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Offer) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Source) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (cr APIContractRepository) Delete(ctx context.Context, projectID uint, clusterID uint, revisionID uuid.UUID) error {\n\tconf := models.APIContractRevision{\n\t\tID: revisionID,\n\t\tProjectID: int(projectID),\n\t}\n\n\tif clusterID != 0 {\n\t\tconf.ClusterID = int(clusterID)\n\t}\n\n\ttx := cr.db.Delete(&conf)\n\tif tx.Error != nil {\n\t\treturn tx.Error\n\t}\n\treturn nil\n}", "func (o *Jet) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *ExchangeCurrency) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Inventory) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *AutomodRuleDatum) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *InstrumentClass) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Stock) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RSSAnnouncement) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Vote) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Failure) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *APIKey) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *BlackCard) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BlackCard provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Item) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *RecordMeasure) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RowerGroup) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Skin) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Skin provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Address) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o AssetRevisionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (og *OrderGood) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif og._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetOrderGoodTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE order_gid = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, og.OrderGid)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, og.OrderGid)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, og.OrderGid)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tog._deleted = true\n\n\treturn nil\n}", "func (o *Comment) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Ticket) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Store) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Shelf) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Shelf provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *DMessageEmbed) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no DMessageEmbed provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *ClaimInList) DeleteG() error {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Notification) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *AuthMessage) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *PremiumCode) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Auth) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Storestate) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Tenant) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *AuthToken) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Author) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (p Database) Delete(id, rev string) error {\n\theaders := map[string][]string{\n\t\t\"If-Match\": []string{rev},\n\t}\n\tu := fmt.Sprintf(\"%s/%s\", p.DBURL(), id)\n\tir := Response{}\n\tif _, err := interact(\"DELETE\", u, headers, nil, &ir); err != nil {\n\t\treturn err\n\t}\n\tif !ir.Ok {\n\t\treturn fmt.Errorf(\"%s: %s\", ir.Error, ir.Reason)\n\t}\n\treturn nil\n}", "func (o *Picture) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Organism) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *FeatureRelationship) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureRelationship provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (db *DB) Delete(id, rev string) (newrev string, err error) {\n\tpath := revpath(rev, db.name, id)\n\treturn responseRev(db.closedRequest(db.ctx, \"DELETE\", path, nil))\n}", "func (o *ItemSide) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (r *pgRepository) Delete(ctx context.Context, tenantID string, id string) error {\n\treturn r.deleter.DeleteOne(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (db *DB) Delete(ctx context.Context, docID, rev string) (newRev string, err error) {\n\treturn db.driverDB.Delete(ctx, docID, rev)\n}", "func (o *PremiumSlot) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *BookCategoryAssign) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Building) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"record: no Building provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), buildingPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"buildings\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to delete from buildings\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: failed to get rows affected by delete for buildings\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (d *DbBackendCouch) delete(database string, id string,\n\trev string) error {\n\t_, err := d.client.DB(database).Delete(id, rev)\n\treturn err\n}", "func (o *AuthUser) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func Deleting(c *gin.Context) {\n\tid := c.Param(\"id\")\n\terr := db.GetRepo().Delete(id)\n\n\tif err != nil && err.Error() == \"record not found\" {\n\t\tc.Writer.WriteHeader(404)\n\t\treturn\n\t}\n\n\tif err != nil {\n\t\tc.Writer.WriteHeader(500)\n\t\treturn\n\t}\n\tc.Writer.WriteHeader(204)\n\treturn\n}", "func (c *GsRegroup) Delete(id string) error {\n\n\treturn crud.Delete(id).Error()\n}", "func Delete(c *gin.Context) {\n\tid, err := book.IDFromString(c.Param(\"id\"))\n\tif err != nil {\n\t\tlog.Println(\"DeleteBook IDFromString\", err)\n\t\tc.JSON(\n\t\t\thttp.StatusBadRequest,\n\t\t\tgin.H{\"error\": fmt.Sprintf(\"not found book to delete by id %q\", id)},\n\t\t)\n\t\treturn\n\t}\n\tdb.Delete(id)\n\tc.JSON(http.StatusOK, gin.H{\"data\": id})\n}", "func (o *FeatureCvtermDbxref) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Where(where, args).Delete()\n}", "func (o *BookCategory) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func FindAssetRevisionG(orig int, selectCols ...string) (*AssetRevision, error) {\n\treturn FindAssetRevision(boil.GetDB(), orig, selectCols...)\n}", "func (tbl DbCompoundTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := tbl.deleteRows(wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (o *Tenant) DeleteGP(ctx context.Context) int64 {\n\trowsAff, err := o.Delete(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *AuthUserUserPermission) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (client *GCSBlobstore) Delete(dest string) error {\n\tif client.readOnly() {\n\t\treturn ErrInvalidROWriteOperation\n\t}\n\n\terr := client.getObjectHandle(client.authenticatedGCS, dest).Delete(context.Background())\n\tif err == storage.ErrObjectNotExist {\n\t\treturn nil\n\t}\n\treturn err\n}", "func (tbl AssociationTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := deleteRowsAssociationTableSql(tbl, wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (o *UsernameListing) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (mu mainUC) Delete(id uint) (interface{}, error) {\n\tcat, err := mu.recipeRepo.Find(id)\n\tif err != nil {\n\t\treturn nil, errors.New(\"recipe data not found\")\n\t}\n\tnow := time.Now()\n\tcat.DeletedAt = &now\n\n\ttx := mu.recipeRepo.DBConn().Begin()\n\terr = mu.recipeRepo.Update(cat, tx)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn nil, errors.New(\"error while deleting recipe\")\n\t}\n\n\ttx.Commit()\n\n\treturn \"recipe deleted\", err\n}", "func (gc *GoodsCategory) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif gc._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetGoodsCategoryTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE gcid = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, gc.Gcid)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, gc.Gcid)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, gc.Gcid)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tgc._deleted = true\n\n\treturn nil\n}", "func (o *Phenotypeprop) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (g *Group) DELETE(path string, h Handler, gases ...Gas) {\n\tg.Air.DELETE(g.Prefix+path, h, append(g.Gases, gases...)...)\n}", "func AssetRevisionExistsG(orig int) (bool, error) {\n\treturn AssetRevisionExists(boil.GetDB(), orig)\n}", "func (s *Badger) Delete(k []byte) error {\n\terr := s.db.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Delete(k)\n\t})\n\n\treturn err\n}", "func Delete(uid, aid string) error {\n\tdb := pg.Connection()\n\n\tdb.\n\t\tWhere(\"user_id = ? and achievement_id = ?\", uid, aid).\n\t\tDelete(&models.Accomplished{})\n\n\treturn db.Error\n}", "func (o *AssetRevision) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AssetRevision) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (g *Group) Delete(db *gorm.DB) *gorm.DB {\r\n\treturn db.Debug().Where(\"\\\"group\\\"=?\", g.Group).Delete(&g)\r\n}", "func (bs *badgerStore) Delete(key []byte) (err error) {\n\treturn bs.db.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Delete(key)\n\t})\n}", "func (t *Transaction) Delete(db *pg.DB) error {\n\tt.Status = \"canceled\"\n\t_, err := db.Model(t).WherePK().Update()\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn nil\n}", "func (o AssetRevisionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (b *BadgerDB) Delete(key Key) error {\n\tfullKey := append(key.Scope().Bytes(), key.ID()...)\n\terr := b.backend.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Delete(fullKey)\n\t})\n\n\treturn err\n}", "func BookingDelete(c *gin.Context) {\n\tif id, err := strconv.Atoi(c.Param(\"id\")); err == nil {\n\t\tif rowsDeleted, err := repo.RemoveBooking(id); err == nil {\n\t\t\tc.JSON(http.StatusOK, gin.H{\"status\": true, \"rowsDeleted\": rowsDeleted})\n\t\t} else {\n\t\t\tc.JSON(http.StatusInternalServerError, err.Error())\n\t\t}\n\t} else {\n\t\tc.JSON(http.StatusBadRequest, err.Error())\n\t}\n}", "func Delete(txn *badger.Txn, entType schema.EntityType, ID []byte) error {\n\treturn item.Delete(txn, entType.EntityID(ID))\n}", "func (asset Asset) Delete() error {\n\tres, err := asset.model.db.DB.Exec(\"delete from Assets where name = ?\", asset.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trowsDeleted, rowsDeletedErr := res.RowsAffected()\n\tif rowsDeletedErr == nil && rowsDeleted != 1 {\n\t\treturn fmt.Errorf(\"Asset.Delete should delete exactly 1 row. Instead, returned %d\", rowsDeleted)\n\t}\n\n\treturn nil\n}", "func (o *Author) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), authorPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"authors\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete from authors\")\n\t}\n\n\treturn nil\n}", "func Delete(client *gophercloud.ServiceClient, instanceID, dbName string) (r DeleteResult) {\n\tresp, err := client.Delete(dbURL(client, instanceID, dbName), nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func (c ContractValue) Delete(rst itransaction.ReadOnlyStateTrie, inst itransaction.Instruction, trans itransaction.Transaction) (sc []byte, err error) {\n\t// cout = coins\n\n\t// Find the darcID for this instance.\n\t// var darcID darc.ID\n\t// _, _, _, darcID, err = rst.GetValues(inst.InstanceID.Slice())\n\t// if err != nil {\n\t// \treturn\n\t// }\n\n\t// sc = transaction.StateChanges{\n\t// \ttransaction.NewStateChange(transaction.Remove, inst.InstanceID, ContractValueID, nil, darcID),\n\t// }\n\treturn nil, nil\n}" ]
[ "0.6695798", "0.666759", "0.6632062", "0.6631709", "0.6589712", "0.6532215", "0.65121335", "0.64525443", "0.6373659", "0.6317866", "0.6289896", "0.6241697", "0.6239944", "0.62227875", "0.6196733", "0.61940396", "0.61759335", "0.6174497", "0.6168525", "0.6164035", "0.61585027", "0.6133703", "0.61293536", "0.6115112", "0.6062118", "0.6021867", "0.60002846", "0.5995275", "0.5952465", "0.5949735", "0.5946043", "0.59181696", "0.5900075", "0.588747", "0.5883786", "0.587609", "0.58472806", "0.58273834", "0.5813884", "0.5778478", "0.5764436", "0.57543564", "0.5753466", "0.5717226", "0.56962603", "0.5694835", "0.5680073", "0.5643429", "0.56430286", "0.5636408", "0.56289333", "0.56041986", "0.5598451", "0.5572916", "0.5563004", "0.5558708", "0.5518982", "0.54927605", "0.54927605", "0.54927605", "0.54927605", "0.54927605", "0.549152", "0.548177", "0.54680413", "0.5460445", "0.5460326", "0.54494745", "0.5436381", "0.5436381", "0.5435238", "0.54304934", "0.54233706", "0.5417645", "0.5409098", "0.5406554", "0.5406209", "0.5398794", "0.5396721", "0.5396522", "0.5390664", "0.53851753", "0.5351954", "0.5345062", "0.5324031", "0.5313654", "0.53037864", "0.5302019", "0.5300382", "0.5294192", "0.5268103", "0.5258428", "0.5256873", "0.5247953", "0.5244755", "0.5237067", "0.52081865", "0.5202325", "0.5171912", "0.5170978" ]
0.811466
0
DeleteGP deletes a single AssetRevision record. DeleteGP will match against the primary key column to find the record to delete. Panics on error.
func (o *AssetRevision) DeleteGP() { if err := o.DeleteG(); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *AssetRevision) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Author) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) DeleteGP(ctx context.Context) int64 {\n\trowsAff, err := o.Delete(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o *Rental) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *APIKey) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Auth) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) DeleteGP() {\n\terr := o.Delete(boil.GetDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (og *OrderGood) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif og._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetOrderGoodTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE order_gid = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, og.OrderGid)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, og.OrderGid)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, og.OrderGid)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tog._deleted = true\n\n\treturn nil\n}", "func (o *StockCvterm) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *APIKey) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *InstrumentClass) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *AuthMessage) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func deleteGPGKey(ctx context.Context, keyID string) (int64, error) {\n\tif keyID == \"\" {\n\t\treturn 0, fmt.Errorf(\"empty KeyId forbidden\") // Should never happen but just to be sure\n\t}\n\t// Delete imported key\n\tn, err := db.GetEngine(ctx).Where(\"key_id=?\", keyID).Delete(new(GPGKeyImport))\n\tif err != nil {\n\t\treturn n, err\n\t}\n\treturn db.GetEngine(ctx).Where(\"key_id=?\", keyID).Or(\"primary_key_id=?\", keyID).Delete(new(GPGKey))\n}", "func (o *AssetRevision) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (cr APIContractRepository) Delete(ctx context.Context, projectID uint, clusterID uint, revisionID uuid.UUID) error {\n\tconf := models.APIContractRevision{\n\t\tID: revisionID,\n\t\tProjectID: int(projectID),\n\t}\n\n\tif clusterID != 0 {\n\t\tconf.ClusterID = int(clusterID)\n\t}\n\n\ttx := cr.db.Delete(&conf)\n\tif tx.Error != nil {\n\t\treturn tx.Error\n\t}\n\treturn nil\n}", "func (o *Transaction) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RecordMeasure) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Organism) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Jet) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *AuthToken) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AutomodRuleDatum) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Failure) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Author) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Vote) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (p Database) Delete(id, rev string) error {\n\theaders := map[string][]string{\n\t\t\"If-Match\": []string{rev},\n\t}\n\tu := fmt.Sprintf(\"%s/%s\", p.DBURL(), id)\n\tir := Response{}\n\tif _, err := interact(\"DELETE\", u, headers, nil, &ir); err != nil {\n\t\treturn err\n\t}\n\tif !ir.Ok {\n\t\treturn fmt.Errorf(\"%s: %s\", ir.Error, ir.Reason)\n\t}\n\treturn nil\n}", "func (o *Skin) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Inventory) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Item) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Offer) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func Delete(uid, aid string) error {\n\tdb := pg.Connection()\n\n\tdb.\n\t\tWhere(\"user_id = ? and achievement_id = ?\", uid, aid).\n\t\tDelete(&models.Accomplished{})\n\n\treturn db.Error\n}", "func (o *Picture) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RentalRower) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RSSAnnouncement) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (db *DB) Delete(ctx context.Context, docID, rev string) (newRev string, err error) {\n\treturn db.driverDB.Delete(ctx, docID, rev)\n}", "func (o *InstrumentClass) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func DeleteGPGKey(doer *user_model.User, id int64) (err error) {\n\tkey, err := GetGPGKeyByID(id)\n\tif err != nil {\n\t\tif IsErrGPGKeyNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"GetPublicKeyByID: %w\", err)\n\t}\n\n\t// Check if user has access to delete this key.\n\tif !doer.IsAdmin && doer.ID != key.OwnerID {\n\t\treturn ErrGPGKeyAccessDenied{doer.ID, key.ID}\n\t}\n\n\tctx, committer, err := db.TxContext(db.DefaultContext)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer committer.Close()\n\n\tif _, err = deleteGPGKey(ctx, key.KeyID); err != nil {\n\t\treturn err\n\t}\n\n\treturn committer.Commit()\n}", "func (o *Stock) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Rental) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (r *pgRepository) Delete(ctx context.Context, tenantID string, id string) error {\n\treturn r.deleter.DeleteOne(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (o *ExchangeCurrency) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Ticket) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *RowerGroup) DeleteG() (int64, error) {\n\treturn o.Delete(boil.GetDB())\n}", "func (o *PremiumCode) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *Phenotypeprop) DeleteGP() {\n\tif err := o.DeleteG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *BlackCard) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BlackCard provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Auth) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (xpcgi *XPriCompGroupItem) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !xpcgi._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif xpcgi._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM x_showroom.x_pri_comp_group_item WHERE id = ?`\n\n\t// run query\n\tXOLog(sqlstr, xpcgi.ID)\n\t_, err = db.Exec(sqlstr, xpcgi.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\txpcgi._deleted = true\n\n\treturn nil\n}", "func (o *AuthToken) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *RecordMeasure) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Phenotypeprop) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *FeatureCvtermDbxref) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), featureCvtermDbxrefPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"feature_cvterm_dbxref\\\" WHERE \\\"feature_cvterm_dbxref_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from feature_cvterm_dbxref\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *PremiumSlot) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *ItemSide) DeleteG(ctx context.Context) (int64, error) {\n\treturn o.Delete(ctx, boil.GetContextDB())\n}", "func (o *AuthUserUserPermission) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o *Comment) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}", "func (o AssetRevisionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Where(where, args).Delete()\n}", "func (p *glusterBlockProvisioner) Delete(volume *v1.PersistentVolume) error {\n\tann, ok := volume.Annotations[\"glusterBlockProvisionerIdentity\"]\n\tif !ok {\n\t\treturn errors.New(\"identity annotation not found on PV\")\n\t}\n\tif ann != string(p.identity) {\n\t\treturn &controller.IgnoredError{\"identity annotation on PV does not match ours\"}\n\t}\n\n\treturn nil\n}", "func Deleting(c *gin.Context) {\n\tid := c.Param(\"id\")\n\terr := db.GetRepo().Delete(id)\n\n\tif err != nil && err.Error() == \"record not found\" {\n\t\tc.Writer.WriteHeader(404)\n\t\treturn\n\t}\n\n\tif err != nil {\n\t\tc.Writer.WriteHeader(500)\n\t\treturn\n\t}\n\tc.Writer.WriteHeader(204)\n\treturn\n}", "func (c *GsRegroup) Delete(id string) error {\n\n\treturn crud.Delete(id).Error()\n}", "func FindAssetRevisionGP(orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(boil.GetDB(), orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (p Plugin) Delete(ctx context.Context, module string, vsn string) error {\n\t_, err := p.c.Delete(ctx, &stpb.DeleteRequest{Module: module, Version: vsn})\n\treturn err\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (tbl AssociationTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := deleteRowsAssociationTableSql(tbl, wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (o *FeatureRelationship) DeleteG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureRelationship provided for deletion\")\n\t}\n\n\treturn o.Delete(boil.GetDB())\n}" ]
[ "0.7120154", "0.65409833", "0.6519917", "0.6512526", "0.6512526", "0.63427454", "0.6270964", "0.6226949", "0.62237245", "0.61454755", "0.6126065", "0.61060673", "0.60876596", "0.60551214", "0.60372317", "0.6025281", "0.5987789", "0.5870752", "0.58706194", "0.58298934", "0.5818467", "0.5809334", "0.5796717", "0.57951903", "0.57917154", "0.5771915", "0.5771105", "0.57659733", "0.57581127", "0.57437295", "0.57241434", "0.57080466", "0.5695278", "0.5687112", "0.56732357", "0.56723565", "0.56659704", "0.5659814", "0.5650374", "0.5649851", "0.5648678", "0.5634026", "0.5619608", "0.56022996", "0.5591143", "0.55774385", "0.55709344", "0.55685693", "0.55628955", "0.5561473", "0.55558074", "0.5524445", "0.55071473", "0.5497621", "0.5480896", "0.5473562", "0.5467728", "0.54629695", "0.54601", "0.5449653", "0.54486316", "0.53933775", "0.53884375", "0.53822446", "0.53582096", "0.5353121", "0.5345884", "0.53450847", "0.5336698", "0.5331958", "0.5322844", "0.53138673", "0.53136474", "0.53089184", "0.5289362", "0.5282814", "0.5211798", "0.5210785", "0.51957154", "0.5183487", "0.5175669", "0.5174877", "0.51745915", "0.5174015", "0.5163729", "0.51564175", "0.515503", "0.5143988", "0.5137611", "0.51246417", "0.5113249", "0.51037186", "0.51001257", "0.5088307", "0.5088307", "0.5088307", "0.5088307", "0.5088307", "0.5083196", "0.50799954" ]
0.7723151
0
Delete deletes a single AssetRevision record with an executor. Delete will match against the primary key column to find the record to delete.
func (o *AssetRevision) Delete(exec boil.Executor) error { if o == nil { return errors.New("public: no AssetRevision provided for delete") } if err := o.doBeforeDeleteHooks(exec); err != nil { return err } args := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping) query := "DELETE FROM \"asset_revision\" WHERE \"orig\"=$1" if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, query) fmt.Fprintln(boil.DebugWriter, args...) } _, err := exec.Exec(query, args...) if err != nil { return errors.Wrap(err, "public: unable to delete from asset_revision") } if err := o.doAfterDeleteHooks(exec); err != nil { return err } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (cr APIContractRepository) Delete(ctx context.Context, projectID uint, clusterID uint, revisionID uuid.UUID) error {\n\tconf := models.APIContractRevision{\n\t\tID: revisionID,\n\t\tProjectID: int(projectID),\n\t}\n\n\tif clusterID != 0 {\n\t\tconf.ClusterID = int(clusterID)\n\t}\n\n\ttx := cr.db.Delete(&conf)\n\tif tx.Error != nil {\n\t\treturn tx.Error\n\t}\n\treturn nil\n}", "func (p Database) Delete(id, rev string) error {\n\theaders := map[string][]string{\n\t\t\"If-Match\": []string{rev},\n\t}\n\tu := fmt.Sprintf(\"%s/%s\", p.DBURL(), id)\n\tir := Response{}\n\tif _, err := interact(\"DELETE\", u, headers, nil, &ir); err != nil {\n\t\treturn err\n\t}\n\tif !ir.Ok {\n\t\treturn fmt.Errorf(\"%s: %s\", ir.Error, ir.Reason)\n\t}\n\treturn nil\n}", "func (o *Author) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), authorPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"authors\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete from authors\")\n\t}\n\n\treturn nil\n}", "func (o *Vote) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Vote provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), votePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"vote\\\" WHERE \\\"hash\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Vote) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), votePrimaryKeyMapping)\n\tsql := \"DELETE FROM `vote` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from vote\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (tbl DbCompoundTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := tbl.deleteRows(wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (o *Transaction) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), transactionPrimaryKeyMapping)\n\tsql := \"DELETE FROM `transactions` WHERE `transaction_id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from transactions\")\n\t}\n\n\treturn nil\n}", "func (db *DB) Delete(ctx context.Context, id string) error {\n\tc := db.pool.Get()\n\tdefer c.Close()\n\n\tr, err := db.Get(ctx, id)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get record: %w\", err)\n\t}\n\n\tlastVersion, err := redis.Int64(c.Do(\"INCR\", db.lastVersionKey))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.DeletedAt = ptypes.TimestampNow()\n\tr.Version = fmt.Sprintf(\"%012X\", lastVersion)\n\tb, err := proto.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmds := []map[string][]interface{}{\n\t\t{\"MULTI\": nil},\n\t\t{\"HSET\": {db.recordType, id, string(b)}},\n\t\t{\"SADD\": {db.deletedSet, id}},\n\t\t{\"ZADD\": {db.versionSet, lastVersion, id}},\n\t}\n\tif err := db.tx(c, cmds); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *RawVisit) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no RawVisit provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), rawVisitPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"raw_visits\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from raw_visits\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for raw_visits\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *AssetRevision) DeleteP(exec boil.Executor) {\n\tif err := o.Delete(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), transactionPrimaryKeyMapping)\n\tsql := \"DELETE FROM `transaction` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to delete from transaction\")\n\t}\n\n\treturn nil\n}", "func (d *DB) Delete(a *Article) {\n\tst, err := d.db.Prepare(\"update articles set deleted = true where id = ?\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tdefer st.Close()\n\n\tif _, err := st.Exec(a.id); err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func (db *DB) Delete(id, rev string) (newrev string, err error) {\n\tpath := revpath(rev, db.name, id)\n\treturn responseRev(db.closedRequest(db.ctx, \"DELETE\", path, nil))\n}", "func (o *HoldenAt) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no HoldenAt provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), holdenAtPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"HoldenAt\\\" WHERE \\\"Id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\twhere, args, err := gdb.GetWhereConditionOfStruct(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Model.Where(where, args).Delete()\n}", "func (b *Executor) Delete() (err error) {\n\tif b.builder != nil {\n\t\terr = b.builder.Delete()\n\t\tb.builder = nil\n\t}\n\treturn err\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (r *Entity) Delete() (result sql.Result, err error) {\n\treturn Model.Where(gdb.GetWhereConditionOfStruct(r)).Delete()\n}", "func (asset Asset) Delete() error {\n\tres, err := asset.model.db.DB.Exec(\"delete from Assets where name = ?\", asset.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trowsDeleted, rowsDeletedErr := res.RowsAffected()\n\tif rowsDeletedErr == nil && rowsDeleted != 1 {\n\t\treturn fmt.Errorf(\"Asset.Delete should delete exactly 1 row. Instead, returned %d\", rowsDeleted)\n\t}\n\n\treturn nil\n}", "func (o *StockCvterm) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), stockCvtermPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"stock_cvterm\\\" WHERE \\\"stock_cvterm_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from stock_cvterm\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *PaymentStorage) Delete(ctx context.Context, id aggregate.ID) error {\n\tlogger := log.FromContext(ctx)\n\n\tquery := `DELETE FROM %[1]s WHERE id = $1`\n\tquery = fmt.Sprintf(query, s.table)\n\n\tif logger != nil {\n\t\tlogger.Debugf(\"exec in transaction sql %s, values %+v\", query, []interface{}{\n\t\t\tid,\n\t\t})\n\t}\n\n\treturn execInTransaction(s.db, func(tx *sqlx.Tx) error {\n\t\t_, err := tx.ExecContext(ctx, query, id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (i *Index) Delete(tx Tx, foreignKey, primaryKey []byte) error {\n\tbkt, err := i.indexBucket(tx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn bkt.Delete(indexKey(foreignKey, primaryKey))\n}", "func (o *Comment) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), commentPrimaryKeyMapping)\n\tsql := \"DELETE FROM `comment` WHERE `index`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from comment\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *gcBlobTaskStore) Delete(ctx context.Context, b *models.GCBlobTask) error {\n\tdefer metrics.InstrumentQuery(\"gc_blob_task_delete\")()\n\n\tq := \"DELETE FROM gc_blob_review_queue WHERE digest = decode($1, 'hex')\"\n\tdgst, err := NewDigest(b.Digest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tres, err := s.db.ExecContext(ctx, q, dgst)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tcount, err := res.RowsAffected()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"deleting GC blob task: %w\", err)\n\t}\n\tif count == 0 {\n\t\treturn fmt.Errorf(\"GC blob task not found\")\n\t}\n\n\treturn nil\n}", "func (db *DB) Delete(ctx context.Context, docID, rev string) (newRev string, err error) {\n\treturn db.driverDB.Delete(ctx, docID, rev)\n}", "func (d *DbBackendCouch) delete(database string, id string,\n\trev string) error {\n\t_, err := d.client.DB(database).Delete(id, rev)\n\treturn err\n}", "func (o *Cvtermsynonym) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), cvtermsynonymPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"cvtermsynonym\\\" WHERE \\\"cvtermsynonym_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from cvtermsynonym\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (tbl AssociationTable) Delete(req require.Requirement, wh where.Expression) (int64, error) {\n\tquery, args := deleteRowsAssociationTableSql(tbl, wh)\n\treturn tbl.Exec(req, query, args...)\n}", "func (db *DB) Delete(ctx context.Context, id string) (err error) {\n\tc := db.pool.Get()\n\t_, span := trace.StartSpan(ctx, \"databroker.redis.Delete\")\n\tdefer span.End()\n\tdefer recordOperation(ctx, time.Now(), \"delete\", err)\n\tdefer c.Close()\n\n\tr, err := db.Get(ctx, id)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get record: %w\", err)\n\t}\n\n\tlastVersion, err := redis.Int64(c.Do(\"INCR\", db.lastVersionKey))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.DeletedAt = ptypes.TimestampNow()\n\tr.Version = fmt.Sprintf(\"%012X\", lastVersion)\n\tb, err := proto.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmds := []map[string][]interface{}{\n\t\t{\"MULTI\": nil},\n\t\t{\"HSET\": {db.recordType, id, string(b)}},\n\t\t{\"SADD\": {db.deletedSet, id}},\n\t\t{\"ZADD\": {db.versionSet, lastVersion, id}},\n\t}\n\tif err := db.tx(c, cmds); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c ContractValue) Delete(rst itransaction.ReadOnlyStateTrie, inst itransaction.Instruction, trans itransaction.Transaction) (sc []byte, err error) {\n\t// cout = coins\n\n\t// Find the darcID for this instance.\n\t// var darcID darc.ID\n\t// _, _, _, darcID, err = rst.GetValues(inst.InstanceID.Slice())\n\t// if err != nil {\n\t// \treturn\n\t// }\n\n\t// sc = transaction.StateChanges{\n\t// \ttransaction.NewStateChange(transaction.Remove, inst.InstanceID, ContractValueID, nil, darcID),\n\t// }\n\treturn nil, nil\n}", "func (t *Transaction) Delete(db *pg.DB) error {\n\tt.Status = \"canceled\"\n\t_, err := db.Model(t).WherePK().Update()\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn nil\n}", "func (mu mainUC) Delete(id uint) (interface{}, error) {\n\tcat, err := mu.recipeRepo.Find(id)\n\tif err != nil {\n\t\treturn nil, errors.New(\"recipe data not found\")\n\t}\n\tnow := time.Now()\n\tcat.DeletedAt = &now\n\n\ttx := mu.recipeRepo.DBConn().Begin()\n\terr = mu.recipeRepo.Update(cat, tx)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn nil, errors.New(\"error while deleting recipe\")\n\t}\n\n\ttx.Commit()\n\n\treturn \"recipe deleted\", err\n}", "func (o *Offer) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"stellarcore: no Offer provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), offerPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"offers\\\" WHERE \\\"offerid\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete from offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by delete for offers\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (r *repository) Delete(id int) error {\n\tq := \"DELETE FROM records WHERE id = $1;\"\n\t_, err := r.db.Exec(q, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (o *Vendor) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Vendor provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), vendorPrimaryKeyMapping)\n\tsql := \"DELETE FROM `vendors` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from vendors\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for vendors\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Failure) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), failurePrimaryKeyMapping)\n\tsql := \"DELETE FROM `failure` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from failure\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (v *Vessel) Delete() error {\n\tdb := adaptors.DBConnector()\n\tdefer db.Close()\n\n\t_, notFoundErr := v.FetchOne()\n\tif notFoundErr != nil {\n\t\treturn notFoundErr\n\t}\n\n\terr := db.Table(\"trades\").Where(\"id = ?\", v.ID).Update(\"deleted_at\", time.Now()).Error\n\treturn err\n}", "func (o *FeatureCvtermDbxref) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), featureCvtermDbxrefPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"feature_cvterm_dbxref\\\" WHERE \\\"feature_cvterm_dbxref_id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete from feature_cvterm_dbxref\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Rental) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), rentalPrimaryKeyMapping)\n\tsql := \"DELETE FROM `rental` WHERE `rental_id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to delete from rental\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Jet) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), jetPrimaryKeyMapping)\n\tsql := \"DELETE FROM `jets` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from jets\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (resolver *ResolverTODO) Delete(params graphql.ResolveParams) (interface{}, error) {\n\t_, err := resolver.Db.Delete(params.Args)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn nil, nil\n}", "func (o *ContentUnitDerivation) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no ContentUnitDerivation provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), contentUnitDerivationPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \\\"source_id\\\"=$1 AND \\\"derived_id\\\"=$2\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete from content_unit_derivations\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by delete for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (dt DefaultTransactor) Delete(ctx context.Context, m *movie.Movie) error {\n\ttx, err := dt.datastorer.BeginTx(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tresult, execErr := tx.ExecContext(ctx,\n\t\t`DELETE from demo.movie\n\t\t WHERE movie_id = $1`, m.ID)\n\n\tif execErr != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, execErr))\n\t}\n\n\t// Only 1 row should be deleted, check the result count to\n\t// ensure this is correct\n\trowsAffected, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\tif rowsAffected == 0 {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, errors.New(\"No Rows Deleted\")))\n\t} else if rowsAffected > 1 {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, errors.New(\"Too Many Rows Deleted\")))\n\t}\n\n\t// Commit the Transaction\n\tif err := dt.datastorer.CommitTx(tx); err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\n\treturn nil\n}", "func (o *Inventory) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), inventoryPrimaryKeyMapping)\n\tsql := \"DELETE FROM `inventory` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from inventory\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Description) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Description provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), descriptionPrimaryKeyMapping)\n\tsql := \"DELETE FROM `descriptions` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from descriptions\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for descriptions\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *BraceletPhoto) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), braceletPhotoPrimaryKeyMapping)\n\tsql := \"DELETE FROM `bracelet_photo` WHERE `id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from bracelet_photo\")\n\t}\n\n\treturn nil\n}", "func Delete(db gorp.SqlExecutor, i interface{}) error {\n\treturn Mapper.Delete(db, i)\n}", "func (m *Manager) Delete(ctx context.Context, name string) error {\n\tquery := \"select delete_chart_repository($1::uuid, $2::text)\"\n\tuserID := ctx.Value(hub.UserIDKey).(string)\n\t_, err := m.db.Exec(ctx, query, userID, name)\n\treturn err\n}", "func (r *pgRepository) Delete(ctx context.Context, tenantID string, id string) error {\n\treturn r.deleter.DeleteOne(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (_Editable *EditableTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _Editable.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (_BaseContent *BaseContentTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _BaseContent.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (o *Building) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"record: no Building provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), buildingPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"buildings\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to delete from buildings\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: failed to get rows affected by delete for buildings\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (c *Command) Delete(ctx *gin.Context) {\n\ttoken := html.EscapeString(ctx.Param(\"token\"))\n\tname := html.EscapeString(ctx.Param(\"name\"))\n\tfilter := map[string]interface{}{\"token\": token, \"name\": name}\n\tresp, err := c.Conn.GetByFilter(c.Table, filter, 1)\n\n\tif err != nil {\n\t\tutil.NiceError(ctx, err, http.StatusBadRequest)\n\t\treturn\n\t}\n\tif resp == nil {\n\t\t// Resource doesn't exist, return a 404\n\t\tctx.AbortWithStatus(http.StatusNotFound)\n\t\treturn\n\t}\n\n\trs, valid := resp[0].(map[string]interface{})\n\tif !valid {\n\t\tlog.Errorf(\"[%s] - Unable to typecast response to correct type\", c.Table)\n\t\tctx.AbortWithStatus(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t_, err = c.Conn.Disable(c.Table, rs[\"id\"].(string))\n\tif err != nil {\n\t\tutil.NiceError(ctx, err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// Success\n\tctx.Header(\"x-resource-id-removed\", rs[\"id\"].(string))\n\tctx.Status(http.StatusOK)\n}", "func (o *Currency) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Currency provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), currencyPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"currency\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for currency\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (a *Activity) Delete(ctx context.Context, key ...interface{}) error {\n\tvar err error\n\tvar dbConn *sql.DB\n\n\t// if deleted, bail\n\tif a._deleted {\n\t\treturn nil\n\t}\n\n\ttx, err := components.M.GetConnFromCtx(ctx)\n\tif err != nil {\n\t\tdbConn, err = components.M.GetMasterConn()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttableName, err := GetActivityTableName(key...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1\n\n\t// sql query with composite primary key\n\tsqlstr := `UPDATE ` + tableName + ` SET is_del = 1 WHERE id = ?`\n\n\t// run query\n\tutils.GetTraceLog(ctx).Debug(\"DB\", zap.String(\"SQL\", fmt.Sprint(sqlstr, a.ID)))\n\tif tx != nil {\n\t\t_, err = tx.Exec(sqlstr, a.ID)\n\t} else {\n\t\t_, err = dbConn.Exec(sqlstr, a.ID)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\ta._deleted = true\n\n\treturn nil\n}", "func (driver *SQLDriver) Delete(id string) error {\n\t// Execute a DELETE statement to delete the paste\n\t_, err := driver.database.Exec(\"DELETE FROM ? WHERE id = ?\", driver.table, id)\n\treturn err\n}", "func (*ActivityStageDataAccessObject) DeleteByAID(aid int) {\n\tvar buf ActivityStage\n\t_, err := orm.Table(ActivityStageDAO.TableName()).\n\t\tWhere(\"activity_id=?\", aid).Delete(&buf)\n\tlogger.LogIfError(err)\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _BaseAccessControlGroup.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (_BaseLibrary *BaseLibraryTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _BaseLibrary.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (s *DbRecorder) Delete() error {\n\twheres := s.WhereIds()\n\tq := s.builder.Delete(s.table).Where(wheres)\n\t_, err := q.Exec()\n\treturn err\n}", "func (u *App) Delete(c echo.Context, id string) error {\n\tif err := u.rbac.EnforceRole(c, model.AdminRole); err != nil {\n\t\treturn err\n\t}\n\n\tpost, err := u.udb.View(u.db, id)\n\tif err = zaplog.ZLog(err); err != nil {\n\t\treturn err\n\t}\n\n\tif post.Status != model.StatusDraft {\n\t\treturn zaplog.ZLog(errors.New(\"Apenas é possível eliminar artigos em rascunho\"))\n\t}\n\n\treturn u.udb.Delete(u.db, id)\n}", "func (o *PaymentObject) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no PaymentObject provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), paymentObjectPrimaryKeyMapping)\n\tsql := \"DELETE FROM `payment_objects` WHERE `id`=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from payment_objects\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for payment_objects\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *InstrumentClass) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), instrumentClassPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"instruments\\\".\\\"instrument_class\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from instrument_class\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *APIKey) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), apiKeyPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"api_keys\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from api_keys\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *ActivityLog) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"dbmodel: no ActivityLog provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), activityLogPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"activity_logs\\\" WHERE \\\"id\\\"=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to delete from activity_logs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by delete for activity_logs\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (_BaseContentType *BaseContentTypeTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _BaseContentType.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func deleteRow(index int, settings string) {\n\tfmt.Println(\"# Deleting\")\n\n\tdb := connectDB(settings)\n\tstmt, _ := db.Prepare(\"delete from test where test_id=$1\")\n\n\tres, _ := stmt.Exec(index)\n\n\taffect, _ := res.RowsAffected()\n\n\tfmt.Println(affect, \"rows changed\")\n}", "func (o *Segment) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"boiler: no Segment provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), segmentPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"segment\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to delete from segment\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: failed to get rows affected by delete for segment\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (dcc *DoorCompiledContent) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !dcc._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif dcc._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM public.door_compiled_content WHERE id = $1`\n\n\t// run query\n\tXOLog(sqlstr, dcc.ID)\n\t_, err = db.Exec(sqlstr, dcc.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tdcc._deleted = true\n\n\treturn nil\n}", "func (o *PublisherSearchIdx) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no PublisherSearchIdx provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), publisherSearchIdxPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"publisher_search_idx\\\" WHERE \\\"segid\\\"=? AND \\\"term\\\"=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from publisher_search_idx\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for publisher_search_idx\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (fr *FetchResult) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !fr._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif fr._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM fetch_result WHERE id = ?`\n\n\t// run query\n\tXOLog(sqlstr, fr.ID)\n\t_, err = db.Exec(sqlstr, fr.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\tfr._deleted = true\n\n\treturn nil\n}", "func (rest *RestController) Delete(w http.ResponseWriter, r *http.Request) (Response, error) {\n\terr := rest.Table.Delete(models.NewDBQuery(nil, map[string]string{\"id\": getParams(r).ByName(\"id\")}))\n\tif err != nil {\n\t\treturn nil, &httpError{err, \"\", 500}\n\t}\n\treturn &TextResponse{\"\", 204}, nil\n}", "func (o *Stock) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), stockPrimaryKeyMapping)\n\tsql := \"DELETE FROM `stock` WHERE `stock_id`=?\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete from stock\")\n\t}\n\n\treturn nil\n}", "func (o *Repository) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Repository provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), repositoryPrimaryKeyMapping)\n\tsql := \"DELETE FROM `repositories` WHERE `id`=?\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from repositories\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for repositories\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func Delete(table modelResource.Resource, idVal int) (err error){\n\treturn nil\n}", "func (p *planner) Delete(n *parser.Delete, autoCommit bool) (planNode, *roachpb.Error) {\n\ttableDesc, pErr := p.getAliasedTableLease(n.Table)\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\tif err := p.checkPrivilege(tableDesc, privilege.DELETE); err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\t// TODO(tamird,pmattis): avoid going through Select to avoid encoding\n\t// and decoding keys.\n\trows, pErr := p.SelectClause(&parser.SelectClause{\n\t\tExprs: tableDesc.allColumnsSelector(),\n\t\tFrom: []parser.TableExpr{n.Table},\n\t\tWhere: n.Where,\n\t})\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\tsel := rows.(*selectNode)\n\n\trh, err := makeReturningHelper(p, n.Returning, tableDesc.Name, tableDesc.Columns)\n\tif err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\tif p.evalCtx.PrepareOnly {\n\t\t// Return the result column types.\n\t\treturn rh.getResults()\n\t}\n\n\t// Construct a map from column ID to the index the value appears at within a\n\t// row.\n\tcolIDtoRowIndex, err := makeColIDtoRowIndex(rows, tableDesc)\n\tif err != nil {\n\t\treturn nil, roachpb.NewError(err)\n\t}\n\n\tprimaryIndex := tableDesc.PrimaryIndex\n\tprimaryIndexKeyPrefix := MakeIndexKeyPrefix(tableDesc.ID, primaryIndex.ID)\n\n\t// Determine the secondary indexes that need to be updated as well.\n\tindexes := tableDesc.Indexes\n\t// Also include all the indexes under mutation; mutation state is\n\t// irrelevant for deletions.\n\tfor _, m := range tableDesc.Mutations {\n\t\tif index := m.GetIndex(); index != nil {\n\t\t\tindexes = append(indexes, *index)\n\t\t}\n\t}\n\n\tif isSystemConfigID(tableDesc.GetID()) {\n\t\t// Mark transaction as operating on the system DB.\n\t\tp.txn.SetSystemConfigTrigger()\n\t}\n\n\t// Check if we can avoid doing a round-trip to read the values and just\n\t// \"fast-path\" skip to deleting the key ranges without reading them first.\n\t// TODO(dt): We could probably be smarter when presented with an index-join,\n\t// but this goes away anyway once we push-down more of SQL.\n\tif scan, ok := sel.table.node.(*scanNode); ok && canDeleteWithoutScan(n, scan, len(indexes)) {\n\t\tcols, err := rh.getResults()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn p.fastDelete(scan, cols, autoCommit)\n\t}\n\n\tb := p.txn.NewBatch()\n\n\tfor rows.Next() {\n\t\trowVals := rows.Values()\n\n\t\tprimaryIndexKey, _, err := encodeIndexKey(\n\t\t\t&primaryIndex, colIDtoRowIndex, rowVals, primaryIndexKeyPrefix)\n\t\tif err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\n\t\tsecondaryIndexEntries, err := encodeSecondaryIndexes(\n\t\t\ttableDesc.ID, indexes, colIDtoRowIndex, rowVals)\n\t\tif err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\n\t\tfor _, secondaryIndexEntry := range secondaryIndexEntries {\n\t\t\tif log.V(2) {\n\t\t\t\tlog.Infof(\"Del %s\", secondaryIndexEntry.key)\n\t\t\t}\n\t\t\tb.Del(secondaryIndexEntry.key)\n\t\t}\n\n\t\t// Delete the row.\n\t\trowStartKey := roachpb.Key(primaryIndexKey)\n\t\trowEndKey := rowStartKey.PrefixEnd()\n\t\tif log.V(2) {\n\t\t\tlog.Infof(\"DelRange %s - %s\", rowStartKey, rowEndKey)\n\t\t}\n\t\tb.DelRange(rowStartKey, rowEndKey, false)\n\n\t\tif err := rh.append(rowVals); err != nil {\n\t\t\treturn nil, roachpb.NewError(err)\n\t\t}\n\t}\n\n\tif pErr := rows.PErr(); pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\tif autoCommit {\n\t\t// An auto-txn can commit the transaction with the batch. This is an\n\t\t// optimization to avoid an extra round-trip to the transaction\n\t\t// coordinator.\n\t\tpErr = p.txn.CommitInBatch(b)\n\t} else {\n\t\tpErr = p.txn.Run(b)\n\t}\n\tif pErr != nil {\n\t\treturn nil, pErr\n\t}\n\n\treturn rh.getResults()\n}", "func (o *VSP) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no VSP provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), vspPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"vsp\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o *Task) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Task provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), taskPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"tasks\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for tasks\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (t BlobsTable) Delete(ctx context.Context, query string, args ...interface{}) error {\n\treturn t.driver.delete(ctx, query, args...)\n}", "func (_Container *ContainerTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _Container.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (c *Chef) Delete(id string) error {\n\treturn c.db.Exec(fmt.Sprintf(`UPDATE %s SET (\"deleted\", \"deleted_at\") = (TRUE, CURRENT_TIMESTAMP) WHERE \"id\"='%s' AND \"deleted\"=FALSE`, c.table, id))\n}", "func (o *RSSAnnouncement) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no RSSAnnouncement provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), rssAnnouncementPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"rss_announcements\\\" WHERE \\\"guild_id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from rss_announcements\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for rss_announcements\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (repo *sqlRepository) Delete(id int, v interface{}) error{\n\tvar (\n\t\terr error\n\t\tdb *sql.DB\n\t\tstmt *sql.Stmt\n \t )\n db, _ = sql.Open(driver, repo.user + \":\" + repo.password + \"@tcp(\" + repo.host + \":\" + repo.port + \")/\" + repo.database)\n\tstmt, err = db.Prepare(\"delete from users where id=?\")\n checkErr(err)\n\n res, err := stmt.Exec(id)\n checkErr(err)\n\n affect, err := res.RowsAffected()\n checkErr(err)\n\n fmt.Println(affect)\n\n db.Close()\n\treturn nil\n}", "func (_BaseContentSpace *BaseContentSpaceTransactor) DeleteVersion(opts *bind.TransactOpts, _versionHash string) (*types.Transaction, error) {\n\treturn _BaseContentSpace.contract.Transact(opts, \"deleteVersion\", _versionHash)\n}", "func (s *Session) Delete(dest interface{}) (int64, error) {\n\ts.initStatemnt()\n\ts.statement.Delete()\n\tscanner, err := NewScanner(dest)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer scanner.Close()\n\tif s.statement.table == \"\" {\n\t\ts.statement.From(scanner.GetTableName())\n\t}\n\tpks := make([]interface{}, 0)\n\tif scanner.Model.PkName == \"\" {\n\t\treturn 0, ModelMissingPrimaryKey\n\t}\n\tif scanner.entityPointer.Kind() == reflect.Slice {\n\t\tfor i := 0; i < scanner.entityPointer.Len(); i++ {\n\t\t\tsub := scanner.entityPointer.Index(i)\n\t\t\tif sub.Kind() == reflect.Ptr {\n\t\t\t\tpks = append(pks, sub.Elem().Field(scanner.Model.PkIdx).Interface())\n\t\t\t} else {\n\t\t\t\tpks = append(pks, sub.Field(scanner.Model.PkIdx).Interface())\n\t\t\t}\n\t\t}\n\t} else if scanner.entityPointer.Kind() == reflect.Struct {\n\t\tpks = append(pks, scanner.entityPointer.Field(scanner.Model.PkIdx).Interface())\n\t} else {\n\t\treturn 0, DeleteExpectSliceOrStruct\n\t}\n\ts.Where(Eq{scanner.Model.PkName: pks})\n\tsql, args, err := s.statement.ToSQL()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\ts.logger.Debugf(\"[Session Delete] sql: %s, args: %v\", sql, args)\n\ts.initCtx()\n\tsResult, err := s.ExecContext(s.ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn sResult.RowsAffected()\n}", "func (bb *BooktestBook) Delete(ctx context.Context, db DB) error {\n\tswitch {\n\tcase !bb._exists: // doesn't exist\n\t\treturn nil\n\tcase bb._deleted: // deleted\n\t\treturn nil\n\t}\n\t// delete with single primary key\n\tconst sqlstr = `DELETE FROM public.booktest_book ` +\n\t\t`WHERE book_id = $1`\n\t// run\n\tlogf(sqlstr, bb.BookID)\n\tif _, err := db.ExecContext(ctx, sqlstr, bb.BookID); err != nil {\n\t\treturn logerror(err)\n\t}\n\t// set deleted\n\tbb._deleted = true\n\treturn nil\n}", "func (table *Table) Delete(db DB, record Map) (Result, error) {\n\tc := Context{StateDelete, db, table, table, record, Field{}}\n\treturn table.execAction(c, table.OnDelete, table.DefaultDelete)\n}", "func (r *DbAuthorRepository) Delete(author Author) (err error) {\n\terr = r.db.Delete(author).Error\n\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete author with id '%d': %s\", author.ID, err)\n\t} else {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"ID\": author.ID,\n\t\t\t\"Name\": author.Name,\n\t\t}).Debug(\"Deleted author\")\n\t}\n\n\treturn err\n}", "func (hh *HealthCheckHandler) Delete(w http.ResponseWriter, r *http.Request) {\n\tuuid := utils.ExtractUUID(r.URL.String())\n\thh.db.Delete(uuid)\n}", "func (h *EcrHandler) Delete(obj interface{}) error {\n\treturn nil\n}", "func (o *Block) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Block provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), blockPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"block\\\" WHERE \\\"height\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from block\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (f *FakeTable) DeleteRow(ovs *libovsdb.OvsdbClient, condition []string) error {\n\treturn nil\n}", "func (o *ExchangeCurrency) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no ExchangeCurrency provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), exchangeCurrencyPrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"exchange_currency\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from exchange_currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for exchange_currency\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func deleteRow(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\n\tvar numberOfArgs int = 2\n\tif len(args) != numberOfArgs {\n\t\treturn nil, errors.New(\"Incorrect number of arguments. Expecting: \" + strconv.Itoa(numberOfArgs))\n\t}\n\n\ttableName, keyValue := args[0], args[1]\n\tvar cols []shim.Column\n\tcol := shim.Column{Value: &shim.Column_String_{String_: keyValue}}\n\tcols = append(cols, col)\n\n\terr := stub.DeleteRow(tableName, cols)\n\n\tif err != nil {\n\t\treturn nil, errors.New(\"Failed to delete row with key '\" + keyValue + \"' from '\" + tableName + \"' table: \" + err.Error())\n\t}\n\n\tfmt.Println(\"Successfuly deleted row with key '\" + keyValue + \"' from '\" + tableName + \"' table if any exists\")\n\treturn nil, nil\n}", "func (rd *Deleter) DeleteRow(\n\tctx context.Context, b *kv.Batch, values []tree.Datum, pm PartialIndexUpdateHelper, traceKV bool,\n) error {\n\n\t// Delete the row from any secondary indices.\n\tfor i := range rd.Helper.Indexes {\n\t\t// If the index ID exists in the set of indexes to ignore, do not\n\t\t// attempt to delete from the index.\n\t\tif pm.IgnoreForDel.Contains(int(rd.Helper.Indexes[i].ID)) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// We want to include empty k/v pairs because we want to delete all k/v's for this row.\n\t\tentries, err := rowenc.EncodeSecondaryIndex(\n\t\t\trd.Helper.Codec,\n\t\t\trd.Helper.TableDesc,\n\t\t\t&rd.Helper.Indexes[i],\n\t\t\trd.FetchColIDtoRowIndex,\n\t\t\tvalues,\n\t\t\ttrue, /* includeEmpty */\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, e := range entries {\n\t\t\tif traceKV {\n\t\t\t\tlog.VEventf(ctx, 2, \"Del %s\", keys.PrettyPrint(rd.Helper.secIndexValDirs[i], e.Key))\n\t\t\t}\n\t\t\tb.Del(&e.Key)\n\t\t}\n\t}\n\n\tprimaryIndexKey, err := rd.Helper.encodePrimaryIndex(rd.FetchColIDtoRowIndex, values)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Delete the row.\n\tvar called bool\n\treturn rd.Helper.TableDesc.ForeachFamily(func(family *descpb.ColumnFamilyDescriptor) error {\n\t\tif called {\n\t\t\t// HACK: MakeFamilyKey appends to its argument, so on every loop iteration\n\t\t\t// after the first, trim primaryIndexKey so nothing gets overwritten.\n\t\t\t// TODO(dan): Instead of this, use something like engine.ChunkAllocator.\n\t\t\tprimaryIndexKey = primaryIndexKey[:len(primaryIndexKey):len(primaryIndexKey)]\n\t\t} else {\n\t\t\tcalled = true\n\t\t}\n\t\tfamilyID := family.ID\n\t\trd.key = keys.MakeFamilyKey(primaryIndexKey, uint32(familyID))\n\t\tif traceKV {\n\t\t\tlog.VEventf(ctx, 2, \"Del %s\", keys.PrettyPrint(rd.Helper.primIndexValDirs, rd.key))\n\t\t}\n\t\tb.Del(&rd.key)\n\t\trd.key = nil\n\t\treturn nil\n\t})\n}", "func (m *LogDAO) Delete(logMessage model.LogMessage) error {\n\t_, err := db.Model(logMessage).WherePK().Delete()\n\treturn err\n}", "func (r *TaskRepository) Delete(db db.DB, Task *entities.Task) error {\n\t_, err := db.Exec(`UPDATE tasks SET status=$1, deleted_at=now() WHERE id=$2`, constants.Deleted, Task.ID)\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error deleting task: %w\", err)\n\t}\n\n\treturn nil\n}" ]
[ "0.6670579", "0.6300614", "0.62432146", "0.61708677", "0.60445356", "0.59982854", "0.59461766", "0.59451926", "0.59322256", "0.59234744", "0.59060895", "0.58941245", "0.5886612", "0.5868289", "0.58577126", "0.5826225", "0.58236575", "0.58236575", "0.58236575", "0.58236575", "0.58236575", "0.57877946", "0.57558656", "0.574074", "0.57406527", "0.57203573", "0.5664379", "0.5649883", "0.5649147", "0.56441534", "0.56418127", "0.5625724", "0.5621642", "0.5617487", "0.55837935", "0.5581773", "0.5573443", "0.5568515", "0.55634016", "0.5560558", "0.555935", "0.55253303", "0.55193126", "0.5518619", "0.5517307", "0.5515303", "0.5475209", "0.5470493", "0.54693025", "0.546759", "0.54631704", "0.54611117", "0.545496", "0.5443289", "0.542706", "0.5417105", "0.54153365", "0.5411874", "0.54100776", "0.54033893", "0.5397213", "0.53951615", "0.5393972", "0.5390231", "0.5386659", "0.53832805", "0.53743297", "0.5372227", "0.5372192", "0.5368391", "0.5367038", "0.53665614", "0.5363636", "0.53548723", "0.5349589", "0.5348521", "0.5339764", "0.5333035", "0.5329964", "0.5323048", "0.5322931", "0.5303309", "0.5302565", "0.530118", "0.5299698", "0.52953446", "0.5295108", "0.5294159", "0.52919567", "0.5282174", "0.52668214", "0.5264304", "0.5261647", "0.52607554", "0.5260245", "0.5258617", "0.5254176", "0.5253394", "0.5248191", "0.52447367" ]
0.7904476
0
DeleteAllP deletes all rows, and panics on error.
func (q assetRevisionQuery) DeleteAllP() { if err := q.DeleteAll(); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q failureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q sourceQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q apiKeyQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q voteQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q inventoryQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q skinQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockCvtermQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authorQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := q.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (q authUserQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q shelfQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q instrumentClassQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q organismQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserUserPermissionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authMessageQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authTokenQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q commentQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q blackCardQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q addressQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q braceletPhotoQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cvtermsynonymQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q recordMeasureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q pictureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q dMessageEmbedQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q claimInListQuery) DeleteAllP(exec boil.Executor) {\n\terr := q.DeleteAll(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureCvtermDbxrefQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SkinSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := o.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o AddressSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q featureRelationshipQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AssetRevisionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ClaimInListSlice) DeleteAllP(exec boil.Executor) {\n\terr := o.DeleteAll(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) DeleteAllGP(ctx context.Context) int64 {\n\trowsAff, err := q.DeleteAll(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (q rowerGroupQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rowerGroupQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q vspQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no vspQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vsp\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q currentChartDataMinutelyQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no currentChartDataMinutelyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from current_chart_data_minutely\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for current_chart_data_minutely\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q kvstoreQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no kvstoreQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q offerQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"stellarcore: no offerQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete all from offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by deleteall for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no voteQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q recipeLipidQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no recipeLipidQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from recipe_lipid\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for recipe_lipid\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FailureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no jetQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from jets\")\n\t}\n\n\treturn nil\n}", "func (p *HbaseClient) DeleteAll(tableName Text, row Text, column Text, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAll(tableName, row, column, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAll()\n}", "func (q currencyQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no currencyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func TipoDeleteAll() (err error) {\n\tstatement := \"delete from tipos\"\n\t_, err = Db.Exec(statement)\n\treturn\n}", "func (q premiumCodeQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no premiumCodeQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premium_codes\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_codes\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q sourceQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no sourceQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from sources\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for sources\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q taskQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no taskQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o JetSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q phenotypepropQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"chado: no phenotypepropQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete all from phenotypeprop\")\n\t}\n\n\treturn nil\n}", "func (q apiKeyQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no apiKeyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from api_keys\")\n\t}\n\n\treturn nil\n}", "func (q withdrawalCryptoQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"sqlite3: no withdrawalCryptoQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to delete all from withdrawal_crypto\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: failed to get rows affected by deleteall for withdrawal_crypto\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q holdenAtQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no holdenAtQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q mempoolBinQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no mempoolBinQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from mempool_bin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for mempool_bin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no voteQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from vote\")\n\t}\n\n\treturn nil\n}", "func DeleteAll(db *sqlx.DB) error {\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := tx.Exec(deleteDoc); err != nil {\n\t\tif err := tx.Rollback(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (q illnessQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no illnessQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from illness\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for illness\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q failureQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no failureQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from failure\")\n\t}\n\n\treturn nil\n}", "func (q publisherSearchIdxQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no publisherSearchIdxQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from publisher_search_idx\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for publisher_search_idx\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (t *Table) ForceDeleteAll() error {\n\thashkey := t.design.GetHashKeyName()\n\trangekey := t.design.GetRangeKeyName()\n\n\tresult, err := t.Scan()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terrData := newErrors()\n\tfor _, item := range result.ToSliceMap() {\n\t\tvar e error\n\t\tswitch rangekey {\n\t\tcase \"\":\n\t\t\te = t.Delete(item[hashkey])\n\t\tdefault:\n\t\t\te = t.Delete(item[hashkey], item[rangekey])\n\t\t}\n\n\t\tif e != nil {\n\t\t\terrData.Add(e)\n\t\t}\n\t}\n\n\tif errData.HasError() {\n\t\treturn errData\n\t}\n\treturn nil\n}", "func (q btcTXOutputQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no btcTXOutputQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from btc_tx_output\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.8199568", "0.81305957", "0.81104624", "0.81104624", "0.8091458", "0.7972003", "0.7968659", "0.79678005", "0.7915315", "0.79041773", "0.78835475", "0.7836509", "0.78364974", "0.7796686", "0.7791324", "0.777531", "0.77542335", "0.7753871", "0.7753871", "0.7743663", "0.7722308", "0.7721258", "0.7665055", "0.7657044", "0.7646988", "0.7630666", "0.7608932", "0.75990736", "0.7596724", "0.75941646", "0.7585568", "0.75767297", "0.75740725", "0.756433", "0.7544422", "0.75369245", "0.7536616", "0.75019324", "0.7464392", "0.745582", "0.7448366", "0.7428039", "0.7424906", "0.7392737", "0.73879087", "0.7387138", "0.7367617", "0.73653424", "0.73116744", "0.73072594", "0.7306369", "0.7296678", "0.7291439", "0.7256298", "0.7221939", "0.721576", "0.71945727", "0.7177935", "0.71706337", "0.71696", "0.712953", "0.71209425", "0.71027625", "0.7013937", "0.69789124", "0.6946882", "0.6911634", "0.6906192", "0.6698073", "0.6519254", "0.65120286", "0.65108514", "0.6460746", "0.6460746", "0.6419858", "0.6415332", "0.64071", "0.6387773", "0.63849753", "0.63826776", "0.6366181", "0.63575864", "0.63557553", "0.63492805", "0.6302561", "0.62919545", "0.62721235", "0.6263598", "0.6260361", "0.6252477", "0.62521994", "0.62469053", "0.6245065", "0.62378705", "0.62373465", "0.6237187", "0.62250334", "0.62248075", "0.62215686", "0.6198849" ]
0.75360316
37
DeleteAll deletes all matching rows.
func (q assetRevisionQuery) DeleteAll() error { if q.Query == nil { return errors.New("public: no assetRevisionQuery provided for delete all") } queries.SetDelete(q.Query) _, err := q.Query.Exec() if err != nil { return errors.Wrap(err, "public: unable to delete all from asset_revision") } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *HbaseClient) DeleteAll(tableName Text, row Text, column Text, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAll(tableName, row, column, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAll()\n}", "func (q rowerGroupQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rowerGroupQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q holdenAtQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no holdenAtQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from HoldenAt\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q cmfTurntableQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no cmfTurntableQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmf_turntable\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_turntable\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (t *Table) ForceDeleteAll() error {\n\thashkey := t.design.GetHashKeyName()\n\trangekey := t.design.GetRangeKeyName()\n\n\tresult, err := t.Scan()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terrData := newErrors()\n\tfor _, item := range result.ToSliceMap() {\n\t\tvar e error\n\t\tswitch rangekey {\n\t\tcase \"\":\n\t\t\te = t.Delete(item[hashkey])\n\t\tdefault:\n\t\t\te = t.Delete(item[hashkey], item[rangekey])\n\t\t}\n\n\t\tif e != nil {\n\t\t\terrData.Add(e)\n\t\t}\n\t}\n\n\tif errData.HasError() {\n\t\treturn errData\n\t}\n\treturn nil\n}", "func (q jetQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no jetQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from jets\")\n\t}\n\n\treturn nil\n}", "func (q weatherQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"db: no weatherQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to delete all from weather\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by deleteall for weather\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q kvstoreQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no kvstoreQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for kvstore\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q recipeLipidQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no recipeLipidQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from recipe_lipid\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for recipe_lipid\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currencyQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no currencyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (d *Demo) DeleteAll(g *gom.Gom) {\n\ttoolkit.Println(\"===== Delete All =====\")\n\n\tvar err error\n\tif d.useParams {\n\t\t_, err = g.Set(&gom.SetParams{\n\t\t\tTableName: \"hero\",\n\t\t\tFilter: gom.EndWith(\"Name\", \"man\"),\n\t\t\tTimeout: 10,\n\t\t}).Cmd().DeleteAll()\n\t} else {\n\t\t_, err = g.Set(nil).Table(\"hero\").Timeout(10).Filter(gom.EndWith(\"Name\", \"man\")).Cmd().DeleteAll()\n\t}\n\n\tif err != nil {\n\t\ttoolkit.Println(err.Error())\n\t\treturn\n\t}\n}", "func (q rentalRowerQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rentalRowerQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rental_rowers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rental_rowers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q buildingQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"record: no buildingQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to delete all from buildings\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: failed to get rows affected by deleteall for buildings\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q sourceQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no sourceQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from sources\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for sources\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no voteQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vote\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (table *ClassTable) DeleteAll() (err error) {\n\tquery := fmt.Sprintf(\"SELECT id FROM %s\", CLASS_TABLE)\n\n\tutilities.Sugar.Infof(\"SQL Query: %v\", query)\n\trows, err := table.connection.Pool.Query(query)\n\tif err != nil {\n\t\terr = errors.Wrapf(err, \"Delete all query failed\")\n\t\treturn\n\t}\n\t// Delete all the classes by calling the relational delete\n\tfor rows.Next() {\n\t\tvar id int64\n\t\tif err = rows.Scan(&id); err != nil {\n\t\t\terr = errors.Wrapf(err, \"Failed to scan into id\")\n\t\t\treturn\n\t\t}\n\t\tif err = table.Delete(strconv.FormatInt(id, 10)); err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (q withdrawalCryptoQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"sqlite3: no withdrawalCryptoQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to delete all from withdrawal_crypto\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: failed to get rows affected by deleteall for withdrawal_crypto\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q currentChartDataMinutelyQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no currentChartDataMinutelyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from current_chart_data_minutely\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for current_chart_data_minutely\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q btcTXOutputQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no btcTXOutputQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from btc_tx_output\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q jobQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no jobQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from jobs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for jobs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q illnessQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no illnessQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from illness\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for illness\")\n\t}\n\n\treturn rowsAff, nil\n}", "func DeleteAll(db *sql.DB, table string) error {\n if _, err := db.Exec(fmt.Sprintf(\"DELETE FROM %s\", table)); err != nil {\n return err\n }\n return nil\n}", "func (q apiKeyQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no apiKeyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from api_keys\")\n\t}\n\n\treturn nil\n}", "func DeleteAll(db *sqlx.DB) error {\n\ttx, err := db.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := tx.Exec(deleteDoc); err != nil {\n\t\tif err := tx.Rollback(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (q oauthClientQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no oauthClientQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from oauth_clients\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for oauth_clients\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q mempoolBinQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no mempoolBinQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from mempool_bin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for mempool_bin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q *Q) DeleteAll(dest api.IRecord) (affected int, err error) {\n\tif q.Mock != nil && q.Mock.Enabled() {\n\t\treturn q.Mock.Int(), q.Mock.Error()\n\t}\n\n\tresult, err := q.db.Exec(fmt.Sprintf(`DELETE FROM %s`, dest.Table()))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn q.db.AffectedRows(result), err\n}", "func (p *HbaseClient) DeleteAllTs(tableName Text, row Text, column Text, timestamp int64, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAllTs(tableName, row, column, timestamp, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAllTs()\n}", "func (q chatQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"dal: no chatQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to delete all from chat\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: failed to get rows affected by deleteall for chat\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q activityLogQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"dbmodel: no activityLogQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to delete all from activity_logs\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by deleteall for activity_logs\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (h *DBHandle) DeleteAll() error {\n\titer, err := h.GetIterator(nil, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer iter.Release()\n\n\t// use leveldb iterator directly to be more efficient\n\tdbIter := iter.Iterator\n\n\t// This is common code shared by all the leveldb instances. Because each leveldb has its own key size pattern,\n\t// each batch is limited by memory usage instead of number of keys. Once the batch memory usage reaches maxBatchSize,\n\t// the batch will be committed.\n\tnumKeys := 0\n\tbatchSize := 0\n\tbatch := &leveldb.Batch{}\n\tfor dbIter.Next() {\n\t\tif err := dbIter.Error(); err != nil {\n\t\t\treturn errors.Wrap(err, \"internal leveldb error while retrieving data from db iterator\")\n\t\t}\n\t\tkey := dbIter.Key()\n\t\tnumKeys++\n\t\tbatchSize = batchSize + len(key)\n\t\tbatch.Delete(key)\n\t\tif batchSize >= maxBatchSize {\n\t\t\tif err := h.db.WriteBatch(batch, true); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tlogger.Infof(\"Have removed %d entries for channel %s in leveldb %s\", numKeys, h.dbName, h.db.conf.DBPath)\n\t\t\tbatchSize = 0\n\t\t\tbatch = &leveldb.Batch{}\n\t\t}\n\t}\n\tif batch.Len() > 0 {\n\t\treturn h.db.WriteBatch(batch, true)\n\t}\n\treturn nil\n}", "func (q emailQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"mysql: no emailQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to delete all from email\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: failed to get rows affected by deleteall for email\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q instrumentClassQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no instrumentClassQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from instrument_class\")\n\t}\n\n\treturn nil\n}", "func (q automodRuleDatumQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no automodRuleDatumQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from automod_rule_data\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for automod_rule_data\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q sourceQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q apiKeyQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"sqlboiler: no rentalQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to delete all from rental\")\n\t}\n\n\treturn nil\n}", "func (q itemQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no itemQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from items\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for items\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q blockQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no blockQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from block\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q publisherSearchIdxQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no publisherSearchIdxQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from publisher_search_idx\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for publisher_search_idx\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q utxoQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no utxoQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from utxo\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for utxo\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q voteQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no voteQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from vote\")\n\t}\n\n\treturn nil\n}", "func (q offerQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"stellarcore: no offerQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete all from offers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by deleteall for offers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q customerQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no customerQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from customers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for customers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q contentUnitDerivationQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"mdbmodels: no contentUnitDerivationQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from content_unit_derivations\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q *Queue) DeleteAll() error {\n\tq.Lock()\n\tdefer q.Unlock()\n\n\titer := q.db.NewIterator(util.BytesPrefix(q.opts.KeyPrefix), nil)\n\tdefer iter.Release()\n\tvar err error\n\n\tbatch := new(leveldb.Batch)\n\n\tfor iter.Next() {\n\t\tbatch.Delete(iter.Key())\n\t}\n\terr = q.db.Write(batch, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn q.initialize()\n}", "func (q storeQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no storeQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from stores\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for stores\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q rawVisitQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rawVisitQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from raw_visits\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for raw_visits\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q skinQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q taskQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no taskQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from tasks\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tasks\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q sourceQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no sourceQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete all from sources\")\n\t}\n\n\treturn nil\n}", "func (q authorQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"model: no transactionQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to delete all from transaction\")\n\t}\n\n\treturn nil\n}", "func (o TransactionSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `transaction` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to delete all from transaction slice\")\n\t}\n\n\treturn nil\n}", "func (q inventoryQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q subscriberQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no subscriberQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from subscribers\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for subscribers\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o HoldenAtSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from holdenAt slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q friendshipQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no friendshipQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from friendship\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for friendship\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (svc *DynamoAccessor) DeleteAll(ctx context.Context) error {\n\tvar deleteErr error\n\tinput := &dynamodb.ScanInput{\n\t\tTableName: aws.String(svc.dynamoTableName()),\n\t}\n\n\tscanHandler := func(output *dynamodb.ScanOutput, lastPage bool) bool {\n\t\twriteDeleteRequests := make([]*dynamodb.WriteRequest, len(output.Items))\n\t\tfor index, eachItem := range output.Items {\n\t\t\tkeyID := \"\"\n\t\t\tstringVal, stringValOk := eachItem[attrID]\n\t\t\tif stringValOk && stringVal.S != nil {\n\t\t\t\tkeyID = *(stringVal.S)\n\t\t\t}\n\t\t\twriteDeleteRequests[index] = &dynamodb.WriteRequest{\n\t\t\t\tDeleteRequest: &dynamodb.DeleteRequest{\n\t\t\t\t\tKey: dynamoKeyValueAttrMap(keyID),\n\t\t\t\t},\n\t\t\t}\n\t\t}\n\t\tinput := &dynamodb.BatchWriteItemInput{\n\t\t\tRequestItems: map[string][]*dynamodb.WriteRequest{\n\t\t\t\tsvc.dynamoTableName(): writeDeleteRequests,\n\t\t\t},\n\t\t}\n\t\t_, deleteErr = svc.dynamoSvc(ctx).BatchWriteItem(input)\n\t\treturn deleteErr == nil\n\t}\n\n\tscanErr := svc.dynamoSvc(ctx).ScanPagesWithContext(ctx, input, scanHandler)\n\tif scanErr != nil {\n\t\treturn scanErr\n\t}\n\treturn deleteErr\n}", "func (o TransactionSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `transactions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from transaction slice\")\n\t}\n\n\treturn nil\n}", "func (db RDB) DeleteAll(o DBObject) error {\n\treturn db.DeleteByID(o, 0)\n}", "func (o ContentUnitDerivationSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from contentUnitDerivation slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q notificationQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no notificationQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from notification\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for notification\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q stockKeepingUnitContentQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"db: no stockKeepingUnitContentQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to delete all from stock_keeping_unit_content\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by deleteall for stock_keeping_unit_content\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q smallblogQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no smallblogQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from smallblog\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for smallblog\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (p *HbaseClient) DeleteAllRow(tableName Text, row Text, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAllRow(tableName, row, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAllRow()\n}", "func (c *Command) DeleteAll() (int64, error) {\n\tclient := c.set.gom.GetClient()\n\n\tcollection := client.Database(c.set.gom.GetDatabase()).Collection(c.set.tableName)\n\n\tctx, cancelFunc := c.set.GetContext()\n\tdefer cancelFunc()\n\n\tres, err := collection.DeleteMany(ctx, c.set.filter)\n\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn res.DeletedCount, nil\n}", "func (q rentalQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q exchangeCurrencyQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no exchangeCurrencyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from exchange_currency\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for exchange_currency\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthorSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"authors\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authorPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete all from author slice\")\n\t}\n\n\treturn nil\n}", "func (q braceletPhotoQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no braceletPhotoQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from bracelet_photo\")\n\t}\n\n\treturn nil\n}", "func (q cvtermsynonymQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q authQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no transactionQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from transactions\")\n\t}\n\n\treturn nil\n}", "func (q stockQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no stockQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from stock\")\n\t}\n\n\treturn nil\n}", "func (q recipeAdditiveQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no recipeAdditiveQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from recipe_additive\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for recipe_additive\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q cvtermsynonymQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"chado: no cvtermsynonymQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete all from cvtermsynonym\")\n\t}\n\n\treturn nil\n}", "func (q userGoogleQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"model2: no userGoogleQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to delete all from user_google\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: failed to get rows affected by deleteall for user_google\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthMessageSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authMessagePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM `auth_message` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authMessagePrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(authMessagePrimaryKeyColumns), 1, len(authMessagePrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from authMessage slice\")\n\t}\n\n\treturn nil\n}", "func (q authUserQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q skinQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no skinQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from skin\")\n\t}\n\n\treturn nil\n}", "func (q channelQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no channelQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from channels\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for channels\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q premiumSlotQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no premiumSlotQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premium_slots\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_slots\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q inventoryQuery) DeleteAll() error {\n\tif q.Query == nil {\n\t\treturn errors.New(\"models: no inventoryQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\t_, err := q.Query.Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from inventory\")\n\t}\n\n\treturn nil\n}", "func (q cmfUserExperienceLogQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no cmfUserExperienceLogQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmf_user_experience_log\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_user_experience_log\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (d *PostgresClient) RemoveAll() error {\n\t_, err := d.DB.Exec(`TRUNCATE TABLE people_publications`)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = d.DB.Exec(`TRUNCATE TABLE people`)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = d.DB.Exec(`TRUNCATE TABLE publications`)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = d.DB.Exec(`TRUNCATE TABLE organizations`)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = d.DB.Exec(`TRUNCATE TABLE concepts`)\n\treturn err\n}", "func (q stockCvtermQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q filesStorageQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no filesStorageQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from files_storages\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for files_storages\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q ticketQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no ticketQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from tickets\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tickets\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q latencyQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no latencyQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from latencies\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for latencies\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q repositoryQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no repositoryQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from repositories\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for repositories\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q recordMeasureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cmfBalanceChargeAdminQuery) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no cmfBalanceChargeAdminQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.ExecContext(ctx, exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmf_balance_charge_admin\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_balance_charge_admin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (g *Group) DeleteAll() {\n\tg.Equivalents = list.New()\n\tg.Fingerprints = make(map[string]*list.Element)\n\tg.FirstExpr = make(map[Operand]*list.Element)\n\tg.SelfFingerprint = \"\"\n}", "func (r Repository) DeleteAll(ctx context.Context, uniqueID string, action authentity.Action) error {\n\totpKey := generateOtpKey(uniqueID, action)\n\t_, err := r.redis.Delete(ctx, otpKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}" ]
[ "0.7441922", "0.7309374", "0.72752523", "0.7194121", "0.71529055", "0.70903033", "0.70787144", "0.706517", "0.6974174", "0.69680166", "0.69609714", "0.69567597", "0.693858", "0.6937204", "0.693068", "0.6928459", "0.69269633", "0.69241506", "0.69185144", "0.6917855", "0.6915302", "0.69074285", "0.69016564", "0.68851924", "0.68764955", "0.6874583", "0.6860475", "0.68559223", "0.6846026", "0.6843207", "0.6832002", "0.68315697", "0.6830261", "0.68288016", "0.6825338", "0.6824772", "0.6823308", "0.6823118", "0.68191123", "0.68095297", "0.67992765", "0.67975205", "0.6792225", "0.6788427", "0.6780914", "0.6780286", "0.67767364", "0.6776105", "0.6776105", "0.6775164", "0.67696476", "0.67680466", "0.6755945", "0.67435974", "0.67369837", "0.67368364", "0.6735771", "0.6723077", "0.67150635", "0.6709151", "0.6702711", "0.6691274", "0.66900074", "0.66866744", "0.6677726", "0.6677412", "0.6676335", "0.6676138", "0.6675099", "0.6674568", "0.6674397", "0.666813", "0.6666615", "0.6665268", "0.66636586", "0.66584694", "0.6656636", "0.6653682", "0.6650683", "0.6649335", "0.6645485", "0.6640976", "0.6640328", "0.6638464", "0.663029", "0.66296196", "0.6620472", "0.6616741", "0.6616216", "0.6614543", "0.6612483", "0.6612344", "0.661052", "0.660545", "0.6601784", "0.6596616", "0.65954363", "0.65875685", "0.65837777", "0.65785563", "0.6570687" ]
0.0
-1
DeleteAllGP deletes all rows in the slice, and panics on error.
func (o AssetRevisionSlice) DeleteAllGP() { if err := o.DeleteAllG(); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o SourceSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) DeleteAllGP(ctx context.Context) int64 {\n\trowsAff, err := o.DeleteAll(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o AuthMessageSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o SkinSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthUserUserPermissionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ClaimInListSlice) DeleteAllGP() {\n\terr := o.DeleteAll(boil.GetDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AddressSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockCvtermSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RentalSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o DMessageEmbedSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OfferSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (q tenantQuery) DeleteAllGP(ctx context.Context) int64 {\n\trowsAff, err := q.DeleteAll(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o PhenotypepropSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AutomodRuleDatumSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o FailureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RowerGroupSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RecordMeasureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalRowerSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o TransactionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ItemSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o TransactionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (q sourceQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RentalSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PremiumSlotSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o InventorySlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o BraceletPhotoSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (q stockCvtermQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthUserUserPermissionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o InstrumentClassSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o APIKeySlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthorSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o OrganismSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (q failureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ItemSideSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o NotificationSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthMessageSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ClaimInListSlice) DeleteAllG() error {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (q jetQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o UserGoogleSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(userGoogleBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to delete all from userGoogle slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: failed to get rows affected by deleteall for user_google\")\n\t}\n\n\tif len(userGoogleAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o InstrumentClassSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ShelfSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StoreSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o AuthTokenSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PremiumCodeSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}" ]
[ "0.77667785", "0.7753922", "0.77429754", "0.77429754", "0.7673432", "0.76400656", "0.76025355", "0.75305206", "0.752892", "0.7524689", "0.7495944", "0.748702", "0.74533933", "0.74273586", "0.7419781", "0.7349038", "0.73384917", "0.7335055", "0.7330366", "0.73022217", "0.73010504", "0.72772443", "0.72609466", "0.7255925", "0.7253976", "0.72430813", "0.723946", "0.7234152", "0.7233469", "0.72041595", "0.71904725", "0.7165236", "0.71584314", "0.7113571", "0.7104267", "0.7090639", "0.7086532", "0.7073168", "0.7071755", "0.7068478", "0.7067407", "0.70524544", "0.70461744", "0.7035344", "0.7035344", "0.7031439", "0.7015649", "0.6998969", "0.69936043", "0.6993228", "0.6984596", "0.6979018", "0.6975407", "0.6960289", "0.6954655", "0.6950123", "0.69467247", "0.69441706", "0.69337493", "0.6922687", "0.6912966", "0.6907395", "0.689695", "0.6896302", "0.68947405", "0.68872106", "0.688529", "0.68703073", "0.68589234", "0.68533045", "0.6850551", "0.6849044", "0.6840225", "0.68194145", "0.6817429", "0.68169075", "0.68112135", "0.68067783", "0.680542", "0.6799648", "0.6790506", "0.67715096", "0.67705405", "0.67610973", "0.67528856", "0.674702", "0.67397803", "0.6739366", "0.67351717", "0.67339516", "0.6718482", "0.67166185", "0.67162555", "0.6708969", "0.67061156", "0.6704461", "0.66997373", "0.6692812", "0.6688721", "0.6688665" ]
0.72000575
30
DeleteAllG deletes all rows in the slice.
func (o AssetRevisionSlice) DeleteAllG() error { if o == nil { return errors.New("public: no AssetRevision slice provided for delete all") } return o.DeleteAll(boil.GetDB()) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o RentalSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RecordMeasureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o JetSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o SourceSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ItemSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o RowerGroupSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RentalRowerSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o OfferSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockCvtermSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o CvtermsynonymSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AutomodRuleDatumSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o ItemSideSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o TransactionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthorSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StoreSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o PremiumSlotSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o TransactionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AddressSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o VoteSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o OrganismSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o InventorySlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ClaimInListSlice) DeleteAllG() error {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ShelfSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Shelf slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o CommentSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o NotificationSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RecordMeasureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StorestateSlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o JetSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o RentalSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o SkinSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Skin slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o ExchangeCurrencySlice) DeleteAllG() (int64, error) {\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o TenantSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o PremiumCodeSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TicketSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o ShelfSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PhenotypepropSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthorSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BlackCard slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o APIKeySlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o InventorySlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o SkinSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o UsernameListingSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o RSSAnnouncementSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o PictureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o FailureSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthMessageSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o FeatureRelationshipSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureRelationship slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o AuthUserUserPermissionSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o StockCvtermSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no DMessageEmbed slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o BookCategorySlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o OrganismSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) DeleteAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken slice provided for delete all\")\n\t}\n\treturn o.DeleteAll(boil.GetDB())\n}", "func (o VoteSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CvtermsynonymSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) DeleteAllGP(ctx context.Context) int64 {\n\trowsAff, err := o.DeleteAll(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o AddressSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BookCategoryAssignSlice) DeleteAllG(ctx context.Context) (int64, error) {\n\treturn o.DeleteAll(ctx, boil.GetContextDB())\n}", "func (o AssetRevisionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ClaimInListSlice) DeleteAllGP() {\n\terr := o.DeleteAll(boil.GetDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) DeleteAllGP(ctx context.Context) int64 {\n\trowsAff, err := q.DeleteAll(ctx, boil.GetContextDB())\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o PhenotypepropSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o UserGoogleSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(userGoogleBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to delete all from userGoogle slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: failed to get rows affected by deleteall for user_google\")\n\t}\n\n\tif len(userGoogleAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rowerGroupQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rowerGroupQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ContentUnitDerivationSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from contentUnitDerivation slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.7766479", "0.77193296", "0.7709203", "0.7701425", "0.76778716", "0.7677664", "0.76609135", "0.76570415", "0.7644065", "0.7615155", "0.7604686", "0.7529517", "0.7525105", "0.7521399", "0.7495394", "0.74887747", "0.74811274", "0.7469117", "0.74638724", "0.7450275", "0.74208987", "0.73628557", "0.7361532", "0.73575455", "0.7347331", "0.7340262", "0.7336428", "0.7335308", "0.7324811", "0.7288266", "0.7277566", "0.7256648", "0.72554433", "0.72551394", "0.7252976", "0.72383964", "0.7236412", "0.72357523", "0.72251016", "0.72251016", "0.7218032", "0.7204201", "0.7169754", "0.7167054", "0.71658576", "0.71572834", "0.71428853", "0.7138021", "0.71289647", "0.7123595", "0.70997477", "0.70991653", "0.70881253", "0.70735955", "0.70730495", "0.70702344", "0.70548195", "0.7042559", "0.7042032", "0.70184237", "0.6993478", "0.6980425", "0.6975322", "0.6969404", "0.69562274", "0.6940147", "0.6931149", "0.6884367", "0.6859013", "0.68531686", "0.6820479", "0.6762563", "0.67305696", "0.67183137", "0.6711814", "0.6710565", "0.6706255", "0.6701978", "0.6700887", "0.66930336", "0.66409874", "0.66322386", "0.6583912", "0.65131325", "0.64845884", "0.64759916", "0.6474566", "0.64739454", "0.6453167", "0.6441977", "0.6437278", "0.6423437", "0.6401605", "0.6397725", "0.63689286", "0.636591", "0.63525593", "0.63525593", "0.63346714", "0.63249624" ]
0.7236869
36
DeleteAllP deletes all rows in the slice, using an executor, and panics on error.
func (o AssetRevisionSlice) DeleteAllP(exec boil.Executor) { if err := o.DeleteAll(exec); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o StockSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthorSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TenantSlice) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := o.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o CvtermsynonymSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ShelfSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RecordMeasureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlackCardSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthTokenSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthMessageSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InstrumentClassSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CommentSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SkinSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OrganismSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PictureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthUserUserPermissionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BraceletPhotoSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q tenantQuery) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := q.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o ClaimInListSlice) DeleteAllP(exec boil.Executor) {\n\terr := o.DeleteAll(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureCvtermDbxrefSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q sourceQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q failureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q jetQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o DMessageEmbedSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q transactionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q stockCvtermQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q voteQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q claimInListQuery) DeleteAllP(exec boil.Executor) {\n\terr := q.DeleteAll(exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TaskSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(taskBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from task slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tasks\")\n\t}\n\n\tif len(taskAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (q stockQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q inventoryQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authorQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o OfferSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"stellarcore: no Offer slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(offerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete all from offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by deleteall for offers\")\n\t}\n\n\tif len(offerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (q apiKeyQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q instrumentClassQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q assetRevisionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q shelfQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authTokenQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q skinQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CurrentChartDataMinutelySlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no CurrentChartDataMinutely slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(currentChartDataMinutelyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currentChartDataMinutely slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for current_chart_data_minutely\")\n\t}\n\n\tif len(currentChartDataMinutelyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o EmailSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"mysql: no Email slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(emailBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to delete all from email slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: failed to get rows affected by deleteall for email\")\n\t}\n\n\tif len(emailAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (q blackCardQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FeatureRelationshipSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q organismQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o VoteSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(voteBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from vote slice\")\n\t}\n\n\tif len(voteAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o VoteSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q pictureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q recordMeasureQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserUserPermissionQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q rentalQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q cvtermsynonymQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q commentQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q addressQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BlockSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from block slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PhenotypepropSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q braceletPhotoQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (q authUserQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RawVisitSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(rawVisitBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rawVisit slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for raw_visits\")\n\t}\n\n\tif len(rawVisitAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o WeatherSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"db: no Weather slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(weatherBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"prh\\\".\\\"weather\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, weatherPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to delete all from weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by deleteall for weather\")\n\t}\n\n\tif len(weatherAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RecipeLipidSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(recipeLipidBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from recipeLipid slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for recipe_lipid\")\n\t}\n\n\tif len(recipeLipidAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PremiumSlotSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_slots\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VSPSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"vsp\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, vspPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vsp slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vsp\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFSlideItemSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfSlideItemBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfSlideItemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_slide_item` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfSlideItemPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfSlideItem slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_slide_item\")\n\t}\n\n\tif len(cmfSlideItemAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o JetSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(jetBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from jet slice\")\n\t}\n\n\tif len(jetAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (q dMessageEmbedQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o RentalSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(rentalBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to delete all from rental slice\")\n\t}\n\n\tif len(rentalAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o RepositorySlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(repositoryBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from repository slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for repositories\")\n\t}\n\n\tif len(repositoryAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (q featureCvtermDbxrefQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BTCTXOutputSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), btcTXOutputPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `btc_tx_output` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, btcTXOutputPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from btcTXOutput slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (q authMessageQuery) DeleteAllP() {\n\tif err := q.DeleteAll(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SegmentSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(segmentBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to delete all from segment slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: failed to get rows affected by deleteall for segment\")\n\t}\n\n\tif len(segmentAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PictureSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(pictureBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from picture slice\")\n\t}\n\n\tif len(pictureAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o PublisherSearchIdxSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(publisherSearchIdxBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), publisherSearchIdxPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"publisher_search_idx\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, publisherSearchIdxPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from publisherSearchIdx slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for publisher_search_idx\")\n\t}\n\n\tif len(publisherSearchIdxAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FailureSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o PremiumCodeSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premiumCode slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_codes\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RowerGroupSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no RowerGroup slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(rowerGroupBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rowerGroup slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\tif len(rowerGroupAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ChannelSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(channelBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from channel slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for channels\")\n\t}\n\n\tif len(channelAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PostSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"orm: no Post slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(postBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to delete all from post slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: failed to get rows affected by deleteall for posts\")\n\t}\n\n\tif len(postAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PeerSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(peerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to delete all from peer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: failed to get rows affected by deleteall for peers\")\n\t}\n\n\tif len(peerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CurrencySlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Currency slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(currencyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for currency\")\n\t}\n\n\tif len(currencyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o UtxoSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(utxoBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from utxo slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for utxo\")\n\t}\n\n\tif len(utxoAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ItemSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from item slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for items\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllGP() {\n\tif err := o.DeleteAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ContentUnitDerivationSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from contentUnitDerivation slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o MempoolBinSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), mempoolBinPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"mempool_bin\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, mempoolBinPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from mempoolBin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for mempool_bin\")\n\t}\n\n\treturn rowsAff, nil\n}" ]
[ "0.74493456", "0.7428404", "0.7428404", "0.73839015", "0.7359511", "0.7260627", "0.7240982", "0.7173056", "0.71394", "0.7129205", "0.7090685", "0.707032", "0.70192295", "0.701109", "0.70030135", "0.70024973", "0.6955495", "0.6941525", "0.6939464", "0.6935739", "0.69269454", "0.69048643", "0.6899374", "0.6895781", "0.6890209", "0.6820162", "0.67938954", "0.67819715", "0.6720327", "0.67036825", "0.66882646", "0.6681637", "0.667913", "0.6656338", "0.6636185", "0.66141886", "0.652908", "0.652908", "0.65283036", "0.651045", "0.64731294", "0.646862", "0.64591306", "0.6418168", "0.6408204", "0.63806784", "0.6376354", "0.6344543", "0.63110167", "0.6303098", "0.62540704", "0.62221617", "0.6217357", "0.62160623", "0.6214865", "0.61946934", "0.619136", "0.6189598", "0.6176142", "0.61663127", "0.6151326", "0.614933", "0.6134408", "0.6129034", "0.61137813", "0.6110271", "0.61078453", "0.61077887", "0.6091087", "0.60547924", "0.60524243", "0.60454476", "0.6042072", "0.60098964", "0.5984953", "0.5960331", "0.59580857", "0.5957285", "0.59571457", "0.595339", "0.5948413", "0.5944103", "0.59393585", "0.59392786", "0.5929778", "0.5925611", "0.5917832", "0.591585", "0.59107274", "0.59085363", "0.59044266", "0.59012914", "0.59002614", "0.58835274", "0.5877411", "0.5873015", "0.5867326", "0.5867326", "0.58673143", "0.5864927" ]
0.6721531
28
DeleteAll deletes all rows in the slice, using an executor.
func (o AssetRevisionSlice) DeleteAll(exec boil.Executor) error { if o == nil { return errors.New("public: no AssetRevision slice provided for delete all") } if len(o) == 0 { return nil } if len(assetRevisionBeforeDeleteHooks) != 0 { for _, obj := range o { if err := obj.doBeforeDeleteHooks(exec); err != nil { return err } } } var args []interface{} for _, obj := range o { pkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), assetRevisionPrimaryKeyMapping) args = append(args, pkeyArgs...) } query := fmt.Sprintf( "DELETE FROM \"asset_revision\" WHERE (%s) IN (%s)", strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, assetRevisionPrimaryKeyColumns), ","), strmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(assetRevisionPrimaryKeyColumns), 1, len(assetRevisionPrimaryKeyColumns)), ) if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, query) fmt.Fprintln(boil.DebugWriter, args) } _, err := exec.Exec(query, args...) if err != nil { return errors.Wrap(err, "public: unable to delete all from assetRevision slice") } if len(assetRevisionAfterDeleteHooks) != 0 { for _, obj := range o { if err := obj.doAfterDeleteHooks(exec); err != nil { return err } } } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o TaskSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(taskBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from task slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tasks\")\n\t}\n\n\tif len(taskAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o BlockSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from block slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for block\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OfferSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"stellarcore: no Offer slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(offerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete all from offer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by deleteall for offers\")\n\t}\n\n\tif len(offerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o EmailSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"mysql: no Email slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(emailBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: unable to delete all from email slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mysql: failed to get rows affected by deleteall for email\")\n\t}\n\n\tif len(emailAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CurrentChartDataMinutelySlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no CurrentChartDataMinutely slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(currentChartDataMinutelyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currentChartDataMinutely slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for current_chart_data_minutely\")\n\t}\n\n\tif len(currentChartDataMinutelyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RawVisitSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(rawVisitBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rawVisit slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for raw_visits\")\n\t}\n\n\tif len(rawVisitAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o WeatherSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"db: no Weather slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(weatherBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"prh\\\".\\\"weather\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, weatherPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to delete all from weather slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by deleteall for weather\")\n\t}\n\n\tif len(weatherAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(voteBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from vote slice\")\n\t}\n\n\tif len(voteAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o ChannelSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(channelBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from channel slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for channels\")\n\t}\n\n\tif len(channelAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o SegmentSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(segmentBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: unable to delete all from segment slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"boiler: failed to get rows affected by deleteall for segment\")\n\t}\n\n\tif len(segmentAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from vote slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for vote\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o HoldenAtSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from holdenAt slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for HoldenAt\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o UtxoSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(utxoBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from utxo slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for utxo\")\n\t}\n\n\tif len(utxoAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TransactionSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o JetSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(jetBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from jet slice\")\n\t}\n\n\tif len(jetAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o ContentUnitDerivationSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from contentUnitDerivation slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for content_unit_derivations\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TransactionSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `transaction` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to delete all from transaction slice\")\n\t}\n\n\treturn nil\n}", "func (o StockCvtermSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(stockCvtermBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockCvtermPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM \\\"stock_cvterm\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, stockCvtermPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(stockCvtermPrimaryKeyColumns), 1, len(stockCvtermPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete all from stockCvterm slice\")\n\t}\n\n\tif len(stockCvtermAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o TransactionSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `transactions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from transaction slice\")\n\t}\n\n\treturn nil\n}", "func (o RowerGroupSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no RowerGroup slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(rowerGroupBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rowerGroup slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\tif len(rowerGroupAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OauthClientSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(oauthClientBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), oauthClientPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `oauth_clients` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, oauthClientPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from oauthClient slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for oauth_clients\")\n\t}\n\n\tif len(oauthClientAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RentalSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(rentalBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to delete all from rental slice\")\n\t}\n\n\tif len(rentalAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o AuthorSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"authors\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authorPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete all from author slice\")\n\t}\n\n\treturn nil\n}", "func (o AddressSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(addressBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), addressPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `address` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, addressPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to delete all from address slice\")\n\t}\n\n\tif len(addressAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o StockSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BTCTXOutputSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), btcTXOutputPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `btc_tx_output` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, btcTXOutputPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from btcTXOutput slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for btc_tx_output\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StockCvtermSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o SourceSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to delete all from source slice\")\n\t}\n\n\treturn nil\n}", "func (o CurrencySlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Currency slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(currencyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from currency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for currency\")\n\t}\n\n\tif len(currencyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OriginSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Origin slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(originBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), originPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"origins\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, originPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from origin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for origins\")\n\t}\n\n\tif len(originAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RepositorySlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(repositoryBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from repository slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for repositories\")\n\t}\n\n\tif len(repositoryAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o NodeSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), nodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"node\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, nodePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from node slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for node\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFTurntableSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfTurntableBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_turntable` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfTurntable slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_turntable\")\n\t}\n\n\tif len(cmfTurntableAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StockKeepingUnitContentSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(stockKeepingUnitContentBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockKeepingUnitContentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"stock_keeping_unit_content\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, stockKeepingUnitContentPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: unable to delete all from stockKeepingUnitContent slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"db: failed to get rows affected by deleteall for stock_keeping_unit_content\")\n\t}\n\n\tif len(stockKeepingUnitContentAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o MempoolBinSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), mempoolBinPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"mempool_bin\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, mempoolBinPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from mempoolBin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for mempool_bin\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o SubscriberSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Subscriber slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(subscriberBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), subscriberPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `subscribers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, subscriberPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from subscriber slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for subscribers\")\n\t}\n\n\tif len(subscriberAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o JobSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Job slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(jobBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"jobs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, jobPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from job slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for jobs\")\n\t}\n\n\tif len(jobAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ItemSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from item slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for items\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o BoardSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(boardBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to delete all from board slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: failed to get rows affected by deleteall for boards\")\n\t}\n\n\tif len(boardAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o StockSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `stock` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from stock slice\")\n\t}\n\n\treturn nil\n}", "func (o ActivityLogSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(activityLogBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), activityLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"activity_logs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, activityLogPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to delete all from activityLog slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by deleteall for activity_logs\")\n\t}\n\n\tif len(activityLogAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CommentSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(commentBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), commentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `comment` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, commentPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from comment slice\")\n\t}\n\n\tif len(commentAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o PremiumSlotSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premiumSlot slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_slots\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (p *HbaseClient) DeleteAll(tableName Text, row Text, column Text, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAll(tableName, row, column, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAll()\n}", "func (o AuthorSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TicketSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"tickets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, ticketPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from ticket slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for tickets\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PublisherSearchIdxSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(publisherSearchIdxBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), publisherSearchIdxPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"publisher_search_idx\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, publisherSearchIdxPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from publisherSearchIdx slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for publisher_search_idx\")\n\t}\n\n\tif len(publisherSearchIdxAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o KvstoreSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(kvstoreBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"kvstore\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, kvstorePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from kvstore slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for kvstore\")\n\t}\n\n\tif len(kvstoreAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o TenantSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(tenantBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `tenants` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: unable to delete all from tenant slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dbmodel: failed to get rows affected by deleteall for tenants\")\n\t}\n\n\tif len(tenantAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CvtermsynonymSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(cvtermsynonymBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cvtermsynonymPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM \\\"cvtermsynonym\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, cvtermsynonymPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(cvtermsynonymPrimaryKeyColumns), 1, len(cvtermsynonymPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete all from cvtermsynonym slice\")\n\t}\n\n\tif len(cvtermsynonymAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o InventorySlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o CustomerSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Customer slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(customerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"customers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, customerPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from customer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for customers\")\n\t}\n\n\tif len(customerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o SourceSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: unable to delete all from source slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"mdbmodels: failed to get rows affected by deleteall for sources\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o RecipeLipidSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(recipeLipidBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from recipeLipid slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for recipe_lipid\")\n\t}\n\n\tif len(recipeLipidAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o FilesStorageSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no FilesStorage slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(filesStorageBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), filesStoragePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `files_storages` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, filesStoragePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from filesStorage slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for files_storages\")\n\t}\n\n\tif len(filesStorageAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o VoteSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o InventorySlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(inventoryBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), inventoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `inventory` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, inventoryPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from inventory slice\")\n\t}\n\n\tif len(inventoryAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o CMFUserExperienceLogSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfUserExperienceLogBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserExperienceLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_user_experience_log` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserExperienceLogPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfUserExperienceLog slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_user_experience_log\")\n\t}\n\n\tif len(cmfUserExperienceLogAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ShelfSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o FailureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o TrainingCostSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(trainingCostBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"training_costs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, trainingCostPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from trainingCost slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for training_costs\")\n\t}\n\n\tif len(trainingCostAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PictureSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(pictureBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from picture slice\")\n\t}\n\n\tif len(pictureAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o TenantSlice) DeleteAllP(ctx context.Context, exec boil.ContextExecutor) int64 {\n\trowsAff, err := o.DeleteAll(ctx, exec)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn rowsAff\n}", "func (o IllnessSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(illnessBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"illness\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, illnessPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from illness slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for illness\")\n\t}\n\n\tif len(illnessAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o DescriptionSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Description slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(descriptionBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), descriptionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `descriptions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, descriptionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from description slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for descriptions\")\n\t}\n\n\tif len(descriptionAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (p *HbaseClient) DeleteAllTs(tableName Text, row Text, column Text, timestamp int64, attributes map[string]Text) (err error) {\n\tif err = p.sendDeleteAllTs(tableName, row, column, timestamp, attributes); err != nil {\n\t\treturn\n\t}\n\treturn p.recvDeleteAllTs()\n}", "func (o ChatSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"chat\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, chatPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: unable to delete all from chat slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"dal: failed to get rows affected by deleteall for chat\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o WithdrawalCryptoSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(withdrawalCryptoBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"withdrawal_crypto\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: unable to delete all from withdrawalCrypto slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"sqlite3: failed to get rows affected by deleteall for withdrawal_crypto\")\n\t}\n\n\tif len(withdrawalCryptoAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o ShelfSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Shelf slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(shelfBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), shelfPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM `shelf` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, shelfPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(shelfPrimaryKeyColumns), 1, len(shelfPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from shelf slice\")\n\t}\n\n\tif len(shelfAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o RentalRowerSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no RentalRower slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(rentalRowerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"rental_rowers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rentalRowerPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rentalRower slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rental_rowers\")\n\t}\n\n\tif len(rentalRowerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o InstrumentClassSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(instrumentClassBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), instrumentClassPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"instruments\\\".\\\"instrument_class\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, instrumentClassPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from instrumentClass slice\")\n\t}\n\n\tif len(instrumentClassAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o CMFSlideItemSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfSlideItemBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfSlideItemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_slide_item` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfSlideItemPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfSlideItem slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_slide_item\")\n\t}\n\n\tif len(cmfSlideItemAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o OrganismSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(organismBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), organismPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM \\\"organism\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, organismPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(organismPrimaryKeyColumns), 1, len(organismPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to delete all from organism slice\")\n\t}\n\n\tif len(organismAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o RecordMeasureSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o ExchangeCurrencySlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no ExchangeCurrency slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(exchangeCurrencyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), exchangeCurrencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"exchange_currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, exchangeCurrencyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from exchangeCurrency slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for exchange_currency\")\n\t}\n\n\tif len(exchangeCurrencyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CvtermsynonymSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AddressSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o APIKeySlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(apiKeyBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"api_keys\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, apiKeyPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from apiKey slice\")\n\t}\n\n\tif len(apiKeyAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o NotificationSlice) DeleteAll(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Notification slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(notificationBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"notification\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, notificationPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from notification slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for notification\")\n\t}\n\n\tif len(notificationAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFBalanceChargeAdminSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfBalanceChargeAdminBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_balance_charge_admin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfBalanceChargeAdmin slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_balance_charge_admin\")\n\t}\n\n\tif len(cmfBalanceChargeAdminAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthUserUserPermissionSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserUserPermissionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM `auth_user_user_permissions` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserUserPermissionPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(authUserUserPermissionPrimaryKeyColumns), 1, len(authUserUserPermissionPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from authUserUserPermission slice\")\n\t}\n\n\treturn nil\n}", "func (o AuthUserSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o BuildingSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"record: no Building slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(buildingBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"buildings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, buildingPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: unable to delete all from building slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"record: failed to get rows affected by deleteall for buildings\")\n\t}\n\n\tif len(buildingAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o InstrumentClassSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o AuthSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(authBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"auths\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from auth slice\")\n\t}\n\n\tif len(authAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o PremiumCodeSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from premiumCode slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for premium_codes\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o BoardsSectionsPositionSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(boardsSectionsPositionBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardsSectionsPositionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `boards_sections_positions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardsSectionsPositionPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: unable to delete all from boardsSectionsPosition slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"rdb: failed to get rows affected by deleteall for boards_sections_positions\")\n\t}\n\n\tif len(boardsSectionsPositionAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o UserGoogleSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(userGoogleBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: unable to delete all from userGoogle slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model2: failed to get rows affected by deleteall for user_google\")\n\t}\n\n\tif len(userGoogleAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (q rowerGroupQuery) DeleteAll(exec boil.Executor) (int64, error) {\n\tif q.Query == nil {\n\t\treturn 0, errors.New(\"models: no rowerGroupQuery provided for delete all\")\n\t}\n\n\tqueries.SetDelete(q.Query)\n\n\tresult, err := q.Query.Exec(exec)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from rower_group\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for rower_group\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthUserSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"DELETE FROM `auth_user` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(o)*len(authUserPrimaryKeyColumns), 1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from authUser slice\")\n\t}\n\n\treturn nil\n}", "func (o PostSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"orm: no Post slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(postBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: unable to delete all from post slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"orm: failed to get rows affected by deleteall for posts\")\n\t}\n\n\tif len(postAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthItemGroupSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(authItemGroupBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authItemGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `auth_item_groups` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, authItemGroupPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from authItemGroup slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for auth_item_groups\")\n\t}\n\n\tif len(authItemGroupAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o AuthTokenSlice) DeleteAll(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken slice provided for delete all\")\n\t}\n\n\tif len(o) == 0 {\n\t\treturn nil\n\t}\n\n\tif len(authTokenBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authTokenPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"auth_tokens\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authTokenPrimaryKeyColumns, len(o))\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args)\n\t}\n\n\t_, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to delete all from authToken slice\")\n\t}\n\n\tif len(authTokenAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(exec); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o ItemSideSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemSidePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM \\\"item_sides\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemSidePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from itemSide slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for item_sides\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (o PeerSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(peerBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: unable to delete all from peer slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"model: failed to get rows affected by deleteall for peers\")\n\t}\n\n\tif len(peerAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CMFFamilyUserPoliciesTakeSlice) DeleteAll(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif len(o) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tif len(cmfFamilyUserPoliciesTakeBeforeDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\tvar args []interface{}\n\tfor _, obj := range o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"DELETE FROM `cmf_family_user_policies_take` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(o))\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete all from cmfFamilyUserPoliciesTake slice\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by deleteall for cmf_family_user_policies_take\")\n\t}\n\n\tif len(cmfFamilyUserPoliciesTakeAfterDeleteHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn rowsAff, nil\n}", "func (o CommentSlice) DeleteAllP(exec boil.Executor) {\n\tif err := o.DeleteAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.6663268", "0.64429116", "0.64330864", "0.64252645", "0.63758576", "0.63507766", "0.6332601", "0.62535197", "0.62500495", "0.6248866", "0.62392336", "0.62334", "0.62305635", "0.6216409", "0.6216409", "0.6215515", "0.6199186", "0.61769515", "0.6163286", "0.615327", "0.6139643", "0.6136154", "0.61287916", "0.61244655", "0.6122491", "0.61199296", "0.6114271", "0.61054325", "0.61042833", "0.6092421", "0.608694", "0.60651344", "0.6064213", "0.6063116", "0.606129", "0.6059165", "0.6051474", "0.60488886", "0.60346746", "0.60329264", "0.6025356", "0.602238", "0.6012955", "0.60041887", "0.5993896", "0.59770167", "0.59708816", "0.5970798", "0.5968736", "0.59629554", "0.5958269", "0.59550625", "0.5947824", "0.5941009", "0.5934488", "0.5932542", "0.59319824", "0.59265", "0.5926492", "0.5923122", "0.5901797", "0.5896551", "0.5893802", "0.5885367", "0.58845395", "0.58798665", "0.5872926", "0.5872294", "0.5868009", "0.58544534", "0.5847685", "0.5844515", "0.584306", "0.58374256", "0.58277917", "0.58228207", "0.5820315", "0.5812187", "0.5807397", "0.5806311", "0.5805872", "0.58042264", "0.5802178", "0.57944924", "0.57930666", "0.57885104", "0.5787585", "0.5777219", "0.5773673", "0.5772302", "0.57683015", "0.5765593", "0.5763279", "0.5762107", "0.575982", "0.5759629", "0.5758498", "0.57582587", "0.5756114", "0.57517356", "0.5751605" ]
0.0
-1
ReloadGP refetches the object from the database and panics on error.
func (o *AssetRevision) ReloadGP() { if err := o.ReloadG(); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Picture) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Failure) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Jet) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Failure) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Rental) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Vote) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Transaction) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AuthUser) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BlackCard provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureRelationship) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureRelationship provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *ClaimInList) ReloadGP() {\n\tif err := o.Reload(boil.GetDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Item) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Item provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Stock) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *APIKey) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Stock) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Phenotypeprop) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Offer) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no Offer provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Picture) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Ticket) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Ticket provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Shelf) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RentalRower) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RentalRower provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Notification) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Notification provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Jet) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Shelf) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Shelf provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *RowerGroup) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RowerGroup provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Inventory) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Author) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *DMessageEmbed) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no DMessageEmbed provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Comment) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Auth) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *StockCvterm) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) ReloadGP(ctx context.Context) {\n\tif err := o.Reload(ctx, boil.GetContextDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *PremiumCode) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no PremiumCode provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Cvtermsynonym) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Address) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Cvtermsynonym) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AuthUserUserPermission) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermission) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AutomodRuleDatum) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AutomodRuleDatum provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *BraceletPhoto) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *UserGoogle) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUserGoogle(ctx, exec, o.GoogleID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *InstrumentClass) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (instance *Host) Reload(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\n\treturn instance.unsafeReload(ctx)\n}", "func (o *Inventory) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AuthToken) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Store) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Store provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *APIKey) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *ExchangeCurrency) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no ExchangeCurrency provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Transaction) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *PremiumSlot) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no PremiumSlot provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Organism) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *ItemSide) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no ItemSide provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *PaymentObject) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPaymentObject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Storestate) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no Storestate provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureCvtermDbxref) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureCvtermDbxref(exec, o.FeatureCvtermDbxrefID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureRelationship) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureRelationship(exec, o.FeatureRelationshipID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *InstrumentClass) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasure) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Friendship) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindFriendship(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Organism) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) Reload(exec boil.Executor) error {\n\tret, err := FindPicture(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RecordMeasure) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *CMFBalanceChargeAdmin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFBalanceChargeAdmin(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Utxo) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUtxo(ctx, exec, o.Rowid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Phenotypeprop) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInList) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no ClaimInList provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AssetRevision) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *UsernameListing) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no UsernameListing provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Tenant) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"dbmodel: no Tenant provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Post) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Peer) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPeer(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ForeignLegalResource) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindForeignLegalResource(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Repository) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRepository(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthToken) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Skin) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Skin provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Skin) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *VSP) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVSP(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Building) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBuilding(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RSSAnnouncement) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RSSAnnouncement provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *FeatureRelationship) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.73883665", "0.718418", "0.7140237", "0.71308863", "0.70993954", "0.70905745", "0.70836407", "0.70558965", "0.701964", "0.7013682", "0.69744664", "0.69744664", "0.6963713", "0.6957918", "0.69560236", "0.6950663", "0.6946262", "0.694259", "0.69402635", "0.69253695", "0.69237393", "0.69236934", "0.69057703", "0.690192", "0.68873674", "0.6861243", "0.6840077", "0.6838047", "0.6834538", "0.6830256", "0.6817831", "0.67910045", "0.6787312", "0.67758304", "0.67711383", "0.6765378", "0.676387", "0.67624784", "0.6761941", "0.6752809", "0.6750247", "0.6750213", "0.6744761", "0.67416704", "0.67182", "0.6717114", "0.6710386", "0.6706953", "0.66985124", "0.6673781", "0.6673136", "0.6672832", "0.66705036", "0.6667419", "0.6661676", "0.66610694", "0.66600037", "0.6657274", "0.6644831", "0.66321445", "0.66307455", "0.6617754", "0.6603984", "0.66019195", "0.6584102", "0.6582693", "0.65804654", "0.65703076", "0.65582556", "0.65558535", "0.6555386", "0.6548918", "0.65481", "0.6506113", "0.6483893", "0.64815116", "0.6480527", "0.6476959", "0.6449841", "0.64418584", "0.6438085", "0.6427292", "0.6425445", "0.6393357", "0.6378419", "0.6356925", "0.6354305", "0.63507706", "0.63343966", "0.63286537", "0.63162696", "0.6282448", "0.627583", "0.62632954", "0.625632", "0.6248441", "0.6245829", "0.6224225", "0.6200251", "0.61838204" ]
0.66558313
58
ReloadP refetches the object from the database with an executor. Panics on error.
func (o *AssetRevision) ReloadP(exec boil.Executor) { if err := o.Reload(exec); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Transaction) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Source) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Rental) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxref) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUser) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockCvterm) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Shelf) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Auth) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Picture) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlackCard) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Author) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureRelationship) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Utxo) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUtxo(ctx, exec, o.Rowid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Cvtermsynonym) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *InstrumentClass) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Tenant) ReloadP(ctx context.Context, exec boil.ContextExecutor) {\n\tif err := o.Reload(ctx, exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Task) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTask(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Inventory) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BraceletPhoto) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (instance *Host) Reload(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\n\treturn instance.unsafeReload(ctx)\n}", "func (o *Post) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PaymentObject) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPaymentObject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Address) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Organism) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Repository) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRepository(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *APIKey) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Comment) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthMessage) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *DMessageEmbed) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Jet) Reload(exec boil.Executor) error {\n\tret, err := FindJet(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Peer) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPeer(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *VSP) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVSP(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthToken) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFBalanceChargeAdmin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFBalanceChargeAdmin(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PremiumCode) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPremiumCode(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUserUserPermission) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Vote) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVote(ctx, exec, o.Hash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) Reload(exec boil.Executor) error {\n\tret, err := FindVote(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Board) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBoard(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Picture) Reload(exec boil.Executor) error {\n\tret, err := FindPicture(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ClaimInList) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Failure) Reload(exec boil.Executor) error {\n\tret, err := FindFailure(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Skin) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFTurntable) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFTurntable(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PremiumSlot) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPremiumSlot(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Origin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindOrigin(ctx, exec, o.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Offer) Reload(exec boil.Executor) error {\n\tret, err := FindOffer(exec, o.Offerid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RecipeLipid) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRecipeLipid(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Job) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindJob(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Item) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindItem(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ForeignLegalResource) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindForeignLegalResource(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureCvtermDbxref(exec, o.FeatureCvtermDbxrefID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Channel) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindChannel(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Source) Reload(exec boil.Executor) error {\n\tret, err := FindSource(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Source) Reload(exec boil.Executor) error {\n\tret, err := FindSource(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RestaurantRank) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRestaurantRank(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Employee) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindEmployee(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Phenotypeprop) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Stock) Reload(exec boil.Executor) error {\n\tret, err := FindStock(exec, o.StockID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RecordMeasure) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *HoldenAt) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindHoldenAt(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Kvstore) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindKvstore(ctx, exec, o.Member, o.Key)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureRelationship) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureRelationship(exec, o.FeatureRelationshipID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Friendship) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindFriendship(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AutomodRuleDatum) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindAutomodRuleDatum(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *DestinationRank) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDestinationRank(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (instance *Host) unsafeReload(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\txerr := instance.MetadataCore.Reload(ctx)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *retry.ErrTimeout: // If retry timed out, log it and return error ErrNotFound\n\t\t\treturn fail.NotFoundError(\"metadata of Host '%s' not found; Host deleted?\", instance.GetName())\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t}\n\n\thid, err := instance.GetID()\n\tif err != nil {\n\t\treturn fail.ConvertError(err)\n\t}\n\n\t// Request Host inspection from provider\n\tahf, xerr := instance.Service().InspectHost(ctx, hid)\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tcache, xerr := instance.Service().GetCache(ctx)\n\tif xerr != nil {\n\t\treturn xerr\n\t}\n\n\tif cache != nil {\n\t\thid, err := instance.GetID()\n\t\tif err != nil {\n\t\t\treturn fail.ConvertError(err)\n\t\t}\n\n\t\tthing, err := cache.Get(ctx, hid)\n\t\tif err != nil || thing == nil { // usually notfound\n\t\t\terr = cache.Set(ctx, hid, instance, &store.Options{Expiration: 120 * time.Minute})\n\t\t\tif err != nil {\n\t\t\t\treturn fail.ConvertError(err)\n\t\t\t}\n\t\t\ttime.Sleep(50 * time.Millisecond) // consolidate cache.Set\n\t\t} else if _, ok := thing.(*Host); !ok {\n\t\t\treturn fail.NewError(\"cache stored the wrong type\")\n\t\t}\n\t}\n\n\t// Updates the Host metadata\n\txerr = instance.Alter(ctx, func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\tahc, ok := clonable.(*abstract.HostCore)\n\t\tif !ok {\n\t\t\treturn fail.InconsistentError(\"'*abstract.HostCore' expected, '%s' received\", reflect.TypeOf(clonable).String())\n\t\t}\n\n\t\tchanged := false\n\t\tif ahc.LastState != ahf.CurrentState {\n\t\t\tahf.CurrentState = ahc.LastState\n\t\t\tchanged = true\n\t\t}\n\n\t\tinnerXErr := props.Alter(hostproperty.SizingV2, func(clonable data.Clonable) fail.Error {\n\t\t\thostSizingV2, ok := clonable.(*propertiesv2.HostSizing)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv2.HostSizing' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tallocated := converters.HostEffectiveSizingFromAbstractToPropertyV2(ahf.Sizing)\n\t\t\tif !reflect.DeepEqual(*allocated, *hostSizingV2.AllocatedSize) {\n\t\t\t\t*hostSizingV2.AllocatedSize = *allocated\n\t\t\t\tchanged = true\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif innerXErr != nil {\n\t\t\treturn innerXErr\n\t\t}\n\n\t\t// Updates Host property propertiesv1.HostNetworking from \"ground\" (Cloud Provider side)\n\t\tinnerXErr = props.Alter(hostproperty.NetworkV2, func(clonable data.Clonable) fail.Error {\n\t\t\thnV2, ok := clonable.(*propertiesv2.HostNetworking)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv2.HostNetworking' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\n\t\t\tif len(ahf.Networking.IPv4Addresses) > 0 {\n\t\t\t\thnV2.IPv4Addresses = ahf.Networking.IPv4Addresses\n\t\t\t}\n\t\t\tif len(ahf.Networking.IPv6Addresses) > 0 {\n\t\t\t\thnV2.IPv6Addresses = ahf.Networking.IPv6Addresses\n\t\t\t}\n\t\t\tif len(ahf.Networking.SubnetsByID) > 0 {\n\t\t\t\thnV2.SubnetsByID = ahf.Networking.SubnetsByID\n\t\t\t}\n\t\t\tif len(ahf.Networking.SubnetsByName) > 0 {\n\t\t\t\thnV2.SubnetsByName = ahf.Networking.SubnetsByName\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t\tif innerXErr != nil {\n\t\t\treturn innerXErr\n\t\t}\n\n\t\tif !changed {\n\t\t\treturn fail.AlteredNothingError()\n\t\t}\n\t\treturn nil\n\t})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\tswitch xerr.(type) {\n\t\tcase *fail.ErrAlteredNothing:\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn xerr\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *RentalRower) Reload(exec boil.Executor) error {\n\tret, err := FindRentalRower(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Project) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindProject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Project) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindProject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Store) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindStore(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Node) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindNode(ctx, exec, o.Address)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Illness) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindIllness(ctx, exec, o.IllnessID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFUserExperienceLog) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFUserExperienceLog(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *TrainingCost) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTrainingCost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Email) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindEmail(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Ticket) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTicket(ctx, exec, o.GuildID, o.LocalID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Description) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDescription(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Building) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBuilding(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ItemSide) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindItemSide(ctx, exec, o.ItemID, o.SideItemID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Rental) Reload(exec boil.Executor) error {\n\tret, err := FindRental(exec, o.RentalID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) Reload(exec boil.Executor) error {\n\tret, err := FindTransaction(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Address) Reload(exec boil.Executor) error {\n\tret, err := FindAddress(exec, o.AddressID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Tenant) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTenant(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Phenotypeprop) Reload(exec boil.Executor) error {\n\tret, err := FindPhenotypeprop(exec, o.PhenotypepropID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFUserSuper) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFUserSuper(ctx, exec, o.UID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) Reload(exec boil.Executor) error {\n\tret, err := FindTransaction(exec, o.TransactionID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *BraceletPhoto) Reload(exec boil.Executor) error {\n\tret, err := FindBraceletPhoto(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PeerProperty) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPeerProperty(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Tree) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTree(ctx, exec, o.Ref, o.Path, o.SetID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *StockCvterm) Reload(exec boil.Executor) error {\n\tret, err := FindStockCvterm(exec, o.StockCvtermID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Weather) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindWeather(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Currency) Reload(exec boil.Executor) error {\n\tret, err := FindCurrency(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CurrentChartDataMinutely) Reload(exec boil.Executor) error {\n\tret, err := FindCurrentChartDataMinutely(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ExchangeCurrency) Reload(exec boil.Executor) error {\n\tret, err := FindExchangeCurrency(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}" ]
[ "0.70584446", "0.70584446", "0.6943949", "0.69314134", "0.68533075", "0.6842804", "0.68383735", "0.6823546", "0.67940766", "0.6758427", "0.67310107", "0.66879946", "0.6674701", "0.66546804", "0.6642879", "0.6634314", "0.66272503", "0.655952", "0.6544683", "0.6535437", "0.6513671", "0.6501457", "0.6498456", "0.64794844", "0.64772797", "0.6467967", "0.64421576", "0.64384776", "0.64116704", "0.640926", "0.6408364", "0.640494", "0.63768", "0.6363684", "0.6357548", "0.63563997", "0.6333464", "0.63272834", "0.6320609", "0.63168216", "0.6305328", "0.62677133", "0.6253689", "0.62512267", "0.6216402", "0.6203797", "0.61771953", "0.61597043", "0.6142243", "0.6129151", "0.6100364", "0.6079017", "0.60708827", "0.6069485", "0.60185736", "0.5985868", "0.5981788", "0.594716", "0.59415996", "0.59415996", "0.5937218", "0.5936638", "0.5931553", "0.5928939", "0.5901349", "0.5901136", "0.588196", "0.58803743", "0.5876188", "0.5873759", "0.5872649", "0.58723706", "0.5857063", "0.5843288", "0.5843288", "0.5842702", "0.5842638", "0.58405083", "0.5839827", "0.5835218", "0.5816443", "0.58089244", "0.58078194", "0.5804109", "0.5799777", "0.5796754", "0.57745236", "0.57701033", "0.5714839", "0.5700775", "0.5699546", "0.5660015", "0.5657361", "0.56537485", "0.5646463", "0.56460077", "0.5644654", "0.5642908", "0.5635949", "0.5632964" ]
0.63634545
34
ReloadG refetches the object from the database using the primary keys.
func (o *AssetRevision) ReloadG() error { if o == nil { return errors.New("public: no AssetRevision provided for reload") } return o.Reload(boil.GetDB()) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Author) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Author provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Rental) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Rental provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *BlackCard) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BlackCard provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Shelf) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Shelf provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureCvtermDbxref) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureCvtermDbxref provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Stock) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Stock provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureRelationship) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no FeatureRelationship provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *APIKey) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no APIKey provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Jet) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Jet provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Source) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: no Source provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Picture) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Picture provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Utxo) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUtxo(ctx, exec, o.Rowid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Vote provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *BraceletPhoto) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BraceletPhoto provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *RentalRower) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RentalRower provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureRelationship) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureRelationship(exec, o.FeatureRelationshipID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Inventory) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Inventory provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Cvtermsynonym) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Cvtermsynonym provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *CMFTurntable) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFTurntable(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *StockCvterm) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no StockCvterm provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Item) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Item provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Ticket) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Ticket provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *PaymentObject) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPaymentObject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureCvtermDbxref(exec, o.FeatureCvtermDbxrefID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no Transaction provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Offer) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no Offer provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Transaction) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Transaction provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Friendship) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindFriendship(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *DMessageEmbed) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no DMessageEmbed provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Rental) Reload(exec boil.Executor) error {\n\tret, err := FindRental(exec, o.RentalID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Stock) Reload(exec boil.Executor) error {\n\tret, err := FindStock(exec, o.StockID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFBalanceChargeAdmin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFBalanceChargeAdmin(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Phenotypeprop) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Phenotypeprop provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *ExchangeCurrency) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no ExchangeCurrency provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *UserGoogle) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUserGoogle(ctx, exec, o.GoogleID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Address) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: no Address provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *InstrumentClass) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no InstrumentClass provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *PremiumCode) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no PremiumCode provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Kvstore) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindKvstore(ctx, exec, o.Member, o.Key)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Comment) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Comment provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Store) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Store provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *ClaimInList) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"model: no ClaimInList provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Auth) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Auth provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *UsernameListing) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no UsernameListing provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Post) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Building) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBuilding(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUser) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUser provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *RowerGroup) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RowerGroup provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Skin) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Skin provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *ItemSide) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no ItemSide provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Notification) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Notification provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Organism) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: no Organism provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Repository) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRepository(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RentalRower) Reload(exec boil.Executor) error {\n\tret, err := FindRentalRower(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PremiumSlot) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no PremiumSlot provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *HoldenAt) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindHoldenAt(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Customer) Reload(exec boil.Executor) error {\n\tret, err := FindCustomer(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Tenant) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"dbmodel: no Tenant provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *ForeignLegalResource) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindForeignLegalResource(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Shelf) Reload(exec boil.Executor) error {\n\tret, err := FindShelf(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthMessage) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthMessage provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *AutomodRuleDatum) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AutomodRuleDatum provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Storestate) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: no Storestate provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *FeatureRelationshipSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty FeatureRelationshipSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *Failure) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no Failure provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *RecordMeasure) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no RecordMeasure provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *BlackCard) Reload(exec boil.Executor) error {\n\tret, err := FindBlackCard(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ClaimInList) ReloadGP() {\n\tif err := o.Reload(boil.GetDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RestaurantRank) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRestaurantRank(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Store) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindStore(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) Reload(exec boil.Executor) error {\n\tret, err := FindVote(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Picture) Reload(exec boil.Executor) error {\n\tret, err := FindPicture(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthToken) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthToken provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *Doc) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDoc(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Author) Reload(exec boil.Executor) error {\n\tret, err := FindAuthor(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Smallblog) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindSmallblog(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RSSAnnouncement) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no RSSAnnouncement provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *Board) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBoard(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) Reload(exec boil.Executor) error {\n\tret, err := FindTransaction(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Item) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindItem(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUserUserPermission) ReloadG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no AuthUserUserPermission provided for reload\")\n\t}\n\n\treturn o.Reload(boil.GetDB())\n}", "func (o *TrainingCost) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTrainingCost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Auth) Reload(exec boil.Executor) error {\n\tret, err := FindAuth(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty FeatureCvtermDbxrefSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *UserGoogleSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UserGoogleSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `user_google`.* FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to reload all in UserGoogleSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (e *Department) Reload() error { return ent.ReloadEnt(e) }", "func (o *VSP) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVSP(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *JetSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: empty JetSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *Jet) Reload(exec boil.Executor) error {\n\tret, err := FindJet(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *APIKey) Reload(exec boil.Executor) error {\n\tret, err := FindAPIKey(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVote(ctx, exec, o.Hash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (e *Account) Reload() error { return ent.ReloadEnt(e) }", "func (e *Account) Reload() error { return ent.ReloadEnt(e) }", "func (o *DMessageEmbed) Reload(exec boil.Executor) error {\n\tret, err := FindDMessageEmbed(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUser) Reload(exec boil.Executor) error {\n\tret, err := FindAuthUser(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) ReloadGP() {\n\tif err := o.ReloadG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BookCategory) ReloadG(ctx context.Context) error {\n\tif o == nil {\n\t\treturn errors.New(\"models: no BookCategory provided for reload\")\n\t}\n\n\treturn o.Reload(ctx, boil.GetContextDB())\n}", "func (o *RecipeLipid) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRecipeLipid(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Currency) Reload(exec boil.Executor) error {\n\tret, err := FindCurrency(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: empty APIKeySlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}" ]
[ "0.7040214", "0.6944757", "0.69406617", "0.6929727", "0.69204175", "0.6895933", "0.68729824", "0.6868286", "0.6843644", "0.6802108", "0.6793841", "0.67771703", "0.6759134", "0.67538744", "0.67365855", "0.6720421", "0.6706144", "0.66955596", "0.6681266", "0.66777396", "0.66706216", "0.66577077", "0.6644464", "0.6629285", "0.66266185", "0.65889883", "0.6584061", "0.6576456", "0.6558972", "0.65571266", "0.65471375", "0.6544121", "0.654129", "0.65398806", "0.6535707", "0.652856", "0.65124387", "0.6498307", "0.6488521", "0.64735526", "0.64716864", "0.6470518", "0.6459579", "0.6451501", "0.6442404", "0.64372617", "0.64294183", "0.64246744", "0.6409913", "0.6407655", "0.6398208", "0.63900226", "0.63769233", "0.6356474", "0.63546115", "0.6345587", "0.634038", "0.6339029", "0.63385105", "0.6337342", "0.6335043", "0.63322604", "0.63238114", "0.6308184", "0.63055664", "0.6300583", "0.62874275", "0.62857276", "0.62850535", "0.6282944", "0.62788033", "0.6276014", "0.6266963", "0.6266456", "0.6266078", "0.62508196", "0.62414473", "0.6238181", "0.6228094", "0.62253124", "0.6200041", "0.6196866", "0.6196267", "0.6193572", "0.6186264", "0.6179794", "0.6173417", "0.61611724", "0.61557645", "0.6152283", "0.61429083", "0.6142884", "0.6142884", "0.61368155", "0.61341333", "0.61328304", "0.61317", "0.6131052", "0.6127643", "0.6123474" ]
0.62984484
66
Reload refetches the object from the database using the primary keys with an executor.
func (o *AssetRevision) Reload(exec boil.Executor) error { ret, err := FindAssetRevision(exec, o.Orig) if err != nil { return err } *o = *ret return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *Utxo) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUtxo(ctx, exec, o.Rowid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PaymentObject) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPaymentObject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Post) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Board) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBoard(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFTurntable) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFTurntable(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Repository) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRepository(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Task) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTask(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Jet) Reload(exec boil.Executor) error {\n\tret, err := FindJet(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) Reload(exec boil.Executor) error {\n\tret, err := FindVote(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *HoldenAt) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindHoldenAt(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Stock) Reload(exec boil.Executor) error {\n\tret, err := FindStock(exec, o.StockID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Vote) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVote(ctx, exec, o.Hash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Employee) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindEmployee(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Kvstore) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindKvstore(ctx, exec, o.Member, o.Key)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFBalanceChargeAdmin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFBalanceChargeAdmin(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) Reload(exec boil.Executor) error {\n\tret, err := FindTransaction(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PremiumCode) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPremiumCode(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Customer) Reload(exec boil.Executor) error {\n\tret, err := FindCustomer(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Job) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindJob(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureCvtermDbxref) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureCvtermDbxref(exec, o.FeatureCvtermDbxrefID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Store) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindStore(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RestaurantRank) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRestaurantRank(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Item) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindItem(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RecipeLipid) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRecipeLipid(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Offer) Reload(exec boil.Executor) error {\n\tret, err := FindOffer(exec, o.Offerid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Rental) Reload(exec boil.Executor) error {\n\tret, err := FindRental(exec, o.RentalID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Ticket) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTicket(ctx, exec, o.GuildID, o.LocalID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RentalRower) Reload(exec boil.Executor) error {\n\tret, err := FindRentalRower(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *VSP) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVSP(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *FeatureRelationship) Reload(exec boil.Executor) error {\n\tret, err := FindFeatureRelationship(exec, o.FeatureRelationshipID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Building) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBuilding(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) Reload(exec boil.Executor) error {\n\tret, err := FindTransaction(exec, o.TransactionID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Shelf) Reload(exec boil.Executor) error {\n\tret, err := FindShelf(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AutomodRuleDatum) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindAutomodRuleDatum(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Peer) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPeer(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PremiumSlot) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindPremiumSlot(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *APIKey) Reload(exec boil.Executor) error {\n\tret, err := FindAPIKey(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Currency) Reload(exec boil.Executor) error {\n\tret, err := FindCurrency(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *BlackCard) Reload(exec boil.Executor) error {\n\tret, err := FindBlackCard(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Tenant) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTenant(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Picture) Reload(exec boil.Executor) error {\n\tret, err := FindPicture(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Author) Reload(exec boil.Executor) error {\n\tret, err := FindAuthor(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Origin) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindOrigin(ctx, exec, o.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ForeignLegalResource) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindForeignLegalResource(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ExchangeCurrency) Reload(exec boil.Executor) error {\n\tret, err := FindExchangeCurrency(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *TrainingCost) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTrainingCost(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Smallblog) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindSmallblog(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Illness) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindIllness(ctx, exec, o.IllnessID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Friendship) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindFriendship(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *StockCvterm) Reload(exec boil.Executor) error {\n\tret, err := FindStockCvterm(exec, o.StockCvtermID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *WithdrawalCrypto) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindWithdrawalCrypto(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *InstrumentClass) Reload(exec boil.Executor) error {\n\tret, err := FindInstrumentClass(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Source) Reload(exec boil.Executor) error {\n\tret, err := FindSource(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Source) Reload(exec boil.Executor) error {\n\tret, err := FindSource(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUser) Reload(exec boil.Executor) error {\n\tret, err := FindAuthUser(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CurrentChartDataMinutely) Reload(exec boil.Executor) error {\n\tret, err := FindCurrentChartDataMinutely(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Doc) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDoc(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Description) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDescription(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Auth) Reload(exec boil.Executor) error {\n\tret, err := FindAuth(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ItemSide) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindItemSide(ctx, exec, o.ItemID, o.SideItemID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Project) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindProject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Project) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindProject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Weather) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindWeather(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFUserSuper) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFUserSuper(ctx, exec, o.UID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *DestinationRank) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindDestinationRank(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RecipeAdditive) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRecipeAdditive(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Channel) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindChannel(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ScheduleSubject) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindScheduleSubject(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Email) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindEmail(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *MultiAddressesSet) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindMultiAddressesSet(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (instance *Host) Reload(ctx context.Context) (ferr fail.Error) {\n\tdefer fail.OnPanic(&ferr)\n\n\tif valid.IsNil(instance) {\n\t\treturn fail.InvalidInstanceError()\n\t}\n\n\treturn instance.unsafeReload(ctx)\n}", "func (o *Address) Reload(exec boil.Executor) error {\n\tret, err := FindAddress(exec, o.AddressID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Transaction) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *Transaction) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *UsernameListing) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUsernameListing(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Cvtermsynonym) Reload(exec boil.Executor) error {\n\tret, err := FindCvtermsynonym(exec, o.CvtermsynonymID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Tree) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindTree(ctx, exec, o.Ref, o.Path, o.SetID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Node) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindNode(ctx, exec, o.Address)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *OauthClient) Reload(exec boil.Executor) error {\n\tret, err := FindOauthClient(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Segment) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindSegment(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFFamilyUserPoliciesTake) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFFamilyUserPoliciesTake(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Inventory) Reload(exec boil.Executor) error {\n\tret, err := FindInventory(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *BoardSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards`.* FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BraceletPhoto) Reload(exec boil.Executor) error {\n\tret, err := FindBraceletPhoto(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *StockKeepingUnitContent) Reload(exec boil.Executor) error {\n\tret, err := FindStockKeepingUnitContent(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *BookCategoryAssign) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindBookCategoryAssign(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *CMFUserExperienceLog) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindCMFUserExperienceLog(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *ContentUnitDerivation) Reload(exec boil.Executor) error {\n\tret, err := FindContentUnitDerivation(exec, o.SourceID, o.DerivedID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *RawVisit) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindRawVisit(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Stock) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ActivityLog) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindActivityLog(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *PaymentObjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PaymentObjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `payment_objects`.* FROM `payment_objects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PaymentObjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *Vendor) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindVendor(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *UserGoogle) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindUserGoogle(ctx, exec, o.GoogleID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *AuthUserUserPermission) Reload(exec boil.Executor) error {\n\tret, err := FindAuthUserUserPermission(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Subscriber) Reload(ctx context.Context, exec boil.ContextExecutor) error {\n\tret, err := FindSubscriber(ctx, exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *DMessageEmbed) Reload(exec boil.Executor) error {\n\tret, err := FindDMessageEmbed(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Failure) Reload(exec boil.Executor) error {\n\tret, err := FindFailure(exec, o.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*o = *ret\n\treturn nil\n}", "func (o *Author) ReloadP(exec boil.Executor) {\n\tif err := o.Reload(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *JetSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tjets := JetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `jets`.* FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&jets)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JetSlice\")\n\t}\n\n\t*o = jets\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vote\\\".* FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}" ]
[ "0.6983497", "0.66200763", "0.65347505", "0.6530822", "0.6524277", "0.65104795", "0.64694655", "0.6401659", "0.6392273", "0.6389676", "0.63832426", "0.6382202", "0.63485044", "0.633335", "0.6314799", "0.6286139", "0.6273836", "0.62624055", "0.62482136", "0.6243617", "0.62367785", "0.6230149", "0.6228272", "0.62120515", "0.6209666", "0.6176749", "0.61435616", "0.6128641", "0.61262685", "0.6122979", "0.61215365", "0.6112485", "0.61065394", "0.6101816", "0.6092357", "0.60820585", "0.6077376", "0.60696816", "0.606701", "0.6058028", "0.6057523", "0.6055738", "0.6054766", "0.60308444", "0.60254395", "0.6020521", "0.6002179", "0.59972596", "0.5994659", "0.59926015", "0.5974468", "0.5972955", "0.59659266", "0.59659266", "0.5963832", "0.59622425", "0.59484416", "0.59433", "0.5938979", "0.59384704", "0.59336156", "0.59336156", "0.59192806", "0.5898401", "0.5897635", "0.58723295", "0.58643055", "0.58630496", "0.58610815", "0.584435", "0.5842114", "0.5840867", "0.58366376", "0.58366376", "0.5815776", "0.58082986", "0.58034563", "0.5801415", "0.5790616", "0.5787762", "0.57770514", "0.5764935", "0.5762196", "0.5739", "0.5733856", "0.5733641", "0.57307446", "0.5712127", "0.5708562", "0.57084155", "0.57055914", "0.5700446", "0.5689971", "0.5677355", "0.56702816", "0.5643935", "0.56436807", "0.5642724", "0.5629475", "0.56271243", "0.562108" ]
0.0
-1
ReloadAllGP refetches every row with matching primary key column values and overwrites the original object slice with the newly updated slice. Panics on error.
func (o *AssetRevisionSlice) ReloadAllGP() { if err := o.ReloadAllG(); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *UserGoogleSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UserGoogleSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `user_google`.* FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to reload all in UserGoogleSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ClaimInListSlice) ReloadAllGP() {\n\tif err := o.ReloadAll(boil.GetDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PaymentObjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PaymentObjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `payment_objects`.* FROM `payment_objects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PaymentObjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeLipidSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeLipidSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_lipid\\\".* FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeLipidSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TransactionSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TransactionSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *KvstoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := KvstoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"kvstore\\\".* FROM \\\"kvstore\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, kvstorePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in KvstoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FriendshipSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := FriendshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `friendship`.* FROM `friendship` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in FriendshipSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tapiKeys := APIKeySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"api_keys\\\".* FROM \\\"api_keys\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, apiKeyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&apiKeys)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in APIKeySlice\")\n\t}\n\n\t*o = apiKeys\n\n\treturn nil\n}", "func (o *PeerSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `peers`.* FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in PeerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VSPSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VSPSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vsp\\\".* FROM \\\"vsp\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, vspPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VSPSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumSlotSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumSlotSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_slots\\\".* FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumSlotSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *EmailSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmailSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `email`.* FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to reload all in EmailSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *HoldenAtSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := HoldenAtSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"HoldenAt\\\".* FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in HoldenAtSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PictureSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tpictures := PictureSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `pictures`.* FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&pictures)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PictureSlice\")\n\t}\n\n\t*o = pictures\n\n\treturn nil\n}", "func (o *AuthUserSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TrainingCostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TrainingCostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"training_costs\\\".* FROM \\\"training_costs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, trainingCostPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TrainingCostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PeerPropertySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerPropertySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"peer_properties\\\".* FROM \\\"peer_properties\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, peerPropertyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PeerPropertySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *APIKeySlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: empty APIKeySlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *PremiumCodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumCodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_codes\\\".* FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumCodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TaskSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TaskSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tasks\\\".* FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TaskSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BuildingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BuildingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"buildings\\\".* FROM \\\"buildings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, buildingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to reload all in BuildingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamilyUserPoliciesTakeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamilyUserPoliciesTakeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_family_user_policies_take`.* FROM `cmf_family_user_policies_take` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamilyUserPoliciesTakeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFTurntableSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFTurntableSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_turntable`.* FROM `cmf_turntable` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFTurntableSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *JobSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := JobSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"jobs\\\".* FROM \\\"jobs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, jobPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JobSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vote\\\".* FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthItemGroupSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := AuthItemGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authItemGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `auth_item_groups`.* FROM `auth_item_groups` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, authItemGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthItemGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WithdrawalCryptoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WithdrawalCryptoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"withdrawal_crypto\\\".* FROM \\\"withdrawal_crypto\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlite3: unable to reload all in WithdrawalCryptoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tphenotypeprops := PhenotypepropSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"phenotypeprop\\\".* FROM \\\"phenotypeprop\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, phenotypepropPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(phenotypepropPrimaryKeyColumns), 1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&phenotypeprops)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in PhenotypepropSlice\")\n\t}\n\n\t*o = phenotypeprops\n\n\treturn nil\n}", "func (o *RowerGroupSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RowerGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rower_group\\\".* FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RowerGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"items\\\".* FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UtxoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UtxoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"utxo\\\".* FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UtxoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauths := AuthSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"auths\\\".* FROM \\\"auths\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&auths)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthSlice\")\n\t}\n\n\t*o = auths\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFFamiliesPolicySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamiliesPolicySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamiliesPolicyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_families_policies`.* FROM `cmf_families_policies` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamiliesPolicyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamiliesPolicySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `project`.* FROM `project` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrentChartDataMinutelySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrentChartDataMinutelySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"current_chart_data_minutely\\\".* FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrentChartDataMinutelySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `projects`.* FROM `projects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *VendorSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VendorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vendors`.* FROM `vendors` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VendorSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserUserPermissionSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *IllnessSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := IllnessSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"illness\\\".* FROM \\\"illness\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, illnessPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in IllnessSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AddressSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ForeignLegalResourceSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ForeignLegalResourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), foreignLegalResourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"ForeignLegalResources\\\".* FROM \\\"ForeignLegalResources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, foreignLegalResourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ForeignLegalResourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RawVisitSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RawVisitSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"raw_visits\\\".* FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RawVisitSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TenantSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TenantSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `tenants`.* FROM `tenants` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to reload all in TenantSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecordMeasureSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TenantSlice) ReloadAllGP(ctx context.Context) {\n\tif err := o.ReloadAll(ctx, boil.GetContextDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CMFBalanceChargeAdminSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFBalanceChargeAdminSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_balance_charge_admin`.* FROM `cmf_balance_charge_admin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFBalanceChargeAdminSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tfeatureRelationships := FeatureRelationshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), featureRelationshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"feature_relationship\\\".* FROM \\\"feature_relationship\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, featureRelationshipPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(featureRelationshipPrimaryKeyColumns), 1, len(featureRelationshipPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&featureRelationships)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in FeatureRelationshipSlice\")\n\t}\n\n\t*o = featureRelationships\n\n\treturn nil\n}", "func (o *NotificationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NotificationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"notification\\\".* FROM \\\"notification\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, notificationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NotificationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CustomerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CustomerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"customers\\\".* FROM \\\"customers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, customerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CustomerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tjets := JetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `jets`.* FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&jets)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JetSlice\")\n\t}\n\n\t*o = jets\n\n\treturn nil\n}", "func (o *StockCvtermSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ShelfSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TicketSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TicketSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tickets\\\".* FROM \\\"tickets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, ticketPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TicketSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *InventorySlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"posts\\\".* FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to reload all in PostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *DocSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DocSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), docPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `doc`.* FROM `doc` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, docPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DocSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty PhenotypepropSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *AuthUserSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthorSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tsources := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&sources)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = sources\n\n\treturn nil\n}", "func (o *ContentUnitDerivationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ContentUnitDerivationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"content_unit_derivations\\\".* FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in ContentUnitDerivationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PhenotypepropSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUsers := AuthUserSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user`.* FROM `auth_user` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserPrimaryKeyColumns), 1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUsers)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserSlice\")\n\t}\n\n\t*o = authUsers\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *NodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), nodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"node\\\".* FROM \\\"node\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, nodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RepositorySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RepositorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `repositories`.* FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RepositorySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BlockSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BlockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"block\\\".* FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlockSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tstocks := StockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `stock`.* FROM `stock` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&stocks)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StockSlice\")\n\t}\n\n\t*o = stocks\n\n\treturn nil\n}", "func (o *ChannelSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChannelSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"channels\\\".* FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ChannelSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty FeatureCvtermDbxrefSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *RecipeAdditiveSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeAdditiveSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeAdditivePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_additive\\\".* FROM \\\"recipe_additive\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeAdditivePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeAdditiveSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CvtermsynonymSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BoardSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards`.* FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFUserSuperSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFUserSuperSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserSuperPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_user_super`.* FROM `cmf_user_super` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserSuperPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFUserSuperSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrencySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrencySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"currency\\\".* FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrencySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tvotes := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vote`.* FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&votes)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = votes\n\n\treturn nil\n}", "func (o *SegmentSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SegmentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"segment\\\".* FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to reload all in SegmentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: empty SourceSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *ChatSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChatSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"chat\\\".* FROM \\\"chat\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, chatPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dal: unable to reload all in ChatSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\trentals := RentalSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `rental`.* FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&rentals)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in RentalSlice\")\n\t}\n\n\t*o = rentals\n\n\treturn nil\n}", "func (o *BlackCardSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *OfferSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OfferSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"offers\\\".* FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to reload all in OfferSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PictureSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TreeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TreeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), treePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"trees\\\".* FROM \\\"trees\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, treePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TreeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalRowerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RentalRowerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rental_rowers\\\".* FROM \\\"rental_rowers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rentalRowerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RentalRowerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BraceletPhotoSlice) ReloadAllGP() {\n\tif err := o.ReloadAllG(); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}" ]
[ "0.7537837", "0.74087936", "0.74004203", "0.7378555", "0.737151", "0.7350787", "0.7350787", "0.7317934", "0.731113", "0.7309129", "0.7305479", "0.7304253", "0.72957605", "0.7287472", "0.72815377", "0.7261435", "0.7238059", "0.7230279", "0.7227206", "0.722411", "0.72151506", "0.7202241", "0.7198719", "0.7191872", "0.71905816", "0.71850634", "0.7175391", "0.7174688", "0.71736073", "0.7170256", "0.7169946", "0.7158005", "0.71557784", "0.7152914", "0.7149527", "0.71423656", "0.7140975", "0.71382546", "0.7133576", "0.7129356", "0.71264565", "0.71259254", "0.7119327", "0.7108996", "0.71064025", "0.71050346", "0.7103827", "0.71016824", "0.7100539", "0.7088512", "0.7087915", "0.70861983", "0.70826304", "0.7079401", "0.7072262", "0.70635563", "0.7057077", "0.7053124", "0.7048203", "0.7046668", "0.7041132", "0.7040438", "0.70343447", "0.7027436", "0.7026015", "0.70190704", "0.7015033", "0.70043075", "0.7000397", "0.69989085", "0.6996267", "0.69947654", "0.69914293", "0.6983164", "0.6976738", "0.6973647", "0.6962919", "0.69557065", "0.6955282", "0.6954477", "0.69533294", "0.6948967", "0.6947638", "0.69445884", "0.6943966", "0.6941411", "0.693997", "0.6937846", "0.6937799", "0.69346774", "0.6932624", "0.69293374", "0.69216394", "0.6921509", "0.69207937", "0.6906056", "0.68951017", "0.6886717", "0.6885673", "0.6881702" ]
0.6899029
96
ReloadAllP refetches every row with matching primary key column values and overwrites the original object slice with the newly updated slice. Panics on error.
func (o *AssetRevisionSlice) ReloadAllP(exec boil.Executor) { if err := o.ReloadAll(exec); err != nil { panic(boil.WrapErr(err)) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *APIKeySlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecipeLipidSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeLipidSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_lipid\\\".* FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeLipidSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumSlotSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumSlotSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_slots\\\".* FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumSlotSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PaymentObjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PaymentObjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `payment_objects`.* FROM `payment_objects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PaymentObjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PeerPropertySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerPropertySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"peer_properties\\\".* FROM \\\"peer_properties\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, peerPropertyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PeerPropertySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PeerSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `peers`.* FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in PeerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TransactionSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TransactionSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *HoldenAtSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := HoldenAtSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"HoldenAt\\\".* FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in HoldenAtSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VSPSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VSPSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vsp\\\".* FROM \\\"vsp\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, vspPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VSPSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumCodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumCodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_codes\\\".* FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumCodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tapiKeys := APIKeySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"api_keys\\\".* FROM \\\"api_keys\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, apiKeyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&apiKeys)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in APIKeySlice\")\n\t}\n\n\t*o = apiKeys\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tphenotypeprops := PhenotypepropSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"phenotypeprop\\\".* FROM \\\"phenotypeprop\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, phenotypepropPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(phenotypepropPrimaryKeyColumns), 1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&phenotypeprops)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in PhenotypepropSlice\")\n\t}\n\n\t*o = phenotypeprops\n\n\treturn nil\n}", "func (o *CMFTurntableSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFTurntableSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_turntable`.* FROM `cmf_turntable` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFTurntableSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *PostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"posts\\\".* FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to reload all in PostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vote\\\".* FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CurrentChartDataMinutelySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrentChartDataMinutelySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"current_chart_data_minutely\\\".* FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrentChartDataMinutelySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ShelfSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ClaimInListSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *EmailSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmailSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `email`.* FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to reload all in EmailSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ForeignLegalResourceSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ForeignLegalResourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), foreignLegalResourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"ForeignLegalResources\\\".* FROM \\\"ForeignLegalResources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, foreignLegalResourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ForeignLegalResourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *StockSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *SourceSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *KvstoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := KvstoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"kvstore\\\".* FROM \\\"kvstore\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, kvstorePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in KvstoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *OfferSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OfferSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"offers\\\".* FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to reload all in OfferSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TaskSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TaskSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tasks\\\".* FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TaskSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PictureSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tpictures := PictureSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `pictures`.* FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&pictures)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PictureSlice\")\n\t}\n\n\t*o = pictures\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\trentals := RentalSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `rental`.* FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&rentals)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in RentalSlice\")\n\t}\n\n\t*o = rentals\n\n\treturn nil\n}", "func (o *AuthorSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"items\\\".* FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *InventorySlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *FriendshipSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := FriendshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `friendship`.* FROM `friendship` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in FriendshipSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *EmployeeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmployeeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), employeePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"employee\\\".* FROM \\\"employee\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, employeePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in EmployeeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockCvtermSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauths := AuthSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"auths\\\".* FROM \\\"auths\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&auths)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthSlice\")\n\t}\n\n\t*o = auths\n\n\treturn nil\n}", "func (o *TicketSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TicketSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tickets\\\".* FROM \\\"tickets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, ticketPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TicketSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *IllnessSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := IllnessSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"illness\\\".* FROM \\\"illness\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, illnessPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in IllnessSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JobSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := JobSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"jobs\\\".* FROM \\\"jobs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, jobPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JobSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tvotes := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vote`.* FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&votes)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = votes\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CustomerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CustomerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"customers\\\".* FROM \\\"customers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, customerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CustomerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AddressSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RecordMeasureSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *TenantSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TenantSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `tenants`.* FROM `tenants` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to reload all in TenantSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UtxoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UtxoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"utxo\\\".* FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UtxoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tjets := JetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `jets`.* FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&jets)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JetSlice\")\n\t}\n\n\t*o = jets\n\n\treturn nil\n}", "func (o *NotificationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NotificationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"notification\\\".* FROM \\\"notification\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, notificationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NotificationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamilyUserPoliciesTakeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamilyUserPoliciesTakeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_family_user_policies_take`.* FROM `cmf_family_user_policies_take` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamilyUserPoliciesTakeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CvtermsynonymSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RentalRowerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RentalRowerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rental_rowers\\\".* FROM \\\"rental_rowers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rentalRowerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RentalRowerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamiliesPolicySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamiliesPolicySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamiliesPolicyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_families_policies`.* FROM `cmf_families_policies` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamiliesPolicyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamiliesPolicySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WithdrawalCryptoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WithdrawalCryptoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"withdrawal_crypto\\\".* FROM \\\"withdrawal_crypto\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlite3: unable to reload all in WithdrawalCryptoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ShelfSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tshelves := ShelfSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), shelfPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `shelf`.* FROM `shelf` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, shelfPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(shelfPrimaryKeyColumns), 1, len(shelfPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&shelves)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ShelfSlice\")\n\t}\n\n\t*o = shelves\n\n\treturn nil\n}", "func (o *BlackCardSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *AuthUserUserPermissionSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CurrencySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrencySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"currency\\\".* FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrencySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WeatherSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WeatherSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"prh\\\".\\\"weather\\\".* FROM \\\"prh\\\".\\\"weather\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, weatherPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to reload all in WeatherSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *BlockSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BlockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"block\\\".* FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlockSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `project`.* FROM `project` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `projects`.* FROM `projects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ChatSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChatSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"chat\\\".* FROM \\\"chat\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, chatPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dal: unable to reload all in ChatSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TenantSlice) ReloadAllP(ctx context.Context, exec boil.ContextExecutor) {\n\tif err := o.ReloadAll(ctx, exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tsources := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&sources)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = sources\n\n\treturn nil\n}", "func (o *BookCategoryAssignSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BookCategoryAssignSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), bookCategoryAssignPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `book_category_assign`.* FROM `book_category_assign` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, bookCategoryAssignPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BookCategoryAssignSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *DocSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DocSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), docPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `doc`.* FROM `doc` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, docPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DocSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BoardSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards`.* FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RepositorySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RepositorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `repositories`.* FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RepositorySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUsers := AuthUserSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user`.* FROM `auth_user` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserPrimaryKeyColumns), 1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUsers)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserSlice\")\n\t}\n\n\t*o = authUsers\n\n\treturn nil\n}", "func (o *PublisherSearchIdxSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PublisherSearchIdxSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), publisherSearchIdxPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"publisher_search_idx\\\".* FROM \\\"publisher_search_idx\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, publisherSearchIdxPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PublisherSearchIdxSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VendorSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VendorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vendors`.* FROM `vendors` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VendorSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AutomodRuleDatumSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := AutomodRuleDatumSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), automodRuleDatumPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"automod_rule_data\\\".* FROM \\\"automod_rule_data\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, automodRuleDatumPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AutomodRuleDatumSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFSlideItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFSlideItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfSlideItemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_slide_item`.* FROM `cmf_slide_item` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfSlideItemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFSlideItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ClaimInListSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ClaimInListSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), claimInListPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `claim_in_list`.* FROM `claim_in_list` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, claimInListPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in ClaimInListSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BraceletPhotoSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RawVisitSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RawVisitSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"raw_visits\\\".* FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RawVisitSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BuildingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BuildingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"buildings\\\".* FROM \\\"buildings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, buildingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to reload all in BuildingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transaction`.* FROM `transaction` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *SegmentSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SegmentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"segment\\\".* FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to reload all in SegmentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *DescriptionSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DescriptionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), descriptionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `descriptions`.* FROM `descriptions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, descriptionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DescriptionSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transactions`.* FROM `transactions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *CMFBalanceChargeAdminSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFBalanceChargeAdminSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_balance_charge_admin`.* FROM `cmf_balance_charge_admin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFBalanceChargeAdminSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserUserPermissionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUserUserPermissions := AuthUserUserPermissionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserUserPermissionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user_user_permissions`.* FROM `auth_user_user_permissions` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserUserPermissionPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserUserPermissionPrimaryKeyColumns), 1, len(authUserUserPermissionPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUserUserPermissions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserUserPermissionSlice\")\n\t}\n\n\t*o = authUserUserPermissions\n\n\treturn nil\n}", "func (o *BraceletPhotoSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tbraceletPhotos := BraceletPhotoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), braceletPhotoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `bracelet_photo`.* FROM `bracelet_photo` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, braceletPhotoPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(braceletPhotoPrimaryKeyColumns), 1, len(braceletPhotoPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&braceletPhotos)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BraceletPhotoSlice\")\n\t}\n\n\t*o = braceletPhotos\n\n\treturn nil\n}", "func (o *UsernameListingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UsernameListingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), usernameListingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"username_listings\\\".* FROM \\\"username_listings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, usernameListingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UsernameListingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tstocks := StockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `stock`.* FROM `stock` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&stocks)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StockSlice\")\n\t}\n\n\t*o = stocks\n\n\treturn nil\n}", "func (o *AuthMessageSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *ChannelSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChannelSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"channels\\\".* FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ChannelSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeAdditiveSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeAdditiveSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeAdditivePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_additive\\\".* FROM \\\"recipe_additive\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeAdditivePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeAdditiveSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFUserExperienceLogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFUserExperienceLogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserExperienceLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_user_experience_log`.* FROM `cmf_user_experience_log` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserExperienceLogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFUserExperienceLogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthorSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthors := AuthorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"authors\\\".* FROM \\\"authors\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authors)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in AuthorSlice\")\n\t}\n\n\t*o = authors\n\n\treturn nil\n}", "func (o *AuthTokenSlice) ReloadAllP(exec boil.Executor) {\n\tif err := o.ReloadAll(exec); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *RowerGroupSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RowerGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rower_group\\\".* FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RowerGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SmallblogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SmallblogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), smallblogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `smallblog`.* FROM `smallblog` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, smallblogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SmallblogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}" ]
[ "0.77445567", "0.7700082", "0.7694786", "0.7645146", "0.7617879", "0.75965595", "0.7586945", "0.7570969", "0.7549929", "0.7549929", "0.7548364", "0.75394225", "0.7538638", "0.7500682", "0.74998564", "0.7494313", "0.74897176", "0.74837875", "0.74810207", "0.7475843", "0.7463085", "0.7458494", "0.7457746", "0.7456265", "0.74461794", "0.7431938", "0.74315524", "0.7427084", "0.7422317", "0.7418239", "0.7396974", "0.73935556", "0.73701316", "0.73577654", "0.7357012", "0.73505485", "0.7349761", "0.7336572", "0.73300314", "0.73293984", "0.7324557", "0.73178816", "0.7317095", "0.73146695", "0.7310914", "0.72993445", "0.7296132", "0.7278225", "0.72779834", "0.72735906", "0.72668266", "0.72561765", "0.7253588", "0.72519326", "0.72300124", "0.72299385", "0.72235525", "0.7218094", "0.7217141", "0.72150904", "0.7212376", "0.7210634", "0.72083795", "0.72034824", "0.72008014", "0.7188891", "0.71868354", "0.7184732", "0.7176993", "0.7174567", "0.7167506", "0.7163674", "0.71605074", "0.7153362", "0.71530193", "0.7144909", "0.7144397", "0.7143071", "0.7136956", "0.7136036", "0.7134099", "0.713332", "0.71289885", "0.7124769", "0.71218026", "0.71205187", "0.71180046", "0.71174663", "0.71153224", "0.7109601", "0.71083146", "0.7107196", "0.7094121", "0.708145", "0.70813304", "0.7074756", "0.7072767", "0.7071936", "0.7068223", "0.70627314" ]
0.71602744
73
ReloadAllG refetches every row with matching primary key column values and overwrites the original object slice with the newly updated slice.
func (o *AssetRevisionSlice) ReloadAllG() error { if o == nil { return errors.New("public: empty AssetRevisionSlice provided for reload all") } return o.ReloadAll(boil.GetDB()) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *HoldenAtSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := HoldenAtSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"HoldenAt\\\".* FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in HoldenAtSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UserGoogleSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UserGoogleSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `user_google`.* FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to reload all in UserGoogleSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFTurntableSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFTurntableSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_turntable`.* FROM `cmf_turntable` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFTurntableSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FriendshipSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := FriendshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `friendship`.* FROM `friendship` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in FriendshipSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *KvstoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := KvstoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"kvstore\\\".* FROM \\\"kvstore\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, kvstorePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in KvstoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PaymentObjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PaymentObjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `payment_objects`.* FROM `payment_objects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PaymentObjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeLipidSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeLipidSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_lipid\\\".* FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeLipidSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BuildingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BuildingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"buildings\\\".* FROM \\\"buildings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, buildingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to reload all in BuildingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrentChartDataMinutelySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrentChartDataMinutelySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"current_chart_data_minutely\\\".* FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrentChartDataMinutelySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RowerGroupSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RowerGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rower_group\\\".* FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RowerGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *EmailSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmailSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `email`.* FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to reload all in EmailSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SegmentSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SegmentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"segment\\\".* FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to reload all in SegmentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tfeatureRelationships := FeatureRelationshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), featureRelationshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"feature_relationship\\\".* FROM \\\"feature_relationship\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, featureRelationshipPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(featureRelationshipPrimaryKeyColumns), 1, len(featureRelationshipPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&featureRelationships)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in FeatureRelationshipSlice\")\n\t}\n\n\t*o = featureRelationships\n\n\treturn nil\n}", "func (o *TrainingCostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TrainingCostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"training_costs\\\".* FROM \\\"training_costs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, trainingCostPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TrainingCostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CustomerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CustomerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"customers\\\".* FROM \\\"customers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, customerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CustomerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumSlotSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumSlotSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_slots\\\".* FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumSlotSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TenantSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TenantSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `tenants`.* FROM `tenants` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to reload all in TenantSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vote\\\".* FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UtxoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UtxoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"utxo\\\".* FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UtxoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JobSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := JobSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"jobs\\\".* FROM \\\"jobs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, jobPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JobSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *IllnessSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := IllnessSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"illness\\\".* FROM \\\"illness\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, illnessPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in IllnessSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumCodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumCodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_codes\\\".* FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumCodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ContentUnitDerivationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ContentUnitDerivationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"content_unit_derivations\\\".* FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in ContentUnitDerivationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *DocSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DocSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), docPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `doc`.* FROM `doc` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, docPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DocSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WithdrawalCryptoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WithdrawalCryptoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"withdrawal_crypto\\\".* FROM \\\"withdrawal_crypto\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlite3: unable to reload all in WithdrawalCryptoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VendorSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VendorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vendors`.* FROM `vendors` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VendorSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RawVisitSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RawVisitSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"raw_visits\\\".* FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RawVisitSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tstocks := StockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `stock`.* FROM `stock` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&stocks)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StockSlice\")\n\t}\n\n\t*o = stocks\n\n\treturn nil\n}", "func (o *ItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"items\\\".* FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamilyUserPoliciesTakeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamilyUserPoliciesTakeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_family_user_policies_take`.* FROM `cmf_family_user_policies_take` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamilyUserPoliciesTakeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauths := AuthSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"auths\\\".* FROM \\\"auths\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&auths)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthSlice\")\n\t}\n\n\t*o = auths\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tapiKeys := APIKeySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"api_keys\\\".* FROM \\\"api_keys\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, apiKeyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&apiKeys)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in APIKeySlice\")\n\t}\n\n\t*o = apiKeys\n\n\treturn nil\n}", "func (o *OauthClientSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OauthClientSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), oauthClientPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `oauth_clients`.* FROM `oauth_clients` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, oauthClientPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in OauthClientSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeAdditiveSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeAdditiveSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeAdditivePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_additive\\\".* FROM \\\"recipe_additive\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeAdditivePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeAdditiveSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TaskSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TaskSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tasks\\\".* FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TaskSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VSPSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VSPSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vsp\\\".* FROM \\\"vsp\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, vspPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VSPSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\trentals := RentalSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `rental`.* FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&rentals)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in RentalSlice\")\n\t}\n\n\t*o = rentals\n\n\treturn nil\n}", "func (o *ShelfSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tshelves := ShelfSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), shelfPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `shelf`.* FROM `shelf` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, shelfPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(shelfPrimaryKeyColumns), 1, len(shelfPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&shelves)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ShelfSlice\")\n\t}\n\n\t*o = shelves\n\n\treturn nil\n}", "func (o *ChannelSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChannelSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"channels\\\".* FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ChannelSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PeerSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `peers`.* FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in PeerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrencySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrencySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"currency\\\".* FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrencySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SmallblogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SmallblogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), smallblogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `smallblog`.* FROM `smallblog` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, smallblogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SmallblogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TicketSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TicketSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tickets\\\".* FROM \\\"tickets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, ticketPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TicketSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"models: empty APIKeySlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *BoardSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards`.* FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ChatSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChatSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"chat\\\".* FROM \\\"chat\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, chatPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dal: unable to reload all in ChatSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFBalanceChargeAdminSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFBalanceChargeAdminSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_balance_charge_admin`.* FROM `cmf_balance_charge_admin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFBalanceChargeAdminSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ForeignLegalResourceSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ForeignLegalResourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), foreignLegalResourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"ForeignLegalResources\\\".* FROM \\\"ForeignLegalResources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, foreignLegalResourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ForeignLegalResourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PictureSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tpictures := PictureSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `pictures`.* FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&pictures)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PictureSlice\")\n\t}\n\n\t*o = pictures\n\n\treturn nil\n}", "func (o *RentalRowerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RentalRowerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rental_rowers\\\".* FROM \\\"rental_rowers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rentalRowerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RentalRowerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tjets := JetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `jets`.* FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&jets)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JetSlice\")\n\t}\n\n\t*o = jets\n\n\treturn nil\n}", "func (o *BlockSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BlockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"block\\\".* FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlockSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureCvtermDbxrefSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty FeatureCvtermDbxrefSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *CMFUserSuperSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFUserSuperSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserSuperPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_user_super`.* FROM `cmf_user_super` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserSuperPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFUserSuperSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RepositorySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RepositorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `repositories`.* FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RepositorySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthItemGroupSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := AuthItemGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authItemGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `auth_item_groups`.* FROM `auth_item_groups` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, authItemGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthItemGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PublisherSearchIdxSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PublisherSearchIdxSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), publisherSearchIdxPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"publisher_search_idx\\\".* FROM \\\"publisher_search_idx\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, publisherSearchIdxPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PublisherSearchIdxSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *NotificationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NotificationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"notification\\\".* FROM \\\"notification\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, notificationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NotificationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tsources := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&sources)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = sources\n\n\treturn nil\n}", "func (o *PostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"posts\\\".* FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to reload all in PostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamiliesPolicySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamiliesPolicySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamiliesPolicyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_families_policies`.* FROM `cmf_families_policies` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamiliesPolicyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamiliesPolicySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RestaurantRankSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RestaurantRankSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), restaurantRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"restaurant_rank\\\".* FROM \\\"restaurant_rank\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, restaurantRankPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RestaurantRankSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SubscriberSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SubscriberSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), subscriberPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `subscribers`.* FROM `subscribers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, subscriberPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SubscriberSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *EmployeeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmployeeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), employeePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"employee\\\".* FROM \\\"employee\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, employeePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in EmployeeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tvotes := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vote`.* FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&votes)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = votes\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: empty SourceSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *UsernameListingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UsernameListingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), usernameListingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"username_listings\\\".* FROM \\\"username_listings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, usernameListingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UsernameListingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty FeatureRelationshipSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *OfferSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OfferSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"offers\\\".* FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to reload all in OfferSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `project`.* FROM `project` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *NodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), nodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"node\\\".* FROM \\\"node\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, nodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *MultiAddressesSetSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := MultiAddressesSetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), multiAddressesSetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"multi_addresses_sets\\\".* FROM \\\"multi_addresses_sets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, multiAddressesSetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in MultiAddressesSetSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockKeepingUnitContentSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := StockKeepingUnitContentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockKeepingUnitContentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"stock_keeping_unit_content\\\".* FROM \\\"stock_keeping_unit_content\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, stockKeepingUnitContentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to reload all in StockKeepingUnitContentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TreeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TreeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), treePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"trees\\\".* FROM \\\"trees\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, treePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TreeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `projects`.* FROM `projects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BlackCardSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tblackCards := BlackCardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blackCardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"black_cards\\\".* FROM \\\"black_cards\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blackCardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&blackCards)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlackCardSlice\")\n\t}\n\n\t*o = blackCards\n\n\treturn nil\n}", "func (o *DescriptionSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DescriptionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), descriptionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `descriptions`.* FROM `descriptions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, descriptionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DescriptionSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transaction`.* FROM `transaction` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *WeatherSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WeatherSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"prh\\\".\\\"weather\\\".* FROM \\\"prh\\\".\\\"weather\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, weatherPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to reload all in WeatherSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthorSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthors := AuthorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"authors\\\".* FROM \\\"authors\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authors)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in AuthorSlice\")\n\t}\n\n\t*o = authors\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tphenotypeprops := PhenotypepropSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"phenotypeprop\\\".* FROM \\\"phenotypeprop\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, phenotypepropPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(phenotypepropPrimaryKeyColumns), 1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&phenotypeprops)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in PhenotypepropSlice\")\n\t}\n\n\t*o = phenotypeprops\n\n\treturn nil\n}", "func (o *PeerPropertySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerPropertySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"peer_properties\\\".* FROM \\\"peer_properties\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, peerPropertyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PeerPropertySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transactions`.* FROM `transactions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *BraceletPhotoSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tbraceletPhotos := BraceletPhotoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), braceletPhotoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `bracelet_photo`.* FROM `bracelet_photo` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, braceletPhotoPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(braceletPhotoPrimaryKeyColumns), 1, len(braceletPhotoPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&braceletPhotos)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BraceletPhotoSlice\")\n\t}\n\n\t*o = braceletPhotos\n\n\treturn nil\n}", "func (o *OriginSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OriginSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), originPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"origins\\\".* FROM \\\"origins\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, originPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in OriginSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"sqlboiler: empty RentalSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *AuthUserSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUsers := AuthUserSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user`.* FROM `auth_user` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserPrimaryKeyColumns), 1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUsers)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserSlice\")\n\t}\n\n\t*o = authUsers\n\n\treturn nil\n}", "func (o *StoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := StoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"stores\\\".* FROM \\\"stores\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, storePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ClaimInListSlice) ReloadAllGP() {\n\tif err := o.ReloadAll(boil.GetDB()); err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n}", "func (o *CommentSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tcomments := CommentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), commentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `comment`.* FROM `comment` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, commentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&comments)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CommentSlice\")\n\t}\n\n\t*o = comments\n\n\treturn nil\n}", "func (o *AddressSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\taddresses := AddressSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), addressPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `address`.* FROM `address` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, addressPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&addresses)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in AddressSlice\")\n\t}\n\n\t*o = addresses\n\n\treturn nil\n}", "func (o *AutomodRuleDatumSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := AutomodRuleDatumSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), automodRuleDatumPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"automod_rule_data\\\".* FROM \\\"automod_rule_data\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, automodRuleDatumPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AutomodRuleDatumSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockCvtermSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tstockCvterms := StockCvtermSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockCvtermPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"stock_cvterm\\\".* FROM \\\"stock_cvterm\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, stockCvtermPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(stockCvtermPrimaryKeyColumns), 1, len(stockCvtermPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&stockCvterms)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in StockCvtermSlice\")\n\t}\n\n\t*o = stockCvterms\n\n\treturn nil\n}", "func (o *BookCategoryAssignSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BookCategoryAssignSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), bookCategoryAssignPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `book_category_assign`.* FROM `book_category_assign` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, bookCategoryAssignPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BookCategoryAssignSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"chado: empty PhenotypepropSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *DestinationRankSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DestinationRankSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), destinationRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"destination_rank\\\".* FROM \\\"destination_rank\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, destinationRankPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DestinationRankSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthorSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"mdbmdbmodels: empty AuthorSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}", "func (o *ItemSideSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSideSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemSidePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"item_sides\\\".* FROM \\\"item_sides\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemSidePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSideSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BoardsSectionsPositionSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardsSectionsPositionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardsSectionsPositionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards_sections_positions`.* FROM `boards_sections_positions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardsSectionsPositionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardsSectionsPositionSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *OfferSlice) ReloadAllG() error {\n\tif o == nil {\n\t\treturn errors.New(\"stellarcore: empty OfferSlice provided for reload all\")\n\t}\n\n\treturn o.ReloadAll(boil.GetDB())\n}" ]
[ "0.77024573", "0.7653973", "0.7619304", "0.7584887", "0.75712883", "0.7532565", "0.752322", "0.74860036", "0.7454503", "0.7445939", "0.7439031", "0.7432364", "0.7416415", "0.741579", "0.7414969", "0.74109316", "0.74105793", "0.7407815", "0.7398735", "0.7394498", "0.7393", "0.73705393", "0.73670495", "0.7359358", "0.73572224", "0.73463374", "0.7333893", "0.73338896", "0.7332559", "0.73319626", "0.7330145", "0.73244107", "0.732214", "0.7319303", "0.73169625", "0.731596", "0.7309425", "0.73067945", "0.73022527", "0.72990346", "0.7286069", "0.7280824", "0.72740304", "0.7271395", "0.726808", "0.72665006", "0.72650415", "0.7258541", "0.72547513", "0.72546786", "0.72532487", "0.72522366", "0.7251542", "0.7251538", "0.7244716", "0.7242998", "0.7241213", "0.723629", "0.7233941", "0.7224986", "0.72224236", "0.72100556", "0.71900123", "0.71886843", "0.7187992", "0.7178721", "0.7163131", "0.7162738", "0.7161049", "0.71544105", "0.71529245", "0.7150345", "0.714849", "0.7142324", "0.71417516", "0.71402484", "0.7132186", "0.7130318", "0.7129706", "0.71278185", "0.7120645", "0.711956", "0.7118409", "0.71158063", "0.7104653", "0.70997673", "0.7091966", "0.7089386", "0.70615166", "0.7057202", "0.70568305", "0.7047887", "0.704365", "0.7029594", "0.7027116", "0.70244396", "0.7023988", "0.70219636", "0.70130634", "0.70129144", "0.70088404" ]
0.0
-1
ReloadAll refetches every row with matching primary key column values and overwrites the original object slice with the newly updated slice.
func (o *AssetRevisionSlice) ReloadAll(exec boil.Executor) error { if o == nil || len(*o) == 0 { return nil } assetRevisions := AssetRevisionSlice{} var args []interface{} for _, obj := range *o { pkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), assetRevisionPrimaryKeyMapping) args = append(args, pkeyArgs...) } query := fmt.Sprintf( "SELECT \"asset_revision\".* FROM \"asset_revision\" WHERE (%s) IN (%s)", strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, assetRevisionPrimaryKeyColumns), ","), strmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(assetRevisionPrimaryKeyColumns), 1, len(assetRevisionPrimaryKeyColumns)), ) q := queries.Raw(exec, query, args...) err := q.Bind(&assetRevisions) if err != nil { return errors.Wrap(err, "public: unable to reload all in AssetRevisionSlice") } *o = assetRevisions return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *HoldenAtSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := HoldenAtSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), holdenAtPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"HoldenAt\\\".* FROM \\\"HoldenAt\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, holdenAtPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in HoldenAtSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFTurntableSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFTurntableSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfTurntablePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_turntable`.* FROM `cmf_turntable` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfTurntablePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFTurntableSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PaymentObjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PaymentObjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), paymentObjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `payment_objects`.* FROM `payment_objects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, paymentObjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PaymentObjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeLipidSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeLipidSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeLipidPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_lipid\\\".* FROM \\\"recipe_lipid\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeLipidPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeLipidSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CustomerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CustomerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), customerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"customers\\\".* FROM \\\"customers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, customerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CustomerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrentChartDataMinutelySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrentChartDataMinutelySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currentChartDataMinutelyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"current_chart_data_minutely\\\".* FROM \\\"current_chart_data_minutely\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currentChartDataMinutelyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrentChartDataMinutelySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumSlotSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumSlotSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumSlotPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_slots\\\".* FROM \\\"premium_slots\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumSlotPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumSlotSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vote\\\".* FROM \\\"vote\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *KvstoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := KvstoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), kvstorePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"kvstore\\\".* FROM \\\"kvstore\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, kvstorePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in KvstoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *APIKeySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tapiKeys := APIKeySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), apiKeyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"api_keys\\\".* FROM \\\"api_keys\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, apiKeyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&apiKeys)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in APIKeySlice\")\n\t}\n\n\t*o = apiKeys\n\n\treturn nil\n}", "func (o *TicketSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TicketSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), ticketPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tickets\\\".* FROM \\\"tickets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, ticketPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TicketSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RentalSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\trentals := RentalSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `rental`.* FROM `rental` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, rentalPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&rentals)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in RentalSlice\")\n\t}\n\n\t*o = rentals\n\n\treturn nil\n}", "func (o *EmployeeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmployeeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), employeePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"employee\\\".* FROM \\\"employee\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, employeePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in EmployeeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *EmailSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := EmailSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), emailPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `email`.* FROM `email` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, emailPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mysql: unable to reload all in EmailSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *FriendshipSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := FriendshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), friendshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `friendship`.* FROM `friendship` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, friendshipPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in FriendshipSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TenantSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TenantSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), tenantPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `tenants`.* FROM `tenants` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, tenantPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to reload all in TenantSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauths := AuthSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"auths\\\".* FROM \\\"auths\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&auths)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthSlice\")\n\t}\n\n\t*o = auths\n\n\treturn nil\n}", "func (o *PostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), postPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"posts\\\".* FROM \\\"posts\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, postPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"orm: unable to reload all in PostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *IllnessSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := IllnessSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), illnessPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"illness\\\".* FROM \\\"illness\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, illnessPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in IllnessSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"items\\\".* FROM \\\"items\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PeerSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `peers`.* FROM `peers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, peerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in PeerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WeatherSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WeatherSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), weatherPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"prh\\\".\\\"weather\\\".* FROM \\\"prh\\\".\\\"weather\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, weatherPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to reload all in WeatherSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *OfferSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OfferSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), offerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"offers\\\".* FROM \\\"offers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, offerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to reload all in OfferSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CurrencySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CurrencySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), currencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"currency\\\".* FROM \\\"currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, currencyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CurrencySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UtxoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UtxoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), utxoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"utxo\\\".* FROM \\\"utxo\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, utxoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UtxoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ForeignLegalResourceSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ForeignLegalResourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), foreignLegalResourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"ForeignLegalResources\\\".* FROM \\\"ForeignLegalResources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, foreignLegalResourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ForeignLegalResourceSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecipeAdditiveSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RecipeAdditiveSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recipeAdditivePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"recipe_additive\\\".* FROM \\\"recipe_additive\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, recipeAdditivePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RecipeAdditiveSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PremiumCodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PremiumCodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), premiumCodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"premium_codes\\\".* FROM \\\"premium_codes\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, premiumCodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PremiumCodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ShelfSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tshelves := ShelfSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), shelfPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `shelf`.* FROM `shelf` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, shelfPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(shelfPrimaryKeyColumns), 1, len(shelfPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&shelves)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ShelfSlice\")\n\t}\n\n\t*o = shelves\n\n\treturn nil\n}", "func (o *TaskSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TaskSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), taskPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"tasks\\\".* FROM \\\"tasks\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, taskPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TaskSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JobSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := JobSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jobPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"jobs\\\".* FROM \\\"jobs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, jobPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JobSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthorSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthors := AuthorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"authors\\\".* FROM \\\"authors\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, authorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authors)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in AuthorSlice\")\n\t}\n\n\t*o = authors\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transactions`.* FROM `transactions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *DocSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DocSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), docPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `doc`.* FROM `doc` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, docPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DocSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TransactionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\ttransactions := TransactionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), transactionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `transaction`.* FROM `transaction` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, transactionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&transactions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in TransactionSlice\")\n\t}\n\n\t*o = transactions\n\n\treturn nil\n}", "func (o *BlockSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BlockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"block\\\".* FROM \\\"block\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlockSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VoteSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tvotes := VoteSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), votePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vote`.* FROM `vote` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, votePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&votes)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VoteSlice\")\n\t}\n\n\t*o = votes\n\n\treturn nil\n}", "func (o *RawVisitSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RawVisitSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rawVisitPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"raw_visits\\\".* FROM \\\"raw_visits\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rawVisitPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RawVisitSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ChatSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChatSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), chatPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"chat\\\".* FROM \\\"chat\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, chatPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dal: unable to reload all in ChatSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SmallblogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SmallblogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), smallblogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `smallblog`.* FROM `smallblog` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, smallblogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SmallblogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VSPSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VSPSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vspPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"vsp\\\".* FROM \\\"vsp\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, vspPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VSPSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tstocks := StockSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `stock`.* FROM `stock` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, stockPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&stocks)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StockSlice\")\n\t}\n\n\t*o = stocks\n\n\treturn nil\n}", "func (o *RentalRowerSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RentalRowerSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rentalRowerPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rental_rowers\\\".* FROM \\\"rental_rowers\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rentalRowerPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RentalRowerSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *WithdrawalCryptoSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := WithdrawalCryptoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), withdrawalCryptoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"withdrawal_crypto\\\".* FROM \\\"withdrawal_crypto\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, withdrawalCryptoPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlite3: unable to reload all in WithdrawalCryptoSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BuildingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BuildingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), buildingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"buildings\\\".* FROM \\\"buildings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, buildingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"record: unable to reload all in BuildingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SegmentSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SegmentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), segmentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"segment\\\".* FROM \\\"segment\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, segmentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"boiler: unable to reload all in SegmentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AutomodRuleDatumSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := AutomodRuleDatumSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), automodRuleDatumPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"automod_rule_data\\\".* FROM \\\"automod_rule_data\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, automodRuleDatumPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AutomodRuleDatumSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BoardSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards`.* FROM `boards` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *JetSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tjets := JetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), jetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `jets`.* FROM `jets` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, jetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&jets)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in JetSlice\")\n\t}\n\n\t*o = jets\n\n\treturn nil\n}", "func (o *RepositorySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RepositorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), repositoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `repositories`.* FROM `repositories` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, repositoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RepositorySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PeerPropertySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PeerPropertySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), peerPropertyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"peer_properties\\\".* FROM \\\"peer_properties\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, peerPropertyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PeerPropertySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ChannelSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ChannelSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), channelPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"channels\\\".* FROM \\\"channels\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, channelPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ChannelSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SourceSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tsources := SourceSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), sourcePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"sources\\\".* FROM \\\"sources\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, sourcePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&sources)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmdbmodels: unable to reload all in SourceSlice\")\n\t}\n\n\t*o = sources\n\n\treturn nil\n}", "func (o *NotificationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NotificationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), notificationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"notification\\\".* FROM \\\"notification\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, notificationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NotificationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *OriginSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OriginSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), originPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"origins\\\".* FROM \\\"origins\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, originPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in OriginSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUsers := AuthUserSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user`.* FROM `auth_user` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserPrimaryKeyColumns), 1, len(authUserPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUsers)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserSlice\")\n\t}\n\n\t*o = authUsers\n\n\treturn nil\n}", "func (o *CMFFamilyUserPoliciesTakeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamilyUserPoliciesTakeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamilyUserPoliciesTakePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_family_user_policies_take`.* FROM `cmf_family_user_policies_take` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamilyUserPoliciesTakePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamilyUserPoliciesTakeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *DescriptionSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := DescriptionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), descriptionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `descriptions`.* FROM `descriptions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, descriptionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in DescriptionSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *VendorSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := VendorSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), vendorPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `vendors`.* FROM `vendors` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, vendorPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in VendorSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ClaimInListSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ClaimInListSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), claimInListPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `claim_in_list`.* FROM `claim_in_list` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, claimInListPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model: unable to reload all in ClaimInListSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UsernameListingSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UsernameListingSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), usernameListingPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"username_listings\\\".* FROM \\\"username_listings\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, usernameListingPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in UsernameListingSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SubscriberSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := SubscriberSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), subscriberPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `subscribers`.* FROM `subscribers` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, subscriberPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SubscriberSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BlackCardSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tblackCards := BlackCardSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), blackCardPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"black_cards\\\".* FROM \\\"black_cards\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, blackCardPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&blackCards)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BlackCardSlice\")\n\t}\n\n\t*o = blackCards\n\n\treturn nil\n}", "func (o *AddressSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\taddresses := AddressSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), addressPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `address`.* FROM `address` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, addressPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&addresses)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"sqlboiler: unable to reload all in AddressSlice\")\n\t}\n\n\t*o = addresses\n\n\treturn nil\n}", "func (o *FeatureRelationshipSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tfeatureRelationships := FeatureRelationshipSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), featureRelationshipPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"feature_relationship\\\".* FROM \\\"feature_relationship\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, featureRelationshipPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(featureRelationshipPrimaryKeyColumns), 1, len(featureRelationshipPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&featureRelationships)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in FeatureRelationshipSlice\")\n\t}\n\n\t*o = featureRelationships\n\n\treturn nil\n}", "func (o *OauthClientSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := OauthClientSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), oauthClientPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `oauth_clients`.* FROM `oauth_clients` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, oauthClientPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in OauthClientSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PictureSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tpictures := PictureSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), picturePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `pictures`.* FROM `pictures` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, picturePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&pictures)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PictureSlice\")\n\t}\n\n\t*o = pictures\n\n\treturn nil\n}", "func (o *PublisherSearchIdxSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := PublisherSearchIdxSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), publisherSearchIdxPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"publisher_search_idx\\\".* FROM \\\"publisher_search_idx\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, publisherSearchIdxPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in PublisherSearchIdxSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *InventorySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tinventories := InventorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), inventoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `inventory`.* FROM `inventory` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, inventoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&inventories)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in InventorySlice\")\n\t}\n\n\t*o = inventories\n\n\treturn nil\n}", "func (o *CMFUserSuperSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFUserSuperSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserSuperPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_user_super`.* FROM `cmf_user_super` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserSuperPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFUserSuperSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *TrainingCostSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := TrainingCostSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), trainingCostPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"training_costs\\\".* FROM \\\"training_costs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, trainingCostPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in TrainingCostSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ContentUnitDerivationSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ContentUnitDerivationSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), contentUnitDerivationPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"content_unit_derivations\\\".* FROM \\\"content_unit_derivations\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, contentUnitDerivationPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"mdbmodels: unable to reload all in ContentUnitDerivationSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RowerGroupSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RowerGroupSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rowerGroupPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rower_group\\\".* FROM \\\"rower_group\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rowerGroupPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RowerGroupSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RestaurantRankSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RestaurantRankSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), restaurantRankPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"restaurant_rank\\\".* FROM \\\"restaurant_rank\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, restaurantRankPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RestaurantRankSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *MultiAddressesSetSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := MultiAddressesSetSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), multiAddressesSetPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"multi_addresses_sets\\\".* FROM \\\"multi_addresses_sets\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, multiAddressesSetPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in MultiAddressesSetSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ActivityLogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ActivityLogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), activityLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"activity_logs\\\".* FROM \\\"activity_logs\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, activityLogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"dbmodel: unable to reload all in ActivityLogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BookCategoryAssignSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BookCategoryAssignSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), bookCategoryAssignPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `book_category_assign`.* FROM `book_category_assign` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, bookCategoryAssignPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BookCategoryAssignSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ScheduleSubjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ScheduleSubjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), scheduleSubjectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"schedule_subject\\\".* FROM \\\"schedule_subject\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, scheduleSubjectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ScheduleSubjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFSlideItemSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFSlideItemSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfSlideItemPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_slide_item`.* FROM `cmf_slide_item` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfSlideItemPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFSlideItemSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CommentSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tcomments := CommentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), commentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `comment`.* FROM `comment` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, commentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&comments)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CommentSlice\")\n\t}\n\n\t*o = comments\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `project`.* FROM `project` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *ProjectSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ProjectSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), projectPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `projects`.* FROM `projects` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, projectPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ProjectSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RSSAnnouncementSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := RSSAnnouncementSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), rssAnnouncementPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"rss_announcements\\\".* FROM \\\"rss_announcements\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, rssAnnouncementPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in RSSAnnouncementSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StockKeepingUnitContentSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := StockKeepingUnitContentSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), stockKeepingUnitContentPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"stock_keeping_unit_content\\\".* FROM \\\"stock_keeping_unit_content\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, stockKeepingUnitContentPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"db: unable to reload all in StockKeepingUnitContentSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StoreSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := StoreSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"stores\\\".* FROM \\\"stores\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, storePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in StoreSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFFamiliesPolicySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFFamiliesPolicySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfFamiliesPolicyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_families_policies`.* FROM `cmf_families_policies` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfFamiliesPolicyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFFamiliesPolicySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *CMFBalanceChargeAdminSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFBalanceChargeAdminSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfBalanceChargeAdminPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_balance_charge_admin`.* FROM `cmf_balance_charge_admin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfBalanceChargeAdminPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFBalanceChargeAdminSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *RecordMeasureSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\trecordMeasures := RecordMeasureSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), recordMeasurePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tquery := fmt.Sprintf(\n\t\t\"SELECT \\\"record_measures\\\".* FROM \\\"record_measures\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, recordMeasurePrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(recordMeasurePrimaryKeyColumns), 1, len(recordMeasurePrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, query, args...)\n\n\terr := q.Bind(&recordMeasures)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to reload all in RecordMeasureSlice\")\n\t}\n\n\t*o = recordMeasures\n\n\treturn nil\n}", "func (o *ExchangeCurrencySlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ExchangeCurrencySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), exchangeCurrencyPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"exchange_currency\\\".* FROM \\\"exchange_currency\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, exchangeCurrencyPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ExchangeCurrencySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BraceletPhotoSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tbraceletPhotos := BraceletPhotoSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), braceletPhotoPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `bracelet_photo`.* FROM `bracelet_photo` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, braceletPhotoPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(braceletPhotoPrimaryKeyColumns), 1, len(braceletPhotoPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&braceletPhotos)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BraceletPhotoSlice\")\n\t}\n\n\t*o = braceletPhotos\n\n\treturn nil\n}", "func (o *CMFUserExperienceLogSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := CMFUserExperienceLogSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), cmfUserExperienceLogPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `cmf_user_experience_log`.* FROM `cmf_user_experience_log` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, cmfUserExperienceLogPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in CMFUserExperienceLogSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *PhenotypepropSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tphenotypeprops := PhenotypepropSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), phenotypepropPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT \\\"phenotypeprop\\\".* FROM \\\"phenotypeprop\\\" WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, phenotypepropPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(phenotypepropPrimaryKeyColumns), 1, len(phenotypepropPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&phenotypeprops)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"chado: unable to reload all in PhenotypepropSlice\")\n\t}\n\n\t*o = phenotypeprops\n\n\treturn nil\n}", "func (o *BookCategorySlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BookCategorySlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), bookCategoryPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `book_category`.* FROM `book_category` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, bookCategoryPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in BookCategorySlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *UserGoogleSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := UserGoogleSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), userGooglePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `user_google`.* FROM `user_google` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, userGooglePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"model2: unable to reload all in UserGoogleSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *AuthUserUserPermissionSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tauthUserUserPermissions := AuthUserUserPermissionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), authUserUserPermissionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := fmt.Sprintf(\n\t\t\"SELECT `auth_user_user_permissions`.* FROM `auth_user_user_permissions` WHERE (%s) IN (%s)\",\n\t\tstrings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, authUserUserPermissionPrimaryKeyColumns), \",\"),\n\t\tstrmangle.Placeholders(dialect.IndexPlaceholders, len(*o)*len(authUserUserPermissionPrimaryKeyColumns), 1, len(authUserUserPermissionPrimaryKeyColumns)),\n\t)\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&authUserUserPermissions)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in AuthUserUserPermissionSlice\")\n\t}\n\n\t*o = authUserUserPermissions\n\n\treturn nil\n}", "func (o *ItemSideSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := ItemSideSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), itemSidePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"item_sides\\\".* FROM \\\"item_sides\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, itemSidePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in ItemSideSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *NodeSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := NodeSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), nodePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"node\\\".* FROM \\\"node\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, nodePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in NodeSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *BoardsSectionsPositionSlice) ReloadAll(ctx context.Context, exec boil.ContextExecutor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := BoardsSectionsPositionSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), boardsSectionsPositionPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `boards_sections_positions`.* FROM `boards_sections_positions` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, boardsSectionsPositionPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(ctx, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"rdb: unable to reload all in BoardsSectionsPositionSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *StorestateSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tslice := StorestateSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), storestatePrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT \\\"storestate\\\".* FROM \\\"storestate\\\" WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 1, storestatePrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(sql, args...)\n\n\terr := q.Bind(nil, exec, &slice)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"stellarcore: unable to reload all in StorestateSlice\")\n\t}\n\n\t*o = slice\n\n\treturn nil\n}", "func (o *SkinSlice) ReloadAll(exec boil.Executor) error {\n\tif o == nil || len(*o) == 0 {\n\t\treturn nil\n\t}\n\n\tskins := SkinSlice{}\n\tvar args []interface{}\n\tfor _, obj := range *o {\n\t\tpkeyArgs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(obj)), skinPrimaryKeyMapping)\n\t\targs = append(args, pkeyArgs...)\n\t}\n\n\tsql := \"SELECT `skin`.* FROM `skin` WHERE \" +\n\t\tstrmangle.WhereClauseRepeated(string(dialect.LQ), string(dialect.RQ), 0, skinPrimaryKeyColumns, len(*o))\n\n\tq := queries.Raw(exec, sql, args...)\n\n\terr := q.Bind(&skins)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"models: unable to reload all in SkinSlice\")\n\t}\n\n\t*o = skins\n\n\treturn nil\n}" ]
[ "0.8102668", "0.801014", "0.7990356", "0.7936499", "0.79038185", "0.78607833", "0.7850483", "0.7832766", "0.7829125", "0.7812598", "0.7791408", "0.7788089", "0.77733654", "0.77730197", "0.7765408", "0.77640593", "0.7761979", "0.7739838", "0.77378994", "0.7728349", "0.7709511", "0.7705431", "0.7704801", "0.7703807", "0.770121", "0.77003974", "0.7699225", "0.76892775", "0.76861596", "0.7677904", "0.76756585", "0.76755095", "0.76692456", "0.7658582", "0.7656212", "0.7650345", "0.7648034", "0.7645144", "0.7631876", "0.7629506", "0.76281244", "0.7626196", "0.7618208", "0.7617872", "0.7611633", "0.76113224", "0.7609315", "0.7595481", "0.75836927", "0.75783503", "0.7576096", "0.7569899", "0.7568418", "0.7563061", "0.7562207", "0.75598186", "0.7558134", "0.7551896", "0.75496787", "0.7541688", "0.7534716", "0.7514477", "0.7512702", "0.75059885", "0.7501661", "0.75006324", "0.74967897", "0.74896747", "0.74819535", "0.74716705", "0.746254", "0.74574846", "0.74542755", "0.7450195", "0.7447768", "0.7443383", "0.74369556", "0.7427738", "0.74219155", "0.74168545", "0.7416244", "0.7412284", "0.7408835", "0.7408212", "0.7397679", "0.73830116", "0.73757064", "0.73678553", "0.7364207", "0.73634744", "0.7361958", "0.73563427", "0.7341806", "0.73313665", "0.732919", "0.73228353", "0.7318306", "0.73042095", "0.7295867", "0.7295511", "0.7285546" ]
0.0
-1
AssetRevisionExists checks if the AssetRevision row exists.
func AssetRevisionExists(exec boil.Executor, orig int) (bool, error) { var exists bool query := "select exists(select 1 from \"asset_revision\" where \"orig\"=$1 limit 1)" if boil.DebugMode { fmt.Fprintln(boil.DebugWriter, query) fmt.Fprintln(boil.DebugWriter, orig) } row := exec.QueryRow(query, orig) err := row.Scan(&exists) if err != nil { return false, errors.Wrap(err, "public: unable to check if asset_revision exists") } return exists, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (q assetRevisionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: failed to check if asset_revision exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func AssetRevisionExistsP(exec boil.Executor, orig int) bool {\n\te, err := AssetRevisionExists(exec, orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func AssetRevisionExistsGP(orig int) bool {\n\te, err := AssetRevisionExists(boil.GetDB(), orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func AssetRevisionExistsG(orig int) (bool, error) {\n\treturn AssetRevisionExists(boil.GetDB(), orig)\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (asset Asset) Exists() bool {\n\treturn asset.model.AssetExists(asset.Name)\n}", "func (model *GrogModel) AssetExists(assetName string) bool {\n\tdoesExist := false\n\n\trow := model.db.DB.QueryRow(\"select count(1) from Assets where name = ?\", assetName)\n\tvar count int\n\trow.Scan(&count)\n\tif count > 0 {\n\t\tdoesExist = true\n\t}\n\n\treturn doesExist\n}", "func (q assetQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if asset exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func FindAssetRevision(exec boil.Executor, orig int, selectCols ...string) (*AssetRevision, error) {\n\tassetRevisionObj := &AssetRevision{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"asset_revision\\\" where \\\"orig\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(exec, query, orig)\n\n\terr := q.Bind(assetRevisionObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"public: unable to select from asset_revision\")\n\t}\n\n\treturn assetRevisionObj, nil\n}", "func (c *MyAssetContract) MyAssetExists(ctx contractapi.TransactionContextInterface, myAssetID string) (bool, error) {\n\tdata, err := ctx.GetStub().GetState(myAssetID)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn data != nil, nil\n}", "func (o *ReportingTaskEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func AuthorExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"authors\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if authors exists\")\n\t}\n\n\treturn exists, nil\n}", "func (version *Version) CheckExists() (bool, error)", "func (a *Analysis) Exists() bool {\n\treturn a.AssetExists(a.assetType)\n}", "func ReleaseExists(actionConfig action.Configuration, name string) (bool, error) {\n\thistClient := action.NewHistory(&actionConfig)\n\thistClient.Max = 1\n\n\t_, err := histClient.Run(name)\n\tif err != nil && err != driver.ErrReleaseNotFound {\n\t\treturn false, fmt.Errorf(\"failed checking for chart history: %w\", err)\n\t}\n\n\treturn err != driver.ErrReleaseNotFound, nil\n}", "func HasRevision(ctx context.Context) bool {\n\tsv, ok := ctx.Value(ctxKeyRevision).(string)\n\n\treturn ok && sv != \"\"\n}", "func (q assetRevisionQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (o *ControllerConfigurationEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func TransactionExists(exec boil.Executor, transactionID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transactions` where `transaction_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, transactionID)\n\t}\n\n\trow := exec.QueryRow(sql, transactionID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if transactions exists\")\n\t}\n\n\treturn exists, nil\n}", "func ActivityLogExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"activity_logs\\\" where \\\"id\\\"=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: unable to check if activity_logs exists\")\n\t}\n\n\treturn exists, nil\n}", "func (t *Transcript) Exists() bool {\n\treturn t.AssetExists(whodunit.AssetTypeTranscript)\n}", "func TransactionExists(exec boil.Executor, id uint64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transaction` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"model: unable to check if transaction exists\")\n\t}\n\n\treturn exists, nil\n}", "func (v *VersionFile) Exists() (bool, error) {\n\treturn exists(v.path)\n}", "func InventoryExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `inventory` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if inventory exists\")\n\t}\n\n\treturn exists, nil\n}", "func (br *builderRev) snapshotExists() (ok bool) {\n\t// If we already know this snapshot exists, don't check again.\n\tsnapshotExistsCache.Lock()\n\texists := snapshotExistsCache.m[*br]\n\tsnapshotExistsCache.Unlock()\n\tif exists {\n\t\treturn true\n\t}\n\n\t// When the function exits, cache an affirmative result.\n\tdefer func() {\n\t\tif ok {\n\t\t\tsnapshotExistsCache.Lock()\n\t\t\tsnapshotExistsCache.m[*br] = true\n\t\t\tsnapshotExistsCache.Unlock()\n\t\t}\n\t}()\n\n\tresp, err := http.Head(br.snapshotURL())\n\tif err != nil || resp.StatusCode != http.StatusOK {\n\t\treturn false\n\t}\n\n\t// If the snapshot is newer than the point at which we fixed writing\n\t// potentially-truncated snapshots to GCS, then stop right here.\n\t// See history in golang.org/issue/12671\n\tlastMod, err := http.ParseTime(resp.Header.Get(\"Last-Modified\"))\n\tif err == nil && lastMod.After(timeSnapshotCorruptionFixed) {\n\t\tlog.Printf(\"Snapshot exists for %v (without truncate checks)\", br)\n\t\treturn true\n\t}\n\n\t// Otherwise, if the snapshot is too old, verify it.\n\t// This path is slow.\n\t// TODO(bradfitz): delete this in 6 months or so? (around 2016-06-01)\n\tresp, err = http.Get(br.snapshotURL())\n\tif err != nil {\n\t\treturn false\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\treturn false\n\t}\n\t// Verify the .tar.gz is valid.\n\tgz, err := gzip.NewReader(resp.Body)\n\tif err != nil {\n\t\tlog.Printf(\"corrupt snapshot? %s gzip.NewReader: %v\", br.snapshotURL(), err)\n\t\treturn false\n\t}\n\ttr := tar.NewReader(gz)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s tar.Next: %v\", br.snapshotURL(), err)\n\t\t\treturn false\n\t\t}\n\t\tif _, err := io.Copy(ioutil.Discard, tr); err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s reading contents of %s: %v\", br.snapshotURL(), hdr.Name, err)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn gz.Close() == nil\n}", "func (o *NetworkElementSummaryAllOf) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *pgRepository) Exists(ctx context.Context, tenantID, id string) (bool, error) {\n\treturn r.existQuerier.Exists(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (i *IndexBuilder) exists(metadata *helmchart.Metadata) bool {\n\tfor m := range i.metadataCommit {\n\t\tif metadata.Name == m.Name && metadata.Version == m.Version {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionNo() bool {\n\tif o != nil && o.RevisionNo != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (q assetRevisionQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"public: failed to count asset_revision rows\")\n\t}\n\n\treturn count, nil\n}", "func (q activityLogQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: failed to check if activity_logs exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func FindAssetRevisionP(exec boil.Executor, orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(exec, orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (v *Version) IsRevision() bool {\n\treturn v.isRevision\n}", "func StockCvtermExists(exec boil.Executor, stockCvtermID int) (bool, error) {\n\tvar exists bool\n\n\tsql := \"select exists(select 1 from \\\"stock_cvterm\\\" where \\\"stock_cvterm_id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockCvtermID)\n\t}\n\n\trow := exec.QueryRow(sql, stockCvtermID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: unable to check if stock_cvterm exists\")\n\t}\n\n\treturn exists, nil\n}", "func (vrsConnection *VRSConnection) CheckEntityExists(id string) (bool, error) {\n\treadRowArgs := ovsdb.ReadRowArgs{\n\t\tCondition: []string{ovsdb.NuageVMTableColumnVMUUID, \"==\", id},\n\t\tColumns: []string{ovsdb.NuageVMTableColumnVMUUID},\n\t}\n\n\tvar idRows []map[string]interface{}\n\tvar err error\n\tif idRows, err = vrsConnection.vmTable.ReadRows(vrsConnection.ovsdbClient, readRowArgs); err != nil {\n\t\treturn false, fmt.Errorf(\"OVSDB read error %v\", err)\n\t}\n\n\tvar ids []string\n\tfor _, row := range idRows {\n\t\tids = append(ids, row[ovsdb.NuageVMTableColumnVMUUID].(string))\n\t}\n\n\tif len(ids) == 1 && id == ids[0] {\n\t\treturn true, err\n\t}\n\treturn false, err\n}", "func SourceExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func (db *DB) DoesArticleExist(id uint64) bool {\n\terr := db.View(func(txn *badger.Txn) error {\n\t\tarticle, err := txn.Get(helpers.I64tob(id))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = article.Value(func(val []byte) error {\n\t\t\treturn nil\n\t\t})\n\t\treturn err\n\t})\n\n\treturn err == nil\n}", "func (o *NormalizedProjectRevisionHook) HasProjectRevisionId() bool {\n\tif o != nil && o.ProjectRevisionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *ModuleStore) Exists(ctx context.Context, module, vsn string) bool {\n\tresult := models.Module{}\n\tquery := r.conn.Where(\"module = ?\", module).Where(\"version = ?\", vsn)\n\tcount, err := query.Count(&result)\n\n\treturn err == nil && count > 0\n}", "func (q inventoryQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if inventory exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func SkinExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `skin` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if skin exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q vendorQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if vendors exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q rssAnnouncementQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if rss_announcements exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *ResourceVersion) HasResourceVersionId() bool {\n\tif o != nil && o.ResourceVersionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func StockExists(exec boil.Executor, stockID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `stock` where `stock_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockID)\n\t}\n\n\trow := exec.QueryRow(sql, stockID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if stock exists\")\n\t}\n\n\treturn exists, nil\n}", "func AssetAlreadyUploaded(hash string, url string) bool {\n\tfmt.Println(\"Checking for hash\", hash, \"and url\", url, \"on database\")\n\trows, err := database.Query(\"SELECT media_hash, media_url FROM media_info;\")\n\tif shared.CheckError(err) {\n\t\treturn false\n\t}\n\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tvar stored_hash, stored_url string\n\t\trows.Scan(&stored_hash, &stored_url)\n\n\t\tif hash == stored_hash || url == stored_url {\n\t\t\tfmt.Println(\"Database match found, aborting upload\")\n\t\t\treturn true\n\t\t}\n\t}\n\n\tfmt.Println(\"Coudln't find a match, proceding!\")\n\treturn false\n}", "func (sm *SourceMgr) RevisionPresentIn(id ProjectIdentifier, r Revision) (bool, error) {\n\tsrc, err := sm.getSourceFor(id)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn src.revisionPresentIn(r)\n}", "func (m *GormResourceRepository) CheckExists(ctx context.Context, id string) error {\n\tdefer goa.MeasureSince([]string{\"goa\", \"db\", \"resource\", \"exists\"}, time.Now())\n\treturn base.CheckExistsWithCustomIDColumn(ctx, m.db, m.TableName(), \"resource_id\", id)\n}", "func VoteExists(ctx context.Context, exec boil.ContextExecutor, hash string) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"vote\\\" where \\\"hash\\\"=$1 limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, hash)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, hash)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vote exists\")\n\t}\n\n\treturn exists, nil\n}", "func SourceExists(exec boil.Executor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\trow := exec.QueryRow(sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func (p *DeliveryConfigProcessor) ResourceExists(search *ExportableResource) bool {\n\tfor eix := range p.deliveryConfig.Environments {\n\t\trix := p.findResourceIndex(search, eix)\n\t\tif rix >= 0 {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func TagExists(projectID int64, repositoryID int64, version string) (bool, error) {\n\ttags, err := GetTagByVersion(projectID, repositoryID, version)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif tags == nil {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func VendorExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `vendors` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vendors exists\")\n\t}\n\n\treturn exists, nil\n}", "func RowExists(query string, args ...interface{}) bool {\n\tvar exists bool\n\tquery = fmt.Sprintf(\"SELECT exists (%s)\", query)\n\terr := Db.QueryRow(query, args...).Scan(&exists)\n\tif err != nil && err != sql.ErrNoRows {\n\t\tfmt.Errorf(\"error checking if row exists '%s' %v\", args, err)\n\t}\n\treturn exists\n}", "func (o *AssetRevision) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tassetRevisionInsertCacheMut.RLock()\n\tcache, cached := assetRevisionInsertCache[key]\n\tassetRevisionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"asset_revision\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"asset_revision\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to insert into asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionInsertCacheMut.Lock()\n\t\tassetRevisionInsertCache[key] = cache\n\t\tassetRevisionInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func HasAssets() predicate.GithubRelease {\n\treturn predicate.GithubRelease(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.O2M, false, AssetsTable, AssetsColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func (ar AlbumDbRepository) Exists(id int) bool {\n\t_, err := ar.Get(id)\n\treturn err == nil\n}", "func (i Item) Exists() bool {\n\tif i.UUID == \"\" {\n\t\treturn false\n\t}\n\tuuid, err := db.SelectFirst(\"SELECT `uuid` FROM `items` WHERE `uuid`=?\", i.UUID)\n\n\tif err != nil {\n\t\tLog(err)\n\t\treturn false\n\t}\n\tLog(\"Exists:\", uuid)\n\treturn uuid != \"\"\n}", "func (s *SpendJournalPruner) DependencyExists(blockHash *chainhash.Hash) bool {\n\ts.dependentsMtx.RLock()\n\tdefer s.dependentsMtx.RUnlock()\n\n\t_, ok := s.dependents[*blockHash]\n\treturn ok\n}", "func CommentExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `comment` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if comment exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q mempoolBinQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if mempool_bin exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (v *Version) Exists(ctx context.Context) bool {\n\treturn keyExists(ctx, v.Backend, v.Prefix)\n}", "func (i *Item) Exists() (bool, error) {\n\tif i.UUID == \"\" {\n\t\treturn false, nil\n\t}\n\tvar id string\n\treturn db.SelectExists(\n\t\t&id,\n\t\t\"SELECT uuid FROM items WHERE uuid=?\",\n\t\ti.UUID,\n\t)\n}", "func (q blockQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if block exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func EntryExists(url string) bool {\n\tresp, err := http.Head(url)\n\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\n\treturn resp.StatusCode == http.StatusOK\n}", "func HasRevision() predicate.Ref {\n\treturn predicate.Ref(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(RevisionTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, RevisionTable, RevisionColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func (q braceletPhotoQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if bracelet_photo exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (r *ClipRepository) Exists(clip *decryptor.Clip) bool {\n\tcPath := filepath.Join(r.Path, clip.Module.Course.ID, computeModuleHash(clip.Module), fmt.Sprintf(\"%v.psv\", clip.ID))\n\treturn file.Exists(cPath)\n}", "func (q holdenAtQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if HoldenAt exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q sourceQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (q authorQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: failed to check if authors exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q stockCvtermQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if stock_cvterm exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (in *PublicAccessBlockClient) SubresourceExists(cr *v1beta1.Bucket) bool {\n\treturn cr.Spec.ForProvider.PublicAccessBlockConfiguration != nil\n}", "func (q repositoryQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if repositories exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (m *InMemoryRepository) Exists(u fyne.URI) (bool, error) {\n\tpath := u.Path()\n\tif path == \"\" {\n\t\treturn false, fmt.Errorf(\"invalid path '%s'\", path)\n\t}\n\n\t_, ok := m.Data[path]\n\treturn ok, nil\n}", "func (q recipeAdditiveQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if recipe_additive exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (a *Activity) Exists() bool { //activity\n\treturn a._exists\n}", "func (sm *SourceMgr) RevisionPresentIn(n ProjectRoot, r Revision) (bool, error) {\n\tpmc, err := sm.getProjectManager(n)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn pmc.pm.RevisionPresentIn(r)\n}", "func RSSAnnouncementExists(ctx context.Context, exec boil.ContextExecutor, guildID int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"rss_announcements\\\" where \\\"guild_id\\\"=$1 limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, guildID)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, guildID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if rss_announcements exists\")\n\t}\n\n\treturn exists, nil\n}", "func (o *Service) HasServiceAssetId() bool {\n\tif o != nil && o.ServiceAssetId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (fh *FilesystemHandler) BlobExists(container models.SimpleContainer, blobName string) (bool, error) {\n\n\tif blobName[0] == os.PathSeparator {\n\t\tblobName = blobName[1:]\n\t}\n\n\tfullPath := fh.generateFullPath(&container) + blobName\n\n\tlog.Printf(\"FH %s\", fullPath)\n\n\treturn false, nil\n}", "func (q featureCvtermDbxrefQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if feature_cvterm_dbxref exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q sourceQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionComment() bool {\n\tif o != nil && o.RevisionComment != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *TransactionResult) HasVersion() bool {\n\tif o != nil && o.Version.IsSet() {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func TenantExists(ctx context.Context, exec boil.ContextExecutor, iD string) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `tenants` where `id`=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: unable to check if tenants exists\")\n\t}\n\n\treturn exists, nil\n}", "func (l *levelDBRepo) Exists(tenantID, id []byte) bool {\n\tkey := getKey(tenantID, id)\n\tres, err := l.db.Has(key, nil)\n\t// TODO check this\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn res\n}", "func (q projectQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if project exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (i IndexFile) Has(name, version string) bool {\n\t_, err := i.Get(name, version)\n\treturn err == nil\n}", "func (xpss *XPipelineSalesStage) Exists() bool {\n\treturn xpss._exists\n}", "func (lir *LookupInResult) Exists(idx int) bool {\n\treturn lir.contents[idx].exists()\n}", "func (r *ClipRepository) Exists(clip *decryptor.Clip) (bool, error) {\n\tif clip == nil {\n\t\treturn false, ErrClipUndefined\n\t}\n\tif clip.Module == nil {\n\t\treturn false, ErrModuleUndefined\n\t}\n\tif clip.Module.Course == nil {\n\t\treturn false, ErrCourseUndefined\n\t}\n\tcPath := filepath.Join(r.Path, clip.Module.Course.ID, computeModuleHash(clip.Module), fmt.Sprintf(\"%v.psv\", clip.ID))\n\treturn r.FileExists(cPath), nil\n}", "func (q stockKeepingUnitContentQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"db: failed to check if stock_keeping_unit_content exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q transactionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"model: failed to check if transaction exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (p Plugin) Exists(ctx context.Context, module string, version string) (bool, error) {\n\tresp, err := p.c.Exists(ctx, &stpb.ExistsRequest{Module: module, Version: version})\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn resp.GetExists(), nil\n}", "func (q buildingQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"record: failed to check if buildings exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (snapshot *DataSnapshot) Exists() bool {\n\treturn bytes.Compare(snapshot.raw, []byte(\"null\")) != 0\n}", "func (q transactionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if transactions exists\")\n\t}\n\n\treturn count > 0, nil\n}" ]
[ "0.7504592", "0.7246322", "0.676757", "0.63569015", "0.6229071", "0.6220883", "0.6145029", "0.6115493", "0.6086223", "0.57564026", "0.55941087", "0.55532306", "0.5548378", "0.5517553", "0.5448903", "0.5428508", "0.5417559", "0.53198934", "0.52905136", "0.528684", "0.5232201", "0.5206095", "0.51660377", "0.51378715", "0.50999993", "0.50724316", "0.5070514", "0.50498104", "0.5037542", "0.5025905", "0.50124544", "0.49858078", "0.4975227", "0.49620548", "0.49345273", "0.49265584", "0.49259773", "0.49254325", "0.49233392", "0.49069887", "0.49064606", "0.49041617", "0.49029022", "0.49009523", "0.4893595", "0.48738408", "0.48655978", "0.4862478", "0.48620358", "0.48591107", "0.48517522", "0.48500982", "0.484606", "0.48421866", "0.484062", "0.4836653", "0.48286426", "0.4827159", "0.48264846", "0.48254353", "0.48252687", "0.48200938", "0.47957373", "0.47750872", "0.47643703", "0.47524244", "0.47401768", "0.4739672", "0.47338235", "0.4727236", "0.4722885", "0.472024", "0.47136697", "0.47019786", "0.47011244", "0.46995872", "0.46890333", "0.46889365", "0.46867546", "0.4682214", "0.46770495", "0.46746475", "0.4669614", "0.46582156", "0.46432438", "0.4634942", "0.46320662", "0.4630673", "0.46295032", "0.4625229", "0.46200758", "0.461382", "0.46124294", "0.46100742", "0.46067736", "0.46046665", "0.46045062", "0.4600504", "0.4592249", "0.45891356" ]
0.8474628
0
AssetRevisionExistsG checks if the AssetRevision row exists.
func AssetRevisionExistsG(orig int) (bool, error) { return AssetRevisionExists(boil.GetDB(), orig) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func AssetRevisionExists(exec boil.Executor, orig int) (bool, error) {\n\tvar exists bool\n\n\tquery := \"select exists(select 1 from \\\"asset_revision\\\" where \\\"orig\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, orig)\n\t}\n\n\trow := exec.QueryRow(query, orig)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: unable to check if asset_revision exists\")\n\t}\n\n\treturn exists, nil\n}", "func AssetRevisionExistsGP(orig int) bool {\n\te, err := AssetRevisionExists(boil.GetDB(), orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func AssetRevisionExistsP(exec boil.Executor, orig int) bool {\n\te, err := AssetRevisionExists(exec, orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q assetRevisionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: failed to check if asset_revision exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (model *GrogModel) AssetExists(assetName string) bool {\n\tdoesExist := false\n\n\trow := model.db.DB.QueryRow(\"select count(1) from Assets where name = ?\", assetName)\n\tvar count int\n\trow.Scan(&count)\n\tif count > 0 {\n\t\tdoesExist = true\n\t}\n\n\treturn doesExist\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (q assetQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if asset exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (asset Asset) Exists() bool {\n\treturn asset.model.AssetExists(asset.Name)\n}", "func (o *ReportingTaskEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func FindAssetRevisionG(orig int, selectCols ...string) (*AssetRevision, error) {\n\treturn FindAssetRevision(boil.GetDB(), orig, selectCols...)\n}", "func (q assetRevisionQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func FindAssetRevisionGP(orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(boil.GetDB(), orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func HasRevision(ctx context.Context) bool {\n\tsv, ok := ctx.Value(ctxKeyRevision).(string)\n\n\treturn ok && sv != \"\"\n}", "func (o *ControllerConfigurationEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func ReleaseExists(actionConfig action.Configuration, name string) (bool, error) {\n\thistClient := action.NewHistory(&actionConfig)\n\thistClient.Max = 1\n\n\t_, err := histClient.Run(name)\n\tif err != nil && err != driver.ErrReleaseNotFound {\n\t\treturn false, fmt.Errorf(\"failed checking for chart history: %w\", err)\n\t}\n\n\treturn err != driver.ErrReleaseNotFound, nil\n}", "func (c *MyAssetContract) MyAssetExists(ctx contractapi.TransactionContextInterface, myAssetID string) (bool, error) {\n\tdata, err := ctx.GetStub().GetState(myAssetID)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn data != nil, nil\n}", "func (br *builderRev) snapshotExists() (ok bool) {\n\t// If we already know this snapshot exists, don't check again.\n\tsnapshotExistsCache.Lock()\n\texists := snapshotExistsCache.m[*br]\n\tsnapshotExistsCache.Unlock()\n\tif exists {\n\t\treturn true\n\t}\n\n\t// When the function exits, cache an affirmative result.\n\tdefer func() {\n\t\tif ok {\n\t\t\tsnapshotExistsCache.Lock()\n\t\t\tsnapshotExistsCache.m[*br] = true\n\t\t\tsnapshotExistsCache.Unlock()\n\t\t}\n\t}()\n\n\tresp, err := http.Head(br.snapshotURL())\n\tif err != nil || resp.StatusCode != http.StatusOK {\n\t\treturn false\n\t}\n\n\t// If the snapshot is newer than the point at which we fixed writing\n\t// potentially-truncated snapshots to GCS, then stop right here.\n\t// See history in golang.org/issue/12671\n\tlastMod, err := http.ParseTime(resp.Header.Get(\"Last-Modified\"))\n\tif err == nil && lastMod.After(timeSnapshotCorruptionFixed) {\n\t\tlog.Printf(\"Snapshot exists for %v (without truncate checks)\", br)\n\t\treturn true\n\t}\n\n\t// Otherwise, if the snapshot is too old, verify it.\n\t// This path is slow.\n\t// TODO(bradfitz): delete this in 6 months or so? (around 2016-06-01)\n\tresp, err = http.Get(br.snapshotURL())\n\tif err != nil {\n\t\treturn false\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\treturn false\n\t}\n\t// Verify the .tar.gz is valid.\n\tgz, err := gzip.NewReader(resp.Body)\n\tif err != nil {\n\t\tlog.Printf(\"corrupt snapshot? %s gzip.NewReader: %v\", br.snapshotURL(), err)\n\t\treturn false\n\t}\n\ttr := tar.NewReader(gz)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s tar.Next: %v\", br.snapshotURL(), err)\n\t\t\treturn false\n\t\t}\n\t\tif _, err := io.Copy(ioutil.Discard, tr); err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s reading contents of %s: %v\", br.snapshotURL(), hdr.Name, err)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn gz.Close() == nil\n}", "func VendorExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `vendors` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vendors exists\")\n\t}\n\n\treturn exists, nil\n}", "func (version *Version) CheckExists() (bool, error)", "func FindAssetRevision(exec boil.Executor, orig int, selectCols ...string) (*AssetRevision, error) {\n\tassetRevisionObj := &AssetRevision{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"asset_revision\\\" where \\\"orig\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(exec, query, orig)\n\n\terr := q.Bind(assetRevisionObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"public: unable to select from asset_revision\")\n\t}\n\n\treturn assetRevisionObj, nil\n}", "func (o *NetworkElementSummaryAllOf) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (a *Analysis) Exists() bool {\n\treturn a.AssetExists(a.assetType)\n}", "func AuthorExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"authors\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if authors exists\")\n\t}\n\n\treturn exists, nil\n}", "func TransactionExists(exec boil.Executor, transactionID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transactions` where `transaction_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, transactionID)\n\t}\n\n\trow := exec.QueryRow(sql, transactionID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if transactions exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q assetRevisionQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"public: failed to count asset_revision rows\")\n\t}\n\n\treturn count, nil\n}", "func AuthorExistsG(id int64) (bool, error) {\n\treturn AuthorExists(boil.GetDB(), id)\n}", "func StockCvtermExists(exec boil.Executor, stockCvtermID int) (bool, error) {\n\tvar exists bool\n\n\tsql := \"select exists(select 1 from \\\"stock_cvterm\\\" where \\\"stock_cvterm_id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockCvtermID)\n\t}\n\n\trow := exec.QueryRow(sql, stockCvtermID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: unable to check if stock_cvterm exists\")\n\t}\n\n\treturn exists, nil\n}", "func (i *IndexBuilder) exists(metadata *helmchart.Metadata) bool {\n\tfor m := range i.metadataCommit {\n\t\tif metadata.Name == m.Name && metadata.Version == m.Version {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionNo() bool {\n\tif o != nil && o.RevisionNo != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func InventoryExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `inventory` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if inventory exists\")\n\t}\n\n\treturn exists, nil\n}", "func TransactionExistsG(transactionID int) (bool, error) {\n\treturn TransactionExists(boil.GetDB(), transactionID)\n}", "func TransactionExists(exec boil.Executor, id uint64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transaction` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"model: unable to check if transaction exists\")\n\t}\n\n\treturn exists, nil\n}", "func HasRevision() predicate.Ref {\n\treturn predicate.Ref(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.To(RevisionTable, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, RevisionTable, RevisionColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func (r *pgRepository) Exists(ctx context.Context, tenantID, id string) (bool, error) {\n\treturn r.existQuerier.Exists(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func ActivityLogExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"activity_logs\\\" where \\\"id\\\"=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: unable to check if activity_logs exists\")\n\t}\n\n\treturn exists, nil\n}", "func HasAssets() predicate.GithubRelease {\n\treturn predicate.GithubRelease(func(s *sql.Selector) {\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(Table, FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.O2M, false, AssetsTable, AssetsColumn),\n\t\t)\n\t\tsqlgraph.HasNeighbors(s, step)\n\t})\n}", "func SourceExists(exec boil.Executor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\trow := exec.QueryRow(sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func AssetAlreadyUploaded(hash string, url string) bool {\n\tfmt.Println(\"Checking for hash\", hash, \"and url\", url, \"on database\")\n\trows, err := database.Query(\"SELECT media_hash, media_url FROM media_info;\")\n\tif shared.CheckError(err) {\n\t\treturn false\n\t}\n\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tvar stored_hash, stored_url string\n\t\trows.Scan(&stored_hash, &stored_url)\n\n\t\tif hash == stored_hash || url == stored_url {\n\t\t\tfmt.Println(\"Database match found, aborting upload\")\n\t\t\treturn true\n\t\t}\n\t}\n\n\tfmt.Println(\"Coudln't find a match, proceding!\")\n\treturn false\n}", "func TransactionExistsG(id uint64) (bool, error) {\n\treturn TransactionExists(boil.GetDB(), id)\n}", "func AuthorExistsGP(id int64) bool {\n\te, err := AuthorExists(boil.GetDB(), id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q vendorQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if vendors exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func FindAssetRevisionP(exec boil.Executor, orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(exec, orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (r *ModuleStore) Exists(ctx context.Context, module, vsn string) bool {\n\tresult := models.Module{}\n\tquery := r.conn.Where(\"module = ?\", module).Where(\"version = ?\", vsn)\n\tcount, err := query.Count(&result)\n\n\treturn err == nil && count > 0\n}", "func RowExists(query string, args ...interface{}) bool {\n\tvar exists bool\n\tquery = fmt.Sprintf(\"SELECT exists (%s)\", query)\n\terr := Db.QueryRow(query, args...).Scan(&exists)\n\tif err != nil && err != sql.ErrNoRows {\n\t\tfmt.Errorf(\"error checking if row exists '%s' %v\", args, err)\n\t}\n\treturn exists\n}", "func StockExists(exec boil.Executor, stockID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `stock` where `stock_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockID)\n\t}\n\n\trow := exec.QueryRow(sql, stockID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if stock exists\")\n\t}\n\n\treturn exists, nil\n}", "func existsProject(gh_id int64) bool {\n\terr := db.QueryRow(\"SELECT gh_id FROM projects WHERE gh_id = $1\", gh_id).\n\t\tScan(&gh_id)\n\treturn err != sql.ErrNoRows\n}", "func (t *Transcript) Exists() bool {\n\treturn t.AssetExists(whodunit.AssetTypeTranscript)\n}", "func SourceExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func (o *VersionedFlowUpdateRequestEntity) HasProcessGroupRevision() bool {\n\tif o != nil && o.ProcessGroupRevision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (snapshot *DataSnapshot) Exists() bool {\n\treturn bytes.Compare(snapshot.raw, []byte(\"null\")) != 0\n}", "func (v *Version) IsRevision() bool {\n\treturn v.isRevision\n}", "func TagExists(projectID int64, repositoryID int64, version string) (bool, error) {\n\ttags, err := GetTagByVersion(projectID, repositoryID, version)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif tags == nil {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (q stockCvtermQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if stock_cvterm exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *NormalizedProjectRevisionHook) HasProjectRevisionId() bool {\n\tif o != nil && o.ProjectRevisionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func SkinExistsG(index int) (bool, error) {\n\treturn SkinExists(boil.GetDB(), index)\n}", "func cephRBDSnapshotExists(clusterName string, poolName string,\n\tvolumeName string, volumeType string, snapshotName string,\n\tuserName string) bool {\n\t_, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"info\",\n\t\tfmt.Sprintf(\"%s_%s@%s\", volumeType, volumeName, snapshotName))\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (q buildingQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"record: failed to check if buildings exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func CommentExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `comment` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if comment exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q inventoryQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if inventory exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (v *VersionFile) Exists() (bool, error) {\n\treturn exists(v.path)\n}", "func (q braceletPhotoQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if bracelet_photo exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func TransactionExistsGP(transactionID int) bool {\n\te, err := TransactionExists(boil.GetDB(), transactionID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (i Item) Exists() bool {\n\tif i.UUID == \"\" {\n\t\treturn false\n\t}\n\tuuid, err := db.SelectFirst(\"SELECT `uuid` FROM `items` WHERE `uuid`=?\", i.UUID)\n\n\tif err != nil {\n\t\tLog(err)\n\t\treturn false\n\t}\n\tLog(\"Exists:\", uuid)\n\treturn uuid != \"\"\n}", "func (q featureCvtermDbxrefQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if feature_cvterm_dbxref exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionComment() bool {\n\tif o != nil && o.RevisionComment != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func SkinExistsGP(index int) bool {\n\te, err := SkinExists(boil.GetDB(), index)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func StockCvtermExistsG(stockCvtermID int) (bool, error) {\n\treturn StockCvtermExists(boil.GetDB(), stockCvtermID)\n}", "func SkinExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `skin` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if skin exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q assetQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (ar AlbumDbRepository) Exists(id int) bool {\n\t_, err := ar.Get(id)\n\treturn err == nil\n}", "func InventoryExistsG(id int64) (bool, error) {\n\treturn InventoryExists(boil.GetDB(), id)\n}", "func (q repositoryQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if repositories exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (a *Activity) Exists() bool { //activity\n\treturn a._exists\n}", "func ItemExistsG(ctx context.Context, iD null.Int64) (bool, error) {\n\treturn ItemExists(ctx, boil.GetContextDB(), iD)\n}", "func BuildingExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"buildings\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"record: unable to check if buildings exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q blockQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if block exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q recipeAdditiveQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if recipe_additive exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (client *GCSBlobstore) Exists(dest string) (exists bool, err error) {\n\tif exists, err = client.exists(client.publicGCS, dest); err == nil {\n\t\treturn exists, nil\n\t}\n\n\t// If the public client fails, try using it as an authenticated actor\n\tif client.authenticatedGCS != nil {\n\t\treturn client.exists(client.authenticatedGCS, dest)\n\t}\n\n\treturn\n}", "func (sm *SourceMgr) RevisionPresentIn(id ProjectIdentifier, r Revision) (bool, error) {\n\tsrc, err := sm.getSourceFor(id)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn src.revisionPresentIn(r)\n}", "func (o *AssetRevision) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tassetRevisionInsertCacheMut.RLock()\n\tcache, cached := assetRevisionInsertCache[key]\n\tassetRevisionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"asset_revision\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"asset_revision\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to insert into asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionInsertCacheMut.Lock()\n\t\tassetRevisionInsertCache[key] = cache\n\t\tassetRevisionInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func ProjectExists(ctx context.Context, exec boil.ContextExecutor, iD uint) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `projects` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if projects exists\")\n\t}\n\n\treturn exists, nil\n}", "func (region *Region) Exists() bool {\n\tfor _, r := range AllRegions {\n\t\tif r == *region {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (v *Version) Exists(ctx context.Context) bool {\n\treturn keyExists(ctx, v.Backend, v.Prefix)\n}", "func (_Storage *StorageCaller) AccountExists(opts *bind.CallOpts, addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"accountExists\", addr)\n\treturn *ret0, err\n}", "func (q activityLogQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: failed to check if activity_logs exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (vrsConnection *VRSConnection) CheckEntityExists(id string) (bool, error) {\n\treadRowArgs := ovsdb.ReadRowArgs{\n\t\tCondition: []string{ovsdb.NuageVMTableColumnVMUUID, \"==\", id},\n\t\tColumns: []string{ovsdb.NuageVMTableColumnVMUUID},\n\t}\n\n\tvar idRows []map[string]interface{}\n\tvar err error\n\tif idRows, err = vrsConnection.vmTable.ReadRows(vrsConnection.ovsdbClient, readRowArgs); err != nil {\n\t\treturn false, fmt.Errorf(\"OVSDB read error %v\", err)\n\t}\n\n\tvar ids []string\n\tfor _, row := range idRows {\n\t\tids = append(ids, row[ovsdb.NuageVMTableColumnVMUUID].(string))\n\t}\n\n\tif len(ids) == 1 && id == ids[0] {\n\t\treturn true, err\n\t}\n\treturn false, err\n}", "func (sm *SourceMgr) RevisionPresentIn(n ProjectRoot, r Revision) (bool, error) {\n\tpmc, err := sm.getProjectManager(n)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn pmc.pm.RevisionPresentIn(r)\n}", "func (q rawVisitQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if raw_visits exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (l *levelDBRepo) Exists(tenantID, id []byte) bool {\n\tkey := getKey(tenantID, id)\n\tres, err := l.db.Has(key, nil)\n\t// TODO check this\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn res\n}", "func (q sourceQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (mvq *ModuleVersionQuery) Exist(ctx context.Context) (bool, error) {\n\treturn mvq.sqlExist(ctx)\n}", "func (m *InMemoryRepository) Exists(u fyne.URI) (bool, error) {\n\tpath := u.Path()\n\tif path == \"\" {\n\t\treturn false, fmt.Errorf(\"invalid path '%s'\", path)\n\t}\n\n\t_, ok := m.Data[path]\n\treturn ok, nil\n}", "func (_Contract *ContractCaller) Exists(opts *bind.CallOpts, _tokenId *big.Int) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Contract.contract.Call(opts, out, \"exists\", _tokenId)\n\treturn *ret0, err\n}", "func (p Plugin) Exists(ctx context.Context, module string, version string) (bool, error) {\n\tresp, err := p.c.Exists(ctx, &stpb.ExistsRequest{Module: module, Version: version})\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn resp.GetExists(), nil\n}", "func (q sourceQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func ProjectExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `project` where `ID`=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if project exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q stockKeepingUnitContentQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"db: failed to check if stock_keeping_unit_content exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func TransactionExistsGP(id uint64) bool {\n\te, err := TransactionExists(boil.GetDB(), id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func instanceExists(name string) bool {\n\tvar exists bool\n\terr := pool.QueryRow(\"SELECT EXISTS (SELECT FROM PROVISION WHERE name = $1)\", name).Scan(&exists)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\treturn exists\n}" ]
[ "0.8322645", "0.80098724", "0.74902403", "0.7193994", "0.6208003", "0.5885187", "0.58735114", "0.5861673", "0.5836511", "0.57838506", "0.5648545", "0.56017876", "0.55470455", "0.55416286", "0.54620993", "0.54539984", "0.54059815", "0.53788596", "0.5308358", "0.5303938", "0.5302599", "0.52749586", "0.5230464", "0.5207047", "0.5186182", "0.5155849", "0.51406753", "0.5137525", "0.51289517", "0.51211697", "0.51176393", "0.51043236", "0.51026225", "0.5094433", "0.50718796", "0.50709903", "0.5062445", "0.50394344", "0.50341153", "0.50019217", "0.49827418", "0.49637637", "0.49405167", "0.4939844", "0.49358484", "0.49280035", "0.4924172", "0.4916206", "0.49071714", "0.48948693", "0.48827773", "0.4878517", "0.48752558", "0.48728767", "0.48695856", "0.48694953", "0.4864971", "0.48486632", "0.48467082", "0.48447964", "0.4835842", "0.48349574", "0.48199248", "0.48199022", "0.481427", "0.48037577", "0.48018607", "0.4794084", "0.4779717", "0.47637165", "0.47627503", "0.47587234", "0.4757444", "0.4754863", "0.47548282", "0.47527283", "0.4740579", "0.47315395", "0.47272843", "0.4725792", "0.47214594", "0.471805", "0.47170046", "0.47158492", "0.47158158", "0.47116867", "0.47084913", "0.47079274", "0.47044486", "0.4694644", "0.46936256", "0.46920514", "0.46870244", "0.46811923", "0.4677047", "0.4672849", "0.46702728", "0.4667045", "0.46589", "0.4654487" ]
0.80171955
1
AssetRevisionExistsGP checks if the AssetRevision row exists. Panics on error.
func AssetRevisionExistsGP(orig int) bool { e, err := AssetRevisionExists(boil.GetDB(), orig) if err != nil { panic(boil.WrapErr(err)) } return e }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func AssetRevisionExists(exec boil.Executor, orig int) (bool, error) {\n\tvar exists bool\n\n\tquery := \"select exists(select 1 from \\\"asset_revision\\\" where \\\"orig\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, orig)\n\t}\n\n\trow := exec.QueryRow(query, orig)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: unable to check if asset_revision exists\")\n\t}\n\n\treturn exists, nil\n}", "func AssetRevisionExistsP(exec boil.Executor, orig int) bool {\n\te, err := AssetRevisionExists(exec, orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func AssetRevisionExistsG(orig int) (bool, error) {\n\treturn AssetRevisionExists(boil.GetDB(), orig)\n}", "func (q assetRevisionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: failed to check if asset_revision exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q assetRevisionQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (model *GrogModel) AssetExists(assetName string) bool {\n\tdoesExist := false\n\n\trow := model.db.DB.QueryRow(\"select count(1) from Assets where name = ?\", assetName)\n\tvar count int\n\trow.Scan(&count)\n\tif count > 0 {\n\t\tdoesExist = true\n\t}\n\n\treturn doesExist\n}", "func (q assetQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if asset exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (version *Version) CheckExists() (bool, error)", "func (asset Asset) Exists() bool {\n\treturn asset.model.AssetExists(asset.Name)\n}", "func ReleaseExists(actionConfig action.Configuration, name string) (bool, error) {\n\thistClient := action.NewHistory(&actionConfig)\n\thistClient.Max = 1\n\n\t_, err := histClient.Run(name)\n\tif err != nil && err != driver.ErrReleaseNotFound {\n\t\treturn false, fmt.Errorf(\"failed checking for chart history: %w\", err)\n\t}\n\n\treturn err != driver.ErrReleaseNotFound, nil\n}", "func FindAssetRevisionGP(orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(boil.GetDB(), orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (o *ReportingTaskEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *pgRepository) Exists(ctx context.Context, tenantID, id string) (bool, error) {\n\treturn r.existQuerier.Exists(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (br *builderRev) snapshotExists() (ok bool) {\n\t// If we already know this snapshot exists, don't check again.\n\tsnapshotExistsCache.Lock()\n\texists := snapshotExistsCache.m[*br]\n\tsnapshotExistsCache.Unlock()\n\tif exists {\n\t\treturn true\n\t}\n\n\t// When the function exits, cache an affirmative result.\n\tdefer func() {\n\t\tif ok {\n\t\t\tsnapshotExistsCache.Lock()\n\t\t\tsnapshotExistsCache.m[*br] = true\n\t\t\tsnapshotExistsCache.Unlock()\n\t\t}\n\t}()\n\n\tresp, err := http.Head(br.snapshotURL())\n\tif err != nil || resp.StatusCode != http.StatusOK {\n\t\treturn false\n\t}\n\n\t// If the snapshot is newer than the point at which we fixed writing\n\t// potentially-truncated snapshots to GCS, then stop right here.\n\t// See history in golang.org/issue/12671\n\tlastMod, err := http.ParseTime(resp.Header.Get(\"Last-Modified\"))\n\tif err == nil && lastMod.After(timeSnapshotCorruptionFixed) {\n\t\tlog.Printf(\"Snapshot exists for %v (without truncate checks)\", br)\n\t\treturn true\n\t}\n\n\t// Otherwise, if the snapshot is too old, verify it.\n\t// This path is slow.\n\t// TODO(bradfitz): delete this in 6 months or so? (around 2016-06-01)\n\tresp, err = http.Get(br.snapshotURL())\n\tif err != nil {\n\t\treturn false\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\treturn false\n\t}\n\t// Verify the .tar.gz is valid.\n\tgz, err := gzip.NewReader(resp.Body)\n\tif err != nil {\n\t\tlog.Printf(\"corrupt snapshot? %s gzip.NewReader: %v\", br.snapshotURL(), err)\n\t\treturn false\n\t}\n\ttr := tar.NewReader(gz)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s tar.Next: %v\", br.snapshotURL(), err)\n\t\t\treturn false\n\t\t}\n\t\tif _, err := io.Copy(ioutil.Discard, tr); err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s reading contents of %s: %v\", br.snapshotURL(), hdr.Name, err)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn gz.Close() == nil\n}", "func HasRevision(ctx context.Context) bool {\n\tsv, ok := ctx.Value(ctxKeyRevision).(string)\n\n\treturn ok && sv != \"\"\n}", "func FindAssetRevisionP(exec boil.Executor, orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(exec, orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func (r *ModuleStore) Exists(ctx context.Context, module, vsn string) bool {\n\tresult := models.Module{}\n\tquery := r.conn.Where(\"module = ?\", module).Where(\"version = ?\", vsn)\n\tcount, err := query.Count(&result)\n\n\treturn err == nil && count > 0\n}", "func (q assetQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func FindAssetRevision(exec boil.Executor, orig int, selectCols ...string) (*AssetRevision, error) {\n\tassetRevisionObj := &AssetRevision{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"asset_revision\\\" where \\\"orig\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(exec, query, orig)\n\n\terr := q.Bind(assetRevisionObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"public: unable to select from asset_revision\")\n\t}\n\n\treturn assetRevisionObj, nil\n}", "func (p Plugin) Exists(ctx context.Context, module string, version string) (bool, error) {\n\tresp, err := p.c.Exists(ctx, &stpb.ExistsRequest{Module: module, Version: version})\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn resp.GetExists(), nil\n}", "func AuthorExistsGP(id int64) bool {\n\te, err := AuthorExists(boil.GetDB(), id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func VendorExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `vendors` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vendors exists\")\n\t}\n\n\treturn exists, nil\n}", "func InventoryExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `inventory` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if inventory exists\")\n\t}\n\n\treturn exists, nil\n}", "func StockCvtermExists(exec boil.Executor, stockCvtermID int) (bool, error) {\n\tvar exists bool\n\n\tsql := \"select exists(select 1 from \\\"stock_cvterm\\\" where \\\"stock_cvterm_id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockCvtermID)\n\t}\n\n\trow := exec.QueryRow(sql, stockCvtermID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: unable to check if stock_cvterm exists\")\n\t}\n\n\treturn exists, nil\n}", "func instanceExists(name string) bool {\n\tvar exists bool\n\terr := pool.QueryRow(\"SELECT EXISTS (SELECT FROM PROVISION WHERE name = $1)\", name).Scan(&exists)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\treturn exists\n}", "func TransactionExistsGP(transactionID int) bool {\n\te, err := TransactionExists(boil.GetDB(), transactionID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func existsProject(gh_id int64) bool {\n\terr := db.QueryRow(\"SELECT gh_id FROM projects WHERE gh_id = $1\", gh_id).\n\t\tScan(&gh_id)\n\treturn err != sql.ErrNoRows\n}", "func AuthorExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"authors\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if authors exists\")\n\t}\n\n\treturn exists, nil\n}", "func (a *Analysis) Exists() bool {\n\treturn a.AssetExists(a.assetType)\n}", "func (c *MyAssetContract) MyAssetExists(ctx contractapi.TransactionContextInterface, myAssetID string) (bool, error) {\n\tdata, err := ctx.GetStub().GetState(myAssetID)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn data != nil, nil\n}", "func (o *ControllerConfigurationEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionNo() bool {\n\tif o != nil && o.RevisionNo != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func SourceExists(exec boil.Executor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\trow := exec.QueryRow(sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func TransactionExistsP(exec boil.Executor, transactionID int) bool {\n\te, err := TransactionExists(exec, transactionID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func IsErrReleaseAlreadyExist(err error) bool {\n\t_, ok := err.(ErrReleaseAlreadyExist)\n\treturn ok\n}", "func (v *VersionFile) Exists() (bool, error) {\n\treturn exists(v.path)\n}", "func (q vendorQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if vendors exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (_Contract *ContractCaller) Exists(opts *bind.CallOpts, _tokenId *big.Int) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Contract.contract.Call(opts, out, \"exists\", _tokenId)\n\treturn *ret0, err\n}", "func TransactionExists(exec boil.Executor, transactionID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transactions` where `transaction_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, transactionID)\n\t}\n\n\trow := exec.QueryRow(sql, transactionID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if transactions exists\")\n\t}\n\n\treturn exists, nil\n}", "func FindAssetRevisionG(orig int, selectCols ...string) (*AssetRevision, error) {\n\treturn FindAssetRevision(boil.GetDB(), orig, selectCols...)\n}", "func TagExists(projectID int64, repositoryID int64, version string) (bool, error) {\n\ttags, err := GetTagByVersion(projectID, repositoryID, version)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif tags == nil {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func TransactionExistsP(exec boil.Executor, id uint64) bool {\n\te, err := TransactionExists(exec, id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func StockExists(exec boil.Executor, stockID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `stock` where `stock_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockID)\n\t}\n\n\trow := exec.QueryRow(sql, stockID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if stock exists\")\n\t}\n\n\treturn exists, nil\n}", "func ProjectExists(ctx context.Context, exec boil.ContextExecutor, iD uint) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `projects` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if projects exists\")\n\t}\n\n\treturn exists, nil\n}", "func CommentExistsGP(index int) bool {\n\te, err := CommentExists(boil.GetDB(), index)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (i *IndexBuilder) exists(metadata *helmchart.Metadata) bool {\n\tfor m := range i.metadataCommit {\n\t\tif metadata.Name == m.Name && metadata.Version == m.Version {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func IsExist(err error) bool", "func SourceExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func Exists(IDval int) (err error){\n\treturn\n}", "func issueDepExists(e Engine, issueID int64, depID int64) (bool, error) {\n\treturn e.Where(\"(issue_id = ? AND dependency_id = ?)\", issueID, depID).Exist(&IssueDependency{})\n}", "func (o *NetworkElementSummaryAllOf) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func TransactionExistsGP(id uint64) bool {\n\te, err := TransactionExists(boil.GetDB(), id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (v *Version) Exists(ctx context.Context) bool {\n\treturn keyExists(ctx, v.Backend, v.Prefix)\n}", "func SkinExistsGP(index int) bool {\n\te, err := SkinExists(boil.GetDB(), index)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q featureCvtermDbxrefQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if feature_cvterm_dbxref exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (r *ClipRepository) Exists(clip *decryptor.Clip) bool {\n\tcPath := filepath.Join(r.Path, clip.Module.Course.ID, computeModuleHash(clip.Module), fmt.Sprintf(\"%v.psv\", clip.ID))\n\treturn file.Exists(cPath)\n}", "func AssetAlreadyUploaded(hash string, url string) bool {\n\tfmt.Println(\"Checking for hash\", hash, \"and url\", url, \"on database\")\n\trows, err := database.Query(\"SELECT media_hash, media_url FROM media_info;\")\n\tif shared.CheckError(err) {\n\t\treturn false\n\t}\n\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tvar stored_hash, stored_url string\n\t\trows.Scan(&stored_hash, &stored_url)\n\n\t\tif hash == stored_hash || url == stored_url {\n\t\t\tfmt.Println(\"Database match found, aborting upload\")\n\t\t\treturn true\n\t\t}\n\t}\n\n\tfmt.Println(\"Coudln't find a match, proceding!\")\n\treturn false\n}", "func InventoryExistsGP(id int64) bool {\n\te, err := InventoryExists(boil.GetDB(), id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func CommentExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `comment` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if comment exists\")\n\t}\n\n\treturn exists, nil\n}", "func TransactionExists(exec boil.Executor, id uint64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transaction` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"model: unable to check if transaction exists\")\n\t}\n\n\treturn exists, nil\n}", "func (ar AlbumDbRepository) Exists(id int) bool {\n\t_, err := ar.Get(id)\n\treturn err == nil\n}", "func (o *VersionedFlowUpdateRequestEntity) HasProcessGroupRevision() bool {\n\tif o != nil && o.ProcessGroupRevision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *Release) remoteExist() error {\n\tvar (\n\t\turl string = fmt.Sprintf(PathTerraform.toString(), r.Version, r.Version, runtime.GOOS, runtime.GOARCH)\n\t\tresp *http.Response\n\t\terr error\n\t)\n\n\tif resp, err = r.HTTPclient.Get(url); err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\t// Verify code equal 200\n\tif resp.StatusCode == http.StatusOK {\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (q braceletPhotoQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if bracelet_photo exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func ProjectExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `project` where `ID`=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if project exists\")\n\t}\n\n\treturn exists, nil\n}", "func StockCvtermExistsP(exec boil.Executor, stockCvtermID int) bool {\n\te, err := StockCvtermExists(exec, stockCvtermID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (vrsConnection *VRSConnection) CheckEntityExists(id string) (bool, error) {\n\treadRowArgs := ovsdb.ReadRowArgs{\n\t\tCondition: []string{ovsdb.NuageVMTableColumnVMUUID, \"==\", id},\n\t\tColumns: []string{ovsdb.NuageVMTableColumnVMUUID},\n\t}\n\n\tvar idRows []map[string]interface{}\n\tvar err error\n\tif idRows, err = vrsConnection.vmTable.ReadRows(vrsConnection.ovsdbClient, readRowArgs); err != nil {\n\t\treturn false, fmt.Errorf(\"OVSDB read error %v\", err)\n\t}\n\n\tvar ids []string\n\tfor _, row := range idRows {\n\t\tids = append(ids, row[ovsdb.NuageVMTableColumnVMUUID].(string))\n\t}\n\n\tif len(ids) == 1 && id == ids[0] {\n\t\treturn true, err\n\t}\n\treturn false, err\n}", "func VoteExists(ctx context.Context, exec boil.ContextExecutor, hash string) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"vote\\\" where \\\"hash\\\"=$1 limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, hash)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, hash)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vote exists\")\n\t}\n\n\treturn exists, nil\n}", "func CheckPostExists(postId string, session *r.Session) bool {\n\tvar post interface{}\n\tdb := os.Getenv(\"DB\")\n\tpostTable := os.Getenv(\"POSTTABLE\")\n\tcur, _ := r.DB(db).Table(postTable).Get(postId).Run(session)\n\tcur.One(&post)\n\tfmt.Println(post)\n\tif post != nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func (q repositoryQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if repositories exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func StockCvtermExistsGP(stockCvtermID int) bool {\n\te, err := StockCvtermExists(boil.GetDB(), stockCvtermID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q buildingQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"record: failed to check if buildings exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (l *levelDBRepo) Exists(tenantID, id []byte) bool {\n\tkey := getKey(tenantID, id)\n\tres, err := l.db.Has(key, nil)\n\t// TODO check this\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn res\n}", "func (q rawVisitQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if raw_visits exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (m *InMemoryRepository) Exists(u fyne.URI) (bool, error) {\n\tpath := u.Path()\n\tif path == \"\" {\n\t\treturn false, fmt.Errorf(\"invalid path '%s'\", path)\n\t}\n\n\t_, ok := m.Data[path]\n\treturn ok, nil\n}", "func InventoryExistsP(exec boil.Executor, id int64) bool {\n\te, err := InventoryExists(exec, id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (s *SpendJournalPruner) dependencyExistsInternal(blockHash *chainhash.Hash, consumerId string) bool {\n\ts.dependentsMtx.RLock()\n\tdependents, ok := s.dependents[*blockHash]\n\ts.dependentsMtx.RUnlock()\n\tif !ok {\n\t\t// The dependency does not exist if the block hash is not\n\t\t// a key for dependents.\n\t\treturn false\n\t}\n\n\tfor _, id := range dependents {\n\t\tif consumerId == id {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (q stockCvtermQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: failed to check if stock_cvterm exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q projectQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if project exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func RowExists(query string, args ...interface{}) bool {\n\tvar exists bool\n\tquery = fmt.Sprintf(\"SELECT exists (%s)\", query)\n\terr := Db.QueryRow(query, args...).Scan(&exists)\n\tif err != nil && err != sql.ErrNoRows {\n\t\tfmt.Errorf(\"error checking if row exists '%s' %v\", args, err)\n\t}\n\treturn exists\n}", "func (_Storage *StorageCaller) AccountExists(opts *bind.CallOpts, addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"accountExists\", addr)\n\treturn *ret0, err\n}", "func (q inventoryQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if inventory exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (m *manifestService) Exists(ctx context.Context, dgst digest.Digest) (bool, error) {\n\tcontext.GetLogger(ctx).Debugf(\"(*manifestService).Exists\")\n\n\timage, _, err := m.repo.getImageOfImageStream(dgst)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn image != nil, nil\n}", "func (q blockQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if block exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (session KeyValueSession) Exists(key DatabaseKey) bool {\n\tversion := session.store.tree.Version64()\n\tindex, _ := session.store.tree.GetVersioned(key, version)\n\tif index == -1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (o *NormalizedProjectRevisionHook) HasProjectRevisionId() bool {\n\tif o != nil && o.ProjectRevisionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (q sourceQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (m *GormResourceRepository) CheckExists(ctx context.Context, id string) error {\n\tdefer goa.MeasureSince([]string{\"goa\", \"db\", \"resource\", \"exists\"}, time.Now())\n\treturn base.CheckExistsWithCustomIDColumn(ctx, m.db, m.TableName(), \"resource_id\", id)\n}", "func (i Item) Exists() bool {\n\tif i.UUID == \"\" {\n\t\treturn false\n\t}\n\tuuid, err := db.SelectFirst(\"SELECT `uuid` FROM `items` WHERE `uuid`=?\", i.UUID)\n\n\tif err != nil {\n\t\tLog(err)\n\t\treturn false\n\t}\n\tLog(\"Exists:\", uuid)\n\treturn uuid != \"\"\n}", "func cephRBDSnapshotExists(clusterName string, poolName string,\n\tvolumeName string, volumeType string, snapshotName string,\n\tuserName string) bool {\n\t_, err := shared.RunCommand(\n\t\t\"rbd\",\n\t\t\"--id\", userName,\n\t\t\"--cluster\", clusterName,\n\t\t\"--pool\", poolName,\n\t\t\"info\",\n\t\tfmt.Sprintf(\"%s_%s@%s\", volumeType, volumeName, snapshotName))\n\tif err != nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func (store KeyValue) Exists(key DatabaseKey) bool {\n\tversion := store.tree.Version64()\n\tindex, _ := store.tree.GetVersioned(key, version)\n\tif index == -1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (r *Repository) CheckPageExists(pageID int) (bool, error) {\n pageExists := false\n psqlStmt := `\n SELECT EXISTS(\n SELECT 1 FROM pages\n WHERE id=$1)`\n err := r.DB.QueryRow(psqlStmt, pageID).Scan(&pageExists)\n if err != nil {\n return false, err\n }\n return pageExists, nil\n}", "func EntryExists(url string) bool {\n\tresp, err := http.Head(url)\n\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\n\treturn resp.StatusCode == http.StatusOK\n}", "func StockExistsP(exec boil.Executor, stockID int) bool {\n\te, err := StockExists(exec, stockID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (snapshot *DataSnapshot) Exists() bool {\n\treturn bytes.Compare(snapshot.raw, []byte(\"null\")) != 0\n}", "func (client *GCSBlobstore) Exists(dest string) (exists bool, err error) {\n\tif exists, err = client.exists(client.publicGCS, dest); err == nil {\n\t\treturn exists, nil\n\t}\n\n\t// If the public client fails, try using it as an authenticated actor\n\tif client.authenticatedGCS != nil {\n\t\treturn client.exists(client.authenticatedGCS, dest)\n\t}\n\n\treturn\n}", "func (q mempoolBinQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if mempool_bin exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func ActivityLogExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"activity_logs\\\" where \\\"id\\\"=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: unable to check if activity_logs exists\")\n\t}\n\n\treturn exists, nil\n}" ]
[ "0.80536807", "0.78159136", "0.73210865", "0.7100514", "0.60784006", "0.5833039", "0.5788161", "0.5773283", "0.57297766", "0.56675017", "0.5660641", "0.5569319", "0.55472803", "0.5503749", "0.5397937", "0.53154546", "0.52949905", "0.5282068", "0.525275", "0.52519023", "0.5243261", "0.5232179", "0.51946753", "0.5183601", "0.5132661", "0.51318663", "0.51133317", "0.51106054", "0.5099786", "0.5091383", "0.5072699", "0.5059395", "0.5043906", "0.50364256", "0.503189", "0.5000047", "0.4999709", "0.4993588", "0.49934745", "0.49860638", "0.49788806", "0.49764335", "0.497619", "0.49744666", "0.49703595", "0.4967039", "0.49643752", "0.49642944", "0.49541092", "0.49522147", "0.49404138", "0.493704", "0.49369296", "0.49351877", "0.49336353", "0.49200875", "0.49160597", "0.49154037", "0.4914208", "0.49108782", "0.49085367", "0.4907445", "0.4901492", "0.4901114", "0.4894628", "0.48930517", "0.4888022", "0.4887633", "0.48816052", "0.48703092", "0.4864692", "0.4864308", "0.48633105", "0.48627567", "0.4851259", "0.48484102", "0.48399532", "0.48385713", "0.4838537", "0.48322138", "0.4829457", "0.48164588", "0.48068172", "0.47977936", "0.47923255", "0.47873715", "0.47847018", "0.47775635", "0.47749513", "0.47683352", "0.47646508", "0.4760215", "0.47572523", "0.47564796", "0.47563997", "0.4755931", "0.47558478", "0.47541106", "0.47528595", "0.47507086" ]
0.8099057
0
AssetRevisionExistsP checks if the AssetRevision row exists. Panics on error.
func AssetRevisionExistsP(exec boil.Executor, orig int) bool { e, err := AssetRevisionExists(exec, orig) if err != nil { panic(boil.WrapErr(err)) } return e }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func AssetRevisionExists(exec boil.Executor, orig int) (bool, error) {\n\tvar exists bool\n\n\tquery := \"select exists(select 1 from \\\"asset_revision\\\" where \\\"orig\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, orig)\n\t}\n\n\trow := exec.QueryRow(query, orig)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: unable to check if asset_revision exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q assetRevisionQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"public: failed to check if asset_revision exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func AssetRevisionExistsGP(orig int) bool {\n\te, err := AssetRevisionExists(boil.GetDB(), orig)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func AssetRevisionExistsG(orig int) (bool, error) {\n\treturn AssetRevisionExists(boil.GetDB(), orig)\n}", "func (q assetRevisionQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q assetQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if asset exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (asset Asset) Exists() bool {\n\treturn asset.model.AssetExists(asset.Name)\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (s *SmartContract) AssetExists(ctx contractapi.TransactionContextInterface, id string) (bool, error) {\n\tassetJSON, err := ctx.GetStub().GetState(id)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"failed to read from world state: %v\", err)\n\t}\n\n\treturn assetJSON != nil, nil\n}", "func (model *GrogModel) AssetExists(assetName string) bool {\n\tdoesExist := false\n\n\trow := model.db.DB.QueryRow(\"select count(1) from Assets where name = ?\", assetName)\n\tvar count int\n\trow.Scan(&count)\n\tif count > 0 {\n\t\tdoesExist = true\n\t}\n\n\treturn doesExist\n}", "func (version *Version) CheckExists() (bool, error)", "func (o *ReportingTaskEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func FindAssetRevisionP(exec boil.Executor, orig int, selectCols ...string) *AssetRevision {\n\tretobj, err := FindAssetRevision(exec, orig, selectCols...)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn retobj\n}", "func FindAssetRevision(exec boil.Executor, orig int, selectCols ...string) (*AssetRevision, error) {\n\tassetRevisionObj := &AssetRevision{}\n\n\tsel := \"*\"\n\tif len(selectCols) > 0 {\n\t\tsel = strings.Join(strmangle.IdentQuoteSlice(dialect.LQ, dialect.RQ, selectCols), \",\")\n\t}\n\tquery := fmt.Sprintf(\n\t\t\"select %s from \\\"asset_revision\\\" where \\\"orig\\\"=$1\", sel,\n\t)\n\n\tq := queries.Raw(exec, query, orig)\n\n\terr := q.Bind(assetRevisionObj)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"public: unable to select from asset_revision\")\n\t}\n\n\treturn assetRevisionObj, nil\n}", "func AuthorExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"authors\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if authors exists\")\n\t}\n\n\treturn exists, nil\n}", "func ReleaseExists(actionConfig action.Configuration, name string) (bool, error) {\n\thistClient := action.NewHistory(&actionConfig)\n\thistClient.Max = 1\n\n\t_, err := histClient.Run(name)\n\tif err != nil && err != driver.ErrReleaseNotFound {\n\t\treturn false, fmt.Errorf(\"failed checking for chart history: %w\", err)\n\t}\n\n\treturn err != driver.ErrReleaseNotFound, nil\n}", "func (r *pgRepository) Exists(ctx context.Context, tenantID, id string) (bool, error) {\n\treturn r.existQuerier.Exists(ctx, resource.API, tenantID, repo.Conditions{repo.NewEqualCondition(\"id\", id)})\n}", "func (c *MyAssetContract) MyAssetExists(ctx contractapi.TransactionContextInterface, myAssetID string) (bool, error) {\n\tdata, err := ctx.GetStub().GetState(myAssetID)\n\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn data != nil, nil\n}", "func (q assetQuery) ExistsP() bool {\n\te, err := q.Exists()\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func HasRevision(ctx context.Context) bool {\n\tsv, ok := ctx.Value(ctxKeyRevision).(string)\n\n\treturn ok && sv != \"\"\n}", "func (o *ControllerConfigurationEntity) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (v *VersionFile) Exists() (bool, error) {\n\treturn exists(v.path)\n}", "func (br *builderRev) snapshotExists() (ok bool) {\n\t// If we already know this snapshot exists, don't check again.\n\tsnapshotExistsCache.Lock()\n\texists := snapshotExistsCache.m[*br]\n\tsnapshotExistsCache.Unlock()\n\tif exists {\n\t\treturn true\n\t}\n\n\t// When the function exits, cache an affirmative result.\n\tdefer func() {\n\t\tif ok {\n\t\t\tsnapshotExistsCache.Lock()\n\t\t\tsnapshotExistsCache.m[*br] = true\n\t\t\tsnapshotExistsCache.Unlock()\n\t\t}\n\t}()\n\n\tresp, err := http.Head(br.snapshotURL())\n\tif err != nil || resp.StatusCode != http.StatusOK {\n\t\treturn false\n\t}\n\n\t// If the snapshot is newer than the point at which we fixed writing\n\t// potentially-truncated snapshots to GCS, then stop right here.\n\t// See history in golang.org/issue/12671\n\tlastMod, err := http.ParseTime(resp.Header.Get(\"Last-Modified\"))\n\tif err == nil && lastMod.After(timeSnapshotCorruptionFixed) {\n\t\tlog.Printf(\"Snapshot exists for %v (without truncate checks)\", br)\n\t\treturn true\n\t}\n\n\t// Otherwise, if the snapshot is too old, verify it.\n\t// This path is slow.\n\t// TODO(bradfitz): delete this in 6 months or so? (around 2016-06-01)\n\tresp, err = http.Get(br.snapshotURL())\n\tif err != nil {\n\t\treturn false\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\treturn false\n\t}\n\t// Verify the .tar.gz is valid.\n\tgz, err := gzip.NewReader(resp.Body)\n\tif err != nil {\n\t\tlog.Printf(\"corrupt snapshot? %s gzip.NewReader: %v\", br.snapshotURL(), err)\n\t\treturn false\n\t}\n\ttr := tar.NewReader(gz)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s tar.Next: %v\", br.snapshotURL(), err)\n\t\t\treturn false\n\t\t}\n\t\tif _, err := io.Copy(ioutil.Discard, tr); err != nil {\n\t\t\tlog.Printf(\"corrupt snapshot? %s reading contents of %s: %v\", br.snapshotURL(), hdr.Name, err)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn gz.Close() == nil\n}", "func (a *Analysis) Exists() bool {\n\treturn a.AssetExists(a.assetType)\n}", "func (p Plugin) Exists(ctx context.Context, module string, version string) (bool, error) {\n\tresp, err := p.c.Exists(ctx, &stpb.ExistsRequest{Module: module, Version: version})\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn resp.GetExists(), nil\n}", "func InventoryExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `inventory` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if inventory exists\")\n\t}\n\n\treturn exists, nil\n}", "func VendorExists(ctx context.Context, exec boil.ContextExecutor, iD int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `vendors` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vendors exists\")\n\t}\n\n\treturn exists, nil\n}", "func (r *ModuleStore) Exists(ctx context.Context, module, vsn string) bool {\n\tresult := models.Module{}\n\tquery := r.conn.Where(\"module = ?\", module).Where(\"version = ?\", vsn)\n\tcount, err := query.Count(&result)\n\n\treturn err == nil && count > 0\n}", "func TransactionExists(exec boil.Executor, transactionID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transactions` where `transaction_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, transactionID)\n\t}\n\n\trow := exec.QueryRow(sql, transactionID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if transactions exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q vendorQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if vendors exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func TransactionExistsP(exec boil.Executor, transactionID int) bool {\n\te, err := TransactionExists(exec, transactionID)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (ar AlbumDbRepository) Exists(id int) bool {\n\t_, err := ar.Get(id)\n\treturn err == nil\n}", "func (t *Transcript) Exists() bool {\n\treturn t.AssetExists(whodunit.AssetTypeTranscript)\n}", "func (_Contract *ContractCaller) Exists(opts *bind.CallOpts, _tokenId *big.Int) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Contract.contract.Call(opts, out, \"exists\", _tokenId)\n\treturn *ret0, err\n}", "func (o *NiaapiRevisionInfoAllOf) HasRevisionNo() bool {\n\tif o != nil && o.RevisionNo != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func TransactionExistsP(exec boil.Executor, id uint64) bool {\n\te, err := TransactionExists(exec, id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func TransactionExists(exec boil.Executor, id uint64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `transaction` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"model: unable to check if transaction exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q mempoolBinQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if mempool_bin exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (sm *SourceMgr) RevisionPresentIn(id ProjectIdentifier, r Revision) (bool, error) {\n\tsrc, err := sm.getSourceFor(id)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn src.revisionPresentIn(r)\n}", "func ActivityLogExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"activity_logs\\\" where \\\"id\\\"=? limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, iD)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: unable to check if activity_logs exists\")\n\t}\n\n\treturn exists, nil\n}", "func (r *ClipRepository) Exists(clip *decryptor.Clip) bool {\n\tcPath := filepath.Join(r.Path, clip.Module.Course.ID, computeModuleHash(clip.Module), fmt.Sprintf(\"%v.psv\", clip.ID))\n\treturn file.Exists(cPath)\n}", "func (v *Version) Exists(ctx context.Context) bool {\n\treturn keyExists(ctx, v.Backend, v.Prefix)\n}", "func (m *InMemoryRepository) Exists(u fyne.URI) (bool, error) {\n\tpath := u.Path()\n\tif path == \"\" {\n\t\treturn false, fmt.Errorf(\"invalid path '%s'\", path)\n\t}\n\n\t_, ok := m.Data[path]\n\treturn ok, nil\n}", "func AuthorExistsP(exec boil.Executor, id int64) bool {\n\te, err := AuthorExists(exec, id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func CommentExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `comment` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if comment exists\")\n\t}\n\n\treturn exists, nil\n}", "func (i *IndexBuilder) exists(metadata *helmchart.Metadata) bool {\n\tfor m := range i.metadataCommit {\n\t\tif metadata.Name == m.Name && metadata.Version == m.Version {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (q repositoryQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if repositories exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (o *NetworkElementSummaryAllOf) HasRevision() bool {\n\tif o != nil && o.Revision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (q assetRevisionQuery) Count() (int64, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"public: failed to count asset_revision rows\")\n\t}\n\n\treturn count, nil\n}", "func (q braceletPhotoQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if bracelet_photo exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func VoteExists(ctx context.Context, exec boil.ContextExecutor, hash string) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"vote\\\" where \\\"hash\\\"=$1 limit 1)\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, hash)\n\t}\n\trow := exec.QueryRowContext(ctx, sql, hash)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vote exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q blockQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if block exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (q inventoryQuery) Exists() (bool, error) {\n\tvar count int64\n\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow().Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if inventory exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (sm *SourceMgr) RevisionPresentIn(n ProjectRoot, r Revision) (bool, error) {\n\tpmc, err := sm.getProjectManager(n)\n\tif err != nil {\n\t\t// TODO(sdboyer) More-er proper-er errors\n\t\treturn false, err\n\t}\n\n\treturn pmc.pm.RevisionPresentIn(r)\n}", "func SourceExists(exec boil.Executor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\trow := exec.QueryRow(sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func (l *levelDBRepo) Exists(tenantID, id []byte) bool {\n\tkey := getKey(tenantID, id)\n\tres, err := l.db.Has(key, nil)\n\t// TODO check this\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn res\n}", "func (q rssAnnouncementQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if rss_announcements exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (vrsConnection *VRSConnection) CheckEntityExists(id string) (bool, error) {\n\treadRowArgs := ovsdb.ReadRowArgs{\n\t\tCondition: []string{ovsdb.NuageVMTableColumnVMUUID, \"==\", id},\n\t\tColumns: []string{ovsdb.NuageVMTableColumnVMUUID},\n\t}\n\n\tvar idRows []map[string]interface{}\n\tvar err error\n\tif idRows, err = vrsConnection.vmTable.ReadRows(vrsConnection.ovsdbClient, readRowArgs); err != nil {\n\t\treturn false, fmt.Errorf(\"OVSDB read error %v\", err)\n\t}\n\n\tvar ids []string\n\tfor _, row := range idRows {\n\t\tids = append(ids, row[ovsdb.NuageVMTableColumnVMUUID].(string))\n\t}\n\n\tif len(ids) == 1 && id == ids[0] {\n\t\treturn true, err\n\t}\n\treturn false, err\n}", "func (m *GormResourceRepository) CheckExists(ctx context.Context, id string) error {\n\tdefer goa.MeasureSince([]string{\"goa\", \"db\", \"resource\", \"exists\"}, time.Now())\n\treturn base.CheckExistsWithCustomIDColumn(ctx, m.db, m.TableName(), \"resource_id\", id)\n}", "func (_ElvTradable *ElvTradableCaller) Exists(opts *bind.CallOpts, tokenId *big.Int) (bool, error) {\n\tvar out []interface{}\n\terr := _ElvTradable.contract.Call(opts, &out, \"exists\", tokenId)\n\n\tif err != nil {\n\t\treturn *new(bool), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\n\n\treturn out0, err\n\n}", "func (o *AssetRevision) Insert(exec boil.Executor, whitelist ...string) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no asset_revision provided for insertion\")\n\t}\n\n\tvar err error\n\n\tif err := o.doBeforeInsertHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\tnzDefaults := queries.NonZeroDefaultSet(assetRevisionColumnsWithDefault, o)\n\n\tkey := makeCacheKey(whitelist, nzDefaults)\n\tassetRevisionInsertCacheMut.RLock()\n\tcache, cached := assetRevisionInsertCache[key]\n\tassetRevisionInsertCacheMut.RUnlock()\n\n\tif !cached {\n\t\twl, returnColumns := strmangle.InsertColumnSet(\n\t\t\tassetRevisionColumns,\n\t\t\tassetRevisionColumnsWithDefault,\n\t\t\tassetRevisionColumnsWithoutDefault,\n\t\t\tnzDefaults,\n\t\t\twhitelist,\n\t\t)\n\n\t\tcache.valueMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, wl)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcache.retMapping, err = queries.BindMapping(assetRevisionType, assetRevisionMapping, returnColumns)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif len(wl) != 0 {\n\t\t\tcache.query = fmt.Sprintf(\"INSERT INTO \\\"asset_revision\\\" (\\\"%s\\\") VALUES (%s)\", strings.Join(wl, \"\\\",\\\"\"), strmangle.Placeholders(dialect.IndexPlaceholders, len(wl), 1, 1))\n\t\t} else {\n\t\t\tcache.query = \"INSERT INTO \\\"asset_revision\\\" DEFAULT VALUES\"\n\t\t}\n\n\t\tif len(cache.retMapping) != 0 {\n\t\t\tcache.query += fmt.Sprintf(\" RETURNING \\\"%s\\\"\", strings.Join(returnColumns, \"\\\",\\\"\"))\n\t\t}\n\t}\n\n\tvalue := reflect.Indirect(reflect.ValueOf(o))\n\tvals := queries.ValuesFromMapping(value, cache.valueMapping)\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, cache.query)\n\t\tfmt.Fprintln(boil.DebugWriter, vals)\n\t}\n\n\tif len(cache.retMapping) != 0 {\n\t\terr = exec.QueryRow(cache.query, vals...).Scan(queries.PtrsFromMapping(value, cache.retMapping)...)\n\t} else {\n\t\t_, err = exec.Exec(cache.query, vals...)\n\t}\n\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to insert into asset_revision\")\n\t}\n\n\tif !cached {\n\t\tassetRevisionInsertCacheMut.Lock()\n\t\tassetRevisionInsertCache[key] = cache\n\t\tassetRevisionInsertCacheMut.Unlock()\n\t}\n\n\treturn o.doAfterInsertHooks(exec)\n}", "func SkinExists(exec boil.Executor, index int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `skin` where `index`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, index)\n\t}\n\n\trow := exec.QueryRow(sql, index)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if skin exists\")\n\t}\n\n\treturn exists, nil\n}", "func SourceExists(exec boil.Executor, id int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"sources\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmdbmodels: unable to check if sources exists\")\n\t}\n\n\treturn exists, nil\n}", "func StockCvtermExists(exec boil.Executor, stockCvtermID int) (bool, error) {\n\tvar exists bool\n\n\tsql := \"select exists(select 1 from \\\"stock_cvterm\\\" where \\\"stock_cvterm_id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockCvtermID)\n\t}\n\n\trow := exec.QueryRow(sql, stockCvtermID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"chado: unable to check if stock_cvterm exists\")\n\t}\n\n\treturn exists, nil\n}", "func (r *ClipRepository) Exists(clip *decryptor.Clip) (bool, error) {\n\tif clip == nil {\n\t\treturn false, ErrClipUndefined\n\t}\n\tif clip.Module == nil {\n\t\treturn false, ErrModuleUndefined\n\t}\n\tif clip.Module.Course == nil {\n\t\treturn false, ErrCourseUndefined\n\t}\n\tcPath := filepath.Join(r.Path, clip.Module.Course.ID, computeModuleHash(clip.Module), fmt.Sprintf(\"%v.psv\", clip.ID))\n\treturn r.FileExists(cPath), nil\n}", "func TagExists(projectID int64, repositoryID int64, version string) (bool, error) {\n\ttags, err := GetTagByVersion(projectID, repositoryID, version)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif tags == nil {\n\t\treturn false, nil\n\t}\n\treturn true, nil\n}", "func (o *AssetRevision) Delete(exec boil.Executor) error {\n\tif o == nil {\n\t\treturn errors.New(\"public: no AssetRevision provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), assetRevisionPrimaryKeyMapping)\n\tquery := \"DELETE FROM \\\"asset_revision\\\" WHERE \\\"orig\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, query)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\t_, err := exec.Exec(query, args...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"public: unable to delete from asset_revision\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func IsExist(err error) bool", "func (o *NormalizedProjectRevisionHook) HasProjectRevisionId() bool {\n\tif o != nil && o.ProjectRevisionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func InventoryExistsP(exec boil.Executor, id int64) bool {\n\te, err := InventoryExists(exec, id)\n\tif err != nil {\n\t\tpanic(boil.WrapErr(err))\n\t}\n\n\treturn e\n}", "func (q activityLogQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"dbmodel: failed to check if activity_logs exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (i Item) Exists() bool {\n\tif i.UUID == \"\" {\n\t\treturn false\n\t}\n\tuuid, err := db.SelectFirst(\"SELECT `uuid` FROM `items` WHERE `uuid`=?\", i.UUID)\n\n\tif err != nil {\n\t\tLog(err)\n\t\treturn false\n\t}\n\tLog(\"Exists:\", uuid)\n\treturn uuid != \"\"\n}", "func (v *Version) IsRevision() bool {\n\treturn v.isRevision\n}", "func (q holdenAtQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if HoldenAt exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func StockExists(exec boil.Executor, stockID int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `stock` where `stock_id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, stockID)\n\t}\n\n\trow := exec.QueryRow(sql, stockID)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if stock exists\")\n\t}\n\n\treturn exists, nil\n}", "func (v Repository) RevisionIsValid(revision string) bool {\n\traw := v.Raw()\n\n\tif raw == nil {\n\t\treturn false\n\t}\n\tif _, err := raw.ResolveRevision(plumbing.Revision(revision)); err == nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func (r *Release) remoteExist() error {\n\tvar (\n\t\turl string = fmt.Sprintf(PathTerraform.toString(), r.Version, r.Version, runtime.GOOS, runtime.GOARCH)\n\t\tresp *http.Response\n\t\terr error\n\t)\n\n\tif resp, err = r.HTTPclient.Get(url); err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\t// Verify code equal 200\n\tif resp.StatusCode == http.StatusOK {\n\t\treturn nil\n\t}\n\n\treturn nil\n}", "func (in *PublicAccessBlockClient) SubresourceExists(cr *v1beta1.Bucket) bool {\n\treturn cr.Spec.ForProvider.PublicAccessBlockConfiguration != nil\n}", "func (_Storage *StorageCaller) AccountExists(opts *bind.CallOpts, addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Storage.contract.Call(opts, out, \"accountExists\", addr)\n\treturn *ret0, err\n}", "func CheckPostExists(postId string, session *r.Session) bool {\n\tvar post interface{}\n\tdb := os.Getenv(\"DB\")\n\tpostTable := os.Getenv(\"POSTTABLE\")\n\tcur, _ := r.DB(db).Table(postTable).Get(postId).Run(session)\n\tcur.One(&post)\n\tfmt.Println(post)\n\tif post != nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func RowExists(query string, args ...interface{}) bool {\n\tvar exists bool\n\tquery = fmt.Sprintf(\"SELECT exists (%s)\", query)\n\terr := Db.QueryRow(query, args...).Scan(&exists)\n\tif err != nil && err != sql.ErrNoRows {\n\t\tfmt.Errorf(\"error checking if row exists '%s' %v\", args, err)\n\t}\n\treturn exists\n}", "func (q sourceQuery) Exists(exec boil.Executor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRow(exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"mdbmodels: failed to check if sources exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (i *Item) Exists() (bool, error) {\n\tif i.UUID == \"\" {\n\t\treturn false, nil\n\t}\n\tvar id string\n\treturn db.SelectExists(\n\t\t&id,\n\t\t\"SELECT uuid FROM items WHERE uuid=?\",\n\t\ti.UUID,\n\t)\n}", "func PostExists(ctx context.Context, exec boil.ContextExecutor, iD int64) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from \\\"posts\\\" where \\\"id\\\"=$1 limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"orm: unable to check if posts exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q projectQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if project exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func EntryExists(url string) bool {\n\tresp, err := http.Head(url)\n\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\n\treturn resp.StatusCode == http.StatusOK\n}", "func (xpss *XPipelineSalesStage) Exists() bool {\n\treturn xpss._exists\n}", "func (q rawVisitQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if raw_visits exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func VoteExists(exec boil.Executor, id int) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `vote` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, id)\n\t}\n\n\trow := exec.QueryRow(sql, id)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if vote exists\")\n\t}\n\n\treturn exists, nil\n}", "func instanceExists(name string) bool {\n\tvar exists bool\n\terr := pool.QueryRow(\"SELECT EXISTS (SELECT FROM PROVISION WHERE name = $1)\", name).Scan(&exists)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false\n\t}\n\treturn exists\n}", "func (p *DeliveryConfigProcessor) ResourceExists(search *ExportableResource) bool {\n\tfor eix := range p.deliveryConfig.Environments {\n\t\trix := p.findResourceIndex(search, eix)\n\t\tif rix >= 0 {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func ProjectExists(ctx context.Context, exec boil.ContextExecutor, iD uint) (bool, error) {\n\tvar exists bool\n\tsql := \"select exists(select 1 from `projects` where `id`=? limit 1)\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, iD)\n\t}\n\n\trow := exec.QueryRowContext(ctx, sql, iD)\n\n\terr := row.Scan(&exists)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: unable to check if projects exists\")\n\t}\n\n\treturn exists, nil\n}", "func (q paymentObjectQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"models: failed to check if payment_objects exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (session KeyValueSession) Exists(key DatabaseKey) bool {\n\tversion := session.store.tree.Version64()\n\tindex, _ := session.store.tree.GetVersioned(key, version)\n\tif index == -1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (q segmentQuery) Exists(ctx context.Context, exec boil.ContextExecutor) (bool, error) {\n\tvar count int64\n\n\tqueries.SetSelect(q.Query, nil)\n\tqueries.SetCount(q.Query)\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Query.QueryRowContext(ctx, exec).Scan(&count)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, \"boiler: failed to check if segment exists\")\n\t}\n\n\treturn count > 0, nil\n}", "func (db *DB) DoesArticleExist(id uint64) bool {\n\terr := db.View(func(txn *badger.Txn) error {\n\t\tarticle, err := txn.Get(helpers.I64tob(id))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = article.Value(func(val []byte) error {\n\t\t\treturn nil\n\t\t})\n\t\treturn err\n\t})\n\n\treturn err == nil\n}", "func (store KeyValue) Exists(key DatabaseKey) bool {\n\tversion := store.tree.Version64()\n\tindex, _ := store.tree.GetVersioned(key, version)\n\tif index == -1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (m *manifestService) Exists(ctx context.Context, dgst digest.Digest) (bool, error) {\n\tcontext.GetLogger(ctx).Debugf(\"(*manifestService).Exists\")\n\n\timage, _, err := m.repo.getImageOfImageStream(dgst)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treturn image != nil, nil\n}", "func (o *VersionedFlowUpdateRequestEntity) HasProcessGroupRevision() bool {\n\tif o != nil && o.ProcessGroupRevision != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (mvq *ModuleVersionQuery) Exist(ctx context.Context) (bool, error) {\n\treturn mvq.sqlExist(ctx)\n}" ]
[ "0.82006913", "0.73832935", "0.7016023", "0.64628965", "0.6251488", "0.5929571", "0.58373654", "0.5832617", "0.5811708", "0.5695554", "0.5651941", "0.55325687", "0.5530963", "0.5495016", "0.54694414", "0.5393326", "0.5384284", "0.5330663", "0.5319994", "0.52559197", "0.5217913", "0.52118236", "0.5186365", "0.51802886", "0.5130519", "0.5121467", "0.5119393", "0.51100135", "0.5103487", "0.5081217", "0.5047781", "0.5044924", "0.5032661", "0.500446", "0.5000129", "0.49992856", "0.49921992", "0.49734205", "0.49568304", "0.49553132", "0.49430043", "0.49425182", "0.4941758", "0.49412167", "0.4938125", "0.49339214", "0.49328268", "0.49290633", "0.49266136", "0.49035814", "0.48968443", "0.4895571", "0.48902693", "0.48839772", "0.4870836", "0.4868525", "0.4867851", "0.48643", "0.4856644", "0.48547235", "0.4854209", "0.4853281", "0.48452264", "0.4841408", "0.48412114", "0.48407584", "0.4840485", "0.48393998", "0.4839275", "0.48250195", "0.48019895", "0.4791842", "0.47850227", "0.47840363", "0.47832778", "0.4776212", "0.47749534", "0.47637978", "0.47633", "0.47474533", "0.47439268", "0.47420543", "0.47410995", "0.47403273", "0.47376302", "0.4732842", "0.47320244", "0.47236434", "0.47226065", "0.4712341", "0.47026423", "0.46989468", "0.46956423", "0.46914282", "0.46834317", "0.46815634", "0.4681276", "0.4679642", "0.46689668", "0.46615154" ]
0.79873574
1
WARNING: `ReparseQuery` might affect nontensorflow clients using S3compatible API with AIStore. To be used with caution.
func ReparseQuery(r *http.Request) { if !strings.ContainsRune(r.URL.Path, '?') { return } q := r.URL.Query() tmpURL, err := url.Parse(r.URL.Path) debug.AssertNoErr(err) for k, v := range tmpURL.Query() { q.Add(k, strings.Join(v, ",")) } r.URL.Path = tmpURL.Path r.URL.RawQuery = q.Encode() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func RewriteQuery(u url.URL) url.URL {\n\t// query is a copy which we will modify using Set() and use in the result\n\tquery := u.Query()\n\n\tquerystr := query.Get(\"q\")\n\tqueryWords := []string{}\n\tfor _, word := range strings.Split(querystr, \" \") {\n\t\tfmt.Printf(\"word = %v\\n\", word)\n\t\tlower := strings.ToLower(word)\n\t\tif strings.HasPrefix(lower, \"filetype:\") {\n\t\t\tquery.Add(\"filetype\", strings.ToLower(word[len(\"filetype:\"):]))\n\t\t} else if strings.HasPrefix(lower, \"-filetype:\") {\n\t\t\tquery.Add(\"nfiletype\", strings.ToLower(word[len(\"-filetype:\"):]))\n\t\t} else if strings.HasPrefix(lower, \"package:\") {\n\t\t\tquery.Set(\"package\", word[len(\"package:\"):])\n\t\t} else if strings.HasPrefix(lower, \"pkg:\") {\n\t\t\tquery.Set(\"package\", word[len(\"pkg:\"):])\n\t\t} else if strings.HasPrefix(lower, \"-package:\") {\n\t\t\tquery.Add(\"npackage\", word[len(\"-package:\"):])\n\t\t} else if strings.HasPrefix(lower, \"-pkg:\") {\n\t\t\tquery.Add(\"npackage\", word[len(\"-pkg:\"):])\n\t\t} else if strings.HasPrefix(lower, \"path:\") || strings.HasPrefix(lower, \"file:\") {\n\t\t\tquery.Add(\"path\", word[len(\"path:\"):])\n\t\t} else if strings.HasPrefix(lower, \"-path:\") || strings.HasPrefix(lower, \"-file:\") {\n\t\t\tquery.Add(\"npath\", word[len(\"-path:\"):])\n\t\t} else {\n\t\t\tqueryWords = append(queryWords, word)\n\t\t}\n\t}\n\tquery.Set(\"q\", strings.Join(queryWords, \" \"))\n\tu.RawQuery = query.Encode()\n\n\treturn u\n}", "func TestParseQuery(t *testing.T) {\n\tunittest.SmallTest(t)\n\n\tq := &Search{}\n\terr := clearParseQuery(q, \"fdiffmax=-1&fref=false&frgbamax=-1&head=true&include=false&issue=2370153003&limit=50&match=gamma_correct&match=name&metric=combined&neg=false&pos=false&query=source_type%3Dgm&sort=desc&unt=true\")\n\trequire.NoError(t, err)\n\n\trequire.Equal(t, &Search{\n\t\tMetric: \"combined\",\n\t\tSort: \"desc\",\n\t\tMatch: []string{\"gamma_correct\", \"name\"},\n\t\tBlameGroupID: \"\",\n\t\tPos: false,\n\t\tNeg: false,\n\t\tHead: true,\n\t\tUnt: true,\n\t\tIncludeIgnores: false,\n\t\tQueryStr: \"\",\n\t\tTraceValues: url.Values{\n\t\t\t\"source_type\": []string{\"gm\"},\n\t\t},\n\t\tRQueryStr: \"\",\n\t\tRTraceValues: paramtools.ParamSet{},\n\t\tChangeListID: \"2370153003\",\n\t\tPatchSetsStr: \"\",\n\t\tPatchSets: []int64(nil),\n\t\tIncludeMaster: false,\n\t\tFCommitBegin: \"\",\n\t\tFCommitEnd: \"\",\n\t\tFRGBAMin: 0,\n\t\tFRGBAMax: -1,\n\t\tFDiffMax: -1,\n\t\tFGroupTest: \"\",\n\t\tFRef: false,\n\t\tOffset: 0,\n\t\tLimit: 50,\n\t\tNoDiff: false,\n\t\tNewCLStore: false,\n\t}, q)\n}", "func ParseQuery(args ...string) Query {\n\tvar cmd string\n\tvar ids []int\n\tvar tags []string\n\tvar antiTags []string\n\tvar project string\n\tvar antiProjects []string\n\tvar priority string\n\tvar template int\n\tvar words []string\n\tvar notesModeActivated bool\n\tvar notes []string\n\tvar ignoreContext bool\n\n\t// something other than an ID has been parsed -- accept no more IDs\n\tvar IDsExhausted bool\n\n\tfor _, item := range args {\n\t\tlcItem := strings.ToLower(item)\n\n\t\tif notesModeActivated {\n\t\t\t// no more parsing syntax\n\t\t\tnotes = append(notes, item)\n\t\t\tcontinue\n\t\t}\n\n\t\tif cmd == \"\" && StrSliceContains(ALL_CMDS, lcItem) {\n\t\t\tcmd = lcItem\n\t\t\tcontinue\n\t\t}\n\n\t\tif s, err := strconv.ParseInt(item, 10, 64); !IDsExhausted && err == nil {\n\t\t\tids = append(ids, int(s))\n\t\t\tcontinue\n\t\t}\n\n\t\tif item == IGNORE_CONTEXT_KEYWORD {\n\t\t\tignoreContext = true\n\t\t} else if item == NOTE_MODE_KEYWORD {\n\t\t\tnotesModeActivated = true\n\t\t} else if project == \"\" && strings.HasPrefix(lcItem, \"project:\") {\n\t\t\tproject = lcItem[8:]\n\t\t\t// invalid project filter, but a common mistake and it's obvious what\n\t\t\t// the user really means.\n\t\t} else if project == \"\" && strings.HasPrefix(lcItem, \"+project:\") {\n\t\t\tproject = lcItem[9:]\n\t\t} else if strings.HasPrefix(lcItem, \"-project:\") {\n\t\t\tantiProjects = append(antiProjects, lcItem[9:])\n\t\t} else if strings.HasPrefix(lcItem, \"template:\") {\n\t\t\tif s, err := strconv.ParseInt(lcItem[9:], 10, 64); err == nil {\n\t\t\t\ttemplate = int(s)\n\t\t\t}\n\t\t} else if len(item) > 1 && lcItem[0:1] == \"+\" {\n\t\t\ttags = append(tags, lcItem[1:])\n\t\t} else if len(item) > 1 && lcItem[0:1] == \"-\" {\n\t\t\tantiTags = append(antiTags, lcItem[1:])\n\t\t} else if priority == \"\" && IsValidPriority(item) {\n\t\t\tpriority = item\n\t\t} else {\n\t\t\twords = append(words, item)\n\t\t}\n\n\t\tIDsExhausted = true\n\t}\n\n\treturn Query{\n\t\tCmd: cmd,\n\t\tIDs: ids,\n\t\tTags: tags,\n\t\tAntiTags: antiTags,\n\t\tProject: project,\n\t\tAntiProjects: antiProjects,\n\t\tPriority: priority,\n\t\tTemplate: template,\n\t\tText: strings.Join(words, \" \"),\n\t\tNote: strings.Join(notes, \" \"),\n\t\tIgnoreContext: ignoreContext,\n\t}\n}", "func execAqlQuery(query string) (interface{}, error) {\n\turl := \"rest/v1/query/?aql=\" + url.QueryEscape(query)\n\tdata, err := HttpGetJsonThink(url)\n\terrorHandler(err, \"execAqlQuery()\")\n\n\treturn data, nil\n}", "func parseQuery(owner *user_model.User, query string) *conan_model.RecipeSearchOptions {\n\topts := &conan_model.RecipeSearchOptions{\n\t\tOwnerID: owner.ID,\n\t}\n\n\tif query != \"\" {\n\t\tparts := strings.Split(strings.ReplaceAll(query, \"@\", \"/\"), \"/\")\n\n\t\topts.Name = parts[0]\n\t\tif len(parts) > 1 && parts[1] != \"*\" {\n\t\t\topts.Version = parts[1]\n\t\t}\n\t\tif len(parts) > 2 && parts[2] != \"*\" {\n\t\t\topts.User = parts[2]\n\t\t}\n\t\tif len(parts) > 3 && parts[3] != \"*\" {\n\t\t\topts.Channel = parts[3]\n\t\t}\n\t}\n\n\treturn opts\n}", "func parseQuery(s string) (charName, query string, backup bool) {\n\tif m := rxQuery.FindStringSubmatch(s); m != nil {\n\t\tcharName = m[1]\n\t\tquery = m[2]\n\t\tif query == \"\" {\n\t\t\tbackup = true\n\t\t}\n\t\tquery = strings.TrimLeft(query, \" \")\n\t\treturn\n\t}\n\tquery = s\n\treturn\n}", "func parseUrlEncodedQueryParams(rawQuery string) (bson.M, map[string]interface{}) {\n\t\n\tqueryMap, _ := url.ParseQuery(rawQuery)\n\tvar query map[string]interface{}\n\tfor key, value := range queryMap {\n\t\tswitch key {\n\t\tdefault:\n\t\t\treturn bson.M{}, nil\n\t\tcase \"where\":\n\t\t\tif len(value) == 1 {\n\t\t\t\t\n\t\t\t\terr := json.Unmarshal([]byte(value[0]), &query)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn bson.M{}, map[string]interface{}{\"code\": helpers.INVALID_JSON, \"error\": \"invalid JSON\"}\n\t\t\t\t}\n\n\t\t\t} \n\t\tcase \"order\":\n\t\tcase \"limit\":\n\t\tcase \"skip\":\n\t\tcase \"keys\":\n\t\tcase \"include\":\n\t\t}\n\t}\n\t// fmt.Println(findObjectWithKey(query, \"$select\"))\n\t_ = parseWhereQuery(query)\n\terrMap := formatObjectQuery(query)\n\t// map can be used as bson.M for return\n\treturn query, errMap\n}", "func DeSerializeQuery(bytes []byte) Query {\n if len(bytes) != 32 {\n fmt.Println(\"Error : bytes length is not 32. Its \", len(bytes))\n }\n\n return Query {\n action : bytes[0],\n empty : 0,\n replyIp : binary.BigEndian.Uint32(bytes[2:6]),\n replyPort : binary.BigEndian.Uint16(bytes[6:8]),\n key : binary.BigEndian.Uint64(bytes[8:16]),\n value : binary.BigEndian.Uint64(bytes[16:24]),\n timeToLive: binary.BigEndian.Uint32(bytes[24:28]),\n requestId : binary.BigEndian.Uint32(bytes[28:32]),\n }\n}", "func (a *API) ParseQuery(ctx *fasthttp.RequestCtx) map[string]string {\n\tqs, _ := url.ParseQuery(string(ctx.URI().QueryString()))\n\tvalues := make(map[string]string)\n\tfor key, val := range qs {\n\t\tvalues[key] = val[0]\n\t}\n\n\treturn values\n}", "func ProcessQuery(query cqr.CommonQueryRepresentation, processor QueryProcessor) cqr.CommonQueryRepresentation {\n\tswitch q := query.(type) {\n\tcase cqr.Keyword:\n\t\tq.QueryString = processor(q.QueryString)\n\t\treturn q\n\tcase cqr.BooleanQuery:\n\t\tfor i, child := range q.Children {\n\t\t\tq.Children[i] = ProcessQuery(child, processor)\n\t\t}\n\t\treturn q\n\t}\n\treturn query\n}", "func parseQueryMeta(resp *http.Response, q *QueryMeta) error {\n\theader := resp.Header\n\n\t// Parse the X-Nomad-Index\n\tindex, err := strconv.ParseUint(header.Get(\"X-Nomad-Index\"), 10, 64)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to parse X-Nomad-Index: %v\", err)\n\t}\n\tq.LastIndex = index\n\n\t// Parse the X-Nomad-LastContact\n\tlast, err := strconv.ParseUint(header.Get(\"X-Nomad-LastContact\"), 10, 64)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to parse X-Nomad-LastContact: %v\", err)\n\t}\n\tq.LastContact = time.Duration(last) * time.Millisecond\n\tq.NextToken = header.Get(\"X-Nomad-NextToken\")\n\n\t// Parse the X-Nomad-KnownLeader\n\tswitch header.Get(\"X-Nomad-KnownLeader\") {\n\tcase \"true\":\n\t\tq.KnownLeader = true\n\tdefault:\n\t\tq.KnownLeader = false\n\t}\n\treturn nil\n}", "func promMQuery(prefix string, e *State, metric, groupBy, filter, agType, rateDuration, stepDuration, sdur, edur string) (r *Results, err error) {\n\tr = new(Results)\n\tprefixes := strings.Split(prefix, \",\")\n\tif len(prefixes) == 1 && prefixes[0] == \"\" {\n\t\treturn promQuery(\"default\", e, metric, groupBy, filter, agType, rateDuration, stepDuration, sdur, edur, true)\n\t}\n\n\twg := sync.WaitGroup{}\n\twg.Add(len(prefixes))\n\tresCh := make(chan *Results, len(prefixes))\n\terrCh := make(chan error, len(prefixes))\n\n\tfor _, prefix := range prefixes {\n\t\tgo func(prefix string) {\n\t\t\tdefer wg.Done()\n\t\t\tres, err := promQuery(prefix, e, metric, groupBy, filter, agType, rateDuration, stepDuration, sdur, edur, true)\n\t\t\tresCh <- res\n\t\t\terrCh <- err\n\t\t}(prefix)\n\t}\n\n\twg.Wait()\n\tclose(resCh)\n\tclose(errCh)\n\t// Gather errors from the request and return an error if any of the requests failled\n\terrors := []string{}\n\tfor err := range errCh {\n\t\tif err == nil {\n\t\t\tcontinue\n\t\t}\n\t\terrors = append(errors, err.Error())\n\t}\n\tif len(errors) > 0 {\n\t\treturn r, fmt.Errorf(strings.Join(errors, \" :: \"))\n\t}\n\tresultCollection := []*Results{}\n\tfor res := range resCh {\n\t\tresultCollection = append(resultCollection, res)\n\t}\n\tif len(resultCollection) == 1 { // no need to merge if there is only one item\n\t\treturn resultCollection[0], nil\n\t}\n\t// Merge the query results into a single seriesSet\n\tr, err = Merge(e, resultCollection...)\n\treturn\n}", "func ParseQuery(query string) (QueryParams, error) {\n\tmatch, _ := ValidateQuery(query)\n\tif !match {\n\t\treturn QueryParams{}, errors.New(\"Not a valid SQL query\")\n\t}\n\twords := strings.Fields(query[0 : len(query)-1])\n\tselectVal := words[1]\n\tfromVal := words[3]\n\torderByVal := \"\"\n\tfor i := 0; i < len(words); i++ {\n\t\tif strings.EqualFold(words[i], \"orderby\") {\n\t\t\torderByVal = strings.ToLower(words[i+1])\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn QueryParams{Select: selectVal, From: fromVal, OrderBy: orderByVal}, nil\n}", "func (s *ExpandQuery) Parse(req *http.Request) error {\n\tvar err error\n\n\tswitch req.Header.Get(contentType) {\n\tcase mimeJSON:\n\t\tbody, err := ioutil.ReadAll(req.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = json.Unmarshal(body, s); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase mimeForm:\n\t\tif err = req.ParseForm(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor key, val := range req.Form {\n\t\t\tswitch key {\n\t\t\tcase \"query\":\n\t\t\t\ts.Query = val\n\t\t\tcase \"jsonp\":\n\t\t\t\ts.JSONP = val[0]\n\t\t\tcase \"groupByExpr\":\n\t\t\t\ts.GroupByExpr, err = strconv.Atoi(val[0])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"leavesOnly\":\n\t\t\t\ts.LeavesOnly, err = strconv.Atoi(val[0])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\tparams := req.URL.Query()\n\ts.Query = append(s.Query, params[\"query\"]...)\n\tif len(params.Get(\"jsonp\")) != 0 {\n\t\ts.JSONP = params.Get(\"jsonp\")\n\t}\n\n\tif len(params.Get(\"groupByExpr\")) != 0 {\n\t\ts.GroupByExpr, err = strconv.Atoi(params.Get(\"groupByExpr\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif len(params.Get(\"leavesOnly\")) != 0 {\n\t\ts.LeavesOnly, err = strconv.Atoi(params.Get(\"leavesOnly\"))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func get_s3_select_query_params(query *Query) *s3.SelectObjectContentInput {\n\tparams := &s3.SelectObjectContentInput{\n\t\tBucket: aws.String(query.Bucket),\n\t\tKey: aws.String(query.Path),\n\t\tExpressionType: aws.String(s3.ExpressionTypeSql),\n\t\tExpression: aws.String(query.Query),\n\t\tInputSerialization: &s3.InputSerialization{\n\t\t\tCompressionType: aws.String(query.Compression),\n\t\t},\n\t\tOutputSerialization: &s3.OutputSerialization{\n\t\t\tJSON: &s3.JSONOutput{\n\t\t\t\tRecordDelimiter: aws.String(\",\"),\n\t\t\t},\n\t\t},\n\t}\n\n\tswitch query.Format {\n\tcase \"CSV\":\n\t\tparams.InputSerialization.CSV = &s3.CSVInput{}\n\t\tparams.InputSerialization.CSV.AllowQuotedRecordDelimiter = aws.Bool(query.CSVAllowQuotedRecordDelimiter)\n\t\tparams.InputSerialization.CSV.FileHeaderInfo = aws.String(query.CSVFileHeaderInfo)\n\n\t\tif query.CSVComments != \"\" {\n\t\t\tparams.InputSerialization.CSV.Comments = aws.String(query.CSVComments)\n\t\t}\n\t\tif query.CSVFieldDelimiter != \"\" {\n\t\t\tparams.InputSerialization.CSV.FieldDelimiter = aws.String(query.CSVFieldDelimiter)\n\t\t}\n\t\tif query.CSVQuoteCharacter != \"\" {\n\t\t\tparams.InputSerialization.CSV.QuoteCharacter = aws.String(query.CSVQuoteCharacter)\n\t\t}\n\t\tif query.CSVQuoteEscapeCharacter != \"\" {\n\t\t\tparams.InputSerialization.CSV.QuoteEscapeCharacter = aws.String(query.CSVQuoteEscapeCharacter)\n\t\t}\n\t\tif query.CSVRecordDelimiter != \"\" {\n\t\t\tparams.InputSerialization.CSV.RecordDelimiter = aws.String(query.CSVRecordDelimiter)\n\t\t}\n\tcase \"JSON\":\n\t\tparams.InputSerialization.JSON = &s3.JSONInput{}\n\t\tif query.JSONType != \"\" {\n\t\t\tparams.InputSerialization.JSON.Type = aws.String(query.JSONType)\n\t\t}\n\t}\n\n\treturn params\n}", "func MutateQuery(v map[string]interface{}) {\n\tclient := gql.NewClient(\"http://localhost:8080/v1/graphql\", nil)\n\tvar articles = repository.New()\n\terr := client.Execute(gql.Request{Query: `mutation update_articles($id: Int, $title: String, $body: String, $count: Int) {\n\t\tupdate_articles(\n\t\t\twhere: {id: {_eq: $id}},\n\t\t\t_set: {\n\t\t\t\tcount: $count\n\t\t\t}\n\t\t) {\n\t\t\taffected_rows\n\t\t}\n\t}`, Variables: v}, &articles)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfmt.Println(articles)\n}", "func ParseQuery(query string) (OrderedValues, error) {\n\tov := OrderedValues{}\n\tfor query != \"\" {\n\t\tif err := parseQuery(&ov, &query); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn ov, nil\n}", "func (d *Datastore) Query(q dsq.Query) (dsq.Results, error) {\n\td.l.Infow(\"executing query\", \"query\", q)\n\tif q.Orders != nil || q.Filters != nil {\n\t\treturn nil, fmt.Errorf(\"storj: filters or orders are not supported\")\n\t}\n\n\tlimit := q.Limit + q.Offset\n\t// disabling this makes tests fail, so we should\n\t// investigate what exactly disabling this does\n\tif limit == 0 || limit > listMax {\n\t\tlimit = listMax\n\t}\n\tresp, err := d.S3.ListObjectsV2(&s3.ListObjectsV2Input{\n\t\tBucket: aws.String(d.Bucket),\n\t\tPrefix: aws.String(d.s3Path(q.Prefix)),\n\t\tMaxKeys: aws.Int64(int64(limit)),\n\t})\n\tif err != nil {\n\t\td.l.Errorw(\"failed to list objects while running query\", \"error\", err)\n\t\treturn nil, err\n\t}\n\n\tindex := q.Offset\n\tnextValue := func() (dsq.Result, bool) {\n\t\tfor index >= len(resp.Contents) {\n\t\t\tif !*resp.IsTruncated {\n\t\t\t\treturn dsq.Result{}, false\n\t\t\t}\n\n\t\t\tindex -= len(resp.Contents)\n\n\t\t\tresp, err = d.S3.ListObjectsV2(&s3.ListObjectsV2Input{\n\t\t\t\tBucket: aws.String(d.Bucket),\n\t\t\t\tPrefix: aws.String(d.s3Path(q.Prefix)),\n\t\t\t\tDelimiter: aws.String(\"/\"),\n\t\t\t\tMaxKeys: aws.Int64(listMax),\n\t\t\t\tContinuationToken: resp.NextContinuationToken,\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\td.l.Errorw(\"failed to list objects while running query\", \"error\", err)\n\t\t\t\treturn dsq.Result{Error: err}, false\n\t\t\t}\n\t\t}\n\n\t\tentry := dsq.Entry{\n\t\t\tKey: ds.NewKey(*resp.Contents[index].Key).String(),\n\t\t}\n\t\tif !q.KeysOnly {\n\t\t\tvalue, err := d.Get(ds.NewKey(entry.Key))\n\t\t\tif err != nil {\n\t\t\t\td.l.Errorw(\"failed to get objects while running query\", \"error\", err)\n\t\t\t\treturn dsq.Result{Error: err}, false\n\t\t\t}\n\t\t\tentry.Value = value\n\t\t}\n\n\t\tindex++\n\t\treturn dsq.Result{Entry: entry}, true\n\t}\n\n\treturn dsq.ResultsFromIterator(q, dsq.Iterator{\n\t\tClose: func() error {\n\t\t\treturn nil\n\t\t},\n\t\tNext: nextValue,\n\t}), nil\n}", "func ParseQueryAndCacheContent(data, buffer interface{}) error {\n\tjsonEncoded, _ := json.Marshal(data)\n\treturn json.Unmarshal(jsonEncoded, &buffer)\n}", "func (g *GraphiteProvider) trimQuery(query string) string {\n\tspace := regexp.MustCompile(`\\s+`)\n\treturn space.ReplaceAllString(query, \" \")\n}", "func QueryParser(key string) Parser {\n\tfn := func(r *http.Request) (string, error) {\n\t\treturn internal.ParseQuery(key, r, ErrMissingPin)\n\t}\n\n\treturn pinFn(fn)\n}", "func ParseQuery(schema *graphql.Schema, query string, name string) (*ast.Document, []gqlerrors.FormattedError) {\n\tsource := source.NewSource(&source.Source{\n\t\tBody: []byte(query),\n\t\tName: name,\n\t})\n\tAST, err := parser.Parse(parser.ParseParams{Source: source})\n\tif err != nil {\n\t\treturn nil, gqlerrors.FormatErrors(err)\n\t}\n\n\tvalidationResult := graphql.ValidateDocument(schema, AST, nil)\n\tif validationResult.IsValid {\n\t\treturn AST, nil\n\t}\n\n\treturn nil, validationResult.Errors\n}", "func ParseDeprecatedQuery(pCtx Context, isDeprecated *bool) Func {\n\treturn func(input []rune) Result {\n\t\trootParser := OneOf(\n\t\t\tmatchExpressionParser(pCtx),\n\t\t\tifExpressionParser(pCtx),\n\t\t\tparseWithTails(bracketsExpressionParser(pCtx), pCtx),\n\t\t\tparseWithTails(literalValueParser(pCtx), pCtx),\n\t\t\tparseWithTails(functionParser(pCtx), pCtx),\n\t\t\tparseDeprecatedFunction(isDeprecated),\n\t\t)\n\n\t\tres := SpacesAndTabs()(input)\n\n\t\tres = arithmeticParser(rootParser)(res.Remaining)\n\t\tif res.Err != nil {\n\t\t\treturn Fail(res.Err, input)\n\t\t}\n\n\t\tresult := res.Payload\n\t\tres = SpacesAndTabs()(res.Remaining)\n\t\treturn Success(result, res.Remaining)\n\t}\n}", "func PromMQuery(prefix string, e *State, metric, groupBy, filter, agType, stepDuration, sdur, edur string) (r *Results, err error) {\n\treturn promMQuery(prefix, e, metric, groupBy, filter, agType, \"\", stepDuration, sdur, edur)\n}", "func preprocessQuery(userQuery string) (preprocessedString string) {\n\treg, err := regexp.Compile(\"[^a-zA-Z+#_]+\")\n\tif err != nil {\n\t\t// on failure return original query\n\t\treturn userQuery\n\t}\n\tpreprocessedString = reg.ReplaceAllString(userQuery, \" \")\n\n\treturn preprocessedString\n}", "func simplifyResultsMismatchedQuery(t *testing.T, query string) string {\n\tt.Helper()\n\tmcmp, closer := start(t)\n\tdefer closer()\n\n\t_, err := mcmp.ExecAllowAndCompareError(query)\n\tif err == nil {\n\t\tt.Fatalf(\"query (%s) does not error\", query)\n\t} else if !strings.Contains(err.Error(), \"mismatched\") {\n\t\tt.Fatalf(\"query (%s) does not error with results mismatched\\nError: %v\", query, err)\n\t}\n\n\trequire.NoError(t, utils.WaitForAuthoritative(t, keyspaceName, \"emp\", clusterInstance.VtgateProcess.ReadVSchema))\n\trequire.NoError(t, utils.WaitForAuthoritative(t, keyspaceName, \"dept\", clusterInstance.VtgateProcess.ReadVSchema))\n\n\tformal, err := vindexes.LoadFormal(\"svschema.json\")\n\trequire.NoError(t, err)\n\tvSchema := vindexes.BuildVSchema(formal)\n\tvSchemaWrapper := &vschemawrapper.VSchemaWrapper{\n\t\tV: vSchema,\n\t\tVersion: planbuilder.Gen4,\n\t}\n\n\tstmt, err := sqlparser.Parse(query)\n\trequire.NoError(t, err)\n\n\tsimplified := simplifier.SimplifyStatement(\n\t\tstmt.(sqlparser.SelectStatement),\n\t\tvSchemaWrapper.CurrentDb(),\n\t\tvSchemaWrapper,\n\t\tfunc(statement sqlparser.SelectStatement) bool {\n\t\t\tq := sqlparser.String(statement)\n\t\t\t_, newErr := mcmp.ExecAllowAndCompareError(q)\n\t\t\tif newErr == nil {\n\t\t\t\treturn false\n\t\t\t} else {\n\t\t\t\treturn strings.Contains(newErr.Error(), \"mismatched\")\n\t\t\t}\n\t\t},\n\t)\n\n\treturn sqlparser.String(simplified)\n}", "func (c *QueryClient) queryRequest(nrqlQuery string, queryResult interface{}) (err error) {\n\tvar request *http.Request\n\tvar response *http.Response\n\n\tqueryURL, err := c.generateQueryURL(nrqlQuery)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif queryResult == nil {\n\t\treturn errors.New(\"must have pointer for result\")\n\t}\n\n\trequest, err = http.NewRequest(\"GET\", queryURL, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\trequest.Header.Add(\"X-Query-Key\", c.QueryKey)\n\n\tclient := &http.Client{Timeout: c.RequestTimeout}\n\n\tresponse, err = client.Do(request)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"failed query request for: %v\", err)\n\t\treturn\n\t}\n\tdefer func() {\n\t\trespErr := response.Body.Close()\n\t\tif respErr != nil && err == nil {\n\t\t\terr = respErr // Don't mask previous errors\n\t\t}\n\t}()\n\n\tif response.StatusCode != http.StatusOK {\n\t\terr = fmt.Errorf(\"bad response code: %d\", response.StatusCode)\n\t\treturn\n\t}\n\n\terr = c.parseResponse(response, queryResult)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"failed query: %v\", err)\n\t}\n\n\treturn err\n}", "func Query(q string) func(r *Zego) {\n\treturn func(r *Zego) {\n\t\tr.query = q\n\t}\n}", "func PromQuery(prefix string, e *State, metric, groupBy, filter, agType, stepDuration, sdur, edur string) (r *Results, err error) {\n\treturn promQuery(prefix, e, metric, groupBy, filter, agType, \"\", stepDuration, sdur, edur, false)\n}", "func (r *ProtocolIncus) rawQuery(method string, url string, data any, ETag string) (*api.Response, string, error) {\n\tvar req *http.Request\n\tvar err error\n\n\t// Log the request\n\tlogger.Debug(\"Sending request to Incus\", logger.Ctx{\n\t\t\"method\": method,\n\t\t\"url\": url,\n\t\t\"etag\": ETag,\n\t})\n\n\t// Get a new HTTP request setup\n\tif data != nil {\n\t\tswitch data := data.(type) {\n\t\tcase io.Reader:\n\t\t\t// Some data to be sent along with the request\n\t\t\treq, err = http.NewRequestWithContext(r.ctx, method, url, data)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, \"\", err\n\t\t\t}\n\n\t\t\t// Set the encoding accordingly\n\t\t\treq.Header.Set(\"Content-Type\", \"application/octet-stream\")\n\t\tdefault:\n\t\t\t// Encode the provided data\n\t\t\tbuf := bytes.Buffer{}\n\t\t\terr := json.NewEncoder(&buf).Encode(data)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, \"\", err\n\t\t\t}\n\n\t\t\t// Some data to be sent along with the request\n\t\t\t// Use a reader since the request body needs to be seekable\n\t\t\treq, err = http.NewRequestWithContext(r.ctx, method, url, bytes.NewReader(buf.Bytes()))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, \"\", err\n\t\t\t}\n\n\t\t\t// Set the encoding accordingly\n\t\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\t\t\t// Log the data\n\t\t\tlogger.Debugf(logger.Pretty(data))\n\t\t}\n\t} else {\n\t\t// No data to be sent along with the request\n\t\treq, err = http.NewRequestWithContext(r.ctx, method, url, nil)\n\t\tif err != nil {\n\t\t\treturn nil, \"\", err\n\t\t}\n\t}\n\n\t// Set the ETag\n\tif ETag != \"\" {\n\t\treq.Header.Set(\"If-Match\", ETag)\n\t}\n\n\t// Send the request\n\tresp, err := r.DoHTTP(req)\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\n\tdefer func() { _ = resp.Body.Close() }()\n\n\treturn incusParseResponse(resp)\n}", "func CleanQuery(query string) string {\n\tret := strings.Replace(query, \"\\n\", \"\", -1)\n\tret = reRemoveExtraSpace.ReplaceAllString(ret, \" \")\n\treturn ret\n}", "func applyAdditionalQueryOptions(queryString string, queryLimit, querySkip int) (string, error) {\n\tconst jsonQueryFields = \"fields\"\n\tconst jsonQueryLimit = \"limit\"\n\tconst jsonQuerySkip = \"skip\"\n\t//create a generic map for the query json\n\tjsonQueryMap := make(map[string]interface{})\n\t//unmarshal the selector json into the generic map\n\tdecoder := json.NewDecoder(bytes.NewBuffer([]byte(queryString)))\n\tdecoder.UseNumber()\n\terr := decoder.Decode(&jsonQueryMap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif fieldsJSONArray, ok := jsonQueryMap[jsonQueryFields]; ok {\n\t\tswitch fieldsJSONArray.(type) {\n\t\tcase []interface{}:\n\t\t\t//Add the \"_id\" field, these are needed by default\n\t\t\tjsonQueryMap[jsonQueryFields] = append(fieldsJSONArray.([]interface{}),\n\t\t\t\tidField)\n\t\tdefault:\n\t\t\treturn \"\", fmt.Errorf(\"fields definition must be an array\")\n\t\t}\n\t}\n\t// Add limit\n\t// This will override any limit passed in the query.\n\t// Explicit paging not yet supported.\n\tjsonQueryMap[jsonQueryLimit] = queryLimit\n\t// Add skip\n\t// This will override any skip passed in the query.\n\t// Explicit paging not yet supported.\n\tjsonQueryMap[jsonQuerySkip] = querySkip\n\t//Marshal the updated json query\n\teditedQuery, err := json.Marshal(jsonQueryMap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tlogger.Debugf(\"Rewritten query: %s\", editedQuery)\n\treturn string(editedQuery), nil\n}", "func promQuery(prefix string, e *State, metric, groupBy, filter, agType, rateDuration, stepDuration, sdur, edur string, addPrefixTag bool) (r *Results, err error) {\n\tr = new(Results)\n\tstart, end, err := parseDurationPair(e, sdur, edur)\n\tif err != nil {\n\t\treturn\n\t}\n\tst, err := opentsdb.ParseDuration(stepDuration)\n\tif err != nil {\n\t\treturn\n\t}\n\tstep := time.Duration(st)\n\tqd := promQueryTemplateData{\n\t\tMetric: metric,\n\t\tAgFunc: agType,\n\t\tTags: groupBy,\n\t\tFilter: filter,\n\t\tRateDuration: rateDuration,\n\t}\n\tquery, err := qd.RenderString()\n\tqRes, err := timePromRequest(e, prefix, query, start, end, step)\n\tif err != nil {\n\t\treturn\n\t}\n\tgroupByTagSet := make(opentsdb.TagSet)\n\tfor _, v := range strings.Split(groupBy, \",\") {\n\t\tif v != \"\" {\n\t\t\tgroupByTagSet[v] = \"\"\n\t\t}\n\t}\n\terr = promMatrixToResults(prefix, e, qRes, len(groupByTagSet), addPrefixTag, r)\n\treturn r, err\n}", "func ReviewQuery(ctx context.Context, cfg *config.Config, urlQuery url.Values) (SearchURLParams, error) {\n\tvar validatedQueryParams SearchURLParams\n\n\tvalidatedQueryParams.Query = urlQuery.Get(\"q\")\n\n\terr := reviewPagination(ctx, cfg, urlQuery, &validatedQueryParams)\n\tif err != nil {\n\t\tlog.Event(ctx, \"unable to review pagination\", log.Error(err), log.ERROR)\n\t\treturn validatedQueryParams, err\n\t}\n\n\treviewSort(ctx, cfg, urlQuery, &validatedQueryParams)\n\n\terr = reviewFilters(ctx, urlQuery, &validatedQueryParams)\n\tif err != nil {\n\t\tlog.Event(ctx, \"unable to review filters\", log.Error(err), log.ERROR)\n\t\treturn validatedQueryParams, err\n\t}\n\n\treturn validatedQueryParams, nil\n}", "func (s *IndexQuery) Parse(req *http.Request) error {\n\tswitch req.Header.Get(contentType) {\n\tcase mimeJSON:\n\t\tbody, err := ioutil.ReadAll(req.Body)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := json.Unmarshal(body, s); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase mimeForm:\n\t\tif err := req.ParseForm(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfor key, val := range req.Form {\n\t\t\tswitch key {\n\t\t\tcase \"jsonp\":\n\t\t\t\ts.JSONP = val[0]\n\t\t\t}\n\t\t}\n\t}\n\n\tparams := req.URL.Query()\n\tif len(params.Get(\"jsonp\")) != 0 {\n\t\ts.JSONP = params.Get(\"jsonp\")\n\t}\n\n\treturn nil\n}", "func (k *Kuzzle) Query(query *types.KuzzleRequest, options types.QueryOptions, responseChannel chan<- *types.KuzzleResponse) {\n\tu := uuid.New()\n\trequestId := u.String()\n\n\tif query.RequestId == \"\" {\n\t\tquery.RequestId = requestId\n\t}\n\n\tif query.Body == nil {\n\t\tquery.Body = json.RawMessage(\"{}\")\n\t}\n\n\tif options == nil {\n\t\toptions = types.NewQueryOptions()\n\t}\n\n\tif options.Volatile() != nil {\n\t\tquery.Volatile = options.Volatile()\n\n\t\tmapped := make(map[string]interface{})\n\t\t_ = json.Unmarshal(query.Volatile, &mapped)\n\n\t\tmapped[\"sdkName\"] = version\n\n\t\tquery.Volatile, _ = json.Marshal(mapped)\n\n\t} else {\n\t\tvol := fmt.Sprintf(`{\"sdkName\": \"%s\"}`, version)\n\t\tquery.Volatile = types.VolatileData(vol)\n\t}\n\n\tjsonRequest, _ := json.Marshal(query)\n\n\tout := map[string]interface{}{}\n\terr := json.Unmarshal(jsonRequest, &out)\n\n\tif err != nil {\n\t\tif responseChannel != nil {\n\t\t\tresponseChannel <- &types.KuzzleResponse{Error: types.NewError(err.Error())}\n\t\t}\n\t\treturn\n\t}\n\n\trefresh := options.Refresh()\n\tif refresh != \"\" {\n\t\tout[\"refresh\"] = refresh\n\t}\n\n\tout[\"from\"] = options.From()\n\tout[\"size\"] = options.Size()\n\n\tscroll := options.Scroll()\n\tif scroll != \"\" {\n\t\tout[\"scroll\"] = scroll\n\t}\n\n\tscrollId := options.ScrollId()\n\tif scrollId != \"\" {\n\t\tout[\"scrollId\"] = scrollId\n\t}\n\n\tretryOnConflict := options.RetryOnConflict()\n\tif retryOnConflict > 0 {\n\t\tout[\"retryOnConflict\"] = retryOnConflict\n\t}\n\n\tjwt := k.Jwt()\n\tif jwt != \"\" {\n\t\tout[\"jwt\"] = jwt\n\t}\n\n\tif len(query.CustomArgs) != 0 {\n\t\tfor k, v := range query.CustomArgs {\n\t\t\tout[k] = v\n\t\t}\n\t}\n\n\tfinalRequest, err := json.Marshal(out)\n\n\tif err != nil {\n\t\tif responseChannel != nil {\n\t\t\tresponseChannel <- &types.KuzzleResponse{Error: types.NewError(err.Error())}\n\t\t}\n\t\treturn\n\t}\n\n\tqueueable := options == nil || options.Queuable()\n\tqueueable = queueable && k.queueFilter(finalRequest)\n\n\tif k.queuing {\n\t\tif queueable {\n\t\t\tk.cleanQueue()\n\t\t\tqo := &types.QueryObject{\n\t\t\t\tTimestamp: time.Now(),\n\t\t\t\tResChan: responseChannel,\n\t\t\t\tQuery: finalRequest,\n\t\t\t\tRequestId: requestId,\n\t\t\t\tOptions: options,\n\t\t\t}\n\t\t\tk.offlineQueue = append(k.offlineQueue, qo)\n\t\t\tk.EmitEvent(event.OfflineQueuePush, qo)\n\t\t\treturn\n\t\t}\n\n\t\tk.EmitEvent(event.Discarded, finalRequest)\n\t\tif responseChannel != nil {\n\t\t\tresponseChannel <- &types.KuzzleResponse{Status: 400, Error: types.NewError(\"Unable to execute request: not connected to a Kuzzle server.\\nDiscarded request: \"+string(finalRequest), 400)}\n\t\t}\n\t\treturn\n\t}\n\n\terr = k.protocol.Send(finalRequest, options, responseChannel, requestId)\n\n\tif err != nil {\n\t\tif responseChannel != nil {\n\t\t\tresponseChannel <- &types.KuzzleResponse{Error: types.NewError(err.Error())}\n\t\t}\n\t\treturn\n\t}\n}", "func ParseQuery(query *influxql.Query, now time.Time) (\n\tparsedQueries []tsdbjson.ParsedQuery,\n\tcolumnNameSets [][]string,\n\terr error) {\n\treturn parseQuery(query, now)\n}", "func UpdateSavedQuery(uri string, apikey string, id string, name string, query string) (*SavedQueryResponse, error) {\n\ttype During struct {\n\t\tTo interface{} `json:\"to\"`\n\t\tFrom interface{} `json:\"from\"`\n\t\tTimeRange interface{} `json:\"time_range\"`\n\t}\n\n\ttype Leql struct {\n\t\tDuring During `json:\"during\"`\n\t\tStatement string `json:\"statement\"`\n\t}\n\n\ttype SavedQuery struct {\n\t\tLogs []interface{} `json:\"logs\"`\n\t\tLeql Leql `json:\"leql\"`\n\t\tName string `json:\"name\"`\n\t}\n\n\ttype SavedQueryRequest struct {\n\t\tSavedQuery SavedQuery `json:\"saved_query\"`\n\t}\n\n\tqueryURI := strings.Join([]string{uri, id}, \"/\")\n\n\tbody := SavedQueryRequest{SavedQuery: SavedQuery{Name: name, Leql: Leql{Statement: query}}}\n\n\tbuf := new(bytes.Buffer)\n\tjson.NewEncoder(buf).Encode(body)\n\n\treq, err := http.NewRequest(\"PATCH\", queryURI, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"x-api-key\", apikey)\n\n\tclient := &http.Client{Timeout: time.Second * 10}\n\n\tres, err := client.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif res.StatusCode >= 400 {\n\t\tswitch res.StatusCode {\n\t\tcase http.StatusForbidden:\n\t\t\treturn nil, fmt.Errorf(\"You are not authorised to perform this action. [Status %v]\", http.StatusForbidden)\n\t\tcase http.StatusNotFound:\n\t\t\treturn nil, fmt.Errorf(\"Unable to locate saved query. [Status %v]\", http.StatusNotFound)\n\t\tcase http.StatusInternalServerError:\n\t\t\treturn nil, fmt.Errorf(\"Unable to modify the query at this time. [Status %d]\", http.StatusInternalServerError)\n\t\tcase http.StatusMethodNotAllowed:\n\t\t\tfallthrough\n\t\tcase http.StatusUnsupportedMediaType:\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"Unable to modify the query at this time. Please raise an issue. [Status %d]\", res.StatusCode)\n\t\t}\n\t}\n\n\tdefer res.Body.Close()\n\n\tvar response SavedQueryResponse\n\n\terr = json.NewDecoder(res.Body).Decode(&response)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &response, nil\n}", "func runGraphStoreQuery(orgID, graphName, inContentType string, in io.Reader) (body []byte, statusCode int, contentType string, err error) {\n\treq, err := http.NewRequest(http.MethodPut, c.Config.GetGraphStoreEndpoint(orgID, \"\"), in)\n\tif err != nil {\n\t\tlog.Printf(\"%s\", fmt.Errorf(\"%s; %w \", ErrInvalidSparqlRequest, err))\n\n\t\treturn\n\t}\n\treq.Header.Add(contentTypeKey, inContentType)\n\tparams := req.URL.Query()\n\tparams.Set(\"graph\", graphName)\n\treq.URL.RawQuery = params.Encode()\n\n\treturn makeSparqlRequest(req)\n}", "func TestBuildSimpleQuery(t *testing.T) {\n\tq := []QueryElement{\n\t\tQueryElement{Key: \"key0\", Op: \"op0\", Val: \"val0\"},\n\t\tQueryElement{Key: \"key1\", Op: \"op1\", Val: \"val1\"},\n\t}\n\ts := BuildQuery(\"http://localhost/v1/test\", q)\n\tu, err := url.Parse(s)\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tif u.Query().Get(\"k0\") != \"key0\" {\n\t\tt.Errorf(\"Issue with k0: %v\\n\", u.Query().Get(\"k0\"))\n\t}\n\tif u.Query().Get(\"o0\") != \"op0\" {\n\t\tt.Errorf(\"Issue with o0: %v\\n\", u.Query().Get(\"o0\"))\n\t}\n\tif u.Query().Get(\"v0\") != \"val0\" {\n\t\tt.Errorf(\"Issue with v0: %v\\n\", u.Query().Get(\"v0\"))\n\t}\n\tif u.Query().Get(\"k1\") != \"key1\" {\n\t\tt.Errorf(\"Issue with k1: %v\\n\", u.Query().Get(\"k1\"))\n\t}\n\tif u.Query().Get(\"o1\") != \"op1\" {\n\t\tt.Errorf(\"Issue with o1: %v\\n\", u.Query().Get(\"o1\"))\n\t}\n\tif u.Query().Get(\"v1\") != \"val1\" {\n\t\tt.Errorf(\"Issue with v1: %v\\n\", u.Query().Get(\"v1\"))\n\t}\n}", "func (c *Client) rawQuery(endpoint string, q *QueryOptions) (io.ReadCloser, error) {\n\tr, err := c.newRequest(\"GET\", endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.setQueryOptions(q)\n\t_, resp, err := requireOK(c.doRequest(r))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resp.Body, nil\n}", "func (ix *Reindexer) Query(q Query) *Reindexer {\n\tix.query = q\n\treturn ix\n}", "func parseBeforeQuery(p *parser) parserStateFn {\n\ttok := p.next()\n\tswitch tok.typ {\n\tcase tokTypeError:\n\t\tp.backup(tok)\n\t\treturn parseErrorTok\n\tcase tokTypeEOF:\n\t\tp.backup(tok)\n\t\treturn parseEOFTok\n\tcase tokTypeOpenParen:\n\t\t// Push the '(' onto the ops stack. It will be the marker at which to\n\t\t// stop when the ')' token is parsed.\n\t\tp.stagedOps.Push(tok)\n\t\treturn parseBeforeQuery\n\tcase tokTypeNot:\n\t\tp.stageBoolOp(tok)\n\t\tp.incompleteBoolOp = true\n\t\treturn parseBeforeQuery\n\tcase tokTypeUnquotedLiteral, tokTypeQuotedLiteral:\n\t\tp.incompleteBoolOp = false\n\t\tswitch tok2 := p.peek(); tok2.typ {\n\t\tcase tokTypeError:\n\t\t\treturn parseErrorTok\n\t\tcase tokTypeGt, tokTypeGte, tokTypeLt, tokTypeLte:\n\t\t\t// E.g.: `a.field >= 100`, `some.date.field < \"2021-02\"`\n\t\t\tif tok.typ == tokTypeQuotedLiteral {\n\t\t\t\treturn p.errorfAt(tok.pos, \"a *quoted* field for a range query is not yet supported\")\n\t\t\t}\n\t\t\tp.field = &tok\n\t\t\treturn parseRangeQuery\n\t\tcase tokTypeColon:\n\t\t\t// E.g.: `foo:value1 value2`, `foo:(a or b)`, `foo:(a and b and c)`,\n\t\t\t// `foo:*`\n\t\t\tif tok.typ == tokTypeQuotedLiteral {\n\t\t\t\treturn p.errorfAt(tok.pos, \"a *quoted* field for a term query is not yet supported\")\n\t\t\t}\n\t\t\tp.field = &tok\n\t\t\treturn parseTermsQuery\n\t\tdefault:\n\t\t\t// E.g.: `foo bar baz`\n\t\t\t// No range operator and no colon means this is a query without\n\t\t\t// a field name. In Kibana, this matches against \"default fields\".\n\t\t\ttermTok := tok\n\t\t\tvar terms []term\n\t\t\tfor {\n\t\t\t\tif termTok.typ == tokTypeUnquotedLiteral {\n\t\t\t\t\tterms = append(terms, newTerm(termTok.val))\n\t\t\t\t} else if termTok.typ == tokTypeQuotedLiteral {\n\t\t\t\t\tterms = append(terms, newQuotedTerm(termTok.val))\n\t\t\t\t} else {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\ttermTok = p.next()\n\t\t\t}\n\t\t\tp.backup(termTok)\n\t\t\tp.filter.addStep(&rpnDefaultFieldsTermsQuery{terms: terms})\n\t\t\treturn parseAfterQuery\n\t\t}\n\tdefault:\n\t\treturn p.errorfAt(tok.pos,\n\t\t\t\"expecting a literal, 'not', or '('; got %s\", tok.typ)\n\t}\n}", "func parseQuery(m *dns.Msg) {\n\tfor _, q := range m.Question {\n\t\tswitch q.Qtype {\n\t\tcase dns.TypeTXT:\n\t\t\t// Debug log\n\t\t\tif mConfig.Debug {\n\t\t\t\tlog.Printf(\"TXT Query for %s\\n\", q.Name)\n\t\t\t}\n\n\t\t\t// Get IP\n\t\t\treplacer := strings.NewReplacer(\n\t\t\t\t\".\"+mConfig.Suffix+\".\", \"\",\n\t\t\t\t\"x\", \":\",\n\t\t\t\t\"z\", \".\")\n\t\t\tip := replacer.Replace(q.Name)\n\n\t\t\t// Send response\n\t\t\tfor _, response := range g.GeoHandle(ip) {\n\t\t\t\tr := new(dns.TXT)\n\t\t\t\tr.Hdr = dns.RR_Header{\n\t\t\t\t\tName: q.Name,\n\t\t\t\t\tRrtype: dns.TypeTXT,\n\t\t\t\t\tClass: dns.ClassINET,\n\t\t\t\t\tTtl: 1,\n\t\t\t\t}\n\t\t\t\tr.Txt = []string{response}\n\t\t\t\tm.Answer = append(m.Answer, r)\n\t\t\t}\n\t\t}\n\t}\n}", "func executeQuery(query *mgo.Query, s *mongoRequest, coll *mgo.Collection) (interface{}, error) {\n\tgdata := new([]interface{})\n\tswitch s.Action {\n\tcase \"find\":\n\t\terr := query.All(gdata)\n\t\tif err != nil {\n\t\t\treturn []byte{}, err\n\t\t}\n\t\treturn json.Marshal(gdata)\n\tcase \"insert\":\n\t\tpayloadLen := len(s.JsonPayloadSlice)\n\t\tif payloadLen > 0 {\n\t\t\terr := coll.Insert(s.JsonPayloadSlice...)\n\t\t\tif err != nil {\n\t\t\t\treturn []byte{}, err\n\t\t\t}\n\t\t\tres := fmt.Sprintf(\"{\\\"nInserted\\\":%d}\", payloadLen)\n\t\t\treturn []byte(res), nil\n\t\t} else {\n\t\t\terr := coll.Insert(s.Args1)\n\t\t\tif err != nil {\n\t\t\t\treturn []byte{}, err\n\t\t\t}\n\t\t\treturn []byte(`{\"nInserted\":1}`), nil\n\t\t}\n\tcase \"remove\":\n\t\tif v, ok := s.Args2[\"justOne\"]; ok && v.(float64) == 1 {\n\t\t\terr := coll.Remove(s.Args1)\n\t\t\tif err != nil {\n\t\t\t\treturn []byte{}, err\n\t\t\t}\n\t\t\treturn []byte(`{\"nRemoved\":1}`), nil\n\t\t}\n\t\tinfo, err := coll.RemoveAll(s.Args1)\n\t\tif err != nil {\n\t\t\treturn []byte{}, err\n\t\t}\n\t\treturnString := fmt.Sprintf(\"{\\\"nRemoved\\\":%d}\", info.Removed)\n\t\treturn []byte(returnString), nil\n\tcase \"update\":\n\t\tif v, ok := s.Args3[\"upsert\"]; ok && v.(float64) == 1 {\n\t\t\tvar returnString string\n\t\t\tinfo, err := coll.Upsert(s.Args1, s.Args2)\n\t\t\tif err != nil {\n\t\t\t\treturn []byte{}, err\n\t\t\t}\n\t\t\tif info.Updated != 0 {\n\t\t\t\treturnString = `{\"nModified\":1}`\n\t\t\t} else {\n\t\t\t\treturnString = `{\"nUpserted\":1}`\n\t\t\t}\n\t\t\treturn []byte(returnString), nil\n\t\t}\n\t\tif v, ok := s.Args3[\"multi\"]; ok && v.(float64) == 1 {\n\t\t\tinfo, err := coll.UpdateAll(s.Args1, s.Args2)\n\t\t\tif err != nil {\n\t\t\t\treturn []byte{}, err\n\t\t\t}\n\t\t\treturnString := fmt.Sprintf(\"{\\\"nModified\\\":%d}\", info.Updated)\n\t\t\treturn []byte(returnString), nil\n\t\t}\n\t\terr := coll.Update(s.Args1, s.Args2)\n\t\tif err != nil {\n\t\t\treturn []byte{}, err\n\t\t}\n\t\treturn []byte(`{\"nModified\":1}`), nil\n\tcase \"count\":\n\t\tn, err := coll.Count()\n\t\tif err != nil {\n\t\t\treturn []byte{}, err\n\t\t}\n\t\tnumber := strconv.Itoa(n)\n\t\treturn number, nil\n\tdefault:\n\t\treturn []byte{}, fmt.Errorf(\"Unable to execute %s\", s.Action)\n\t}\n}", "func NewFromRawQuery(rawQuery string) *QueryString {\n\tqs := new(QueryString)\n\tqs.fields = make(map[string]string)\n\n\tfor {\n\t\ti := strings.IndexRune(rawQuery, '=')\n\t\tif i == -1 {\n\t\t\tbreak\n\t\t}\n\t\tname := rawQuery[:i]\n\t\trawQuery = rawQuery[i+1:]\n\n\t\ti = strings.IndexFunc(rawQuery, charClassDetector(1, 1))\n\t\tvar value string\n\t\tif i == -1 {\n\t\t\tvalue = rawQuery\n\t\t} else {\n\t\t\tvalue = rawQuery[:i]\n\t\t\trawQuery = rawQuery[i+1:]\n\t\t}\n\n\t\tqs.fields[name] = value\n\n\t\tif i == -1 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn qs\n}", "func (s *store) Query(query string) ([]*api.KeyValue, error) {\n\tpanic(\"not implemented\")\n}", "func (sem SEManager) Query(query string) (string, error) {\n words := strings.Split(query, \" \")\n\n switch len(words){\n case 0:\n return \"\", errors.New(\"Query strings does not contain any words\")\n case 1:\n return fmt.Sprintf(sem.Default.URL, words[0]), nil\n }\n\n se, err := sem.Get(words[0])\n if err != nil {\n return fmt.Sprintf(sem.Default.URL, strings.Join(words, \" \")), nil \n }\n return fmt.Sprintf(se.URL, strings.Join(words[1:], \" \")), nil\n}", "func (m *Main) Query(index, rawQuery, query string) (string, error) {\n\tresp := MustDo(\"POST\", m.URL()+fmt.Sprintf(\"/index/%s/query?\", index)+rawQuery, query)\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn \"\", fmt.Errorf(\"invalid status: %d, body=%s\", resp.StatusCode, resp.Body)\n\t}\n\treturn resp.Body, nil\n}", "func MetadataSearch(w http.ResponseWriter, r *http.Request) *appError {\n s3, err := getS3(r)\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n decoder := json.NewDecoder(r.Body)\n var query Query\n err = decoder.Decode(&query)\n if err != nil {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Can't decode JSON data\"}\n }\n path := \"/?query=\" + query.Query\n if query.Marker != \"\" {\n path += \"&marker=\" + query.Marker\n }\n if query.MaxKeys != \"\" {\n path += \"&max-keys=\" + query.MaxKeys\n }\n if query.SortedBy != \"\" {\n path += \"&sorted=\" + query.SortedBy\n }\n if query.ReturnAllMetadata {\n path += \"&attributes=ALL\"\n }\n bucketQueryResponse, err := s3Request(s3, query.Bucket, \"GET\", path, make(map[string][]string), \"\")\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n if bucketQueryResponse.Code == 200 {\n bucketQueryResult := &BucketQueryResult{}\n xml.NewDecoder(strings.NewReader(bucketQueryResponse.Body)).Decode(bucketQueryResult)\n // Generate a shared URL for each object returned by the metadata search\n if len(bucketQueryResult.EntryLists) > 0 {\n expires := time.Now().Add(time.Second*24*3600)\n for i, item := range bucketQueryResult.EntryLists {\n if item.ObjectName[len(item.ObjectName)-1:] != \"/\" {\n headers := make(map[string][]string)\n preparedS3Request, _ := prepareS3Request(s3, query.Bucket, \"GET\", query.Bucket + \"/\" + item.ObjectName + \"?Expires=\" + strconv.FormatInt(expires.Unix(), 10), headers, true)\n values := url.Values{}\n values = preparedS3Request.Params\n bucketQueryResult.EntryLists[i].Url = strings.Split(preparedS3Request.Url, \"?\")[0] + \"?\" + values.Encode()\n }\n }\n }\n rendering.JSON(w, http.StatusOK, bucketQueryResult)\n } else {\n return &appError{err: err, status: http.StatusInternalServerError, xml: bucketQueryResponse.Body}\n }\n\n return nil\n}", "func Interpret(toks []FastToken) (*Query, error) {\n\t/*\n\t * Will run the tokens through the rule match to get the rules to be run\n\t * Then will run the rules on the tokens\n\t */\n\t//running through rules for finding matches\n\trules := MatchRules(toks)\n\n\t//iterating through the rules to resolve them\n\tq := &Query{Tables: map[string]TableNode{}}\n\tfor _, rule := range rules {\n\t\tfor _, i := range rule.Matches {\n\t\t\tqu, err := rule.Resolve(*q, toks, i)\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(\"Couldn't apply the rule\", rule.Name, \"to the query at index\", i, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t*q = qu\n\t\t}\n\t}\n\treturn q, nil\n}", "func Search(w http.ResponseWriter, r *http.Request) *appError {\n session, err := store.Get(r, \"session-name\")\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n s3 := S3{\n EndPointString: session.Values[\"Endpoint\"].(string),\n AccessKey: session.Values[\"AccessKey\"].(string),\n SecretKey: session.Values[\"SecretKey\"].(string),\n Namespace: session.Values[\"Namespace\"].(string),\n }\n decoder := json.NewDecoder(r.Body)\n var query Query\n err = decoder.Decode(&query)\n if err != nil {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Can't decode JSON data\"}\n }\n\n imageWidth := \"0\"\n imageHeight := \"0\"\n if(query.Width != \"\") {\n imageWidth = query.Width\n }\n if(query.Height != \"\") {\n imageHeight = query.Height\n }\n path := \"\"\n if query.Area {\n path = \"/?query=x-amz-meta-image-width > \" + imageWidth + \" and x-amz-meta-image-height > \" + imageHeight + \" and x-amz-meta-gps-latitude > \" + query.SWLatitude + \" and x-amz-meta-gps-latitude < \" + query.NELatitude + \" and x-amz-meta-gps-longitude > \" + query.SWLongitude + \" and x-amz-meta-gps-longitude < \" + query.NELongitude + \"&attributes=Retention\"\n } else {\n path = \"/?query=x-amz-meta-image-width > \" + imageWidth + \" and x-amz-meta-image-height > \" + imageHeight + \"&attributes=Retention\"\n }\n bucketQueryResponse, err := s3Request(s3, query.Bucket, \"GET\", path, make(map[string][]string), \"\")\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n } else {\n bucketQueryResult := &BucketQueryResult{}\n xml.NewDecoder(strings.NewReader(bucketQueryResponse.Body)).Decode(bucketQueryResult)\n var pictures []Picture\n if len(bucketQueryResult.EntryLists) > 0 {\n for _, item := range bucketQueryResult.EntryLists {\n if item.ObjectName[len(item.ObjectName)-1:] != \"/\" {\n expires := time.Now().Add(time.Second*24*3600)\n pictureHeaders := make(map[string][]string)\n preparedPictureS3Request, _ := prepareS3Request(s3, query.Bucket, \"GET\", item.ObjectName + \"?Expires=\" + strconv.FormatInt(expires.Unix(), 10), pictureHeaders, true)\n pictureValues := url.Values{}\n pictureValues = preparedPictureS3Request.Params\n thumbnailHeaders := make(map[string][]string)\n preparedThumbnailS3Request, _ := prepareS3Request(s3, query.Bucket, \"GET\", strings.Replace(item.ObjectName, \"pictures/\", \"thumbnails/\", 1) + \"?Expires=\" + strconv.FormatInt(expires.Unix(), 10), thumbnailHeaders, true)\n thumbnailValues := url.Values{}\n thumbnailValues = preparedThumbnailS3Request.Params\n deletePictureHeaders := make(map[string][]string)\n preparedPictureS3DeleteRequest, _ := prepareS3Request(s3, query.Bucket, \"DELETE\", item.ObjectName, deletePictureHeaders, true)\n deleteThumbnailHeaders := make(map[string][]string)\n preparedThumbnailS3DeleteRequest, _ := prepareS3Request(s3, query.Bucket, \"DELETE\", strings.Replace(item.ObjectName, \"pictures/\", \"thumbnails/\", 1), deleteThumbnailHeaders, true)\n metadatas := map[string]string{}\n for _, metadata := range item.Metadatas {\n metadatas[metadata.Key] = metadata.Value\n }\n pictures = append(pictures, Picture{\n PictureUrl: strings.Split(preparedPictureS3Request.Url, \"?\")[0] + \"?\" + pictureValues.Encode(),\n ThumbnailUrl: strings.Split(preparedThumbnailS3Request.Url, \"?\")[0] + \"?\" + thumbnailValues.Encode(),\n PictureKey: item.ObjectName,\n DeleteRequestPictureHeaders: deletePictureHeaders,\n DeleteRequestPictureUrl: preparedPictureS3DeleteRequest.Url,\n DeleteRequestThumbnailHeaders: deleteThumbnailHeaders,\n DeleteRequestThumbnailUrl: preparedThumbnailS3DeleteRequest.Url,\n PictureMetadatas: metadatas,\n })\n }\n }\n } else {\n return &appError{err: err, status: http.StatusBadRequest, json: \"The specified search didn't return any result\"}\n }\n rendering.JSON(w, http.StatusOK, pictures)\n return nil\n }\n}", "func (app *Application) Query(reqQuery types.RequestQuery) (resQuery types.ResponseQuery) {\n lib.Log.Notice(reqQuery)\n// \tif reqQuery.Prove {\n// \t lib.Log.Notice(string(reqQuery.Data))\n\t // 此处修改 value, err := app.state.db.Get(prefixKey(reqQuery.Data))\n// \t\tvalue, err := app.state.db.Get(reqQuery.Data)\n//\n// \t\tif err != nil {\n// \t\t\tpanic(err)\n// \t\t}\n// \t\tif value == nil {\n// \t\t\tresQuery.Log = \"does not exist\"\n// \t\t} else {\n// \t\t\tresQuery.Log = \"exists\"\n// \t\t}\n// \t\tresQuery.Index = -1 // TODO make Proof return index\n// \t\tresQuery.Key = reqQuery.Data\n// \t\tresQuery.Value = value\n// \t\tresQuery.Height = app.state.Height\n//\n// \t\treturn resQuery\n// \t}\n lib.Log.Notice(string(reqQuery.Path))\n if reqQuery.Path == \"\" {\n resQuery.Key = reqQuery.Data\n // 此处修改 value, err := app.state.db.Get(prefixKey(reqQuery.Data))\n value, err := app.state.db.Get(reqQuery.Data)\n\n if err != nil {\n panic(err)\n }\n if value == nil {\n resQuery.Log = \"does not exist\"\n } else {\n resQuery.Log = \"exists\"\n }\n resQuery.Value = value\n resQuery.Height = app.state.Height\n\n\n }else{\n // 迭代器\n itr, e := app.state.db.Iterator(nil, nil)\n // 查询kv获取对应数据\n var build strings.Builder\n build.WriteString(\"[\")\n for ; itr.Valid(); itr.Next() {\n key := itr.Key()\n value := itr.Value()\n if strings.Index(string(key), reqQuery.Path) != -1 && strings.Index(string(value), string(reqQuery.Data)) != -1 {\n build.WriteString(string(value))\n build.WriteString(\",\")\n }\n }\n result := build.String()\n result = strings.TrimRight(result, \",\")\n result = result + \"]\"\n lib.Log.Notice(result)\n lib.Log.Notice(e)\n resQuery.Key = reqQuery.Data\n resQuery.Value = []byte(result)\n }\n\n return resQuery\n}", "func Query(s *vcloud.Session, opts Options) (*ResultRecords, error) {\n\tvar (\n\t\turl string\n\t\tqr ResultRecords\n\t)\n\n\t// opts.Href overrides a query URL\n\tif opts.Href != \"\" {\n\t\turl = opts.Href\n\t} else {\n\t\turl, _ = opts.Url(s)\n\t}\n\n\tbody, err := s.DoRequestGetBody(\"GET\", url, nil)\n\tfmt.Println(string(body))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = xml.Unmarshal(body, &qr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Assigns the Record interface to the actual Record struct\n\t// obtained from running the Query. Switch is\n\t// ordered by most-commonly-requested Query record\n\tswitch opts.Element.(type) {\n\tcase VAppTemplateRecord:\n\t\tqr.Records = qr.VAppTemplateRecords\n\tcase VAppRecord:\n\t\tqr.Records = qr.VAppRecords\n\tcase VAppNetworkRecord:\n\t\tqr.Records = qr.VAppNetworkRecords\n\tcase OrgVdcRecord:\n\t\tqr.Records = qr.OrgVdcRecords\n\tcase UserRecord:\n\t\tqr.Records = qr.UserRecords\n\tcase CatalogItemRecord:\n\t\tqr.Records = qr.CatalogItemRecords\n\tcase CatalogRecord:\n\t\tqr.Records = qr.CatalogRecords\n\tcase VMRecord:\n\t\tqr.Records = qr.VMRecords\n\tcase EventRecord:\n\t\tqr.Records = qr.EventRecords\n\tcase TaskRecord:\n\t\tqr.Records = qr.TaskRecords\n\tcase OrgNetworkRecord:\n\t\tqr.Records = qr.OrgNetworkRecords\n\tcase ApiDefinitionRecord:\n\t\tqr.Records = qr.ApiDefinitionRecords\n\tcase DiskRecord:\n\t\tqr.Records = qr.DiskRecords\n\tcase FileDescriptorRecord:\n\t\tqr.Records = qr.FileDescriptorRecords\n\tcase GroupRecord:\n\t\tqr.Records = qr.GroupRecords\n\tcase MediaRecord:\n\t\tqr.Records = qr.MediaRecords\n\tcase OrgVdcStorageProfileRecord:\n\t\tqr.Records = qr.OrgVdcStorageProfileRecords\n\tcase ServiceRecord:\n\t\tqr.Records = qr.ServiceRecords\n\tcase VmDiskRelationRecord:\n\t\tqr.Records = qr.VmDiskRelationRecords\n\t}\n\tfmt.Printf(\"%#v\\n\",qr)\n\treturn &qr, nil\n}", "func (m *Command) Query(t testing.TB, index, rawQuery, query string) (string, error) {\n\tresp := Do(t, \"POST\", fmt.Sprintf(\"%s/index/%s/query?%s\", m.URL(), index, rawQuery), query)\n\tif resp.StatusCode != gohttp.StatusOK {\n\t\treturn \"\", fmt.Errorf(\"invalid status: %d, body=%s\", resp.StatusCode, resp.Body)\n\t}\n\treturn resp.Body, nil\n}", "func RecommendationQuery(query string, callback func(string, autocompletion.Type, string)) {\n go func(query string) {\n s := &autocompletion.Sparql{ Buffer : query, Scope : scope }\n s.Init()\n autocompletion.Reset(s)\n err := s.Parse()\n if err == nil {\n s.Execute()\n callback(s.RecommendationQuery(), s.RecommendationType(), \"\")\n } else {\n callback(query, autocompletion.NONE, \"Unable to create recommendation query\\n\" + err.Error())\n }\n }(query)\n}", "func decodeRequestQuery(r *http.Request, v interface{}) error {\n\tif err := schema.NewDecoder().Decode(v, r.URL.Query()); err != nil {\n\t\tlog.WithField(\"err\", err).Info(\"Invalid request query\")\n\t\treturn err\n\t}\n\treturn nil\n}", "func parseDNSQuery(req []byte) (que *dnsQuestion, err error) {\n\thead := &dnsHeader{}\n\thead.getHeader(req)\n\tif !head.isAQuery() {\n\t\treturn nil, errors.New(\"not a dns query, ignore\")\n\t}\n\tque = &dnsQuestion{\n\t\tevent: eventNothing,\n\t}\n\t// Generally, when the recursive DNS server requests upward, it may\n\t// initiate a resolution request for multiple aliases/domain names\n\t// at once, Edge DNS does not need to process a message that carries\n\t// multiple questions at a time.\n\tif head.qdCount != 1 {\n\t\tque.event = eventUpstream\n\t\treturn\n\t}\n\n\toffset := uint16(unsafe.Sizeof(dnsHeader{}))\n\t// DNS NS <ROOT> operation\n\tif req[offset] == 0x0 {\n\t\tque.event = eventUpstream\n\t\treturn\n\t}\n\tque.getQuestion(req, offset, head)\n\terr = nil\n\treturn\n}", "func prepareQueryArg(ctx *context.APIContext, name string) (value string, err error) {\n\tvalue, err = url.PathUnescape(ctx.Query(name))\n\tvalue = strings.Trim(value, \" \")\n\treturn\n}", "func unmarshalWavefrontQuery(data []byte) (WavefrontQuery, error) {\n\tvar r WavefrontQuery\n\terr := json.Unmarshal(data, &r)\n\treturn r, err\n}", "func AutoQuery(jsonq *gojsonq.JSONQ, values url.Values, limit, offset int) interface{} {\n\tdata := jsonq.Copy()\n\tfor query, value := range values {\n\t\tif query == \"limit\" || query == \"offset\" {\n\t\t\tcontinue\n\t\t}\n\t\tfor _, el := range value {\n\t\t\top, cleanValue := prefixHandler(el)\n\t\t\tqueryBuilder(data, query, op, cleanValue)\n\t\t}\n\t}\n\tdata.Limit(limit).Offset(offset)\n\treturn data.Get()\n}", "func queryOperationDeserialize(result interface{}, id string, document map[string]interface{}, metadata map[string]interface{}, fieldsToFetch *fieldsToFetchToken, disableEntitiesTracking bool, session *InMemoryDocumentSessionOperations) error {\n\t_, ok := jsonGetAsBool(metadata, MetadataProjection)\n\tif !ok {\n\t\treturn session.TrackEntity(result, id, document, metadata, disableEntitiesTracking)\n\t}\n\ttp := reflect.TypeOf(result)\n\tpanicIf(tp.Kind() != reflect.Ptr, \"result should be a *<type>, is %T\", result)\n\tclazz := tp.Elem()\n\tif fieldsToFetch != nil && len(fieldsToFetch.projections) == 1 {\n\t\t// we only select a single field\n\t\tisString := clazz.Kind() == reflect.String\n\t\tif isString || isPrimitiveOrWrapper(clazz) || typeIsEnum(clazz) {\n\t\t\tprojectionField := fieldsToFetch.projections[0]\n\n\t\t\tif fieldsToFetch.sourceAlias != \"\" {\n\t\t\t\t// remove source-alias from projection name\n\t\t\t\tprojectionField = projectionField[len(fieldsToFetch.sourceAlias)+1:]\n\n\t\t\t}\n\n\t\t\tjsonNode, ok := document[projectionField]\n\t\t\tif ok && jsonIsValueNode(jsonNode) {\n\t\t\t\tres, err := treeToValue(clazz, jsonNode)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif res != nil {\n\t\t\t\t\treturn setInterfaceToValue(result, res)\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\n\t\tinner, ok := document[fieldsToFetch.projections[0]]\n\t\tif !ok {\n\t\t\treturn nil\n\t\t}\n\n\t\tif fieldsToFetch.fieldsToFetch != nil && fieldsToFetch.fieldsToFetch[0] == fieldsToFetch.projections[0] {\n\t\t\tdoc, ok := inner.(map[string]interface{})\n\t\t\tif ok {\n\t\t\t\t// extraction from original type\n\t\t\t\tdocument = doc\n\t\t\t}\n\t\t}\n\t}\n\n\tres, err := treeToValue(clazz, document)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif stringIsNotEmpty(id) {\n\t\t// we need to make an additional check, since it is possible that a value was explicitly stated\n\t\t// for the identity property, in which case we don't want to override it.\n\n\t\tidentityProperty := session.GetConventions().GetIdentityProperty(clazz)\n\t\tif identityProperty != \"\" {\n\t\t\tif _, ok := document[identityProperty]; !ok {\n\t\t\t\tsession.generateEntityIDOnTheClient.trySetIdentity(res, id)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn setInterfaceToValue(result, res)\n}", "func ParseQuery(query string, stmtid int) *QueryInfo {\n // Remove any comments that start before the command.\n if strings.HasPrefix(query, \"/*cmd*/\") {\n query = query[len(\"/*cmd*/\"):]\n }\n\n // Get the first word in the query, which is the command.\n words := strings.Fields(query)\n var columns []string\n var numParams int\n var numColumns int\n\n // Handle the command accordingly\n switch strings.ToLower(words[0]) {\n case \"select\":\n // WILL eventually fix to use a regex parser.\n split1 := strings.Split(query, \"from\")\n // If there was no from, then set the numParams to...0?\n // UNCLEAR\n if split1[0] == query {\n // this should be accurate\n numParams = 0\n columns = strings.Split(split1[0], \",\")\n numColumns = len(columns)\n } else {\n // Split the columns (listed after SELECT but before FROM)\n // // by ',''\n // columns = strings.Split(split1[0], \",\")\n // numParams = len(columns)\n\n\n // set temporarily to the com_stmt_prepare\n // expected in coordinator_basic\n numParams = 1\n numColumns = 2\n }\n\n case \"update\":\n // Isolate the string after keyword 'SET'\n split1 := strings.Split(query, \"set\")\n // Isolate the string before keyword 'WHERE'\n split2 := strings.Split(split1[1], \"WHERE\")\n // Split the string of column names separated by ','\n // which are between keywords SET and WHERE\n columns = strings.Split(split2[0], \",\")\n numColumns = len(columns)\n\n case \"insert\":\n // Isolate the string after '('\n split1 := strings.Split(query, \"(\")\n // Isolate the string before ')'\n split2 := strings.Split(split1[1], \")\")\n // Split the string of column names separated by ','\n // which are between table name and keyword VALUES\n columns = strings.Split(split2[0], \",\")\n numParams = len(columns)\n\n }\n\n // Return QueryInfo struct.\n return &QueryInfo{numParams:numParams, colNames:columns,\n stmtid:stmtid, query:query, numColumns:numColumns}\n}", "func XaFromQuery(qs url.Values) (*Xa, error) {\n\txa := &Xa{TransBase: *dtmimp.TransBaseFromQuery(qs)}\n\tif xa.Gid == \"\" || xa.BranchID == \"\" {\n\t\treturn nil, fmt.Errorf(\"bad xa info: gid: %s branchid: %s\", xa.Gid, xa.BranchID)\n\t}\n\treturn xa, nil\n}", "func resultToQueryItems(queryType string, results Results) ([]QueryItem, error) {\n\tresultSize := int64(results.Results.Total)\n\tif resultSize < 1 {\n\t\treturn nil, nil\n\t}\n\tvar items = make([]QueryItem, resultSize)\n\tswitch queryType {\n\tcase types.QtAdminCatalogItem:\n\t\tfor i, item := range results.Results.AdminCatalogItemRecord {\n\t\t\titems[i] = QueryCatalogItem(*item)\n\t\t}\n\tcase types.QtCatalogItem:\n\t\tfor i, item := range results.Results.CatalogItemRecord {\n\t\t\titems[i] = QueryCatalogItem(*item)\n\t\t}\n\tcase types.QtMedia:\n\t\tfor i, item := range results.Results.MediaRecord {\n\t\t\titems[i] = QueryMedia(*item)\n\t\t}\n\tcase types.QtAdminMedia:\n\t\tfor i, item := range results.Results.AdminMediaRecord {\n\t\t\titems[i] = QueryMedia(*item)\n\t\t}\n\tcase types.QtVappTemplate:\n\t\tfor i, item := range results.Results.VappTemplateRecord {\n\t\t\titems[i] = QueryVAppTemplate(*item)\n\t\t}\n\tcase types.QtAdminVappTemplate:\n\t\tfor i, item := range results.Results.AdminVappTemplateRecord {\n\t\t\titems[i] = QueryVAppTemplate(*item)\n\t\t}\n\tcase types.QtEdgeGateway:\n\t\tfor i, item := range results.Results.EdgeGatewayRecord {\n\t\t\titems[i] = QueryEdgeGateway(*item)\n\t\t}\n\tcase types.QtOrgVdcNetwork:\n\t\tfor i, item := range results.Results.OrgVdcNetworkRecord {\n\t\t\titems[i] = QueryOrgVdcNetwork(*item)\n\t\t}\n\tcase types.QtCatalog:\n\t\tfor i, item := range results.Results.CatalogRecord {\n\t\t\titems[i] = QueryCatalog(*item)\n\t\t}\n\tcase types.QtAdminCatalog:\n\t\tfor i, item := range results.Results.AdminCatalogRecord {\n\t\t\titems[i] = QueryAdminCatalog(*item)\n\t\t}\n\tcase types.QtVm:\n\t\tfor i, item := range results.Results.VMRecord {\n\t\t\titems[i] = QueryVm(*item)\n\t\t}\n\tcase types.QtAdminVm:\n\t\tfor i, item := range results.Results.AdminVMRecord {\n\t\t\titems[i] = QueryVm(*item)\n\t\t}\n\tcase types.QtVapp:\n\t\tfor i, item := range results.Results.VAppRecord {\n\t\t\titems[i] = QueryVapp(*item)\n\t\t}\n\tcase types.QtAdminVapp:\n\t\tfor i, item := range results.Results.AdminVAppRecord {\n\t\t\titems[i] = QueryVapp(*item)\n\t\t}\n\tcase types.QtOrgVdc:\n\t\tfor i, item := range results.Results.OrgVdcRecord {\n\t\t\titems[i] = QueryOrgVdc(*item)\n\t\t}\n\tcase types.QtAdminOrgVdc:\n\t\tfor i, item := range results.Results.OrgVdcAdminRecord {\n\t\t\titems[i] = QueryOrgVdc(*item)\n\t\t}\n\tcase types.QtTask:\n\t\tfor i, item := range results.Results.TaskRecord {\n\t\t\titems[i] = QueryTask(*item)\n\t\t}\n\tcase types.QtAdminTask:\n\t\tfor i, item := range results.Results.TaskRecord {\n\t\t\titems[i] = QueryAdminTask(*item)\n\t\t}\n\n\t}\n\tif len(items) > 0 {\n\t\treturn items, nil\n\t}\n\treturn nil, fmt.Errorf(\"unsupported query type %s\", queryType)\n}", "func redactQuerystring(ruleMatch HTTPMatch, u *url.URL) string {\n\tqueryValues := url.Values{}\n\n\tfor k, values := range u.Query() {\n\t\tfor _, v := range values {\n\t\t\tvalue := RedactedStr\n\t\t\tif ruleMatch.HasQuerystringWhitelistMatch(k) {\n\t\t\t\tvalue = v\n\t\t\t}\n\n\t\t\tqueryValues.Add(k, value)\n\t\t}\n\t}\n\n\treturn queryValues.Encode()\n}", "func RawQuery(q []byte) Term {\n\tdata := json.RawMessage(q)\n\treturn Term{\n\t\tname: \"RawQuery\",\n\t\trootTerm: true,\n\t\trawQuery: true,\n\t\tdata: &data,\n\t\targs: []Term{\n\t\t\tTerm{\n\t\t\t\ttermType: p.Term_DATUM,\n\t\t\t\tdata: string(q),\n\t\t\t},\n\t\t},\n\t}\n}", "func UnmarshalQuery(b []byte, v interface{}) *backend.DataResponse {\n\tif err := json.Unmarshal(b, v); err != nil {\n\t\treturn &backend.DataResponse{\n\t\t\tError: errors.Wrap(err, \"failed to unmarshal JSON request into query\"),\n\t\t}\n\t}\n\treturn nil\n}", "func TestQueryIntegration(t *testing.T) {\n\tqueryData := []byte{\n\t\t111, 222, // [0-1] query ID\n\t\t1, 0, // [2-3] flags, RD=1\n\t\t0, 1, // [4-5] QDCOUNT (number of queries) = 1\n\t\t0, 0, // [6-7] ANCOUNT (number of answers) = 0\n\t\t0, 0, // [8-9] NSCOUNT (number of authoritative answers) = 0\n\t\t0, 0, // [10-11] ARCOUNT (number of additional records) = 0\n\t\t// Start of first query\n\t\t7, 'y', 'o', 'u', 't', 'u', 'b', 'e',\n\t\t3, 'c', 'o', 'm',\n\t\t0, // null terminator of FQDN (DNS root)\n\t\t0, 1, // QTYPE = A\n\t\t0, 1, // QCLASS = IN (Internet)\n\t}\n\n\ttestQuery := func(queryData []byte) {\n\n\t\tdoh, err := NewTransport(testURL, ips, nil, nil, nil)\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\tresp, err2 := doh.Query(queryData)\n\t\tif err2 != nil {\n\t\t\tt.Fatal(err2)\n\t\t}\n\t\tif resp[0] != queryData[0] || resp[1] != queryData[1] {\n\t\t\tt.Error(\"Query ID mismatch\")\n\t\t}\n\t\tif len(resp) <= len(queryData) {\n\t\t\tt.Error(\"Response is short\")\n\t\t}\n\t}\n\n\ttestQuery(queryData)\n\n\tpaddedQueryBytes, err := AddEdnsPadding(simpleQueryBytes)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\ttestQuery(paddedQueryBytes)\n}", "func (encryptor *TokenizeQuery) OnQuery(ctx context.Context, query base.OnQueryObject) (base.OnQueryObject, bool, error) {\n\tlogrus.Debugln(\"TokenizeQuery.OnQuery\")\n\tstmt, err := query.Statement()\n\tif err != nil {\n\t\tlogrus.WithError(err).Debugln(\"Can't parse SQL statement\")\n\t\treturn query, false, err\n\t}\n\n\t// Extract the subexpressions that we are interested in for encryption.\n\t// The list might be empty for non-SELECT queries or for non-eligible SELECTs.\n\t// In that case we don't have any more work to do here.\n\titems := encryptor.searchableQueryFilter.FilterSearchableComparisons(stmt)\n\tif len(items) == 0 {\n\t\treturn query, false, nil\n\t}\n\tclientSession := base.ClientSessionFromContext(ctx)\n\tbindSettings := queryEncryptor.PlaceholderSettingsFromClientSession(clientSession)\n\tfor _, item := range items {\n\t\tif !item.Setting.IsTokenized() {\n\t\t\tcontinue\n\t\t}\n\n\t\trightVal, ok := item.Expr.Right.(*sqlparser.SQLVal)\n\t\tif !ok {\n\t\t\tlogrus.Debugln(\"expect SQLVal as Right expression for searchable consistent tokenization\")\n\t\t\tcontinue\n\t\t}\n\n\t\tencryptor.searchableQueryFilter.ChangeSearchableOperator(item.Expr)\n\n\t\terr = queryEncryptor.UpdateExpressionValue(ctx, item.Expr.Right, encryptor.coder, item.Setting, encryptor.getTokenizerDataWithSetting(item.Setting))\n\t\tif err != nil {\n\t\t\tlogrus.WithError(err).Debugln(\"Failed to update expression\")\n\t\t\treturn query, false, err\n\t\t}\n\n\t\tplaceholderIndex, err := queryEncryptor.ParsePlaceholderIndex(rightVal)\n\t\tif err == queryEncryptor.ErrInvalidPlaceholder {\n\t\t\tcontinue\n\t\t} else if err != nil {\n\t\t\treturn query, false, err\n\t\t}\n\t\tbindSettings[placeholderIndex] = item.Setting\n\t}\n\tlogrus.Debugln(\"TokenizeQuery.OnQuery changed query\")\n\treturn base.NewOnQueryObjectFromStatement(stmt, nil), true, nil\n}", "func MockQuery3() *types.Query {\n\tquery := &types.Query{\n\t\tDataSetName: mockWikiStatDataSet,\n\t\tTimeInterval: &types.TimeInterval{Name: \"date\", Start: \"2021-05-06\", End: \"2021-05-08\"},\n\t\tMetrics: []string{\"source_avg\"},\n\t\tDimensions: []string{\"project\", \"class_name\"},\n\t\tFilters: []*types.Filter{\n\t\t\t{OperatorType: types.FilterOperatorTypeNotIn, Name: \"path\", Value: []interface{}{\"*\"}},\n\t\t\t{OperatorType: types.FilterOperatorTypeIn, Name: \"project\", Value: []interface{}{\"city\", \"school\", \"music\"}},\n\t\t},\n\t\tOrders: []*types.OrderBy{\n\t\t\t{Name: \"project\", Direction: types.OrderDirectionTypeDescending},\n\t\t},\n\t}\n\treturn query\n}", "func (query *Query) CleanQuery() *Query {\n\tquery.content = make(map[string]interface{})\n\treturn query\n}", "func Query(w http.ResponseWriter, r *http.Request, format string) error {\n\n\treturn nil\n}", "func (r *Request) Query(q map[string]string) *Request {\n\tr.query = q\n\treturn r\n}", "func LoadRequestFromQuery(query string) Request {\n\tparsedQuery, _ := url.ParseQuery(query)\n\trequest := Request{}\n\trequest.Token = parsedQuery.Get(\"token\")\n\trequest.TeamID = parsedQuery.Get(\"team_id\")\n\trequest.TeamDomain = parsedQuery.Get(\"team_domain\")\n\trequest.ChannelID = parsedQuery.Get(\"channel_id\")\n\trequest.ChannelName = parsedQuery.Get(\"channel_name\")\n\trequest.UserID = parsedQuery.Get(\"user_id\")\n\trequest.UserName = parsedQuery.Get(\"user_name\")\n\trequest.Command = parsedQuery.Get(\"command\")\n\trequest.Text = parsedQuery.Get(\"text\")\n\trequest.ResponseURL = parsedQuery.Get(\"response_url\")\n\treturn request\n}", "func (tsr *TextSearchRequest) Query(qrfj *QueryResult) (string, error) {\n\n\t// Get data\n\tres, err := http.Get(tsr.Url) // post\n\n\tif err != nil {\n\t\treturn \"ERROR: \", err\n\t}\n\n\t// Check resposne code\n\tif res.StatusCode != 200 {\n\t\treturn \"ERROR: \", errors.New(res.Status)\n\t}\n\n\t//DEBUG\n\t// // Output response headers\n\t// for key, ele := range res.Header {\n\t// \tfmt.Print(key + \": \\n\")\n\t// \tfor key1, ele1 := range ele {\n\t// \t\tfmt.Print(\" \" + string(key1) + \":\" + ele1 + \"\\n\")\n\t// \t}\n\t// }\n\t//DEBUG^^^\n\n\t// Store data in it's type\n\tbody, err := ioutil.ReadAll(res.Body)\n\n\tif tsr.ResponseType == \"json\" {\n\t\tjson.Unmarshal(body, qrfj)\n\t} else if tsr.ResponseType == \"xml\" {\n\t\txml.Unmarshal(body, qrfj)\n\t}\n\tres.Body.Close()\n\tif err != nil {\n\t\treturn string(body), err\n\t}\n\n\t// store placaes in memory\t\n\n\t// return json\n\treturn string(body), nil\n\n}", "func (r *AnnounceRequest) FromQuery(vs url.Values) (err error) {\n\tif err = r.InfoHash.FromString(vs.Get(\"info_hash\")); err != nil {\n\t\treturn\n\t}\n\n\tif err = r.PeerID.FromString(vs.Get(\"peer_id\")); err != nil {\n\t\treturn\n\t}\n\n\tv, err := strconv.ParseInt(vs.Get(\"uploaded\"), 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\tr.Uploaded = v\n\n\tv, err = strconv.ParseInt(vs.Get(\"downloaded\"), 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\tr.Downloaded = v\n\n\tv, err = strconv.ParseInt(vs.Get(\"left\"), 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\tr.Left = v\n\n\tif s := vs.Get(\"event\"); s != \"\" {\n\t\tv, err := strconv.ParseUint(s, 10, 64)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.Event = uint32(v)\n\t}\n\n\tif s := vs.Get(\"port\"); s != \"\" {\n\t\tv, err := strconv.ParseUint(s, 10, 64)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.Port = uint16(v)\n\t}\n\n\tif s := vs.Get(\"numwant\"); s != \"\" {\n\t\tv, err := strconv.ParseUint(s, 10, 64)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.NumWant = int32(v)\n\t}\n\n\tif s := vs.Get(\"key\"); s != \"\" {\n\t\tv, err := strconv.ParseInt(s, 10, 64)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr.Key = int32(v)\n\t}\n\n\tr.IP = vs.Get(\"ip\")\n\tswitch vs.Get(\"compact\") {\n\tcase \"1\":\n\t\tr.Compact = true\n\tcase \"0\":\n\t\tr.Compact = false\n\t}\n\n\treturn\n}", "func parseError(err error) error {\n\tif s3Err, ok := err.(awserr.Error); ok && s3Err.Code() == s3.ErrCodeNoSuchKey {\n\t\treturn ds.ErrNotFound\n\t}\n\treturn err\n}", "func Query(\n\tctx context.Context,\n\tl log.Logger,\n\tendpoint *url.URL,\n\tt auth.TokenProvider,\n\tquery options.Query,\n\ttls options.TLS,\n\tdefaultStep time.Duration,\n) (int, promapiv1.Warnings, error) {\n\tvar (\n\t\twarn promapiv1.Warnings\n\t\terr error\n\t\trt *auth.BearerTokenRoundTripper\n\t)\n\n\tlevel.Debug(l).Log(\"msg\", \"running specified query\", \"name\", query.GetName(), \"query\", query.GetQuery())\n\n\t// Copy URL to avoid modifying the passed value.\n\tu := new(url.URL)\n\t*u = *endpoint\n\n\tif u.Scheme == transport.HTTPS {\n\t\ttp, err := transport.NewTLSTransport(l, tls)\n\t\tif err != nil {\n\t\t\treturn 0, warn, errors.Wrap(err, \"create round tripper\")\n\t\t}\n\n\t\trt = auth.NewBearerTokenRoundTripper(l, t, tp)\n\t} else {\n\t\trt = auth.NewBearerTokenRoundTripper(l, t, nil)\n\t}\n\n\tc, err := promapi.NewClient(promapi.Config{\n\t\tAddress: u.String(),\n\t\tRoundTripper: rt,\n\t})\n\tif err != nil {\n\t\terr = fmt.Errorf(\"create new API client: %w\", err)\n\t\treturn 0, warn, err\n\t}\n\n\treturn query.Run(ctx, c, l, rt.TraceID, defaultStep)\n}", "func handleQuery() func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tresult := graphql.Do(graphql.Params{\n\t\t\tSchema: createSchema(),\n\t\t\tRequestString: r.URL.Query().Get(\"query\"),\n\t\t})\n\t\terr := json.NewEncoder(w).Encode(result)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error Serializing result\")\n\t\t\tpanic(err)\n\t\t}\n\t}\n}", "func parse(body io.ReadCloser) ([]Result, error) {\n\tdefer body.Close()\n\n\tdoc, err := goquery.NewDocumentFromReader(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresults := []Result{}\n\tfor _, parserMethod := range defaultParserFunctions {\n\t\tresults = parserMethod(doc)\n\t\tif len(results) != 0 {\n\t\t\treturn results, nil\n\t\t}\n\t}\n\n\treturn results, nil\n}", "func Parse(q string, avail []string) (*Query, error) {\n\tvar i, esc, open, tail int\n\tvar c, seq rune\n\tvar out string\n\n\tif debug.TRACE {\n\t\tfmt.Println(\"<Q>\", q)\n\t}\n\n\tvar cols []string\n\tfor i, c = range q {\n\t\tswitch c {\n\n\t\tcase '\\\\':\n\t\t\tesc++\n\n\t\tcase '\\'', '\"':\n\t\t\tif seq == 0 {\n\t\t\t\tif esc%2 == 0 {\n\t\t\t\t\tif seq == 0 {\n\t\t\t\t\t\tseq = c\n\t\t\t\t\t} else {\n\t\t\t\t\t\tseq = 0\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if seq == c {\n\t\t\t\tif esc%2 == 0 {\n\t\t\t\t\tseq = 0\n\t\t\t\t}\n\t\t\t}\n\t\t\tesc = 0\n\n\t\tcase '{':\n\t\t\tif seq == 0 && esc%2 == 0 {\n\t\t\t\tif open != 0 {\n\t\t\t\t\treturn nil, fmt.Errorf(\"Meta sequence opened more than once\")\n\t\t\t\t}\n\t\t\t\topen = i\n\t\t\t}\n\t\t\tesc = 0\n\n\t\tcase '}':\n\t\t\tif seq == 0 && esc%2 == 0 {\n\t\t\t\tif open == 0 {\n\t\t\t\t\treturn nil, fmt.Errorf(\"Meta sequence closed but never opened\")\n\t\t\t\t}\n\t\t\t\tx, err := parse(q[open+1 : i])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\ts, e, err := emitPropertyList(x, avail)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tout = q[tail:open] + s\n\t\t\t\tcols = append(cols, e...)\n\t\t\t\topen = 0\n\t\t\t\ttail = i + 1\n\t\t\t}\n\t\t\tesc = 0\n\n\t\t}\n\t}\n\n\tif seq != 0 {\n\t\treturn nil, fmt.Errorf(\"Quote sequence never closed\")\n\t}\n\tif i-tail > 0 {\n\t\tout += q[tail:]\n\t}\n\n\tif debug.TRACE {\n\t\tfmt.Println(\"<R>\", out)\n\t}\n\n\treturn &Query{out, cols}, nil\n}", "func RegisterQuery(qr weave.QueryRouter) {\n\tNewBucket().Register(\"escrows\", qr)\n}", "func processQuery(userQuery string) (keywords string) {\n\tcandidates := rake.RunRake(userQuery)\n\tkeywords = \"\"\n\tfor _, candidate := range candidates {\n\t\tkeywords += candidate.Key + \";\"\n\t}\n\treturn keywords\n\n}", "func parseQueryOptions(sqlQuery string) ([]QueryOption, error) {\n\n\t//Parse matchOptions\n\tmatchOptions := regexOptions.FindString(sqlQuery)\n\n\t//If there are no options, return an empty slice\n\tif matchOptions == \"\" {\n\t\treturn make([]QueryOption, 0), nil\n\t}\n\n\t//Remove the beginning and ending identifiers\n\tmatchOptions = regexOptionIdentifier.ReplaceAllString(matchOptions, \"\")\n\n\t//JSON decode the matchOptions\n\toptions := make([]QueryOption, 0)\n\terr := json.Unmarshal([]byte(matchOptions), &options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn options, nil\n}", "func (pp *pipParser) ParseQuery(out *command.Out) ([]*resource.Pkg, error) {\n\treturn parseStream(out, parseQueryOut, pp.hinter.query, \"pip\")\n}", "func HandleRequest(query []byte, conn *DatabaseConnection) {\n\tlog.Printf(\"Handling raw query: %s\", query)\n\tlog.Printf(\"Parsing request...\")\n\trequest, err := grammar.ParseRequest(query)\n\tlog.Printf(\"Parsed request\")\n\tvar response grammar.Response\n\n\tif err != nil {\n\t\tlog.Printf(\"Error in request parsing! %s\", err.Error())\n\t\tresponse.Type = grammar.UNKNOWN_TYPE_RESPONSE\n\t\tresponse.Status = grammar.RESP_STATUS_ERR_INVALID_QUERY\n\t\tresponse.Data = err.Error()\n\t\tconn.Write(grammar.GetBufferFromResponse(response))\n\t}\n\n\tswitch request.Type {\n\tcase grammar.AUTH_REQUEST:\n\t\t// AUTH {username} {password}\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_AUTH_REQUEST, false, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in AUTH request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\t\tusername := request.RequestData[0]\n\t\tpassword := request.RequestData[1]\n\t\t// bucketname := tokens[2]\n\t\tlog.Printf(\"Client wants to authenticate.<username>:<password> %s:%s\", username, password)\n\n\t\tauthRequest := AuthRequest{Username: username, Password: password, Conn: conn}\n\t\tresponse = processAuthRequest(authRequest)\n\tcase grammar.SET_REQUEST:\n\t\t// SET {key} {value} [ttl] [nooverride]\n\t\trequest.Type = grammar.SET_RESPONSE\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_SET_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in SET request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tkey := request.RequestData[0]\n\t\tvalue := request.RequestData[1]\n\t\tlog.Printf(\"Setting %s:%s\", key, value)\n\t\tsetRequest := SetRequest{Key: key, Value: value, Conn: conn}\n\t\tresponse = processSetRequest(setRequest)\n\n\tcase grammar.GET_REQUEST:\n\t\t// GET {key}\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_GET_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in GET request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tkey := request.RequestData[0]\n\t\tlog.Printf(\"Client wants to get key '%s'\", key)\n\t\tgetRequest := GetRequest{Key: key, Conn: conn}\n\t\tresponse = processGetRequest(getRequest)\n\n\tcase grammar.DELETE_REQUEST:\n\t\t// DELETE {key}\n\t\tlog.Println(\"Client wants to delete a bucket/key\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_DELETE_REQUEST, true, true)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in DELETE request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\t\t// TODO implement\n\tcase grammar.CREATE_BUCKET_REQUEST:\n\t\tlog.Println(\"Client wants to create a bucket\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_CREATE_BUCKET_REQUEST, true, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in CREATE bucket request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tbucketName := request.RequestData[0]\n\t\tcreateBucketRequest := CreateBucketRequest{BucketName: bucketName, Conn: conn}\n\n\t\tresponse = processCreateBucketRequest(createBucketRequest)\n\tcase grammar.CREATE_USER_REQUEST:\n\t\tlog.Printf(\"Client wants to create a user\")\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_CREATE_USER_REQUEST, false, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in CREATE user request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tusername := request.RequestData[0]\n\t\tpassword := request.RequestData[1]\n\t\tcreateUserRequest := CreateUserRequest{Username: username, Password: password, Conn: conn}\n\n\t\tresponse = processCreateUserRequest(createUserRequest)\n\tcase grammar.USE_REQUEST:\n\t\terrorStatus := checkRequirements(request, conn, grammar.LENGTH_OF_USE_REQUEST, true, false)\n\t\tif errorStatus != 0 {\n\t\t\tlog.Printf(\"Error in USE request! %d\", errorStatus)\n\t\t\tresponse.Status = errorStatus\n\t\t\tbreak\n\t\t}\n\n\t\tbucketname := request.RequestData[0]\n\t\tif bucketname == SALTS_BUCKET || bucketname == USERS_BUCKET {\n\t\t\tresponse.Status = grammar.RESP_STATUS_ERR_UNAUTHORIZED\n\t\t\tbreak\n\t\t}\n\n\t\tuseRequest := UseRequest{BucketName: bucketname, Conn: conn}\n\t\tresponse = processUseRequest(useRequest)\n\tdefault:\n\t\tlog.Printf(illegalRequestTemplate, request.Type)\n\t\tresponse.Type = grammar.UNKNOWN_TYPE_RESPONSE\n\t\tresponse.Status = grammar.RESP_STATUS_ERR_UNKNOWN_COMMAND\n\t}\n\tif response.Status != 0 {\n\t\tlog.Printf(\"Error in request. status: %d\", response.Status)\n\t}\n\tconn.Write(grammar.GetBufferFromResponse(response))\n\tlog.Printf(\"Wrote buffer: %s to client\", grammar.GetBufferFromResponse(response))\n\n}", "func (c *QueryClient) Query(nrqlQuery string, response interface{}) (err error) {\n\tif response == nil {\n\t\treturn errors.New(\"go-insights: Invalid query response can not be nil\")\n\t}\n\n\terr = c.queryRequest(nrqlQuery, response)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *Command) Queryf(t *testing.T, index, rawQuery, query string, params ...interface{}) (string, error) {\n\tquery = fmt.Sprintf(query, params...)\n\tresp := Do(t, \"POST\", fmt.Sprintf(\"%s/index/%s/query?%s\", m.URL(), index, rawQuery), query)\n\tif resp.StatusCode != gohttp.StatusOK {\n\t\treturn \"\", fmt.Errorf(\"invalid status: %d, body=%s\", resp.StatusCode, resp.Body)\n\t}\n\treturn resp.Body, nil\n}", "func (r *ProtocolIncus) RawQuery(method string, path string, data any, ETag string) (*api.Response, string, error) {\n\t// Generate the URL\n\turl := fmt.Sprintf(\"%s%s\", r.httpBaseURL.String(), path)\n\n\treturn r.rawQuery(method, url, data, ETag)\n}", "func (q queryManager) processQuery(sql string, pubKey []byte, executeifallowed bool) (uint, []byte, []byte, *structures.Transaction, error) {\n\tlocalError := func(err error) (uint, []byte, []byte, *structures.Transaction, error) {\n\t\treturn SQLProcessingResultError, nil, nil, nil, err\n\t}\n\tqp := q.getQueryParser()\n\t// this will get sql type and data from comments. data can be pubkey, txBytes, signature\n\tqparsed, err := qp.ParseQuery(sql)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\t// maybe this query contains signature and txData from previous calls\n\tif len(qparsed.Signature) > 0 && len(qparsed.TransactionBytes) > 0 {\n\t\t// this is a case when signature and txdata were part of SQL comments.\n\t\ttx, err := q.processQueryWithSignature(qparsed.TransactionBytes, qparsed.Signature, executeifallowed)\n\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\treturn SQLProcessingResultTranactionComplete, nil, nil, tx, nil\n\t}\n\n\tneedsTX, err := q.checkQueryNeedsTransaction(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif !needsTX {\n\t\tif !executeifallowed {\n\t\t\t// no need to execute query. just return\n\t\t\treturn SQLProcessingResultExecuted, nil, nil, nil, nil\n\t\t}\n\t\t// no need to have TX\n\t\tif qparsed.IsUpdate() {\n\t\t\t_, err := qp.ExecuteQuery(qparsed.SQL)\n\t\t\tif err != nil {\n\t\t\t\treturn localError(err)\n\t\t\t}\n\t\t}\n\t\treturn SQLProcessingResultExecuted, nil, nil, nil, nil\n\t}\n\t// decide which pubkey to use.\n\n\t// first priority for a key posted as argument, next is the key in SQL comment (parsed) and final is the key\n\t// provided to thi module\n\tif len(pubKey) == 0 {\n\t\tif len(qparsed.PubKey) > 0 {\n\t\t\tpubKey = qparsed.PubKey\n\t\t} else if len(q.pubKey) > 0 {\n\t\t\tpubKey = q.pubKey\n\t\t} else {\n\t\t\t// no pubkey to use. return notice about pubkey required\n\t\t\treturn SQLProcessingResultPubKeyRequired, nil, nil, nil, nil\n\t\t}\n\t}\n\n\t// check if the key has permissions to execute this query\n\thasPerm, err := q.checkExecutePermissions(qparsed, pubKey)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif !hasPerm {\n\t\treturn localError(errors.New(\"No permissions to execute this query\"))\n\t}\n\n\tamount, err := q.checkQueryNeedsPayment(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\t// prepare SQL part of a TX\n\t// this builds RefID for a TX update\n\tsqlUpdate, err := qp.MakeSQLUpdateStructure(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\t// prepare curency TX and add SQL part\n\n\ttxBytes, datatosign, err := q.getTransactionsManager().PrepareNewSQLTransaction(pubKey, sqlUpdate, amount, \"MINTER\")\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\ttx, err := structures.DeserializeTransaction(txBytes)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif len(q.pubKey) > 0 && bytes.Compare(q.pubKey, pubKey) == 0 {\n\t\t// transaction was created by internal pubkey. we have private key for it\n\t\tsignature, err := utils.SignDataByPubKey(q.pubKey, q.privKey, datatosign)\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\ttx, err = q.processQueryWithSignature(txBytes, signature, executeifallowed)\n\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\treturn SQLProcessingResultTranactionCompleteInternally, nil, nil, tx, nil\n\t}\n\treturn SQLProcessingResultSignatureRequired, txBytes, datatosign, nil, nil\n}", "func ReadQueryIntoStruct(request *http.Request, dest interface{}, ignoreMissing bool) error {\n\tquery := request.URL.Query()\n\tif query == nil {\n\t\treturn errors.New(\"Request has no query parameters\")\n\t}\n\tfor k, v := range query {\n\t\tif len(v) == 1 {\n\t\t\tquery[k] = strings.Split(v[0], \",\")\n\t\t}\n\t}\n\treturn ReadMapIntoStruct(query, dest, ignoreMissing)\n}", "func trimToQuery(steps []Step, runQuery string, targets []Target) ([]Step, []TargetStatus) {\n\trunQueryParts := strings.Split(runQuery, \"::\")\n\tif len(runQueryParts) != 2 {\n\t\terr := fmt.Errorf(errorRunQueryArgument)\n\t\treturn nil, makeTargetStatuses(err, targets)\n\t}\n\n\tvar stepName, queryName string = runQueryParts[0], runQueryParts[1]\n\tif stepName == \"\" || queryName == \"\" {\n\t\terr := fmt.Errorf(errorRunQueryArgument)\n\t\treturn nil, makeTargetStatuses(err, targets)\n\t}\n\n\tsteps, trimErr := trimSteps(steps, stepName, targets)\n\tif trimErr != nil {\n\t\treturn nil, trimErr\n\t}\n\n\tstep := steps[0] // safe\n\tqueries := []Query{}\n\tfor _, query := range step.Queries {\n\t\tif query.Name == queryName {\n\t\t\tqueries = append(queries, query)\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif len(queries) == 0 {\n\t\terr := fmt.Errorf(\"%s: '%s'\", errorRunQueryNotFound, queryName)\n\t\treturn nil, makeTargetStatuses(err, targets)\n\t}\n\tstep.Queries = queries\n\n\treturn []Step{step}, nil\n}", "func (w *Wrapper) RawQuery(query string, values ...interface{}) (err error) {\n\tw.query = query\n\tw.params = values\n\t_, err = w.runQuery()\n\treturn\n}", "func QueryInputString(rd *bufio.Reader, query string) string {\n\tfmt.Println(query)\n\tfmt.Printf(\">\")\n\treturn strings.TrimSpace(ReadInputString(rd))\n}", "func (r *request) setQueryOptions(q *QueryOptions) {\n\tif q == nil {\n\t\treturn\n\t}\n\tif q.Region != \"\" {\n\t\tr.params.Set(\"region\", q.Region)\n\t}\n\tif q.Namespace != \"\" {\n\t\tr.params.Set(\"namespace\", q.Namespace)\n\t}\n\tif q.AuthToken != \"\" {\n\t\tr.token = q.AuthToken\n\t}\n\tif q.AllowStale {\n\t\tr.params.Set(\"stale\", \"\")\n\t}\n\tif q.WaitIndex != 0 {\n\t\tr.params.Set(\"index\", strconv.FormatUint(q.WaitIndex, 10))\n\t}\n\tif q.WaitTime != 0 {\n\t\tr.params.Set(\"wait\", durToMsec(q.WaitTime))\n\t}\n\tif q.Prefix != \"\" {\n\t\tr.params.Set(\"prefix\", q.Prefix)\n\t}\n\tif q.Filter != \"\" {\n\t\tr.params.Set(\"filter\", q.Filter)\n\t}\n\tif q.PerPage != 0 {\n\t\tr.params.Set(\"per_page\", fmt.Sprint(q.PerPage))\n\t}\n\tif q.NextToken != \"\" {\n\t\tr.params.Set(\"next_token\", q.NextToken)\n\t}\n\tif q.Reverse {\n\t\tr.params.Set(\"reverse\", \"true\")\n\t}\n\tfor k, v := range q.Params {\n\t\tr.params.Set(k, v)\n\t}\n\tr.ctx = q.Context()\n\n\tfor k, v := range q.Headers {\n\t\tr.header.Set(k, v)\n\t}\n}", "func FromQuery(key string) TokenExtractor {\n\treturn func(r *http.Request) (string, error) {\n\t\treturn r.URL.Query().Get(key), nil\n\t}\n}", "func (t *SimpleChaincode) queryAll(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tif len(args) != 0 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 0\")\n\t}\n\n resultsIterator, err := stub.GetStateByRange(\"\",\"\")\n if err != nil {\n return shim.Error(err.Error())\n }\n defer resultsIterator.Close()\n\n // buffer is a JSON array containing QueryResults\n var buffer bytes.Buffer\n buffer.WriteString(\"\\n[\")\n\n\tbArrayMemberAlreadyWritten := false\n for resultsIterator.HasNext() {\n queryResponse, err := resultsIterator.Next()\n if err != nil {\n return shim.Error(err.Error())\n }\n // Add a comma before array members, suppress it for the first array member\n if bArrayMemberAlreadyWritten == true {\n buffer.WriteString(\",\")\n }\n buffer.WriteString(\"{\\\"Key\\\":\")\n buffer.WriteString(\"\\\"\")\n buffer.WriteString(queryResponse.Key)\n buffer.WriteString(\"\\\"\")\n\n buffer.WriteString(\", \\\"Record\\\":\")\n // Record is a JSON object, so we write as-is\n buffer.WriteString(string(queryResponse.Value))\n buffer.WriteString(\"}\")\n bArrayMemberAlreadyWritten = true\n }\n buffer.WriteString(\"]\\n\")\n return shim.Success(buffer.Bytes())\n}", "func (util copyHandlerUtil) redactSigQueryParam(rawQuery string) (bool, string) {\n\trawQuery = strings.ToLower(rawQuery) // lowercase the string so we can look for ?sig= and &sig=\n\tsigFound := strings.Contains(rawQuery, \"?sig=\")\n\tif !sigFound {\n\t\tsigFound = strings.Contains(rawQuery, \"&sig=\")\n\t\tif !sigFound {\n\t\t\treturn sigFound, rawQuery // [?|&]sig= not found; return same rawQuery passed in (no memory allocation)\n\t\t}\n\t}\n\t// [?|&]sig= found, redact its value\n\tvalues, _ := url.ParseQuery(rawQuery)\n\tfor name := range values {\n\t\tif strings.EqualFold(name, \"sig\") {\n\t\t\tvalues[name] = []string{\"REDACTED\"}\n\t\t}\n\t}\n\treturn sigFound, values.Encode()\n}", "func TestQueryParser(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tinputMV2Query string\n\t\texpectedMetricSelector string\n\t\texpectedEntitySelector string\n\t\texpectedUnit string\n\t\texpectError bool\n\t}{\n\t\t// these should fail\n\t\t{\n\t\t\tname: \"percent unit does not work\",\n\t\t\tinputMV2Query: \"MV2;Percent;metricSelector=builtin:host.cpu.usage:merge(\\\"dt.entity.host\\\"):avg:names&entitySelector=type(HOST)\",\n\t\t\texpectError: true,\n\t\t},\n\t\t{\n\t\t\tname: \"missing microsecond metric unit\",\n\t\t\tinputMV2Query: \"MV2;metricSelector=builtin:service.response.server:filter(and(in(\\\"dt.entity.service\\\",entitySelector(\\\"type(service),tag(~\\\"KeptnQualityGate~\\\")\\\")))):splitBy():percentile(90)\",\n\t\t\texpectError: true,\n\t\t},\n\t\t{\n\t\t\tname: \"missing mv2 prefix\",\n\t\t\tinputMV2Query: \"MicroSecond;metricSelector=builtin:service.response.server:filter(and(in(\\\"dt.entity.service\\\",entitySelector(\\\"type(service),tag(~\\\"KeptnQualityGate~\\\")\\\")))):splitBy():percentile(90)\",\n\t\t\texpectError: true,\n\t\t},\n\t\t{\n\t\t\tname: \"missing mv2 prefix\",\n\t\t\tinputMV2Query: \"MV2;MicroSeconds;metricSelector=builtin:service.response.server:filter(and(in(\\\"dt.entity.service\\\",entitySelector(\\\"type(service),tag(~\\\"KeptnQualityGate~\\\")\\\")))):splitBy():percentile(90)\",\n\t\t\texpectError: true,\n\t\t},\n\t\t// these should not fail\n\t\t{\n\t\t\tname: \"microsecond metric works\",\n\t\t\tinputMV2Query: \"MV2;MicroSecond;metricSelector=builtin:service.response.server:filter(and(in(\\\"dt.entity.service\\\",entitySelector(\\\"type(service),tag(~\\\"KeptnQualityGate~\\\")\\\")))):splitBy():percentile(90)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.server:filter(and(in(\\\"dt.entity.service\\\",entitySelector(\\\"type(service),tag(~\\\"KeptnQualityGate~\\\")\\\")))):splitBy():percentile(90)\",\n\t\t\texpectedUnit: \"MicroSecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works 2\",\n\t\t\tinputMV2Query: \"MV2;MicroSecond;metricSelector=builtin:service.keyRequest.response.server:filter(and(in(\\\"dt.entity.service_method\\\",entitySelector(\\\"type(service_method),entityName(~\\\"/api/ui/v2/bootstrap~\\\")\\\")))):splitBy(\\\"dt.entity.service_method\\\"):percentile(90)\",\n\t\t\texpectedMetricSelector: \"builtin:service.keyRequest.response.server:filter(and(in(\\\"dt.entity.service_method\\\",entitySelector(\\\"type(service_method),entityName(~\\\"/api/ui/v2/bootstrap~\\\")\\\")))):splitBy(\\\"dt.entity.service_method\\\"):percentile(90)\",\n\t\t\texpectedUnit: \"MicroSecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works - metric selector first\",\n\t\t\tinputMV2Query: \"MV2;MicroSecond;metricSelector=builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)&entitySelector=type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedEntitySelector: \"type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedUnit: \"MicroSecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works - entity selector first - MicroSecond unit\",\n\t\t\tinputMV2Query: \"MV2;MicroSecond;entitySelector=type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)&metricSelector=builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedEntitySelector: \"type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedUnit: \"MicroSecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works - entity selector first - Microsecond unit\",\n\t\t\tinputMV2Query: \"MV2;Microsecond;entitySelector=type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)&metricSelector=builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedEntitySelector: \"type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedUnit: \"Microsecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works - entity selector first - microsecond unit\",\n\t\t\tinputMV2Query: \"MV2;microsecond;entitySelector=type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)&metricSelector=builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedEntitySelector: \"type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedUnit: \"microsecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"microsecond metric works - entity selector first - microSecond unit\",\n\t\t\tinputMV2Query: \"MV2;microSecond;entitySelector=type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)&metricSelector=builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedMetricSelector: \"builtin:service.response.time:merge(\\\"dt.entity.service\\\"):percentile(50)\",\n\t\t\texpectedEntitySelector: \"type(SERVICE),tag(keptn_project:project1),tag(keptn_stage:staging),tag(keptn_service:carts),tag(keptn_deployment:direct)\",\n\t\t\texpectedUnit: \"microSecond\",\n\t\t},\n\t\t{\n\t\t\tname: \"byte metric works - Byte unit\",\n\t\t\tinputMV2Query: \"MV2;Byte;metricSelector=builtin:host.disk.avail:merge(\\\"dt.entity.host\\\"):merge(\\\"dt.entity.disk\\\")\",\n\t\t\texpectedMetricSelector: \"builtin:host.disk.avail:merge(\\\"dt.entity.host\\\"):merge(\\\"dt.entity.disk\\\")\",\n\t\t\texpectedUnit: \"Byte\",\n\t\t},\n\t\t{\n\t\t\tname: \"byte metric works - byte unit\",\n\t\t\tinputMV2Query: \"MV2;byte;metricSelector=builtin:host.disk.avail:merge(\\\"dt.entity.host\\\"):merge(\\\"dt.entity.disk\\\")\",\n\t\t\texpectedMetricSelector: \"builtin:host.disk.avail:merge(\\\"dt.entity.host\\\"):merge(\\\"dt.entity.disk\\\")\",\n\t\t\texpectedUnit: \"byte\",\n\t\t},\n\t}\n\tfor _, tc := range tests {\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tquery, err := NewQueryParser(tc.inputMV2Query).Parse()\n\t\t\tif tc.expectError {\n\t\t\t\tassert.Error(t, err)\n\t\t\t\tassert.Nil(t, query)\n\t\t\t} else {\n\t\t\t\tassert.NoError(t, err)\n\t\t\t\tif assert.NotNil(t, query) {\n\t\t\t\t\tassert.EqualValues(t, tc.expectedUnit, query.GetUnit())\n\t\t\t\t\tassert.EqualValues(t, tc.expectedMetricSelector, query.GetQuery().GetMetricSelector())\n\t\t\t\t\tassert.EqualValues(t, tc.expectedEntitySelector, query.GetQuery().GetEntitySelector())\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}" ]
[ "0.63390106", "0.59011483", "0.5840018", "0.56048757", "0.55387545", "0.5483519", "0.5473306", "0.5466787", "0.5458519", "0.542968", "0.5415906", "0.54036045", "0.53993297", "0.53982764", "0.53841496", "0.5381801", "0.53746504", "0.5294112", "0.52432704", "0.5229974", "0.5224964", "0.52011514", "0.5193785", "0.5155027", "0.5128325", "0.5115339", "0.509545", "0.507411", "0.50586027", "0.50533503", "0.50356036", "0.50348353", "0.50325537", "0.5028529", "0.50238746", "0.5017601", "0.5015396", "0.49894184", "0.49821395", "0.49783534", "0.4973729", "0.4972254", "0.49679005", "0.49650344", "0.4953535", "0.4949424", "0.4945278", "0.49420932", "0.49418133", "0.4941046", "0.4934364", "0.4932724", "0.49300188", "0.4925934", "0.4918413", "0.49160782", "0.49155614", "0.48959565", "0.4894722", "0.48778474", "0.48648587", "0.48631597", "0.48593205", "0.48527208", "0.481782", "0.48164648", "0.48117915", "0.4809689", "0.48054457", "0.48039472", "0.48035833", "0.47985637", "0.4792432", "0.47915074", "0.47905824", "0.478666", "0.47865757", "0.47808507", "0.47662848", "0.47543353", "0.47524208", "0.47441056", "0.47375867", "0.47352448", "0.4727649", "0.47271472", "0.47254264", "0.47194386", "0.4717955", "0.47127002", "0.4711177", "0.47106913", "0.4670484", "0.46655625", "0.46596494", "0.46588746", "0.4656119", "0.4651473", "0.46509233", "0.46504956" ]
0.6528821
0
JoinWords uses forward slash to join any number of words into a single path. Returned path is prefixed with a slash.
func JoinWords(w string, words ...string) (path string) { path = w if path[0] != '/' { path = "/" + path } for _, s := range words { path += "/" + s } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Join(words []string) string {\n\tvar buf bytes.Buffer\n\tfor i, w := range words {\n\t\tif i != 0 {\n\t\t\tbuf.WriteByte(' ')\n\t\t}\n\t\tbuf.WriteString(Escape(w))\n\t}\n\treturn buf.String()\n}", "func Join(paths ...string) string {\n\tunnormalized := make([]string, len(paths))\n\tfor i, path := range paths {\n\t\tunnormalized[i] = Unnormalize(path)\n\t}\n\tvalue := filepath.Join(unnormalized...)\n\tif value == \"\" {\n\t\treturn \"\"\n\t}\n\treturn Normalize(value)\n}", "func join(elems ...string) string {\n\tvar result string\n\tfor i, v := range elems {\n\t\tif len(v) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif i == 0 {\n\t\t\tresult = strings.TrimRight(v, \"/\")\n\t\t\tcontinue\n\t\t}\n\t\tresult += \"/\" + strings.Trim(v, \"/\")\n\t}\n\treturn result\n}", "func PathJoin(incoming []string) string { return filesys.PathJoin(incoming) }", "func joinPath(parts ...string) string {\n\treturn strings.Join(parts, \"/\")\n}", "func JoinPath(filenames ...string) string {\n\n\thasSlash := false\n\tresult := \"\"\n\tfor _, str := range filenames {\n\t\tcurrentHasSlash := false\n\t\tif len(result) > 0 {\n\t\t\tcurrentHasSlash = strings.HasPrefix(str, \"/\")\n\t\t\tif hasSlash && currentHasSlash {\n\t\t\t\tstr = strings.TrimLeft(str, \"/\")\n\t\t\t}\n\t\t\tif !(hasSlash || currentHasSlash) {\n\t\t\t\tstr = \"/\" + str\n\t\t\t}\n\t\t}\n\t\thasSlash = strings.HasSuffix(str, \"/\")\n\t\tresult += str\n\t}\n\treturn result\n}", "func JoinPath(segments []string, root string) string {\n\tres := \"\"\n\tif segments != nil {\n\t\tfor i := 0; i < len(segments); i++ {\n\t\t\tres = res + segments[i]\n\t\t\tif i < len(segments)-1 {\n\t\t\t\tres = res + \"/\"\n\t\t\t}\n\t\t}\n\t}\n\tif root != \"\" {\n\t\tif len(segments) > 0 {\n\t\t\tres = root + \"/\" + res\n\t\t} else {\n\t\t\tres = root\n\t\t}\n\t}\n\treturn res\n}", "func Join(fs FileSystem, elem ...string) string {\n\tsep := string(fs.PathSeparator())\n\tfor i, e := range elem {\n\t\tif e != \"\" {\n\t\t\treturn filepath.Clean(strings.Join(elem[i:], sep))\n\t\t}\n\t}\n\treturn \"\"\n}", "func JoinPath(path []string) string {\n\treturn \"/\" + strings.Join(path, \"/\")\n}", "func JoinPath(path_arr []string) string {\n\treturn strings.Join(path_arr, \"/\")\n}", "func simpleJoin(dir, path string) string {\n\treturn dir + string(filepath.Separator) + path\n}", "func JoinPath(base string, rest ...string) string {\n\tresult := base\n\tfor _, next := range rest {\n\t\tif filepath.IsAbs(next) {\n\t\t\tresult = next\n\t\t} else {\n\t\t\tresult = filepath.Join(result, next)\n\t\t}\n\t}\n\treturn result\n}", "func JoinWordSeries(items []string) string {\n\tif len(items) == 0 {\n\t\treturn \"\"\n\t} else if len(items) == 1 {\n\t\treturn items[0]\n\t} else {\n\t\treturn strings.Join(items[:len(items)-1], \", \") + \", and \" + items[len(items)-1]\n\t}\n}", "func (t *DesktopTracer) JoinPath(paths []string) string {\n\tif len(paths) == 0 {\n\t\treturn \"\"\n\t}\n\n\tif t.b.Instance().GetConfiguration().GetOS().GetKind() == device.Windows {\n\t\tif paths[0] == \"\" {\n\t\t\treturn path.Clean(strings.Join(paths[1:], \"\\\\\"))\n\t\t}\n\t\treturn path.Clean(strings.Join(paths, \"\\\\\"))\n\t} else {\n\t\treturn path.Join(paths...)\n\t}\n}", "func join(prefix, suffix string) string {\n\tif prefix == \"/\" {\n\t\treturn suffix\n\t}\n\tif suffix == \"/\" {\n\t\treturn prefix\n\t}\n\treturn prefix + suffix\n}", "func join(elem ...string) string {\n\tresult := path.Join(elem...)\n\tif result == \".\" {\n\t\treturn \"\"\n\t}\n\tif len(result) > 0 && result[0] == '/' {\n\t\tresult = result[1:]\n\t}\n\treturn result\n}", "func Join(paths ...PathRetriever) PathRetriever {\n\tvar components []string\n\tvar err error\n\tfor _, path := range paths {\n\t\tvar component string\n\t\tcomponent, err = path()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tcomponents = append(components, component)\n\t}\n\tpath := filepath.Join(components...)\n\n\treturn func() (string, error) {\n\t\treturn path, err\n\t}\n}", "func JoinPath(ss ...string) string {\n\treturn filepath.Join(ss...)\n}", "func JoinString(p Path, elem ...string) Path {\n\treturn Join(p, ToPaths(elem)...)\n}", "func (*Mysqlfs) Join(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func JoinPath(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func (o *OS) Join(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func singleJoiningSlash(a, b string) string {\n\taslash := strings.HasSuffix(a, \"/\")\n\tbslash := strings.HasPrefix(b, \"/\")\n\tswitch {\n\tcase aslash && bslash:\n\t\treturn a + b[1:]\n\tcase !aslash && !bslash:\n\t\treturn a + \"/\" + b\n\t}\n\treturn a + b\n}", "func singleJoiningSlash(a, b string) string {\n\taslash := strings.HasSuffix(a, \"/\")\n\tbslash := strings.HasPrefix(b, \"/\")\n\tswitch {\n\tcase aslash && bslash:\n\t\treturn a + b[1:]\n\tcase !aslash && !bslash:\n\t\treturn a + \"/\" + b\n\t}\n\treturn a + b\n}", "func NormalizeJoin(l []string, d, f string) string {\n n := len(l)\n var s string\n for i, e := range l {\n if i > 0 {\n if n - (i + 1) == 0 {\n s += f\n }else{\n s += d\n }\n }\n s += e\n }\n return s\n}", "func (rp RelPath) Join(components ...string) RelPath {\n\tpath := append([]string{string(rp)}, components...)\n\treturn RelPath(filepath.Join(path...))\n}", "func WordsShiritoriPath(word string) string {\n\treturn fmt.Sprintf(\"/words/%v\", word)\n}", "func Join(path string, dir string) string {\n\treturn filepath.Join(path, dir)\n}", "func (o Operator) Join(arr []string) string {\n\tvar str string\n\n\tswitch o {\n\tcase \"/\":\n\t\tfor i, v := range arr {\n\t\t\tif v[:1] != \"{\" {\n\t\t\t\tarr[i] = \"/\" + v\n\t\t\t}\n\t\t}\n\t\tstr = filepath.Join(strings.Join(arr, \"\"))\n\n\tcase \"#\", \"?\", \".\", \";\", \"&\":\n\t\tm := opmap[o]\n\t\tfor i, v := range arr {\n\t\t\tif i > 0 && v[:1] != \"{\" {\n\t\t\t\tarr[i] = m[1] + v\n\t\t\t}\n\t\t}\n\t\tstr = m[0] + strings.Join(arr, \"\")\n\n\t\t// TODO revisit, not particularly pretty\n\t\tif str[:2] == \"&{\" {\n\t\t\tstr = str[1:] // remove extra &\n\t\t}\n\n\tdefault: // handles +, `{+var}` and blank, `{var}`\n\t\tstr = strings.Join(arr, \",\")\n\t}\n\n\treturn str\n}", "func StringJoin(a []string, sep string) string { return strings.Join(a, sep) }", "func JoinWithDot(ks []string) string { return strings.Join(ks, \".\") }", "func Join(a []string, sep string) string {\n\treturn strings.Join(a, sep)\n}", "func JoinNormalized(n Normalization, base string, elem ...string) (string, error) {\n\tif n == NoNorm {\n\t\treturn filepath.Join(append([]string{base}, elem...)...), nil\n\t}\n\treturn joinFold(n == FoldPreferExactNorm, base, elem...)\n}", "func Join(sep string, parts ...string) string {\n\treturn strings.Join(parts, sep)\n}", "func Join(s []string, sep string) string {\n\tvar buf string\n\tfor _, v := range s[:len(s)-1] {\n\t\tbuf += fmt.Sprintf(\"%s%s\", v, sep)\n\t}\n\n\tbuf += s[len(s)-1]\n\treturn buf\n}", "func URLJoin(base string, elements ...string) string {\n\tu, err := url.Parse(base)\n\tif err != nil {\n\t\tlog.Error(\"httputil.URLJoin\", fmt.Sprintf(\"base URL '%s' does not parse\", base), err)\n\t\tpanic(err)\n\t}\n\tscrubbed := []string{}\n\tu.Path = strings.TrimRight(u.Path, \"/\")\n\tif u.Path != \"\" {\n\t\tscrubbed = append(scrubbed, u.Path)\n\t}\n\tfor _, s := range elements {\n\t\ts = strings.Trim(s, \"/\")\n\t\tif s != \"\" {\n\t\t\tscrubbed = append(scrubbed, s)\n\t\t}\n\t}\n\tu.Path = strings.Join(scrubbed, \"/\")\n\treturn u.String()\n}", "func JoinPath(url, path string) string {\n\tsuffix := url[len(url)-1] == '/'\n\tprefix := path[0] == '/'\n\tif suffix && prefix {\n\t\treturn url + path[1:]\n\t}\n\tif !suffix && !prefix {\n\t\treturn url + \"/\" + path\n\t}\n\treturn url + path\n}", "func pathJoin(base, leaf string) string {\n\t// Disalllow empty leaf names.\n\tif leaf == \"\" {\n\t\tpanic(\"empty leaf name\")\n\t}\n\n\t// When joining a path to the synchronization root, we don't want to\n\t// concatenate.\n\tif base == \"\" {\n\t\treturn leaf\n\t}\n\n\t// Concatenate the paths.\n\treturn base + \"/\" + leaf\n}", "func Join(elem ...string) string {\n\treturn std.Join(elem...)\n}", "func joinDemo(l []string, s string) string {\n\treturn strings.Join(l, s)\n}", "func Join(sep string, strs ...string) string {\n\tvar buf bytes.Buffer\n\tif len(strs) == 0 {\n\t\treturn \"\"\n\t}\n\tfor _, str := range strs {\n\t\tbuf.WriteString(str + sep)\n\t}\n\treturn strings.TrimRight(buf.String(), sep)\n}", "func Join(p Path, ps ...Path) Path {\n\tif len(ps) == 0 {\n\t\treturn p[:]\n\t}\n\treturn Append(p[:], ps...)\n}", "func JoinURL(base string, paths ...string) string {\n\treturn fmt.Sprintf(\"%s/%s\", strings.TrimRight(base, \"/\"),\n\t\tstrings.TrimLeft(path.Join(paths...), \"/\"))\n}", "func urlJoin(repo string, path string) string {\n\treturn strings.TrimSuffix(repo, \"/\") + \"/\" + strings.TrimPrefix(path, \"/\")\n}", "func spaceJoin(a string, b string) string {\n\treturn join(a, b, ` `)\n}", "func Join(sep string, operand []string) string { return strings.Join(operand, sep) }", "func (lc LowerCaseConvention) Join(names []string) string {\n\treturn strings.Join(names, \"\")\n}", "func JoinPath(base string, other string) string {\n\tif filepath.IsAbs(other) || base == \"\" {\n\t\treturn other\n\t}\n\treturn filepath.Join(base, other)\n}", "func (sc SnakeCaseConvention) Join(names []string) string {\n\treturn strings.Join(names, \"_\")\n}", "func (s *Space) joinPath(path string, defaults string) string {\n\tif path == \"\" {\n\t\tpath = defaults\n\t}\n\treturn s.Join(path)\n}", "func ShellJoin(args ...string) string {\n\tvar buf bytes.Buffer\n\tfor i, arg := range args {\n\t\tif i != 0 {\n\t\t\tbuf.WriteByte(' ')\n\t\t}\n\t\tquote(arg, &buf)\n\t}\n\treturn buf.String()\n}", "func Words(prefix, suffix string, words ...string) string {\n\tfor i, word := range words {\n\t\twords[i] = regexp.QuoteMeta(word)\n\t}\n\treturn prefix + `(` + strings.Join(words, `|`) + `)` + suffix\n}", "func JoinPaths(paths ...*pb.Path) *pb.Path {\n\tvar elems []*pb.PathElem\n\tfor _, path := range paths {\n\t\tif path != nil {\n\t\t\tpath = upgradePath(path)\n\t\t\telems = append(elems, path.Elem...)\n\t\t}\n\t}\n\treturn &pb.Path{Elem: elems}\n}", "func join(sep string, a ...string) string {\n\tswitch len(a) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn a[0]\n\t}\n\n\tres := bytes.NewBufferString(a[0])\n\tfor _, s := range a[1:] {\n\t\tres.WriteString(sep + s)\n\t}\n\n\treturn res.String()\n}", "func (lss *ListStrings) Join(lsep string, ssep string) (s string) {\n\tlsslen := len(*lss) - 1\n\n\tfor x, ls := range *lss {\n\t\ts += strings.Join(ls, ssep)\n\n\t\tif x < lsslen {\n\t\t\ts += lsep\n\t\t}\n\t}\n\treturn\n}", "func JoinStrings(separator string, stringArray ...string) string {\n\n\tvar buffer bytes.Buffer\n\tvar max int = len(stringArray) - 1\n\tfor vi, v := range stringArray {\n\t\tbuffer.WriteString(v)\n\t\tif vi < max {\n\t\t\tbuffer.WriteString(separator)\n\t\t}\n\t}\n\treturn buffer.String()\n\n}", "func (sc SameCaseConvention) Join(names []string) string {\n\treturn strings.Join(names, \"\")\n}", "func Join(sep string, strs ...string) string {\n\tswitch len(strs) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn strs[0]\n\tcase 2:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn strs[0] + sep + strs[1]\n\tcase 3:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn strs[0] + sep + strs[1] + sep + strs[2]\n\t}\n\tn := len(sep) * (len(strs) - 1)\n\tfor i := 0; i < len(strs); i++ {\n\t\tn += len(strs[i])\n\t}\n\n\tb := make([]byte, n)\n\tbp := copy(b, strs[0])\n\tfor _, s := range strs[1:] {\n\t\tbp += copy(b[bp:], sep)\n\t\tbp += copy(b[bp:], s)\n\t}\n\treturn string(b)\n}", "func (p *Path) JoinPath(elem ...string) *Path {\n\ttemp := []string{p.Path}\n\telem = append(temp, elem[0:]...)\n\tnewP := New(path.Join(elem...))\n\treturn newP\n}", "func QuotedJoin(terms []string, conjunction, none string) string {\n\tswitch len(terms) {\n\tcase 0:\n\t\treturn none\n\tcase 1:\n\t\treturn fmt.Sprintf(\"%q\", terms[0])\n\tcase 2:\n\t\treturn fmt.Sprintf(\"%q %s %q\", terms[0], conjunction, terms[1])\n\tdefault:\n\t\ti := 1\n\t\tinner := \"\"\n\t\tfor ; i < len(terms)-1; i++ {\n\t\t\tinner = fmt.Sprintf(\"%s, %q\", inner, terms[i])\n\t\t}\n\t\t// first, inner, inner, and/or last\n\t\treturn fmt.Sprintf(\"%q%s, %s %q\", terms[0], inner, conjunction, terms[i])\n\t}\n}", "func (c StringArrayCollection) Join(delimiter string) string {\n\ts := \"\"\n\tfor i := 0; i < len(c.value); i++ {\n\t\tif i != len(c.value)-1 {\n\t\t\ts += c.value[i] + delimiter\n\t\t} else {\n\t\t\ts += c.value[i]\n\t\t}\n\t}\n\treturn s\n}", "func joinURL(baseURL string, paths ...string) string {\n\n\tu, err := url.Parse(baseURL)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\turl := path.Join(paths...)\n\tu.Path = path.Join(u.Path, url)\n\treturn u.String()\n}", "func (m MultiString) Join(separator string) string {\n\treturn strings.Join(m, separator)\n}", "func JoinString(parts ...string) string {\n\tvar st strings.Builder\n\n\tvar lastIsNewLine bool\n\tfor _, p := range parts {\n\t\tif len(p) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif st.Len() != 0 {\n\t\t\tif !lastIsNewLine {\n\t\t\t\t_, _ = st.WriteString(\"\\n\")\n\t\t\t}\n\t\t\tst.WriteString(yamlSeparator)\n\t\t}\n\t\t_, _ = st.WriteString(p)\n\t\tlastIsNewLine = p[len(p)-1] == '\\n'\n\t}\n\n\treturn st.String()\n}", "func prefixJoin(prefix string, array []string, separator string) (result string) {\n\tif len(array) == 0 {\n\t\treturn\n\t}\n\tfor index, val := range array {\n\t\tif index == 0 {\n\t\t\tresult = val\n\t\t} else {\n\t\t\tresult = join(result, concat(prefix, val), separator)\n\t\t}\n\t}\n\treturn\n}", "func JoinString(array []string, slim string) string {\n\tstr := array[0]\n\tfor index, val := range array {\n\t\tif index != 0 {\n\t\t\tstr = fmt.Sprintf(\"%s%s%s\", str, slim, val)\n\t\t}\n\t}\n\treturn str\n}", "func JoinIDPath(ids ...int64) string {\n\tidStr := make([]string, 0, len(ids))\n\tfor _, id := range ids {\n\t\tidStr = append(idStr, strconv.FormatInt(id, 10))\n\t}\n\treturn path.Join(idStr...)\n}", "func JoinIDPath(ids ...typeutil.UniqueID) string {\n\tidStr := make([]string, 0, len(ids))\n\tfor _, id := range ids {\n\t\tidStr = append(idStr, strconv.FormatInt(id, 10))\n\t}\n\treturn path.Join(idStr...)\n}", "func join(str ...string) string {\n\tvar joined string\n\n\tfor n, s := range str {\n\t\tswitch n {\n\t\t\tcase 0: joined = s\n\t\t\tcase 1: joined = joined + \" \" + s\n\t\t\tdefault: joined = joined + \", \" + s\n\t\t}\n\t}\n\treturn joined\n}", "func joinLeft(g []string) string {\n\tif g == nil || len(g) == 0 {\n\t\treturn \"\"\n\t}\n\tvar bf bytes.Buffer\n\tfor i := range g {\n\t\tc := strings.Index(g[i], \"#\")\n\t\tif c == -1 {\n\t\t\tbf.WriteString(g[i])\n\t\t} else {\n\t\t\tbf.WriteString(g[i][0:c])\n\t\t\tbreak\n\t\t}\n\t}\n\treturn string(bf.Bytes())\n}", "func StringJoin(scope *Scope, inputs []tf.Output, optional ...StringJoinAttr) (output tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\tattrs := map[string]interface{}{}\n\tfor _, a := range optional {\n\t\ta(attrs)\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"StringJoin\",\n\t\tInput: []tf.Input{\n\t\t\ttf.OutputList(inputs),\n\t\t},\n\t\tAttrs: attrs,\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0)\n}", "func SliceJoin(a []Stringer, sep string) string {\n\tswitch len(a) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn a[0].String()\n\tcase 2:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn a[0].String() + sep + a[1].String()\n\tcase 3:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn a[0].String() + sep + a[1].String() + sep + a[2].String()\n\t}\n\tn := len(sep) * (len(a) - 1)\n\tfor i := 0; i < len(a); i++ {\n\t\tn += len(a[i].String())\n\t}\n\n\tb := make([]byte, n)\n\tbp := copy(b, a[0].String())\n\tfor _, s := range a[1:] {\n\t\tbp += copy(b[bp:], sep)\n\t\tbp += copy(b[bp:], s.String())\n\t}\n\treturn string(b)\n}", "func (c *apiClient) joinURLPath(urlPath, resourcePath string) string {\n\tif resourcePath == \"\" {\n\t\tif urlPath == \"\" {\n\t\t\treturn \"/\"\n\t\t}\n\t\treturn urlPath\n\t}\n\tif !strings.HasSuffix(urlPath, \"/\") {\n\t\turlPath += \"/\"\n\t}\n\tresourcePath = strings.TrimPrefix(resourcePath, \"/\")\n\treturn urlPath + resourcePath\n}", "func joinPath(dir, file string) string {\n\tif filepath.IsAbs(file) {\n\t\treturn file\n\t}\n\treturn filepath.Join(dir, file)\n}", "func joinStrings(sep string, a ...interface{}) (o string) {\n\tfor i := range a {\n\t\to += fmt.Sprint(a[i])\n\t\tif i < len(a)-1 {\n\t\t\to += sep\n\t\t}\n\t}\n\treturn\n}", "func joinUrl(url string, endpoint ...string) string {\n\tp := url\n\tfor _, e := range endpoint {\n\t\tp += \"/\" + e\n\t}\n\treturn p\n}", "func Of(s ...string) string { return strings.Join(s, \" \") }", "func (fsr *fsLockedRepo) join(paths ...string) string {\n\treturn filepath.Join(append([]string{fsr.path}, paths...)...)\n}", "func joinStrings(str ...string) string {\n\treturn strings.Join(str, \"\")\n}", "func Join(parts ...[]byte) []byte {\n\tvar b bytes.Buffer\n\n\tvar lastIsNewLine bool\n\tfor _, p := range parts {\n\t\tif len(p) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif b.Len() != 0 {\n\t\t\tif !lastIsNewLine {\n\t\t\t\t_, _ = b.WriteString(\"\\n\")\n\t\t\t}\n\t\t\tb.WriteString(yamlSeparator)\n\t\t}\n\t\t_, _ = b.Write(p)\n\t\ts := string(p)\n\t\tlastIsNewLine = s[len(s)-1] == '\\n'\n\t}\n\n\treturn b.Bytes()\n}", "func joinFilePath(path, file string) string {\n\treturn filepath.Join(path, file)\n}", "func SecureJoin(root, unsafePath string) (string, error) {\n\treturn SecureJoinVFS(root, unsafePath, nil)\n}", "func joinCoverPath(pwd, fileName string) string {\n\tpwdPath := lessEmpty(strings.Split(pwd, \"/\"))\n\tfileDirPath := lessEmpty(strings.Split(filepath.Dir(fileName), \"/\"))\n\n\tfor index, dir := range pwdPath {\n\t\tif dir == first(fileDirPath) {\n\t\t\tpwdPath = pwdPath[:index]\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn filepath.Join(maybePrefix(strings.Join(pwdPath, \"/\"), \"/\"), fileName)\n}", "func QuoteJoin(elems []string, sep string) string {\n\tswitch len(elems) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn elems[0]\n\t}\n\tn := len(sep) * (len(elems) - 1)\n\tfor i := 0; i < len(elems); i++ {\n\t\tn += len(elems[i]) + 2\n\t}\n\n\tvar b strings.Builder\n\tb.Grow(n)\n\tb.WriteByte('\"')\n\tb.WriteString(elems[0])\n\tb.WriteByte('\"')\n\tfor _, s := range elems[1:] {\n\t\tb.WriteString(sep)\n\t\tb.WriteByte('\"')\n\t\tb.WriteString(s)\n\t\tb.WriteByte('\"')\n\t}\n\treturn b.String()\n}", "func UrlJoin(baseUrl string, elem ...string) (string, error) {\n\tu, err := url.Parse(baseUrl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(elem) > 0 {\n\t\telem = append([]string{u.Path}, elem...)\n\t\tu.Path = path.Join(elem...)\n\t}\n\treturn u.String(), nil\n}", "func WebPath(target, start string) string {\n\ttarget = Posixfy(target)\n\tstart = Posixfy(start)\n\turiPrefix := \"\"\n\tisWebroot := IsWebRoot(target)\n\n\tif !isWebroot && start != \"\" {\n\t\ttarget = start + \"/\" + target\n\t\tif strings.Contains(target, \":\") {\n\t\t\tif res := regexps.UriSniffRx.FindStringSubmatchIndex(target); len(res) == 4 {\n\t\t\t\turiPrefix = target[:res[3]]\n\t\t\t\ttarget = target[res[3]:]\n\t\t\t}\n\t\t}\n\t}\n\t// BUG? slash seems to be lost in https://github.com/asciidoctor/asciidoctor/blob/ab1e0b9c45e5138394b089dac205fb6d854e15e6/lib/asciidoctor/path_resolver.rb#L352-L366\n\tisWebroot = IsWebRoot(target)\n\tif testpr == \"test_Webath_uriPrefix\" {\n\t\treturn fmt.Sprintf(\"target='%v', uriPrefix='%v'\", target, uriPrefix)\n\t}\n\ttargetSegments, targetRoot, _ := PartitionPath(target, true)\n\tif testpr == \"test_Webath_partitionTarget\" {\n\t\treturn fmt.Sprintf(\"targetSegments=(%v)'%v', targetRoot='%v'\", len(targetSegments), targetSegments, targetRoot)\n\t}\n\taccum := []string{}\n\tfor _, segment := range targetSegments {\n\t\tif segment == \"..\" {\n\t\t\tif len(accum) == 0 {\n\t\t\t\tif targetRoot == \"\" || targetRoot == \".\" {\n\t\t\t\t\taccum = append(accum, segment)\n\t\t\t\t}\n\t\t\t} else if accum[len(accum)-1] == \"..\" {\n\t\t\t\taccum = append(accum, segment)\n\t\t\t} else {\n\t\t\t\taccum = accum[:len(accum)-1]\n\t\t\t}\n\t\t} else {\n\t\t\taccum = append(accum, segment)\n\t\t}\n\t}\n\tresolvedSegments := accum\n\n\tjoinPath := JoinPath(resolvedSegments, targetRoot)\n\tif uriPrefix != \"\" {\n\t\treturn uriPrefix + joinPath\n\t}\n\tif isWebroot {\n\t\treturn \"/\" + joinPath\n\t}\n\treturn joinPath\n}", "func (s *Scope) appendPaths(path string, parts ...string) string {\n\tpaths := []string{path}\n\tpaths = append(paths, parts...)\n\treturn filepath.Join(paths...)\n}", "func concatURLPaths(host string, path []string) (string, error) {\n\tif host == \"\" {\n\t\treturn \"\", errors.New(\"host can not be empty\")\n\t}\n\n\telements := make([]string, len(path)+1)\n\telements[0] = strings.TrimSuffix(host, \"/\")\n\n\tfor k, v := range path {\n\t\telements[k+1] = strings.Trim(v, \"/\")\n\t}\n\n\treturn strings.Join(elements, \"/\"), nil\n}", "func join(a string, b string, separator string) string {\n\tvals := make([]byte, 0, 10)\n\tvals = append(vals, a...)\n\tvals = append(vals, separator...)\n\tvals = append(vals, b...)\n\treturn string(vals)\n}", "func JoinString(arr []string, delimiter string) string {\n\treturn strings.Join(arr, delimiter)\n}", "func JoinEnv(key string, vals []string) string {\n\treturn key + \"=\" + strings.Join(vals, string(os.PathListSeparator))\n}", "func join(ins []rune, c rune) (result []string) {\n\tfor i := 0; i <= len(ins); i++ {\n\t\tresult = append(result, string(ins[:i])+string(c)+string(ins[i:]))\n\t}\n\treturn\n}", "func Append(p Path, ps ...Path) Path {\n\tfor _, e := range ps {\n\t\t// Drop the leading '/' when appending/joining fully qualified paths.\n\t\tif len(e) > 0 && e[0] == \"/\" {\n\t\t\te = e[1:]\n\t\t}\n\t\tp = append(p, e...)\n\t}\n\treturn p\n}", "func mergeAlternately(word1 string, word2 string) string {\n\tvar buf bytes.Buffer\n\tfor i := range word1 {\n\t\tbuf.WriteByte(word1[i])\n\t\tif i < len(word2) {\n\t\t\tbuf.WriteByte(word2[i])\n\t\t}\n\t}\n\n\tif len(word1) < len(word2) {\n\t\tbuf.WriteString(word2[len(word1):])\n\t}\n\treturn buf.String()\n}", "func joinIDs(ids []WKID, separator string) string {\n\tvar s string\n\n\tfor i, n := range ids {\n\t\tif i != 0 {\n\t\t\ts += \",\"\n\t\t}\n\n\t\ts += strconv.FormatInt(int64(n), 10)\n\t}\n\n\treturn s\n}", "func addLeadSlash(name string) string {\n\tparts := strings.Split(name, \"/\")\n\tif len(parts) == 3 && parts[0] != \"\" {\n\t\tname = \"/\" + name\n\t}\n\treturn name\n}", "func JoinStringsReversed(separator string, stringArray ...string) string {\n\n\tvar buffer bytes.Buffer\n\n\tfor vi := len(stringArray) - 1; vi >= 0; vi-- {\n\t\tbuffer.WriteString(stringArray[vi])\n\t\tif vi > 0 {\n\t\t\tbuffer.WriteString(separator)\n\t\t}\n\t}\n\n\treturn buffer.String()\n\n}", "func (array Array) Join(separator string) string {\n\tstr := fmt.Sprint()\n\tfor i, v := range array {\n\t\tstr += fmt.Sprintf(\"%v\", v)\n\t\tif i != len(array) - 1 {\n\t\t\tstr += fmt.Sprintf(\"%s\", separator)\n\t\t}\n\t}\n\treturn str\n}", "func (w Word) WordString() string {\n\tphoneList := w.PhoneList()\n\tvar buffer bytes.Buffer\n\tvar phone Phone\n\n\tfor len(phoneList) > 0 {\n\t\tphone, phoneList = phoneList[0], phoneList[1:]\n\t\tbuffer.WriteString(string(phone.Char()))\n\t}\n\n\treturn buffer.String()\n}", "func JoinStrings(ss ...string) string {\n\treturn strings.Join(ss, \"\")\n}" ]
[ "0.65263057", "0.65035397", "0.6358016", "0.6277726", "0.61468893", "0.611875", "0.602499", "0.59774566", "0.5975566", "0.5926904", "0.58029735", "0.5750395", "0.5749765", "0.5708375", "0.5669194", "0.5654913", "0.56510186", "0.56203854", "0.5601783", "0.5548136", "0.54925305", "0.54207706", "0.5411077", "0.5411077", "0.53841656", "0.5363367", "0.52917325", "0.52309936", "0.5227133", "0.5201703", "0.5197743", "0.51779574", "0.51739365", "0.5151946", "0.51390904", "0.51322424", "0.51167357", "0.5107331", "0.5092237", "0.50797236", "0.50320995", "0.5030981", "0.5025185", "0.4984795", "0.49774992", "0.49738026", "0.49623236", "0.4958225", "0.49360108", "0.49097204", "0.49066848", "0.48907173", "0.48719278", "0.48584172", "0.4848221", "0.48450872", "0.483075", "0.48249888", "0.48218372", "0.48069695", "0.47872418", "0.47834915", "0.47545958", "0.47422382", "0.47411188", "0.47033587", "0.4683318", "0.46827808", "0.46790263", "0.46704537", "0.46693286", "0.46282172", "0.46235836", "0.4589421", "0.45774385", "0.4571591", "0.45461106", "0.4542558", "0.44982746", "0.44871032", "0.4484628", "0.44823417", "0.44680086", "0.4465698", "0.44499117", "0.44286117", "0.441841", "0.4416839", "0.44054952", "0.43991622", "0.43916774", "0.43843594", "0.43808982", "0.43761218", "0.43741328", "0.43706736", "0.43693906", "0.43667004", "0.43575844", "0.43553266" ]
0.88737655
0
JoinPath joins two path elements that may (or may not) be prefixed/suffixed with a slash.
func JoinPath(url, path string) string { suffix := url[len(url)-1] == '/' prefix := path[0] == '/' if suffix && prefix { return url + path[1:] } if !suffix && !prefix { return url + "/" + path } return url + path }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func joinPath(parts ...string) string {\n\treturn strings.Join(parts, \"/\")\n}", "func (rp RelPath) Join(components ...string) RelPath {\n\tpath := append([]string{string(rp)}, components...)\n\treturn RelPath(filepath.Join(path...))\n}", "func (p *Path) JoinPath(elem ...string) *Path {\n\ttemp := []string{p.Path}\n\telem = append(temp, elem[0:]...)\n\tnewP := New(path.Join(elem...))\n\treturn newP\n}", "func JoinPath(base string, other string) string {\n\tif filepath.IsAbs(other) || base == \"\" {\n\t\treturn other\n\t}\n\treturn filepath.Join(base, other)\n}", "func PathJoin(incoming []string) string { return filesys.PathJoin(incoming) }", "func JoinPath(segments []string, root string) string {\n\tres := \"\"\n\tif segments != nil {\n\t\tfor i := 0; i < len(segments); i++ {\n\t\t\tres = res + segments[i]\n\t\t\tif i < len(segments)-1 {\n\t\t\t\tres = res + \"/\"\n\t\t\t}\n\t\t}\n\t}\n\tif root != \"\" {\n\t\tif len(segments) > 0 {\n\t\t\tres = root + \"/\" + res\n\t\t} else {\n\t\t\tres = root\n\t\t}\n\t}\n\treturn res\n}", "func JoinPath(path []string) string {\n\treturn \"/\" + strings.Join(path, \"/\")\n}", "func pathJoin(base, leaf string) string {\n\t// Disalllow empty leaf names.\n\tif leaf == \"\" {\n\t\tpanic(\"empty leaf name\")\n\t}\n\n\t// When joining a path to the synchronization root, we don't want to\n\t// concatenate.\n\tif base == \"\" {\n\t\treturn leaf\n\t}\n\n\t// Concatenate the paths.\n\treturn base + \"/\" + leaf\n}", "func JoinPath(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func JoinPath(base string, rest ...string) string {\n\tresult := base\n\tfor _, next := range rest {\n\t\tif filepath.IsAbs(next) {\n\t\t\tresult = next\n\t\t} else {\n\t\t\tresult = filepath.Join(result, next)\n\t\t}\n\t}\n\treturn result\n}", "func Join(p Path, ps ...Path) Path {\n\tif len(ps) == 0 {\n\t\treturn p[:]\n\t}\n\treturn Append(p[:], ps...)\n}", "func JoinPath(path_arr []string) string {\n\treturn strings.Join(path_arr, \"/\")\n}", "func join(elem ...string) string {\n\tresult := path.Join(elem...)\n\tif result == \".\" {\n\t\treturn \"\"\n\t}\n\tif len(result) > 0 && result[0] == '/' {\n\t\tresult = result[1:]\n\t}\n\treturn result\n}", "func join(prefix, suffix string) string {\n\tif prefix == \"/\" {\n\t\treturn suffix\n\t}\n\tif suffix == \"/\" {\n\t\treturn prefix\n\t}\n\treturn prefix + suffix\n}", "func Join(fs FileSystem, elem ...string) string {\n\tsep := string(fs.PathSeparator())\n\tfor i, e := range elem {\n\t\tif e != \"\" {\n\t\t\treturn filepath.Clean(strings.Join(elem[i:], sep))\n\t\t}\n\t}\n\treturn \"\"\n}", "func singleJoiningSlash(a, b string) string {\n\taslash := strings.HasSuffix(a, \"/\")\n\tbslash := strings.HasPrefix(b, \"/\")\n\tswitch {\n\tcase aslash && bslash:\n\t\treturn a + b[1:]\n\tcase !aslash && !bslash:\n\t\treturn a + \"/\" + b\n\t}\n\treturn a + b\n}", "func singleJoiningSlash(a, b string) string {\n\taslash := strings.HasSuffix(a, \"/\")\n\tbslash := strings.HasPrefix(b, \"/\")\n\tswitch {\n\tcase aslash && bslash:\n\t\treturn a + b[1:]\n\tcase !aslash && !bslash:\n\t\treturn a + \"/\" + b\n\t}\n\treturn a + b\n}", "func JoinPath(filenames ...string) string {\n\n\thasSlash := false\n\tresult := \"\"\n\tfor _, str := range filenames {\n\t\tcurrentHasSlash := false\n\t\tif len(result) > 0 {\n\t\t\tcurrentHasSlash = strings.HasPrefix(str, \"/\")\n\t\t\tif hasSlash && currentHasSlash {\n\t\t\t\tstr = strings.TrimLeft(str, \"/\")\n\t\t\t}\n\t\t\tif !(hasSlash || currentHasSlash) {\n\t\t\t\tstr = \"/\" + str\n\t\t\t}\n\t\t}\n\t\thasSlash = strings.HasSuffix(str, \"/\")\n\t\tresult += str\n\t}\n\treturn result\n}", "func Join(paths ...string) string {\n\tunnormalized := make([]string, len(paths))\n\tfor i, path := range paths {\n\t\tunnormalized[i] = Unnormalize(path)\n\t}\n\tvalue := filepath.Join(unnormalized...)\n\tif value == \"\" {\n\t\treturn \"\"\n\t}\n\treturn Normalize(value)\n}", "func Join(paths ...PathRetriever) PathRetriever {\n\tvar components []string\n\tvar err error\n\tfor _, path := range paths {\n\t\tvar component string\n\t\tcomponent, err = path()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tcomponents = append(components, component)\n\t}\n\tpath := filepath.Join(components...)\n\n\treturn func() (string, error) {\n\t\treturn path, err\n\t}\n}", "func simpleJoin(dir, path string) string {\n\treturn dir + string(filepath.Separator) + path\n}", "func (s *Space) joinPath(path string, defaults string) string {\n\tif path == \"\" {\n\t\tpath = defaults\n\t}\n\treturn s.Join(path)\n}", "func JoinPath(ss ...string) string {\n\treturn filepath.Join(ss...)\n}", "func JoinString(p Path, elem ...string) Path {\n\treturn Join(p, ToPaths(elem)...)\n}", "func join(elems ...string) string {\n\tvar result string\n\tfor i, v := range elems {\n\t\tif len(v) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif i == 0 {\n\t\t\tresult = strings.TrimRight(v, \"/\")\n\t\t\tcontinue\n\t\t}\n\t\tresult += \"/\" + strings.Trim(v, \"/\")\n\t}\n\treturn result\n}", "func (t *DesktopTracer) JoinPath(paths []string) string {\n\tif len(paths) == 0 {\n\t\treturn \"\"\n\t}\n\n\tif t.b.Instance().GetConfiguration().GetOS().GetKind() == device.Windows {\n\t\tif paths[0] == \"\" {\n\t\t\treturn path.Clean(strings.Join(paths[1:], \"\\\\\"))\n\t\t}\n\t\treturn path.Clean(strings.Join(paths, \"\\\\\"))\n\t} else {\n\t\treturn path.Join(paths...)\n\t}\n}", "func (p *Path) Join(q *Path) *Path {\n\tif q == nil || len(q.d) == 0 {\n\t\treturn p\n\t} else if len(p.d) == 0 {\n\t\treturn q\n\t}\n\ti0 := len(p.d) + q.i0\n\tif q.d[0] == moveToCmd {\n\t\tx0, y0 := p.d[len(p.d)-2], p.d[len(p.d)-1]\n\t\tx1, y1 := q.d[1], q.d[2]\n\t\tif equal(x0, x1) && equal(y0, y1) {\n\t\t\tq.d = q.d[3:]\n\t\t\ti0 -= 3\n\t\t\tif q.i0 == 0 {\n\t\t\t\ti0 = p.i0\n\t\t\t}\n\t\t}\n\t} else if q.i0 == 0 {\n\t\ti0 = p.i0\n\t}\n\treturn &Path{append(p.d, q.d...), i0}\n}", "func (c *apiClient) joinURLPath(urlPath, resourcePath string) string {\n\tif resourcePath == \"\" {\n\t\tif urlPath == \"\" {\n\t\t\treturn \"/\"\n\t\t}\n\t\treturn urlPath\n\t}\n\tif !strings.HasSuffix(urlPath, \"/\") {\n\t\turlPath += \"/\"\n\t}\n\tresourcePath = strings.TrimPrefix(resourcePath, \"/\")\n\treturn urlPath + resourcePath\n}", "func JoinPaths(paths ...*pb.Path) *pb.Path {\n\tvar elems []*pb.PathElem\n\tfor _, path := range paths {\n\t\tif path != nil {\n\t\t\tpath = upgradePath(path)\n\t\t\telems = append(elems, path.Elem...)\n\t\t}\n\t}\n\treturn &pb.Path{Elem: elems}\n}", "func urlJoin(repo string, path string) string {\n\treturn strings.TrimSuffix(repo, \"/\") + \"/\" + strings.TrimPrefix(path, \"/\")\n}", "func URLJoin(base string, elements ...string) string {\n\tu, err := url.Parse(base)\n\tif err != nil {\n\t\tlog.Error(\"httputil.URLJoin\", fmt.Sprintf(\"base URL '%s' does not parse\", base), err)\n\t\tpanic(err)\n\t}\n\tscrubbed := []string{}\n\tu.Path = strings.TrimRight(u.Path, \"/\")\n\tif u.Path != \"\" {\n\t\tscrubbed = append(scrubbed, u.Path)\n\t}\n\tfor _, s := range elements {\n\t\ts = strings.Trim(s, \"/\")\n\t\tif s != \"\" {\n\t\t\tscrubbed = append(scrubbed, s)\n\t\t}\n\t}\n\tu.Path = strings.Join(scrubbed, \"/\")\n\treturn u.String()\n}", "func JoinIDPath(ids ...typeutil.UniqueID) string {\n\tidStr := make([]string, 0, len(ids))\n\tfor _, id := range ids {\n\t\tidStr = append(idStr, strconv.FormatInt(id, 10))\n\t}\n\treturn path.Join(idStr...)\n}", "func joinPath(dir, file string) string {\n\tif filepath.IsAbs(file) {\n\t\treturn file\n\t}\n\treturn filepath.Join(dir, file)\n}", "func JoinURL(base string, paths ...string) string {\n\treturn fmt.Sprintf(\"%s/%s\", strings.TrimRight(base, \"/\"),\n\t\tstrings.TrimLeft(path.Join(paths...), \"/\"))\n}", "func Append(p Path, ps ...Path) Path {\n\tfor _, e := range ps {\n\t\t// Drop the leading '/' when appending/joining fully qualified paths.\n\t\tif len(e) > 0 && e[0] == \"/\" {\n\t\t\te = e[1:]\n\t\t}\n\t\tp = append(p, e...)\n\t}\n\treturn p\n}", "func (o *OS) Join(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func JoinIDPath(ids ...int64) string {\n\tidStr := make([]string, 0, len(ids))\n\tfor _, id := range ids {\n\t\tidStr = append(idStr, strconv.FormatInt(id, 10))\n\t}\n\treturn path.Join(idStr...)\n}", "func (*Mysqlfs) Join(elem ...string) string {\n\treturn filepath.Join(elem...)\n}", "func Join(path string, dir string) string {\n\treturn filepath.Join(path, dir)\n}", "func (v *traversedValues) joinFieldPath(path, partPath string) error {\n\tif partPath == \"\" {\n\t\treturn errReallyWrong\n\t}\n\n\tif path == \"\" {\n\t\tv.fieldPath = partPath\n\t} else {\n\t\tv.fieldPath = path + FieldPathSeparator + partPath\n\t}\n\n\treturn nil\n}", "func joinURL(baseURL string, paths ...string) string {\n\n\tu, err := url.Parse(baseURL)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\turl := path.Join(paths...)\n\tu.Path = path.Join(u.Path, url)\n\treturn u.String()\n}", "func JoinURL(baseURL, subPath string) (string, error) {\n\tparsedURL, err := url.Parse(baseURL)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tparsedURL.RawPath = path.Join(parsedURL.RawPath, subPath)\n\tparsedURL.Path = path.Join(parsedURL.Path, subPath)\n\tURL := parsedURL.String()\n\treturn URL, nil\n}", "func JoinNSPath(path string) (bool, int, error) {\n\treturn false, -1, errors.New(\"this function is not supported on this os3\")\n}", "func (b *Builder) WithPathJoin(basePath, relPath string) *Builder {\n\tb.path = HostPath(filepath.Join(basePath, relPath))\n\treturn b\n}", "func RequireJoinPathSegments(segments ...string) string {\n\tpath, err := JoinPathSegments(segments...)\n\tpanicOnErr(err)\n\treturn path\n}", "func UrlJoin(baseUrl string, elem ...string) (string, error) {\n\tu, err := url.Parse(baseUrl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(elem) > 0 {\n\t\telem = append([]string{u.Path}, elem...)\n\t\tu.Path = path.Join(elem...)\n\t}\n\treturn u.String(), nil\n}", "func joinUrl(url string, endpoint ...string) string {\n\tp := url\n\tfor _, e := range endpoint {\n\t\tp += \"/\" + e\n\t}\n\treturn p\n}", "func JoinWords(w string, words ...string) (path string) {\n\tpath = w\n\tif path[0] != '/' {\n\t\tpath = \"/\" + path\n\t}\n\tfor _, s := range words {\n\t\tpath += \"/\" + s\n\t}\n\treturn\n}", "func xPath(root string, tail string) string {\n\tif !strings.HasPrefix(root, \"/\") {\n\t\troot = \"/\" + root\n\t}\n\treturn path.Join(root, tail)\n}", "func concatURLPaths(host string, path []string) (string, error) {\n\tif host == \"\" {\n\t\treturn \"\", errors.New(\"host can not be empty\")\n\t}\n\n\telements := make([]string, len(path)+1)\n\telements[0] = strings.TrimSuffix(host, \"/\")\n\n\tfor k, v := range path {\n\t\telements[k+1] = strings.Trim(v, \"/\")\n\t}\n\n\treturn strings.Join(elements, \"/\"), nil\n}", "func (fsr *fsLockedRepo) join(paths ...string) string {\n\treturn filepath.Join(append([]string{fsr.path}, paths...)...)\n}", "func (i *Idb) joinBaseURL(p ...string) *url.URL {\n\tu := new(url.URL)\n\t*u = *i.url\t\n\tu.Path = path.Join(append([]string{u.Path}, p...)...)\n\treturn u\n}", "func Concat(path, path2 []string) []string {\n\tpathCopy := make([]string, len(path))\n\tcopy(pathCopy, path)\n\n\tif len(path) == 0 {\n\t\tpath2Copy := make([]string, len(path2))\n\t\tcopy(path2Copy, path)\n\t\treturn path2Copy\n\t}\n\n\tif len(path2) == 0 {\n\t\treturn pathCopy\n\t}\n\n\tif path[len(path)-1] != path2[0] {\n\t\tlog.Fatalln(\"Tried to compute\", path, \"+\", path2, \"but their ends differ\")\n\t}\n\n\tpathCopy = append(pathCopy[:len(pathCopy)-1], path2...)\n\treturn pathCopy\n}", "func SecureJoin(root, unsafePath string) (string, error) {\n\treturn SecureJoinVFS(root, unsafePath, nil)\n}", "func Join(owner, name string) string {\n\treturn owner + \"/\" + name\n}", "func (s *Steps) Join(step Step) Path {\n\tsteps := NewSteps()\n\tsteps.parent = s\n\tsteps.step = step\n\treturn steps\n}", "func WebPath(target, start string) string {\n\ttarget = Posixfy(target)\n\tstart = Posixfy(start)\n\turiPrefix := \"\"\n\tisWebroot := IsWebRoot(target)\n\n\tif !isWebroot && start != \"\" {\n\t\ttarget = start + \"/\" + target\n\t\tif strings.Contains(target, \":\") {\n\t\t\tif res := regexps.UriSniffRx.FindStringSubmatchIndex(target); len(res) == 4 {\n\t\t\t\turiPrefix = target[:res[3]]\n\t\t\t\ttarget = target[res[3]:]\n\t\t\t}\n\t\t}\n\t}\n\t// BUG? slash seems to be lost in https://github.com/asciidoctor/asciidoctor/blob/ab1e0b9c45e5138394b089dac205fb6d854e15e6/lib/asciidoctor/path_resolver.rb#L352-L366\n\tisWebroot = IsWebRoot(target)\n\tif testpr == \"test_Webath_uriPrefix\" {\n\t\treturn fmt.Sprintf(\"target='%v', uriPrefix='%v'\", target, uriPrefix)\n\t}\n\ttargetSegments, targetRoot, _ := PartitionPath(target, true)\n\tif testpr == \"test_Webath_partitionTarget\" {\n\t\treturn fmt.Sprintf(\"targetSegments=(%v)'%v', targetRoot='%v'\", len(targetSegments), targetSegments, targetRoot)\n\t}\n\taccum := []string{}\n\tfor _, segment := range targetSegments {\n\t\tif segment == \"..\" {\n\t\t\tif len(accum) == 0 {\n\t\t\t\tif targetRoot == \"\" || targetRoot == \".\" {\n\t\t\t\t\taccum = append(accum, segment)\n\t\t\t\t}\n\t\t\t} else if accum[len(accum)-1] == \"..\" {\n\t\t\t\taccum = append(accum, segment)\n\t\t\t} else {\n\t\t\t\taccum = accum[:len(accum)-1]\n\t\t\t}\n\t\t} else {\n\t\t\taccum = append(accum, segment)\n\t\t}\n\t}\n\tresolvedSegments := accum\n\n\tjoinPath := JoinPath(resolvedSegments, targetRoot)\n\tif uriPrefix != \"\" {\n\t\treturn uriPrefix + joinPath\n\t}\n\tif isWebroot {\n\t\treturn \"/\" + joinPath\n\t}\n\treturn joinPath\n}", "func Join(sep string, parts ...string) string {\n\treturn strings.Join(parts, sep)\n}", "func joinFilePath(path, file string) string {\n\treturn filepath.Join(path, file)\n}", "func JoinIfNeeded(parent *url.Resource, URI string) (result *url.Resource) {\n\tdefer func() {\n\t\tif parent != nil {\n\t\t\tif result.Credentials == \"\" {\n\t\t\t\tresult.Credentials = parent.Credentials\n\t\t\t}\n\t\t\tif result.CustomKey == nil {\n\t\t\t\tresult.CustomKey = parent.CustomKey\n\t\t\t}\n\t\t}\n\t}()\n\tif strings.Contains(URI, \":/\") {\n\t\tresult = url.NewResource(URI)\n\t} else if !(strings.HasPrefix(URI, \"/\") || strings.HasPrefix(URI, \"$\")) {\n\t\tvar hostname = parent.ParsedURL.Hostname()\n\t\tif hostname == \"\" || hostname == \"127.0.0.1\" || hostname == \"localhost\" {\n\t\t\tvar candidate = url.NewResource(URI)\n\n\t\t\tif toolbox.FileExists(candidate.ParsedURL.Path) {\n\t\t\t\tresult = candidate\n\t\t\t}\n\t\t}\n\t\tif result == nil {\n\t\t\tresult = url.NewResource(toolbox.URLPathJoin(parent.URL, URI))\n\t\t}\n\n\t} else if parent != nil {\n\t\tresult = url.NewResource(toolbox.URLPathJoin(parent.URL, URI))\n\t} else {\n\t\tresult = url.NewResource(URI)\n\t}\n\tif strings.HasPrefix(URI, \"$\") { //has to expand to be re-evaluated\n\t\tresult.URL = URI\n\t}\n\n\treturn result\n}", "func JoinNormalized(n Normalization, base string, elem ...string) (string, error) {\n\tif n == NoNorm {\n\t\treturn filepath.Join(append([]string{base}, elem...)...), nil\n\t}\n\treturn joinFold(n == FoldPreferExactNorm, base, elem...)\n}", "func JoinFromHome(paths ...PathRetriever) PathRetriever {\n\treturn Join(append([]PathRetriever{os.UserHomeDir}, paths...)...)\n}", "func prefixJoin(prefix string, array []string, separator string) (result string) {\n\tif len(array) == 0 {\n\t\treturn\n\t}\n\tfor index, val := range array {\n\t\tif index == 0 {\n\t\t\tresult = val\n\t\t} else {\n\t\t\tresult = join(result, concat(prefix, val), separator)\n\t\t}\n\t}\n\treturn\n}", "func AppendSlashInPathAnnotationIfNotPresent(currentAnnotations map[string]interface{}) map[string]interface{} {\n\tif !strings.HasPrefix(currentAnnotations[constants.INGRESS_URL_PATH].(string), \"/\") {\n\t\tcurrentAnnotations[constants.INGRESS_URL_PATH] = \"/\" + currentAnnotations[constants.INGRESS_URL_PATH].(string)\n\t}\n\n\treturn currentAnnotations\n}", "func Join(elem ...string) string {\n\treturn std.Join(elem...)\n}", "func joinCoverPath(pwd, fileName string) string {\n\tpwdPath := lessEmpty(strings.Split(pwd, \"/\"))\n\tfileDirPath := lessEmpty(strings.Split(filepath.Dir(fileName), \"/\"))\n\n\tfor index, dir := range pwdPath {\n\t\tif dir == first(fileDirPath) {\n\t\t\tpwdPath = pwdPath[:index]\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn filepath.Join(maybePrefix(strings.Join(pwdPath, \"/\"), \"/\"), fileName)\n}", "func cleanPath(p string) string {\n\t// Turn empty string into \"/\"\n\tif p == \"\" {\n\t\treturn \"/\"\n\t}\n\n\tn := len(p)\n\tvar buf []byte\n\n\t// Invariants:\n\t// reading from path; r is index of next byte to process.\n\t// writing to buf; w is index of next byte to write.\n\n\t// path must start with '/'\n\tr := 1\n\tw := 1\n\n\tif p[0] != '/' {\n\t\tr = 0\n\t\tbuf = make([]byte, n+1)\n\t\tbuf[0] = '/'\n\t}\n\n\ttrailing := n > 2 && p[n-1] == '/'\n\n\t// A bit more clunky without a 'lazybuf' like the path package, but the loop\n\t// gets completely inlined (bufApp). So in contrast to the path package this\n\t// loop has no expensive function calls (except 1x make)\n\n\tfor r < n {\n\t\tswitch {\n\t\tcase p[r] == '/':\n\t\t\t// empty path element, trailing slash is added after the end\n\t\t\tr++\n\n\t\tcase p[r] == '.' && r+1 == n:\n\t\t\ttrailing = true\n\t\t\tr++\n\n\t\tcase p[r] == '.' && p[r+1] == '/':\n\t\t\t// . element\n\t\t\tr++\n\n\t\tcase p[r] == '.' && p[r+1] == '.' && (r+2 == n || p[r+2] == '/'):\n\t\t\t// .. element: remove to last /\n\t\t\tr += 2\n\n\t\t\tif w > 1 {\n\t\t\t\t// can backtrack\n\t\t\t\tw--\n\n\t\t\t\tif buf == nil {\n\t\t\t\t\tfor w > 1 && p[w] != '/' {\n\t\t\t\t\t\tw--\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tfor w > 1 && buf[w] != '/' {\n\t\t\t\t\t\tw--\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\tdefault:\n\t\t\t// real path element.\n\t\t\t// add slash if needed\n\t\t\tif w > 1 {\n\t\t\t\tbufApp(&buf, p, w, '/')\n\t\t\t\tw++\n\t\t\t}\n\n\t\t\t// copy element\n\t\t\tfor r < n && p[r] != '/' {\n\t\t\t\tbufApp(&buf, p, w, p[r])\n\t\t\t\tw++\n\t\t\t\tr++\n\t\t\t}\n\t\t}\n\t}\n\n\t// re-append trailing slash\n\tif trailing && w > 1 {\n\t\tbufApp(&buf, p, w, '/')\n\t\tw++\n\t}\n\n\tif buf == nil {\n\t\treturn p[:w]\n\t}\n\treturn string(buf[:w])\n}", "func (o Operator) Join(arr []string) string {\n\tvar str string\n\n\tswitch o {\n\tcase \"/\":\n\t\tfor i, v := range arr {\n\t\t\tif v[:1] != \"{\" {\n\t\t\t\tarr[i] = \"/\" + v\n\t\t\t}\n\t\t}\n\t\tstr = filepath.Join(strings.Join(arr, \"\"))\n\n\tcase \"#\", \"?\", \".\", \";\", \"&\":\n\t\tm := opmap[o]\n\t\tfor i, v := range arr {\n\t\t\tif i > 0 && v[:1] != \"{\" {\n\t\t\t\tarr[i] = m[1] + v\n\t\t\t}\n\t\t}\n\t\tstr = m[0] + strings.Join(arr, \"\")\n\n\t\t// TODO revisit, not particularly pretty\n\t\tif str[:2] == \"&{\" {\n\t\t\tstr = str[1:] // remove extra &\n\t\t}\n\n\tdefault: // handles +, `{+var}` and blank, `{var}`\n\t\tstr = strings.Join(arr, \",\")\n\t}\n\n\treturn str\n}", "func JoinRootPath(root, unsafePath string) string {\n\tunsafePath = path.Clean(unsafePath)\n\tif unsafePath[0] != '/' {\n\t\tunsafePath = \"/\" + unsafePath\n\t}\n\tfor strings.HasPrefix(unsafePath, \"/..\") {\n\t\tunsafePath = unsafePath[3:]\n\t}\n\treturn path.Join(root, unsafePath)\n}", "func join(sep string, a ...string) string {\n\tswitch len(a) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn a[0]\n\t}\n\n\tres := bytes.NewBufferString(a[0])\n\tfor _, s := range a[1:] {\n\t\tres.WriteString(sep + s)\n\t}\n\n\treturn res.String()\n}", "func (s *Scope) appendPaths(path string, parts ...string) string {\n\tpaths := []string{path, s.Vendor, s.App}\n\tpaths = append(paths, parts...)\n\treturn filepath.Join(paths...)\n}", "func shiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\treturn p[1:], \"/\"\n\t}\n\treturn p[1:i], p[i:]\n}", "func AppendString(p Path, elem ...string) Path {\n\treturn Append(p, ToPaths(elem)...)\n}", "func shiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\treturn p[1:], \"/\"\n\t}\n\n\treturn p[1:i], p[1:]\n}", "func (n UDN) Join(s Name) Name {\n\treturn MustParse(n.String() + \".\" + s.String())\n}", "func Join(trustDomain string, segments ...string) (string, error) {\n\tid, err := New(trustDomain, segments...)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn id.String(), nil\n}", "func strPathV03(path *pb.Path) string {\n\treturn \"/\" + strings.Join(path.Element, \"/\")\n}", "func ShiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\treturn p[1:], \"/\"\n\t}\n\treturn p[1:i], p[i:]\n}", "func ShiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\treturn p[1:], \"/\"\n\t}\n\treturn p[1:i], p[i:]\n}", "func ShiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\treturn p[1:], \"/\"\n\t}\n\treturn p[1:i], p[i:]\n}", "func TestPathPrefixOrEqual(t *testing.T) {\n\tif !isPathPrefixOrEqual(\"foo\", \"foo\") {\n\t\tt.Error(\"Same path should return true\")\n\t}\n\n\tif isPathPrefixOrEqual(\"foo\", \"fooer\") {\n\t\tt.Error(\"foo is not a path-type prefix of fooer\")\n\t}\n\n\tif !isPathPrefixOrEqual(\"foo\", \"foo/bar\") {\n\t\tt.Error(\"foo is a path prefix of foo/bar\")\n\t}\n\n\tif isPathPrefixOrEqual(\"foo\", \"foo/\") {\n\t\tt.Error(\"special case - foo is not a path prefix of foo/\")\n\t}\n}", "func TestPathPrefixOrEqual(t *testing.T) {\n\tif !isPathPrefixOrEqual(\"foo\", \"foo\") {\n\t\tt.Error(\"Same path should return true\")\n\t}\n\n\tif isPathPrefixOrEqual(\"foo\", \"fooer\") {\n\t\tt.Error(\"foo is not a path-type prefix of fooer\")\n\t}\n\n\tif !isPathPrefixOrEqual(\"foo\", \"foo/bar\") {\n\t\tt.Error(\"foo is a path prefix of foo/bar\")\n\t}\n\n\tif isPathPrefixOrEqual(\"foo\", \"foo/\") {\n\t\tt.Error(\"special case - foo is not a path prefix of foo/\")\n\t}\n}", "func join(a string, b string, separator string) string {\n\tvals := make([]byte, 0, 10)\n\tvals = append(vals, a...)\n\tvals = append(vals, separator...)\n\tvals = append(vals, b...)\n\treturn string(vals)\n}", "func absJoin(dir, name string) (string, error) {\n\tif name == \"\" {\n\t\treturn filepath.Abs(dir)\n\t}\n\n\tif filepath.IsAbs(name) {\n\t\treturn name, nil\n\t}\n\n\tif len(name) > 0 && name[0] == '~' && (len(name) == 1 || name[1] == '/' || name[1] == '\\\\') {\n\t\treturn expendHome(name)\n\t}\n\n\treturn absPath(filepath.Join(dir, name))\n}", "func ShiftPath(p string) (head, tail string) {\n\tp = path.Clean(\"/\" + p)\n\ti := strings.Index(p[1:], \"/\") + 1\n\tif i <= 0 {\n\t\t//log.Printf(\"head: %s, tail: /\", p[1:])\n\t\treturn p[1:], \"/\"\n\t}\n\t//log.Printf(\"head: %s, tail: %s\", p[1:i], p[i:])\n\treturn p[1:i], p[i:]\n}", "func (project Project) RelPath(subpath ...string) string {\n\tfullpath := []string{}\n\tfullpath = append(fullpath, project.Path)\n\tfullpath = append(fullpath, subpath...)\n\treturn filepath.Join(fullpath...)\n}", "func (s *Scope) appendPaths(path string, parts ...string) string {\n\tpaths := []string{path}\n\tpaths = append(paths, parts...)\n\treturn filepath.Join(paths...)\n}", "func Join(s []string, sep string) string {\n\tvar buf string\n\tfor _, v := range s[:len(s)-1] {\n\t\tbuf += fmt.Sprintf(\"%s%s\", v, sep)\n\t}\n\n\tbuf += s[len(s)-1]\n\treturn buf\n}", "func Join(sep string, strs ...string) string {\n\tswitch len(strs) {\n\tcase 0:\n\t\treturn \"\"\n\tcase 1:\n\t\treturn strs[0]\n\tcase 2:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn strs[0] + sep + strs[1]\n\tcase 3:\n\t\t// Special case for common small values.\n\t\t// Remove if golang.org/issue/6714 is fixed\n\t\treturn strs[0] + sep + strs[1] + sep + strs[2]\n\t}\n\tn := len(sep) * (len(strs) - 1)\n\tfor i := 0; i < len(strs); i++ {\n\t\tn += len(strs[i])\n\t}\n\n\tb := make([]byte, n)\n\tbp := copy(b, strs[0])\n\tfor _, s := range strs[1:] {\n\t\tbp += copy(b[bp:], sep)\n\t\tbp += copy(b[bp:], s)\n\t}\n\treturn string(b)\n}", "func Join(a, b Pair) Pair {\n\treturn nest{a, b}\n}", "func (d HostDir) Path(elem ...string) string {\n\treturn filepath.Join(append([]string{string(d)}, elem...)...)\n}", "func CommonPath(path1, path2 string) string {\n\tparts1 := strings.Split(path1, \"/\")\n\tparts2 := strings.Split(path2, \"/\")\n\tmaxIdx := len(parts1)\n\tif len(parts2) < maxIdx {\n\t\tmaxIdx = len(parts2)\n\t}\n\n\tcommon := make([]string, 0, maxIdx)\n\tfor i := 0; i < maxIdx; i++ {\n\t\tif parts1[i] == parts2[i] {\n\t\t\tcommon = append(common, parts1[i])\n\t\t}\n\t}\n\n\treturn strings.Join(common, \"/\")\n}", "func Join(a []string, sep string) string {\n\treturn strings.Join(a, sep)\n}", "func Join(sep string, strs ...string) string {\n\tvar buf bytes.Buffer\n\tif len(strs) == 0 {\n\t\treturn \"\"\n\t}\n\tfor _, str := range strs {\n\t\tbuf.WriteString(str + sep)\n\t}\n\treturn strings.TrimRight(buf.String(), sep)\n}", "func SecureJoinVFS(root, unsafePath string, vfs VFS) (string, error) {\n\t// Use the os.* VFS implementation if none was specified.\n\tif vfs == nil {\n\t\tvfs = osVFS{}\n\t}\n\n\tvar path bytes.Buffer\n\tn := 0\n\tfor unsafePath != \"\" {\n\t\tif n > 255 {\n\t\t\treturn \"\", &os.PathError{Op: \"SecureJoin\", Path: root + \"/\" + unsafePath, Err: syscall.ELOOP}\n\t\t}\n\n\t\t// Next path component, p.\n\t\ti := strings.IndexRune(unsafePath, filepath.Separator)\n\t\tvar p string\n\t\tif i == -1 {\n\t\t\tp, unsafePath = unsafePath, \"\"\n\t\t} else {\n\t\t\tp, unsafePath = unsafePath[:i], unsafePath[i+1:]\n\t\t}\n\n\t\t// Create a cleaned path, using the lexical semantics of /../a, to\n\t\t// create a \"scoped\" path component which can safely be joined to fullP\n\t\t// for evaluation. At this point, path.String() doesn't contain any\n\t\t// symlink components.\n\t\tcleanP := filepath.Clean(string(filepath.Separator) + path.String() + p)\n\t\tif cleanP == string(filepath.Separator) {\n\t\t\tpath.Reset()\n\t\t\tcontinue\n\t\t}\n\t\tfullP := filepath.Clean(root + cleanP)\n\n\t\t// Figure out whether the path is a symlink.\n\t\tfi, err := vfs.Lstat(fullP)\n\t\tif err != nil && !IsNotExist(err) {\n\t\t\treturn \"\", err\n\t\t}\n\t\t// Treat non-existent path components the same as non-symlinks (we\n\t\t// can't do any better here).\n\t\tif IsNotExist(err) || fi.Mode()&os.ModeSymlink == 0 {\n\t\t\tpath.WriteString(p)\n\t\t\tpath.WriteRune(filepath.Separator)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Only increment when we actually dereference a link.\n\t\tn++\n\n\t\t// It's a symlink, expand it by prepending it to the yet-unparsed path.\n\t\tdest, err := vfs.Readlink(fullP)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\t// Absolute symlinks reset any work we've already done.\n\t\tif filepath.IsAbs(dest) {\n\t\t\tpath.Reset()\n\t\t}\n\t\tunsafePath = dest + string(filepath.Separator) + unsafePath\n\t}\n\n\t// We have to clean path.String() here because it may contain '..'\n\t// components that are entirely lexical, but would be misleading otherwise.\n\t// And finally do a final clean to ensure that root is also lexically\n\t// clean.\n\tfullP := filepath.Clean(string(filepath.Separator) + path.String())\n\treturn filepath.Clean(root + fullP), nil\n}", "func (dsn *S3DSN) Join(filename string) string {\n\treturn filepath.Join(filepath.Dir(dsn.Key), filename)\n}", "func toPath(path string) string {\n\tpath = strings.TrimSpace(strings.ToLower(path))\n\n\tif !strings.HasPrefix(path, \"/\") {\n\t\tpath += \"/\" + path\n\t}\n\n\treturn path\n}", "func NormalizeJoin(l []string, d, f string) string {\n n := len(l)\n var s string\n for i, e := range l {\n if i > 0 {\n if n - (i + 1) == 0 {\n s += f\n }else{\n s += d\n }\n }\n s += e\n }\n return s\n}", "func Join(sep string, operand []string) string { return strings.Join(operand, sep) }", "func CreatePath(redditPath string) (Path, error) {\n\n\tredditUrl, error := url.Parse(redditPath)\n\tif error != nil {\n\t\treturn \"\", errors.New(\"Bad reddit path: \" + error.Error())\n\t}\n\n\tredditUrl.Scheme = \"\"\n\tredditUrl.Host = \"\"\n\n\tif !strings.HasSuffix(redditUrl.Path, redditJsonSuffix) {\n\t\tredditUrl.Path = redditUrl.Path + redditJsonSuffix\n\t}\n\n\t// before/after is added when path is fetched\n\tif len(redditUrl.RawQuery) != 0 {\n\t\tqueries := redditUrl.Query()\n\t\tqueries.Del(redditBeforeModifier)\n\t\tqueries.Del(redditAfterModifier)\n\t\tredditUrl.RawQuery = queries.Encode()\n\t}\n\n\t// flatten once to (hopefully) reduce garbage\n\tflattenedPath := redditUrl.String()\n\n\t// this is the sketchiest part -- assume dangling ? and & are\n\t// acceptable, and assume they won't be present in these situations ...\n\t// unittests can at least catch unexpected string values.\n\tif len(redditUrl.RawQuery) == 0 { // len can change after removal above\n\t\tflattenedPath = flattenedPath + \"?\"\n\t} else {\n\t\tflattenedPath = flattenedPath + \"&\"\n\t}\n\n\treturn Path(flattenedPath), nil\n}" ]
[ "0.74658626", "0.7450469", "0.74454105", "0.7416059", "0.7216877", "0.714421", "0.7119705", "0.7096432", "0.70933175", "0.7088609", "0.69537205", "0.6947479", "0.6940326", "0.68680114", "0.6841759", "0.6784506", "0.6784506", "0.67776614", "0.67571795", "0.6740385", "0.66987973", "0.66825145", "0.66656315", "0.6627219", "0.65561", "0.6532324", "0.6479602", "0.6436864", "0.6424805", "0.62435645", "0.61490643", "0.61171967", "0.6109114", "0.6085539", "0.6054834", "0.5973523", "0.5915371", "0.5894258", "0.58915895", "0.58808845", "0.5810446", "0.5792829", "0.5756065", "0.5733697", "0.5679467", "0.56673896", "0.5622405", "0.5597211", "0.5522277", "0.55103767", "0.55057555", "0.54949915", "0.548538", "0.54709166", "0.54433787", "0.543808", "0.54375124", "0.5429611", "0.5385146", "0.53815496", "0.5346731", "0.5306816", "0.5288116", "0.52831745", "0.52572674", "0.52421945", "0.5211765", "0.5199959", "0.51869786", "0.51572645", "0.5156773", "0.51563156", "0.5144309", "0.51430726", "0.51386213", "0.51246953", "0.5117493", "0.51160216", "0.51160216", "0.51160216", "0.5113563", "0.5113563", "0.510539", "0.5102043", "0.5100661", "0.508405", "0.5082733", "0.5071263", "0.5059671", "0.4996127", "0.49873096", "0.49832305", "0.4981419", "0.49756393", "0.4955632", "0.49478745", "0.4944064", "0.49240074", "0.4916063", "0.49123198" ]
0.7110337
7
New creates a new extension instance
func New(store IDStore) *Extension { defaultVal := unsupported return &Extension{supportedByServer: &defaultVal, replayStore: store} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *SpotifyExtender) New() {\r\n\r\n}", "func NewExtension(ctx context.Context, _ *controller.Impl) operator.Extension {\n\treturn &extension{\n\t\tkubeclient: kubeclient.Get(ctx),\n\t}\n}", "func NewExt(zwsId, baseUrl string) Zillow {\n\treturn &zillow{zwsId, baseUrl}\n}", "func NewSchemaExtension()(*SchemaExtension) {\n m := &SchemaExtension{\n Entity: *NewEntity(),\n }\n return m\n}", "func NewExtension(ctx context.Context, impl *controller.Impl) operator.Extension {\n\tdeploymentInformer := deploymentinformer.Get(ctx)\n\n\t// We move the Kourier deployments into a different namespace so the usual informer\n\t// that enqueues the OwnerRef doesn't catch those, so we add them here explicitly.\n\tdeploymentInformer.Informer().AddEventHandler(cache.FilteringResourceEventHandler{\n\t\tFilterFunc: reconciler.LabelExistsFilterFunc(socommon.ServingOwnerNamespace),\n\t\tHandler: controller.HandleAll(impl.EnqueueLabelOfNamespaceScopedResource(socommon.ServingOwnerNamespace, socommon.ServingOwnerName)),\n\t})\n\n\treturn &extension{\n\t\tocpclient: ocpclient.Get(ctx),\n\t\tkubeclient: kubeclient.Get(ctx),\n\t}\n}", "func New(parts []interface{}, opts Options) *Blob {\n\tblob := js.Global.Get(\"Blob\").New(parts, opts)\n\treturn &Blob{Object: blob}\n}", "func New(cmd *cobra.Command, args []string) {\n\t// Create object for current working directory\n\tpwd, err := teflon.NewTeflonObject(\".\")\n\tif err != nil {\n\t\tlog.Fatalln(\"Couldn't create object for '.' :\", err)\n\t}\n\n\t// Create a show.\n\tif showFlag {\n\t\tnshws, err := pwd.CreateShow(args[0], newShowProtoFlag)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(\"ABORT: Couldnt create show:\", err)\n\t\t}\n\t\tfor _, shw := range nshws {\n\t\t\tfmt.Println(shw.Path)\n\t\t}\n\t\treturn\n\t}\n\n\t// If nothing else commands otherwise new will create an ordinary file-system\n\t// object.\n\tnobjs, err := pwd.CreateObject(args[0], newFileFlag)\n\tif err != nil {\n\t\tlog.Fatalln(\"ABORT: Couldn't create objects:\", err)\n\t}\n\tclose(teflon.Events)\n\tfor _, obj := range nobjs {\n\t\tfmt.Println(obj.Path)\n\t}\n}", "func New(classArgs ...interface{}) *Ex {\n\treturn &Ex{class: fmt.Sprint(classArgs...), stack: callers()}\n}", "func New() Object {\n\treturn Object{}\n}", "func New() Go { return Go{} }", "func New() {\n\ttypeOfProject()\n}", "func New(w http.ResponseWriter, r *http.Request) {\n\tgetTemplates().ExecuteTemplate(w, \"New\", nil)\n}", "func New(s shorter.Store, c interface{}) (shorter.Shorter, error) {\n\t// load extends configuration\n\t// TODO: implement\n\treturn nil, errors.New(\"unimplemented\")\n}", "func (p *Patch) New() Module {\n\treturn &Patch{}\n}", "func New() *file {\n\treturn &file{}\n}", "func New(s string) *Exp {\n\texp := &Exp{context: s}\n\texp.init()\n\treturn exp\n}", "func new() exampleInterface {\n\treturn config{}\n}", "func NewExtensionManifest(raw *string) graphqlbackend.ExtensionManifest {\n\tif raw == nil {\n\t\treturn nil\n\t}\n\treturn &extensionManifest{raw: *raw}\n}", "func New(v interface{}) (provider.Provider, error) {\n\ts := Spec{}\n\treturn &s, ioutil.Intermarshal(v, &s)\n}", "func New(storage *storage.Storage, execmode bool) ModuleManager {\n\tmm := ModuleManager{}\n\tmm.modules = make(map[string]*shared.Module)\n\n\tmm.storage = storage\n\n\t// set settings\n\tmm.settings.execMode = execmode\n\tlog.Println(mm)\n\treturn mm\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func New(w http.ResponseWriter, r *http.Request) {\r\n\ttmpl.ExecuteTemplate(w, \"New\", nil)\r\n}", "func New(data []byte, filename string) (*Exec, error) {\n\tlog.Tracef(\"Creating new at %v\", filename)\n\treturn loadExecutable(filename, data)\n}", "func NewEx() *Ex {\n\treturn &Ex{input: NewInput()}\n}", "func newObserver(logger *zap.Logger, config *Config) (component.Extension, error) {\n\treturn &dockerObserver{logger: logger, config: config}, nil\n}", "func New(next goproxy.Plugin, cache FileCache) goproxy.Plugin {\n\treturn &plugin{next: next, cache: cache}\n}", "func New() *Entitlements {\n\treturn &Entitlements{\n\t\tAPS: &APS{},\n\t\tDataProtection: &DataProtection{},\n\t\tcustom: map[string]interface{}{},\n\t}\n}", "func (o *JS) New(args []interface{}) Object {\n\treturn &JS{o.object.New(args...)}\n}", "func New() *Command {\n\treturn &Command{}\n}", "func New() base.Mod {\n\treturn &TestMod{\n\t\tname: \"Test\",\n\t\tcommands: make(map[string]*base.ModCommand),\n\t\tallowedTypes: base.MessageTypeCreate,\n\t\tallowDMs: true,\n\t}\n}", "func New(zwsId string) Zillow {\n\treturn NewExt(zwsId, baseUrl)\n}", "func New() Builder {\n\treturn newBuilder()\n}", "func newInstance(x *runtime.Runtime, p *build.Instance, v *adt.Vertex) *Instance {\n\t// TODO: associate root source with structLit.\n\tinst := &Instance{\n\t\troot: v,\n\t\tinst: p,\n\t}\n\tif p != nil {\n\t\tinst.ImportPath = p.ImportPath\n\t\tinst.Dir = p.Dir\n\t\tinst.PkgName = p.PkgName\n\t\tinst.DisplayName = p.ImportPath\n\t\tif p.Err != nil {\n\t\t\tinst.setListOrError(p.Err)\n\t\t}\n\t}\n\n\tx.AddInst(p.ImportPath, v, p)\n\tx.SetBuildData(p, inst)\n\tinst.index = x\n\treturn inst\n}", "func New() (interface{}, error) {\n\treturn Backend(), nil\n}", "func New() *Plugin {\n\treturn &Plugin{}\n}", "func (c MethodsCollection) New() pNew {\n\treturn pNew{\n\t\tMethod: c.MustGet(\"New\"),\n\t}\n}", "func New(config config.Config) Plugins {\n\tpkg := Plugins{\n\t\tclient: http.NewClient(config),\n\t\tpager: &http.LinkHeaderPager{},\n\t}\n\n\treturn pkg\n}", "func New(msg string) error {\n\treturn &basic{msg}\n}", "func (f StorageFileWebp) construct() StorageFileTypeClass { return &f }", "func (o *FakeObject) New(args ...interface{}) Object { return o.Invoke(args) }", "func NewFoo() *Foo {\n return &Foo{}\n}", "func New() (*Plugin, error) {\n\treturn &Plugin{\n\t\tHandler: admission.NewHandler(admission.Create, admission.Update),\n\t}, nil\n}", "func New() {\n\tfmt.Println(POLICY_TEMPLATE_OBJECT)\n}", "func newFileFetcher() *fileFetcherExtender {\n\treturn &fileFetcherExtender{new(DefaultExtender)}\n}", "func New(opts ...Option) (p9.Attacher, error) {\n\ta := &attacher{\n\t\tfiles: make(map[string]string),\n\t\tqids: &p9.QIDGenerator{},\n\t}\n\tfor _, o := range opts {\n\t\tif err := o(a); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn a, nil\n}", "func New(chain []byte) (p *Provider, err error) {\n\tp = &Provider{}\n\tif err = p.Decode(chain); err != nil {\n\t\treturn nil, err\n\t}\n\treturn p, nil\n}", "func New() error {\n\treturn &errorMessage{\"Find the bug\"}\n}", "func New() *Storage {\n\treturn &Storage{\n\t\tstat: &statApp{},\n\t}\n}", "func New() *cobra.Command {\n\tcommand := cobra.Command{\n\t\tUse: \"%template%\",\n\t\tShort: \"%template%\",\n\t\tLong: \"%template%\",\n\n\t\tArgs: cobra.NoArgs,\n\n\t\tSilenceErrors: false,\n\t\tSilenceUsage: true,\n\t}\n\n\t/* configure instance */\n\tcommand.AddCommand(\n\t\tdemo.Panic(),\n\t\tdemo.Stderr(),\n\t\tdemo.Stdout(),\n\t)\n\n\treturn &command\n}", "func (t *Text) New() Module {\n\treturn &Text{}\n}", "func New(settings *environment.Settings) *cobra.Command {\n\treturn newCmd(settings, httpclient.New())\n}", "func NewExtensionModule(ctx *Context, hideOutput bool) Module {\n\twork := newExtWorker(ctx)\n\tif hideOutput {\n\t\treturn newModule(work)\n\t}\n\trend := newExtRenderer(work, ctx.top)\n\treturn newModule(work, rend)\n}", "func New(path string) (Plugin, error) {\n\tplugin, err := plugin.Open(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp, err := plugin.Lookup(\"Plugin\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn p.(Plugin), nil\n}", "func New(p provider) (*Command, error) {\n\tverifiableStore, err := verifiablestore.New(p)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"new vc store : %w\", err)\n\t}\n\n\tdidStore, err := didstore.New(p)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"new did store : %w\", err)\n\t}\n\n\treturn &Command{\n\t\tverifiableStore: verifiableStore,\n\t\tdidStore: didStore,\n\t\tresolver: verifiable.NewVDRKeyResolver(p.VDRegistry()),\n\t\tctx: p,\n\t\tdocumentLoader: p.JSONLDDocumentLoader(),\n\t}, nil\n}", "func New() *Gjango {\n\treturn &Gjango{}\n}", "func New() *M {\n\tc := &M{}\n\tc.Component()\n\tc.items = make([]*js.Object, 0)\n\treturn c\n}", "func Newf(classFormat string, args ...interface{}) *Ex {\n\treturn &Ex{class: fmt.Sprintf(classFormat, args...), stack: callers()}\n}", "func New() *differ {\n\treturn &differ{}\n}", "func New(engine cas.Engine, src casext.DescriptorPath) (*Mutator, error) {\n\t// We currently only support changing a given manifest through a walk.\n\tif mt := src.Descriptor().MediaType; mt != ispec.MediaTypeImageManifest {\n\t\treturn nil, errors.Errorf(\"unsupported source type: %s\", mt)\n\t}\n\n\treturn &Mutator{\n\t\tengine: casext.NewEngine(engine),\n\t\tsource: src,\n\t}, nil\n}", "func New() *component {\n\treturn &component{}\n}", "func New() *component {\n\treturn &component{}\n}", "func New(plArgs *runtime.Unknown, _ framework.FrameworkHandle) (framework.Plugin, error) {\n\targs := &Args{}\n\tif err := framework.DecodeInto(plArgs, args); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &NodeLabel{\n\t\tpredicate: predicates.NewNodeLabelPredicate(args.Labels, args.Presence),\n\t}, nil\n}", "func New(cfg Config) Module { return Module{Factory: cfg} }", "func New(key string) *Dot {\n\tdot := &Dot{\n\t\ttag.New(key), // init key\n\t\tlsm.New(), // init val\n\t\tnew(sync.RWMutex), // private mutex\n\t}\n\treturn dot\n}", "func New(configfile string) *Poloniex {\r\n\treturn NewWithConfig(configfile)\r\n}", "func New() provider.Provider {\n\tp := newProvider()\n\n\treturn p\n}", "func New(m map[string]interface{}) (app.Provider, error) {\n\tc, err := parseConfig(m)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &provider{iframeUIProvider: c.IFrameUIProvider}, nil\n}", "func New(handle interfaces.FrameworkHandle) (interfaces.Plugin, error) {\n\treturn &DefaultBinder{handle: handle}, nil\n}", "func New(name string) *Framework {\n\tfmw.Name = name\n\tfmw.init()\n\treturn fmw\n}", "func New(app *grun.App, newW func() gtk.Widgetter) *Maker {\n\tbaseID := app.ID\n\tif baseID == \"\" {\n\t\tbaseID = \"com.github.gotk4.gtkest.default\"\n\t}\n\treturn &Maker{app: app, newW: newW, baseID: baseID}\n}", "func New(cnf *config.Config) iface.Backend {\n\treturn &Backend{Backend: common.NewBackend(cnf)}\n}", "func New(behavior Func) structmap.Behavior {\n\treturn behavior\n}", "func New(name string) *File {\n\treturn &File{Name: name}\n}", "func NewAddCommand() *cobra.Command {\n\treturn NewCmd(\"add\").\n\t\tWithDescription(\"Add goployer manifest file\").\n\t\tSetFlags().\n\t\tRunWithArgs(funcAdd)\n}", "func New(message string) error {\n\treturn root.New(message)\n}", "func New(_ runtime.Object, h framework.FrameworkHandle) (framework.Plugin, error) {\n\treturn &RepeatPriority{handle: h}, nil\n}", "func New(fileType string, filecontent interface{}, filename string) Replacer {\n\tswitch fileType {\n\tcase \"json\":\n\t\treturn &JSON{\n\t\t\tFileName: filename,\n\t\t\treplace: filecontent.(map[string]interface{}),\n\t\t}\n\tdefault:\n\t\treturn &JSON{\n\t\t\tFileName: filename,\n\t\t\treplace: filecontent.(map[string]interface{}),\n\t\t}\n\t}\n}", "func New() Template {\n\treturn Template{}\n}", "func (l *Loader) newExtSym(name string, ver int) Sym {\n\ti := Sym(len(l.objSyms))\n\tif l.extStart == 0 {\n\t\tl.extStart = i\n\t}\n\tl.growSyms(int(i))\n\tpi := l.newPayload(name, ver)\n\tl.objSyms = append(l.objSyms, objSym{l.extReader, int(pi)})\n\tl.extReader.syms = append(l.extReader.syms, i)\n\treturn i\n}", "func (fgen *funcGen) newZExtInst(ident ir.LocalIdent, old *ast.ZExtInst) (*ir.InstZExt, error) {\n\tto, err := fgen.gen.irType(old.To())\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\treturn &ir.InstZExt{LocalIdent: ident, To: to}, nil\n}", "func New(opts ...Option) goldmark.Extender {\n\treturn &emoji{\n\t\toptions: opts,\n\t}\n}", "func NewTransacter(db *sql.DB) *TransactionExtension {\n\treturn &TransactionExtension{\n\t\tdb: db,\n\t}\n}", "func New(ctx activity.InitContext) (activity.Activity, error) {\n\n//\ts := &Settings{}\n//\terr := metadata.MapToStruct(ctx.Settings(), s, true)\n//\tif err != nil {\n//\t\treturn nil, err\n//\t}\n\n//\tctx.Logger().Debugf(\"Setting: %s\", s.ASetting)\n\n\tact := &Activity{} //add aSetting to instance//nothing to add now\n\n\treturn act, nil\n}", "func (c complexSetupWorker) New(conf complexConfig) *complexSetupWorker {\n\t// Do compex setup...\n\treturn &complexSetupWorker{conf}\n}", "func New() *Meta {\n\treturn &Meta{Version: Version}\n}", "func New(kind string, result metadata.Contract) *instance {\n\treturn &instance{\n\t\tkind: kind,\n\t\tresult: result,\n\t}\n}", "func New(label ...string) (result Metadata) {\n\tresult = &metadata{\n\t\tactions: map[string]Action{},\n\t}\n\n\t// global debug\n\tif debugEnabled {\n\t\tresult.Debug()\n\t}\n\n\tif len(label) > 0 {\n\t\tresult.Name(label[0])\n\t}\n\n\treturn\n}", "func New(capabilities ...string) *Capability {\n\treturn &Capability{\n\t\tCapabilities: capabilities,\n\t}\n}", "func New(name string, shortHelp string, longHelp string, flagFn func(*flag.FlagSet), runFn func(*flag.FlagSet) error) *Subcommand {\n\treturn defaultRegistry.new(name, shortHelp, longHelp, flagFn, runFn)\n}", "func execNew(_ int, p *gop.Context) {\n\targs := p.GetArgs(3)\n\tret := doc.New(args[0].(*ast.Package), args[1].(string), doc.Mode(args[2].(int)))\n\tp.Ret(3, ret)\n}", "func New() MME {\n\tvar m MME = new(mme)\n\treturn m\n}", "func New(data []byte) *suffixarray.Index", "func New() *Meta {\n\treturn &Meta{}\n}", "func New() *Meta {\n\treturn &Meta{}\n}", "func New(filePath string) File {\n\treturn File{\n\t\tFilepath: filePath,\n\t}\n}", "func New(major, minor, patch int) Version {\n\treturn &version{major, minor, patch}\n}", "func New() iface.Backend {\n\treturn &Backend{\n\t\tBackend: common.NewBackend(new(config.Config)),\n\t\tgroups: make(map[string][]string),\n\t\ttasks: make(map[string][]byte),\n\t}\n}", "func New(version string, command *command.Cmd) CLI {\n\tapp := app.App{\n\t\tName: path.Base(os.Args[0]),\n\t\tDescription: command.Description,\n\t\tVersionString: version,\n\t\tHasSubCmds: false,\n\t}\n\treturn mkNew(app, command)\n}", "func New(s string, base int) (v BigNum) {\n\tv.Input(s, base)\n\treturn\n}", "func New(args []string) pakelib.Command {\n\treturn &remove{\n\t\targs: args,\n\t}\n}" ]
[ "0.67842555", "0.6761811", "0.63972706", "0.62526065", "0.61175954", "0.6110591", "0.6066089", "0.6040809", "0.60183716", "0.5942007", "0.59030384", "0.5836826", "0.57957095", "0.5794075", "0.57899344", "0.5774618", "0.57209194", "0.569953", "0.56983095", "0.5694132", "0.5684422", "0.5684422", "0.56804764", "0.56621176", "0.56603414", "0.56423515", "0.5635949", "0.56313235", "0.56298333", "0.5614702", "0.5609926", "0.5602508", "0.5599268", "0.5590294", "0.5583191", "0.5578372", "0.5572266", "0.55568373", "0.55472916", "0.5502676", "0.5484693", "0.5482988", "0.5482303", "0.5479744", "0.5479249", "0.54788464", "0.5476487", "0.54587287", "0.5446364", "0.5444641", "0.54368395", "0.5436252", "0.543149", "0.5430705", "0.5419708", "0.54087704", "0.54029137", "0.5382252", "0.53771144", "0.5376645", "0.5376645", "0.5372976", "0.53634214", "0.535644", "0.53546363", "0.5353516", "0.534201", "0.5319776", "0.53177905", "0.5315147", "0.53149354", "0.531455", "0.53109246", "0.5307635", "0.530663", "0.53047335", "0.5302121", "0.5301437", "0.5295242", "0.5291861", "0.5282533", "0.5281871", "0.52796394", "0.5276207", "0.5274644", "0.5274352", "0.52731884", "0.52695924", "0.5266651", "0.52624965", "0.5255885", "0.52556324", "0.52540815", "0.52540815", "0.52528393", "0.52515274", "0.5246296", "0.524518", "0.52451783", "0.5243556" ]
0.6701476
2
Outgoing attaches any additional metadata to a message
func (e *Extension) Outgoing(ms *bayeux.Message) { switch ms.Channel { case bayeux.MetaHandshake: ext := ms.GetExt(true) ext[ExtensionName] = true case bayeux.MetaSubscribe: if e.isSupported() { ext := ms.GetExt(true) ext[ExtensionName] = e.replayStore.AsMap() } } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *EventMessageDetail) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (basic *BasicMessage) AddMetadata(key string, value string) {\n\tmetadata := NewMetadata(key, value)\n\tbasic.Metadata = append(basic.Metadata, metadata)\n}", "func mergeOutgoingMetadata(ctx context.Context, mds ...metadata.MD) context.Context {\n\t// There may not be metadata in the context, only insert the existing\n\t// metadata if it exists (ok).\n\tctxMD, ok := metadata.FromOutgoingContext(ctx)\n\tif ok {\n\t\t// The ordering matters, hence why ctxMD is added to the front.\n\t\tmds = append([]metadata.MD{ctxMD}, mds...)\n\t}\n\n\treturn metadata.NewOutgoingContext(ctx, metadata.Join(mds...))\n}", "func (c *jsiiProxy_CfnDeliveryStream) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (coll *Collector) Write(msg string, extra map[string]interface{}) (err error) {\n\n\tm := gelf.Message{\n\t\tVersion: \"1.1\",\n\t\tHost: coll.host,\n\t\tShort: msg,\n\t\tTimeUnix: float64(time.Now().Unix()),\n\t\tLevel: 6, // info always\n\t\tFacility: \"drone\",\n\t\tExtra: extra,\n\t}\n\n\tif err = coll.writer.WriteMessage(&m); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n\n}", "func (h Dnstap) TapMessageWithMetadata(ctx context.Context, m *tap.Message, state request.Request) {\n\tif h.ExtraFormat == \"\" {\n\t\th.tapWithExtra(m, nil)\n\t\treturn\n\t}\n\textraStr := h.repl.Replace(ctx, state, nil, h.ExtraFormat)\n\th.tapWithExtra(m, []byte(extraStr))\n}", "func (w *reply) Write(m *Msg) {\n\tw.Client().ReplyChan <- &Exchange{Request: w.req, Reply: m}\n}", "func (this *IdentityProvider) ServeMetadata(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/samlmetadata+xml\")\n\tw.Header().Set(\"Content-Disposition\", \"attachment; filename=metadata.xml\")\n\tw.Write(this.Metadata(r).Bytes(true))\n}", "func (c *peerConn) writeMessage(mType uint8, data []byte) error {\n\tcLen := len(data)\n\tval := make([]byte, cLen+4)\n\tbinary.LittleEndian.PutUint32(val, uint32(cLen))\n\tcopy(val[4:], data)\n\tmsg := []byte{c.version, mType}\n\tmsg = append(msg, val...)\n\t_, err := c.rw.Write(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func persist_outbound(s Store, m proto.Message) {\n\tswitch m := m.(type) {\n\tcase *proto.Publish:\n\t\tif m.Header.QosLevel == proto.QosAtLeastOnce {\n\t\t\t// store in obound until PubAck received\n\t\t\ts.Put(obound_mid2key(m.MessageId), m)\n\t\t}\n\n\tcase *proto.PubAck:\n\t\ts.Del(ibound_mid2key(m.MessageId))\n\n\t}\n}", "func (m *EventMessageDetail) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (bagStorer *BagStorer) SendToMetadataQueue(helper *bagman.IngestHelper) {\n\terr := bagman.Enqueue(helper.ProcUtil.Config.NsqdHttpAddress,\n\t\thelper.ProcUtil.Config.RecordWorker.NsqTopic, helper.Result)\n\tif err != nil {\n\t\terrMsg := fmt.Sprintf(\"Error adding '%s' to metadata queue: %v \",\n\t\t\thelper.Result.S3File.Key.Key, err)\n\t\thelper.ProcUtil.MessageLog.Error(errMsg)\n\t\thelper.Result.ErrorMessage += errMsg\n\t} else {\n\t\thelper.ProcUtil.MessageLog.Debug(\"Sent '%s' to metadata queue\",\n\t\t\thelper.Result.S3File.Key.Key)\n\t}\n}", "func writeEndMessage(currentMessage *bytes.Buffer) *bytes.Buffer {\n\t// 11 -event type - 7 - 3 \"End\"\n\t// 13 -message-type -7 5 \"event\"\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\theaderLen := len(endHeaders)\n\tcurrentMessage.Write(writePayloadSize(0, headerLen))\n\n\tcurrentMessage.Write(writeHeaderSize(headerLen))\n\n\t//Calculate the Prelude CRC here:\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\n\tcurrentMessage.Write(endHeaders)\n\n\t// Now we do a CRC check on the entire messages\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\treturn currentMessage\n\n}", "func (i *invoker) appendMetadata(ctx context.Context) context.Context {\n\treturn metadata.AppendToOutgoingContext(ctx, \"Authorization\", \"Bearer token\")\n}", "func writeStatMessage(payload string, currentMessage *bytes.Buffer) *bytes.Buffer {\n\t// 11 -event type - 7 - 5 \"Stat\" 20\n\t// 13 -content-type -7 -8 \"text/xml\" 25\n\t// 13 -message-type -7 5 \"event\" 22\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\theaderLen := len(statHeaders)\n\n\tcurrentMessage.Write(writePayloadSize(len(payload), headerLen))\n\n\tcurrentMessage.Write(writeHeaderSize(headerLen))\n\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\n\tcurrentMessage.Write(statHeaders)\n\n\t// This part is where the payload is written, this will be only one row, since\n\t// we're sending one message at a types\n\tcurrentMessage.Write(writePayload(payload))\n\n\t// Now we do a CRC check on the entire messages\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\treturn currentMessage\n\n}", "func (m *ChatMessageAttachment) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *metaObject) writeMetadata(pth string, noOverwrite bool, buffer []byte) error {\n\tswitch msCRC := m.meta.(type) {\n\tcase storage.StoreCRC:\n\t\tcrc := crc32.Checksum(buffer, crc32.MakeTable(crc32.Castagnoli))\n\t\treturn msCRC.PutCRC(m.contexter(), pth, bytes.NewReader(buffer), noOverwrite, crc)\n\tdefault:\n\t\treturn msCRC.Put(m.contexter(), pth, bytes.NewReader(buffer), noOverwrite)\n\t}\n}", "func WriteMessage(pb proto.Message, w io.Writer) error {\n\tbuf, err := proto.Marshal(pb)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed marshalling proto message: %v\", err)\n\t}\n\tif _, err := fmt.Fprintf(w, \"\\nContent-Length: %d\\n\\n%s\", len(buf), buf); err != nil {\n\t\treturn fmt.Errorf(\"Failed writing response: %v\", err)\n\t}\n\treturn nil\n}", "func (i Intermediate) Write(w io.Writer, b *bin.Buffer) error {\n\tif err := checkOutgoingMessage(b); err != nil {\n\t\treturn err\n\t}\n\n\tif err := checkAlign(b, 4); err != nil {\n\t\treturn err\n\t}\n\n\tif err := writeIntermediate(w, b); err != nil {\n\t\treturn errors.Wrap(err, \"write intermediate\")\n\t}\n\n\treturn nil\n}", "func WriteMessage(msg Message) {\n\tVac.writeMessage(msg)\n}", "func (h Handler) Add(ctx context.Context, request *proto.Identifier) (*proto.Message, error) {\n\terr := h.meta.Create(ctx, request.UserID)\n\terr = errors.Wrap(err, \"Error while creating and adding new meta data\")\n\treturn &proto.Message{}, err\n}", "func (t *transaction) Write(msg Message) error {\n\tt.out = append(t.out, msg)\n\treturn nil\n}", "func (m *PresenceStatusMessage) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *InvitedUserMessageInfo) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *Message) WriteTo(w io.Writer) (int64, error) {\n\tvar buf [10]byte\n\tvar total int64\n\n\tbinary.BigEndian.PutUint64(buf[:8], uint64(m.Timestamp))\n\tbinary.BigEndian.PutUint16(buf[8:10], uint16(m.Attempts))\n\n\tn, err := w.Write(buf[:])\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\tn, err = w.Write(m.ID[:])\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\tn, err = w.Write(m.Body)\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\treturn total, nil\n}", "func (m *Message) WriteTo(w io.Writer) (int64, error) {\n\tvar buf [10]byte\n\tvar total int64\n\n\tbinary.BigEndian.PutUint64(buf[:8], uint64(m.Timestamp))\n\tbinary.BigEndian.PutUint16(buf[8:10], uint16(m.Attempts))\n\n\tn, err := w.Write(buf[:])\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\tn, err = w.Write(m.ID[:])\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\tn, err = w.Write(m.Body)\n\ttotal += int64(n)\n\tif err != nil {\n\t\treturn total, err\n\t}\n\n\treturn total, nil\n}", "func (m *ChannelIdentity) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (c *jsiiProxy_CfnContactChannel) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (c *Conn) WriteMessage(msg interface{}) error {\n\tdata, err := xml.Marshal(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcolor.Printf(\"@{|}<!-- REQUEST -->\\n%s\\n\\n\", string(data))\n\treturn c.WriteDataUnit(data)\n}", "func (u *UnityServer) SendMessage(msg string) {\n\tu.Logger.Debugf(\"Adding to outgoing: %v\", msg)\n\tu.outgoing <- msg\n}", "func (rb *routerBase) fwdMetadata(meta *execinfrapb.ProducerMetadata) {\n\tif meta == nil {\n\t\tlog.Fatalf(context.TODO(), \"asked to fwd empty metadata\")\n\t\treturn\n\t}\n\n\trb.semaphore <- struct{}{}\n\tdefer func() {\n\t\t<-rb.semaphore\n\t}()\n\tif metaErr := meta.Err; metaErr != nil {\n\t\t// Forward the error to all non-closed streams.\n\t\tif rb.fwdErrMetadata(metaErr) {\n\t\t\treturn\n\t\t}\n\t} else {\n\t\t// Forward the metadata to the first non-closed stream.\n\t\tfor i := range rb.outputs {\n\t\t\tro := &rb.outputs[i]\n\t\t\tro.mu.Lock()\n\t\t\tif ro.mu.streamStatus != execinfra.ConsumerClosed {\n\t\t\t\tro.addMetadataLocked(meta)\n\t\t\t\tro.mu.Unlock()\n\t\t\t\tro.mu.cond.Signal()\n\t\t\t\treturn\n\t\t\t}\n\t\t\tro.mu.Unlock()\n\t\t}\n\t}\n\t// If we got here it means that we couldn't even forward metadata anywhere;\n\t// all streams are closed.\n\tatomic.StoreUint32(&rb.aggregatedStatus, uint32(execinfra.ConsumerClosed))\n}", "func (c *Conn) Write(message proto.Message) (int, error) {\n\tpayload, err := proto.Marshal(message)\n\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\tvar buf bytes.Buffer\n\n\t// Prepare *type* prefix\n\tmtype := messageTypes[reflect.TypeOf(message)]\n\tif _, err := buf.Write(uint16tbs(mtype)); err != nil {\n\t\treturn -1, err\n\t}\n\n\t// Prepare *size* prefix\n\tsize := uint32(len(payload))\n\tif _, err := buf.Write(uint32tbs(size)); err != nil {\n\t\treturn -1, err\n\t}\n\n\t// Prepare *payload* body\n\tif _, err := buf.Write(payload); err != nil {\n\t\treturn -1, err\n\t}\n\n\treturn c.conn.Write(buf.Bytes())\n}", "func MergeToOutgoingMD(ctx context.Context, md metadata.MD) context.Context {\n\tif appended := ctx.Value(RpcPayloadAppended); appended == nil {\n\t\tif _, ok := metadata.FromOutgoingContext(ctx); ok {\n\t\t\tkvs := make([]string, 0)\n\t\t\t// append md into result first\n\t\t\tfor k, v := range md {\n\t\t\t\tfor i := range v {\n\t\t\t\t\tkvs = append(kvs, k, v[i])\n\t\t\t\t}\n\t\t\t}\n\t\t\tctx = context.WithValue(ctx, RpcPayloadAppended, \"\")\n\n\t\t\t// merge incoming MD into outgoing metadata\n\t\t\tctx = metadata.AppendToOutgoingContext(ctx, kvs...)\n\t\t} else {\n\t\t\tctx = metadata.NewOutgoingContext(ctx, md)\n\t\t}\n\t}\n\n\treturn ctx\n}", "func (c *Client) Write(message *Message) {\n\tfmt.Fprintf(c.Conn, message.ToString())\n}", "func (o *PutMeetupOK) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(200)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func (msg *Message) writeHeader(buffer *bytes.Buffer) {\n\tfor key, values := range msg.Headers {\n\t\tfor _, keyval := range values {\n\t\t\tio.WriteString(buffer, key)\n\t\t\tio.WriteString(buffer, \": \")\n\t\t\tswitch {\n\t\t\tcase key == \"Content-Type\" || key == \"Content-Disposition\":\n\t\t\t\tbuffer.Write([]byte(keyval))\n\t\t\tdefault:\n\t\t\t\tbuffer.Write([]byte(mime.QEncoding.Encode(\"UTF-8\", keyval)))\n\t\t\t}\n\t\t\tio.WriteString(buffer, \"\\r\\n\")\n\t\t}\n\t}\n\tio.WriteString(buffer, \"\\r\\n\")\n}", "func (q *queue) persistMetaOfMessage(dataPageIndex int64, dataLen, messageOffset int) error {\n\tq.rwMutex.Lock()\n\tdefer q.rwMutex.Unlock()\n\n\tseq := q.appendedSeq.Load() + 1 // append sequence\n\tindexPageIndex := seq / indexItemsPerPage\n\tif indexPageIndex != q.indexPageIndex {\n\t\t// check size limit before index page acquire\n\t\tif err := q.checkDataSize(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// sync previous data page\n\t\tif err := q.indexPage.Sync(); err != nil {\n\t\t\tqueueLogger.Error(\"sync index page err when alloc\",\n\t\t\t\tlogger.String(\"queue\", q.dirPath), logger.Error(err))\n\t\t}\n\t\tindexPage, err := q.indexPageFct.AcquirePage(indexPageIndex)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tq.indexPage = indexPage\n\t\tq.indexPageIndex = indexPageIndex\n\t}\n\n\t// save index data\n\tindexOffset := int((seq % indexItemsPerPage) * indexItemLength)\n\tq.indexPage.PutUint64(uint64(dataPageIndex), indexOffset+queueDataPageIndexOffset)\n\tq.indexPage.PutUint32(uint32(messageOffset), indexOffset+messageOffsetOffset)\n\tq.indexPage.PutUint32(uint32(dataLen), indexOffset+messageLengthOffset)\n\n\t// save metadata\n\tq.metaPage.PutUint64(uint64(seq), queueAppendedSeqOffset)\n\tq.appendedSeq.Store(seq)\n\n\t// new data written, notify all waiting consumer groups can consume data\n\tq.notEmpty.Broadcast()\n\treturn nil\n}", "func (protocol *HumanReadableJsonProtocol) WriteMessageEnd() error {\n\treturn protocol.oprot.OutputObjectEnd()\n}", "func (m *ChatMessage) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAttachments() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetAttachments())\n err = writer.WriteCollectionOfObjectValues(\"attachments\", cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"body\", m.GetBody())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"channelIdentity\", m.GetChannelIdentity())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"chatId\", m.GetChatId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteTimeValue(\"createdDateTime\", m.GetCreatedDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteTimeValue(\"deletedDateTime\", m.GetDeletedDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"etag\", m.GetEtag())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"eventDetail\", m.GetEventDetail())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"from\", m.GetFrom())\n if err != nil {\n return err\n }\n }\n if m.GetHostedContents() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetHostedContents())\n err = writer.WriteCollectionOfObjectValues(\"hostedContents\", cast)\n if err != nil {\n return err\n }\n }\n if m.GetImportance() != nil {\n cast := (*m.GetImportance()).String()\n err = writer.WriteStringValue(\"importance\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteTimeValue(\"lastEditedDateTime\", m.GetLastEditedDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteTimeValue(\"lastModifiedDateTime\", m.GetLastModifiedDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"locale\", m.GetLocale())\n if err != nil {\n return err\n }\n }\n if m.GetMentions() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetMentions())\n err = writer.WriteCollectionOfObjectValues(\"mentions\", cast)\n if err != nil {\n return err\n }\n }\n if m.GetMessageType() != nil {\n cast := (*m.GetMessageType()).String()\n err = writer.WriteStringValue(\"messageType\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"policyViolation\", m.GetPolicyViolation())\n if err != nil {\n return err\n }\n }\n if m.GetReactions() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetReactions())\n err = writer.WriteCollectionOfObjectValues(\"reactions\", cast)\n if err != nil {\n return err\n }\n }\n if m.GetReplies() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetReplies())\n err = writer.WriteCollectionOfObjectValues(\"replies\", cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"replyToId\", m.GetReplyToId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"subject\", m.GetSubject())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"summary\", m.GetSummary())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"webUrl\", m.GetWebUrl())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (e *agentEndpoint) write(msg *submitws.Message) {\n\te.mutex.Lock()\n\tdefer e.mutex.Unlock()\n\tif e.isClosed {\n\t\treturn\n\t}\n\tif err := e.conn.WriteMessage(websocket.BinaryMessage, msg.ToBinary()); err != nil {\n\t\tlogger.WithError(err).Errorf(\"error sending message to agent with id == %s: %v\", e.id, err)\n\t\tif err := e.conn.Close(); err != nil {\n\t\t\tlogger.WithError(err).Errorf(\"error closing connection to agent with id == %s after write error: %v\", e.id, err)\n\t\t}\n\t\te.isClosed = true\n\t}\n}", "func (c *jsiiProxy_CfnElasticLoadBalancerAttachment) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (m *CommsNotification) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func writeRecordMessage(payload string, currentMessage *bytes.Buffer) *bytes.Buffer {\n\t// The below are the specifications of the header for a \"record\" event\n\t// 11 -event type - 7 - 7 \"Records\"\n\t// 13 -content-type -7 -24 \"application/octet-stream\"\n\t// 13 -message-type -7 5 \"event\"\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\theaderLen := len(recordHeaders)\n\t// Writes the total size of the message.\n\tcurrentMessage.Write(writePayloadSize(len(payload), headerLen))\n\t// Writes the total size of the header.\n\tcurrentMessage.Write(writeHeaderSize(headerLen))\n\t// Writes the CRC of the Prelude\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\tcurrentMessage.Write(recordHeaders)\n\n\t// This part is where the payload is written, this will be only one row, since\n\t// we're sending one message at a types\n\tcurrentMessage.Write(writePayload(payload))\n\n\t// Now we do a CRC check on the entire messages\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\treturn currentMessage\n\n}", "func (d *Doorman) Write(msg audio.Msg) error {\n\n\t// make sure the Metadata dict exists\n\tif msg.Metadata == nil {\n\t\treturn nil\n\t}\n\n\t// make sure the userID key has been set\n\tuserID, ok := msg.Metadata[\"userID\"]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\ttxUser := \"\"\n\n\t// make sure the interface{} can be casted to a string\n\tswitch uID := userID.(type) {\n\tdefault:\n\t\tlog.Println(\"doorman: can not cast userID to string\")\n\t\treturn nil\n\tcase string:\n\t\ttxUser = uID\n\t}\n\n\td.Lock()\n\tlastUser := d.lastUser\n\tlastHeard := d.lastHeard\n\td.Unlock()\n\n\t// in case this is the current txUser forward the msg\n\tif lastUser == txUser {\n\t\td.Lock()\n\t\td.lastHeard = time.Now() //update the timestamp\n\t\tif d.onDataCb != nil {\n\t\t\t// call callback asynchronously to pass the data to the next node\n\t\t\tgo d.onDataCb(msg)\n\t\t}\n\t\td.Unlock()\n\t\treturn nil\n\t}\n\n\t// in case d.lastUser != txUser, but we don't expect any more audio msgs\n\t// from the original txUser\n\tif time.Since(lastHeard) >= time.Duration(time.Millisecond*100) {\n\t\td.Lock()\n\t\td.lastUser = txUser\n\t\td.lastHeard = time.Now()\n\t\tif d.onDataCb != nil {\n\t\t\t// call callback asynchronously to pass the data to the next node\n\t\t\tgo d.onDataCb(msg)\n\t\t}\n\t\tif d.onTxUserChanged != nil {\n\t\t\t// notify application that txUser has changed.\n\t\t\tgo d.onTxUserChanged(txUser)\n\t\t}\n\t\td.Unlock()\n\t}\n\n\t// if d.lastUser != txUser and d.lastUser heard within the last\n\t// 100ms, we drop the msg\n\n\treturn nil\n}", "func (t *cliTransHandler) Write(ctx context.Context, conn net.Conn, sendMsg remote.Message) (err error) {\n\tvar bufWriter remote.ByteBuffer\n\tstats2.Record(ctx, sendMsg.RPCInfo(), stats.WriteStart, nil)\n\tdefer func() {\n\t\tt.ext.ReleaseBuffer(bufWriter, err)\n\t\tstats2.Record(ctx, sendMsg.RPCInfo(), stats.WriteFinish, err)\n\t}()\n\n\tbufWriter = t.ext.NewWriteByteBuffer(ctx, conn, sendMsg)\n\tsendMsg.SetPayloadCodec(t.opt.PayloadCodec)\n\terr = t.codec.Encode(ctx, sendMsg, bufWriter)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn bufWriter.Flush()\n}", "func writeProgressMessage(payload string, currentMessage *bytes.Buffer) *bytes.Buffer {\n\t// The below are the specifications of the header for a \"Progress\" event\n\t// 11 -event type - 7 - 8 \"Progress\" 23\n\t// 13 -content-type -7 -8 \"text/xml\" 25\n\t// 13 -message-type -7 5 \"event\" 22\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\theaderLen := len(progressHeaders)\n\n\tcurrentMessage.Write(writePayloadSize(len(payload), headerLen))\n\n\tcurrentMessage.Write(writeHeaderSize(headerLen))\n\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\n\tcurrentMessage.Write(progressHeaders)\n\n\t// This part is where the payload is written, this will be only one row, since\n\t// we're sending one message at a types\n\tcurrentMessage.Write(writePayload(payload))\n\n\t// Now we do a CRC check on the entire messages\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\treturn currentMessage\n\n}", "func (out *OutBuffer) WriteMessage(message Message) (err error) {\n\tvar n int\n\tn, err = message.MarshalTo(out.GetContainer())\n\tout.pos += n\n\treturn\n}", "func emailWrite(splited []string, length int) {\n if length != 3 { return }\n var new_pack = settings.PackageTCP {\n From: models.From {\n Name: settings.User.Hash,\n },\n To: splited[2],\n Head: models.Head {\n Header: settings.HEAD_EMAIL,\n Mode: settings.MODE_SAVE,\n }, \n Body: \n set_email.title + settings.SEPARATOR +\n set_email.body + settings.SEPARATOR +\n time.Now().Format(time.RFC850),\n }\n connect.CreateRedirectPackage(&new_pack)\n connect.SendInitRedirectPackage(new_pack)\n}", "func (conn *IRODSConnection) PostprocessMessage(msg *message.IRODSMessage) error {\n\tif msg.Body == nil || msg.Body.Message == nil {\n\t\treturn nil\n\t}\n\n\tvar err error\n\n\tmsg.Body.Message, err = conn.PostprocessXML(msg.Body.Message)\n\tmsg.Header.MessageLen = uint32(len(msg.Body.Message))\n\n\treturn err\n}", "func (s *FileSnapshotSink) writeMeta() error {\n\t// Open the meta file\n\tmetaPath := filepath.Join(s.dir, metaFilePath)\n\tfh, err := os.Create(metaPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer fh.Close()\n\n\t// Buffer the file IO\n\tbuffered := bufio.NewWriter(fh)\n\tdefer buffered.Flush()\n\n\t// Write out as JSON\n\tenc := json.NewEncoder(buffered)\n\tif err := enc.Encode(&s.meta); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Server) Publish(message interface{}) error {\n\tvar sb strings.Builder\n\tsb.WriteString(\"data:\")\n\tbytes, err := json.Marshal(message)\n\tif err != nil {\n\t\treturn err\n\t}\n\tsb.Write(bytes)\n\tsb.WriteString(\"\\n\\n\")\n\n\ts.messages <- []byte(sb.String())\n\treturn nil\n}", "func (m *OutterMessage) Respond(envelope *pb.Envelope, successCallBack func(interface{}), errCallBack func(error)) {\n\n\tm.Stream.Send(envelope, successCallBack, errCallBack)\n}", "func (m *Manager) SendMessage(msgData []byte, to ...identity.ID) {\n\tmsg := &pb.Message{Data: msgData}\n\tm.send(marshal(msg), to...)\n}", "func (handler *MessageHandler) PostMessage(w http.ResponseWriter, r *http.Request) {\n\ttokenString := r.Header.Get(\"Authorization\")\n\n\tvar message entities.Message\n\terr := json.NewDecoder(r.Body).Decode(&message)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tjson.NewEncoder(w).Encode(entities.Error{\n\t\t\tError: err.Error(),\n\t\t})\n\t\treturn\n\t}\n\n\tmessageID, timestamp, err := handler.usecases.PostMessage(tokenString, message)\n\tif err != nil && err.Error() == \"not authenticated\" {\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\tjson.NewEncoder(w).Encode(entities.Error{\n\t\t\tError: \"not authenticated\",\n\t\t})\n\t\treturn\n\t}\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tjson.NewEncoder(w).Encode(entities.Error{\n\t\t\tError: err.Error(),\n\t\t})\n\t\treturn\n\t}\n\tnewMessageOutput := struct {\n\t\tID uint `json:\"id\"`\n\t\tTimestamp time.Time `json:\"timestamp\"`\n\t}{\n\t\tID: messageID,\n\t\tTimestamp: timestamp,\n\t}\n\tjson.NewEncoder(w).Encode(&newMessageOutput)\n\tw.WriteHeader(http.StatusOK)\n}", "func (o *AddAttendeeToTalkDefault) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(o._statusCode)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func (m *UserSimulationEventInfo) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func ForwardResponseMessage(ctx context.Context, mux *runtime.ServeMux, marshaler runtime.Marshaler, w http.ResponseWriter, req *http.Request, resp proto.Message, opts ...func(context.Context, http.ResponseWriter, proto.Message) error) {\n\tmd, ok := runtime.ServerMetadataFromContext(ctx)\n\tif !ok {\n\t\tgrpclog.Infof(\"Failed to extract ServerMetadata from context\")\n\t}\n\n\thandleForwardResponseServerMetadata(w, mux, md)\n\thandleForwardResponseTrailerHeader(w, md)\n\tw.Header().Set(\"Content-Type\", marshaler.ContentType())\n\tif err := handleForwardResponseOptions(ctx, w, resp, opts); err != nil {\n\t\truntime.HTTPError(ctx, mux, marshaler, w, req, err)\n\t\treturn\n\t}\n\tvar buf []byte\n\tvar err error\n\tif rb, ok := resp.(responseBody); ok {\n\t\tbuf, err = marshaler.Marshal(rb.XXX_ResponseBody())\n\t} else {\n\t\tbuf, err = marshaler.Marshal(resp)\n\t}\n\tif err != nil {\n\t\tgrpclog.Infof(\"Marshal error: %v\", err)\n\t\truntime.HTTPError(ctx, mux, marshaler, w, req, err)\n\t\treturn\n\t}\n\n\tbody := struct {\n\t\tCode int `json:\"code\"`\n\t\tMsg string `json:\"msg\"`\n\t\tData json.RawMessage `json:\"data,omitempty\"`\n\t}{\n\t\tCode: 200,\n\t\tMsg: \"Success\",\n\t\tData: buf,\n\t}\n\n\tif bts, err := json.Marshal(body); err == nil {\n\t\tif _, err = w.Write(bts); err != nil {\n\t\t\tgrpclog.Infof(\"Failed to write response: %v\", err)\n\t\t}\n\t} else {\n\t\tgrpclog.Infof(\"json marshal err : %v\", err)\n\t}\n\thandleForwardResponseTrailer(w, md)\n}", "func (x *Message) Write(w io.Writer) error {\n\tif x.Compression != NoCompression {\n\t\tpanic(\"message compression not supported\")\n\t}\n\tvar _magic int32\n\tif x.Compression != NoCompression {\n\t\t_magic = 1\n\t}\n\tvar _length int32 = 1 /* magic */ + _magic /* compression */ + 4 /* checksum */ + int32(len(x.Payload))\n\tw.Write(int32Bytes(_length))\n\tw.Write([]byte{byte(_magic)})\n\tif _magic == 1 {\n\t\tw.Write([]byte{byte(x.Compression)})\n\t}\n\tw.Write(uint32Bytes(crc32.ChecksumIEEE(x.Payload)))\n\t_, err := w.Write(x.Payload)\n\treturn err\n}", "func (m *ChatMessageAttachment) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteStringValue(\"content\", m.GetContent())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"contentType\", m.GetContentType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"contentUrl\", m.GetContentUrl())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"id\", m.GetId())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"name\", m.GetName())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"teamsAppId\", m.GetTeamsAppId())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"thumbnailUrl\", m.GetThumbnailUrl())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (c *Client) Send(msg *message.OutMessage) error {\n\tom := translateOutMessage(msg)\n\tif c.config.OverrideUserName != \"\" && om.UserName == \"\" {\n\t\tom.UserName = c.config.OverrideUserName\n\t}\n\tif c.config.IconURL != \"\" && om.IconURL == \"\" {\n\t\tom.IconURL = c.config.IconURL\n\t}\n\n\tbuf, err := toJSON(om)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Printf(\"[Send]\\n%s\\n----------------\\n\", buf)\n\tfmt.Printf(\"mmbot> %s\\n\", om.Text)\n\n\treturn nil\n}", "func printOutgoingRumorMessage(destination string) {\n\tfmt.Println(\"MONGERING with \" + destination)\n}", "func (cr CommandReply) AppendWireMessage([]byte) ([]byte, error) {\r\n\tpanic(\"not implemented\")\r\n}", "func (sp *SyncProtocol) createOutboundWriter(ws *WrappedStream) error {\n\n\t// create stan connection with random client id\n\tpid := string(ws.remotePeerID())\n\tsc, err := NSSConnection(pid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tgo func() {\n\t\terrc := make(chan error)\n\t\tdefer close(errc)\n\t\tdefer sc.Close()\n\n\t\t// main message sending routine\n\t\tsub, err := sc.Subscribe(\"feed\", func(m *stan.Msg) {\n\n\t\t\tsendMessage := true\n\n\t\t\t// get the block from the feed\n\t\t\tblk := DeserializeBlock(m.Data)\n\n\t\t\t// log.Printf(\"\\n\\n\\tauthor: %s\\n\\tremote-peer: %s\", blk.Author, ws.remotePeerID())\n\n\t\t\t// don't send if created by the remote peer\n\t\t\tif blk.Author == ws.remotePeerID() {\n\t\t\t\tsendMessage = false\n\t\t\t\t// log.Println(\"author is remote peer sendmessage:\", sendMessage)\n\t\t\t}\n\n\t\t\t// log.Printf(\"\\n\\n\\tsender: %s\\n\\tremote-peer: %s\", blk.Sender, ws.remotePeerID())\n\t\t\t// don't send if it came from the reomote peer\n\t\t\tif blk.Sender == ws.remotePeerID() {\n\t\t\t\tsendMessage = false\n\t\t\t\t// log.Println(\"sender is remote peer sendmessage:\", sendMessage)\n\t\t\t}\n\n\t\t\t// don't send if the remote peer has seen it before\n\t\t\tif sendMessage {\n\t\t\t\tblk.Receiver = ws.remotePeerID()\n\t\t\t\tif sp.node.msgCMS.Estimate(blk.CmsKey()) > 0 {\n\t\t\t\t\tsendMessage = false\n\t\t\t\t\t// log.Println(\"message:\", blk.BlockId, \" already sent to:\", ws.remotePeerID(), \" sendmaessage:\", sendMessage)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif sendMessage {\n\t\t\t\t// update the sender\n\t\t\t\tblk.Sender = ws.localPeerID()\n\t\t\t\tsendErr := sendBlock(blk, ws)\n\t\t\t\tif sendErr != nil {\n\t\t\t\t\tlog.Println(\"cannot send message to peer: \", sendErr)\n\t\t\t\t\terrc <- sendErr\n\t\t\t\t}\n\t\t\t\t// register that we've sent this message to this peer\n\t\t\t\tsp.node.msgCMS.Update(blk.CmsKey(), 1)\n\t\t\t}\n\n\t\t}, stan.DeliverAllAvailable())\n\t\tif err != nil {\n\t\t\tlog.Println(\"error creating feed subscription: \", err)\n\t\t\treturn\n\t\t}\n\n\t\t// wait for errors on writing, such as stream closing\n\t\t<-errc\n\t\tsub.Close()\n\t}()\n\n\treturn nil\n}", "func (this *SmtpWriter) SayGoodbye() error {\n\treturn this.SendResponse(smtpconstants.SMTP_CLOSING_MESSAGE)\n}", "func (msg *MessageWithMeta) ToMessage() Message {\n\tout := Message{\n\t\tExternalID: msg.ExternalID,\n\t\tExternalDt: msg.ExternalDt,\n\t\tSource: msg.Source,\n\t\tMessage: msg.Message,\n\t\tSegment: msg.Segment,\n\t}\n\n\treturn out\n}", "func Out(ch chan pb.Message, reader *bufio.Reader, authorName string) {\n\tfor {\n\t\tmsg, _ := reader.ReadString('\\n')\n\t\tch <- pb.Message{\n\t\t\tAuthor: &pb.Author{\n\t\t\t\tName: authorName,\n\t\t\t},\n\t\t\tBody: msg,\n\t\t}\n\t}\n}", "func (ms *msgSender) SendMessage(ctx context.Context, p peer.ID, pmes *pb.Message) error {\n\tctx, _ = tag.New(ctx, kadmetrics.UpsertMessageType(pmes))\n\tdefer stats.Record(ctx, kadmetrics.SentMessages.M(1))\n\n\ts, err := ms.h.NewStream(ctx, p, ms.protocols...)\n\tif err != nil {\n\t\tstats.Record(ctx, kadmetrics.SentMessageErrors.M(1))\n\t\treturn err\n\t}\n\tdefer func() { _ = s.Close() }()\n\n\tif err = protoio.NewDelimitedWriter(s).WriteMsg(pmes); err != nil {\n\t\tstats.Record(ctx, kadmetrics.SentMessageErrors.M(1))\n\t\treturn err\n\t}\n\n\tstats.Record(ctx, kadmetrics.SentBytes.M(int64(pmes.Size())))\n\treturn nil\n}", "func (cr *Chat) Publish(ctx context.Context, message string) error {\n\n\tmsg := &pb.Chat2Message{\n\t\tTimestamp: uint64(time.Now().Unix()),\n\t\tNick: cr.nick,\n\t\tPayload: []byte(message),\n\t}\n\n\tmsgBytes, err := proto.Marshal(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t//log.Println(\"Publish\", msg)\n\n\t//var version uint32\n\t// TODO Add support\n\t//var timestamp float64 = float64(time.Now().UnixNano())\n\t// var keyInfo *node.KeyInfo = &node.KeyInfo{}\n\n\t// if cr.useV1Payload { // Use WakuV1 encryption\n\t// \tkeyInfo.Kind = node.Symmetric\n\t// \tkeyInfo.SymKey = generateSymKey(cr.contentTopic)\n\t// \tversion = 1\n\t// } else {\n\t// \tkeyInfo.Kind = node.None\n\t// \tversion = 0\n\t// }\n\n\t// TODO Implement, see what makes sense here vs private API\n\t// p := new(node.Payload)\n\t// p.Data = msgBytes\n\t// p.Key = keyInfo\n\n\t// // XXX Is this right?\n\t// payload, err := p.Encode(version)\n\t// if err != nil {\n\t// \treturn err\n\t// }\n\n\t// For version 0, should get payload.Data, []byte\n\t// TODO Want it hex encoded though\n\tvar payload = msgBytes\n\n\t// wakuMsg := &wpb.WakuMessage{\n\t// \tPayload: payload,\n\t// \tVersion: version,\n\t// \tContentTopic: cr.contentTopic,\n\t// \tTimestamp: timestamp,\n\t// }\n\n\tvar hexEncoded = make([]byte, hex.EncodedLen(len(payload)))\n\thex.Encode(hexEncoded, payload)\n\t//fmt.Println(\"%s\\n\", hexEncoded)\n\n\t// TODO Replace with jSON RPC\n\t//_, err = cr.node.Publish(ctx, wakuMsg, nil)\n\t// NOTE version field support https://rfc.vac.dev/spec/16/#wakurelaymessage\n\tvar wakuMsg = nwaku.WakuRelayMessage{\n\t\tPayload: string(hexEncoded), // \"0x1a2b3c4d5e6f\",\n\t\tContentTopic: cr.contentTopic,\n\t\t//Timestamp: timestamp,\n\t}\n\t// TODO Error handling\n\tvar _, _ = nwaku.PostWakuRelayMessage(cr.client, wakuMsg)\n\n\treturn nil\n}", "func (_BaseLibrary *BaseLibraryTransactor) PutMeta(opts *bind.TransactOpts, key []byte, value []byte) (*types.Transaction, error) {\n\treturn _BaseLibrary.contract.Transact(opts, \"putMeta\", key, value)\n}", "func (c *jsiiProxy_CfnContact) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (log *LogFile) writeMetaData(data metaData) error {\n\t// seek to start of the meta file\n\t_, err := log.metaFile.Seek(0, 0)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn binary.Write(log.metaFile, log.byteOrder, data)\n}", "func (pipe *slimPipe) SendMessage(message string) error {\n\t_, err := io.WriteString(pipe.writer, message)\n\treturn err\n}", "func sendOutgoing(conn net.Conn, message []byte) {\n numWritten, err := conn.Write(message)\n\n if err != nil {\n fmt.Println(\"Error writing outgoing message: \", err.Error())\n os.Exit(1)\n }\n\n if numWritten != len(message) {\n fmt.Println(\"Could not write out the full message.\")\n }\n}", "func (h Response) Tell(message string) {\n\th.emit(\":tell\", strings.Replace(strings.Replace(message, \"\\n\", \" \", -1), \"\\t\", \"\", -1))\n}", "func (cmr *ConsumerMetadataRequest) Write(encoder Encoder) {\n\tencoder.WriteString(cmr.Group)\n}", "func (e *TApplicationException) Write(p TProtocol) (err error) {\n\tif err = p.WriteStructBegin(TStructHeader{\"TApplicationException\"}); err != nil {\n\t\treturn\n\t}\n\tif len(e.Message) > 0 {\n\t\tif err = p.WriteFieldBegin(TFieldHeader{\"message\", STRING, 1}); err != nil {\n\t\t\treturn\n\t\t}\n\t\tif err = p.WriteString(e.Message); err != nil {\n\t\t\treturn\n\t\t}\n\t\tif err = p.WriteFieldEnd(); err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tif err = p.WriteFieldBegin(TFieldHeader{\"type\", I32, 2}); err != nil {\n\t\treturn\n\t}\n\tif err = p.WriteI32(e.Type); err != nil {\n\t\treturn\n\t}\n\tif err = p.WriteFieldEnd(); err != nil {\n\t\treturn\n\t}\n\tif err = p.WriteFieldStop(); err != nil {\n\t\treturn\n\t}\n\terr = p.WriteStructEnd()\n\treturn\n}", "func (o *AddAttendeeToTalkOK) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(200)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func (m *PresenceStatusMessage) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteObjectValue(\"expiryDateTime\", m.GetExpiryDateTime())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteObjectValue(\"message\", m.GetMessage())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteTimeValue(\"publishedDateTime\", m.GetPublishedDateTime())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func writeEndHeader() []byte {\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\tvar currentMessage = &bytes.Buffer{}\n\t// header name\n\tcurrentMessage.Write(encodeHeaderStringName(\":event-type\"))\n\t// header type\n\tcurrentMessage.Write(encodeNumber(7, 1))\n\t// header value and header value length\n\tcurrentMessage.Write(encodeHeaderStringValue(\"End\"))\n\n\t// Creation of the Header for message-type 13 -message-type -7 5 \"event\"\n\t// header name\n\tcurrentMessage.Write(encodeHeaderStringName(\":message-type\"))\n\t// header type\n\tcurrentMessage.Write(encodeNumber(7, 1))\n\t// header value and header value length\n\tcurrentMessage.Write(encodeHeaderStringValue(\"event\"))\n\treturn currentMessage.Bytes()\n}", "func (cr *CommandResponse) AddMeta(name, value string) {\n\tcr.Metadata = append(cr.Metadata, CommandResponseMetadata{\n\t\tName: name,\n\t\tValue: value,\n\t})\n}", "func messageCreate(s *discordgo.Session, m *discordgo.MessageCreate) {\n\n\t// Ignore all messages created by the bot itself\n\t// This isn't required in this specific example but it's a good practice.\n\tif m.Author.ID == s.State.User.ID {\n\t\treturn\n\t}\n\n\tmsg := strings.ToLower(m.Content)\n\n\tfmt.Printf(\"%s: %s\\n\", m.Author.ID, msg)\n\n\tif strings.Contains(msg, \"alice\") || rand.Intn(1000) == 173 {\n\t\tmsg = strings.TrimPrefix(msg, \"alice\")\n\t\tmsg = strings.TrimLeft(msg, \":, \")\n\n\t\tfd, err := net.Dial(\"unix\", \"/tmp/alice\")\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"unix socket connect failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tbuf := m.Author.ID + \"\\007\" + msg\n\t\t_, err = fd.Write([]byte(buf))\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"write failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tout := make([]byte, 8192)\n\t\tn, err := fd.Read(out)\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"read failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tfmt.Printf(\"me: %s\\n\", string(out[:n]))\n\n\t\ts.ChannelMessageSend(m.ChannelID, string(out[:n]))\n\n\t\tfd.Close()\n\t}\n}", "func (w *Writer) Write(key string, req *remote.WriteRequest) error {\n\tdata, err := proto.Marshal(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm := &sarama.ProducerMessage{\n\t\tTopic: w.topic,\n\t\tKey: sarama.StringEncoder(key),\n\t\tValue: sarama.ByteEncoder(data),\n\t}\n\n\tgo func() {\n\t\tw.producer.Input() <- m\n\n\t\tw.queuedForWrites.Inc()\n\t}()\n\n\treturn nil\n}", "func (s *Websocket) direct(msg models.Message) {\n\ts.hmu.RLock()\n\tuserFrom, okFrom := s.hub[msg.Sender]\n\tuserTo, okTo := s.hub[msg.Receiver]\n\ts.hmu.RUnlock()\n\n\tif !okFrom {\n\t\ts.logger.Errorf(\"unnown user from send message: %s\", msg.Sender)\n\t\treturn\n\t}\n\n\tif !okTo {\n\t\ts.logger.Errorf(\"unnown user to send message: %s\", msg.Receiver)\n\t\treturn\n\t}\n\n\t// Write to sender\n\tif err := userFrom.Conn.WriteJSON(msg); err != nil {\n\t\ts.logger.Errorf(\"error sending message: %v\", err)\n\t}\n\n\t// Write to receiver\n\tif err := userTo.Conn.WriteJSON(msg); err != nil {\n\t\ts.logger.Errorf(\"error sending message: %v\", err)\n\t}\n}", "func (o *PutAttachmentOK) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(200)\n\tpayload := o.Payload\n\tif err := producer.Produce(rw, payload); err != nil {\n\t\tpanic(err) // let the recovery middleware deal with this\n\t}\n\n}", "func (_BaseContentSpace *BaseContentSpaceTransactor) PutMeta(opts *bind.TransactOpts, key []byte, value []byte) (*types.Transaction, error) {\n\treturn _BaseContentSpace.contract.Transact(opts, \"putMeta\", key, value)\n}", "func (m *SharePostRequestBody) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func writeMIMEHeader(wr io.Writer, hdr textproto.MIMEHeader) (err error) {\n // write headers\n for k, vals := range(hdr) {\n for _, val := range(vals) {\n _, err = io.WriteString(wr, fmt.Sprintf(\"%s: %s\\n\", k, val))\n }\n }\n // end of headers\n _, err = io.WriteString(wr, \"\\n\")\n return\n}", "func (c *sender) Write(pkt []byte) {\n\tincreaseCounterMetric(c.metrics.IPPktsSent, 1)\n\tincreaseCounterMetric(c.metrics.IPPktBytesSent, float64(len(pkt)))\n\n\tc.encoder.Write(pkt)\n}", "func (h extraHeader) Write(w *bufio.Writer) {\n\tif h.date != nil {\n\t\tw.Write(headerDate)\n\t\tw.Write(h.date)\n\t\tw.Write(crlf)\n\t}\n\tif h.contentLength != nil {\n\t\tw.Write(headerContentLength)\n\t\tw.Write(h.contentLength)\n\t\tw.Write(crlf)\n\t}\n\tfor i, v := range []string{h.contentType, h.connection, h.transferEncoding} {\n\t\tif v != \"\" {\n\t\t\tw.Write(extraHeaderKeys[i])\n\t\t\tw.Write(colonSpace)\n\t\t\tw.WriteString(v)\n\t\t\tw.Write(crlf)\n\t\t}\n\t}\n}", "func (o *PostInteractionCreated) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(201)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func (s *server) WriteReplyMessage(w io.Writer, xid uint32, acceptType AcceptType, ret interface{}) error {\n\tvar buf bytes.Buffer\n\n\t// Header\n\theader := Message{\n\t\tXid: xid,\n\t\tType: Reply,\n\t}\n\n\tif _, err := xdr.Marshal(&buf, header); err != nil {\n\t\treturn err\n\t}\n\n\t// \"Accepted\"\n\tif _, err := xdr.Marshal(&buf, ReplyBody{Type: Accepted}); err != nil {\n\t\treturn err\n\t}\n\n\t// \"Success\"\n\tif _, err := xdr.Marshal(&buf, AcceptedReply{Type: acceptType}); err != nil {\n\t\treturn err\n\t}\n\n\t// Return data\n\tif ret != nil {\n\t\tif _, err := xdr.Marshal(&buf, ret); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t_, err := w.Write(buf.Bytes())\n\treturn err\n}", "func (c *jsiiProxy_CfnQueue) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (pipe *PipeWS) Send(msg *mangos.Message) error {\n\t// if msg.Expired() {\n\t// \tmsg.Free()\n\t// \treturn nil\n\t// }\n\tvar buf []byte\n\tif len(msg.Header) > 0 {\n\t\tbuf = make([]byte, 0, len(msg.Header)+len(msg.Body))\n\t\tbuf = append(buf, msg.Header...)\n\t\tbuf = append(buf, msg.Body...)\n\t} else {\n\t\tbuf = msg.Body\n\t}\n\n\t_, err := (*pipe.conn).Write(buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmsg.Free()\n\treturn nil\n}", "func WriteMetaInfo(fs *formatters.FormatterConfig) {\n sugared.WriteName(fs, \" - \")\n sugared.WriteLevel(fs)\n fs.PoolState.Buffer.WriteString(\": \")\n sugared.WriteLevelTabulation(fs)\n}", "func WrapWithMetadataPIID(msgMap ...service.DIDCommMsg) {\n\tpiID := uuid.New().String()\n\n\tfor _, msg := range msgMap {\n\t\tmsg.Metadata()[metaPIID] = piID\n\t}\n}", "func OutMessage(message string) (responseMsg string, err error) {\n\treturn\n}", "func (me *ChannelLine) Put(message string) {\n\tme.channel <- (me.name + \":\\t\" + message)\n}", "func WriteResponseMessage(w http.ResponseWriter, message string, trace string, status int, success bool) {\n\tmsg := models.Message{MSG: message, Success: success, Trace: trace, Status: status}\n\tgo logger.LogString(msg)\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Write(msg.Bytes())\n}", "func (chs *Chunks) WriteMessage(w io.Writer, message *message.Raw) (err error) {\n\tbasic, err := chs.writeHeader(w, message)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := message.Data\n\tfor len(data) > 0 {\n\t\tvar buf []byte\n\t\tif len(data) > len(chs.DataBuffer) {\n\t\t\tbuf, data = data[:len(chs.DataBuffer)], data[len(chs.DataBuffer):]\n\t\t} else {\n\t\t\tbuf, data = data, nil\n\t\t}\n\n\t\t_, err = w.Write(buf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif len(data) > 0 {\n\t\t\tbasic[0] = (3 << 6) | (basic[0] & 0x3f)\n\n\t\t\t_, err = w.Write(basic)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn\n}", "func writeMessage(c context.Context, w http.ResponseWriter, msg proto.Message, format Format) {\n\tif msg == nil {\n\t\tpanic(\"msg is nil\")\n\t}\n\n\tvar body []byte\n\tvar err error\n\tswitch format {\n\tcase FormatBinary:\n\t\tbody, err = proto.Marshal(msg)\n\n\tcase FormatJSONPB:\n\t\tvar buf bytes.Buffer\n\t\tbuf.WriteString(JSONPBPrefix)\n\t\tm := jsonpb.Marshaler{}\n\t\terr = m.Marshal(&buf, msg)\n\t\tif err == nil {\n\t\t\t_, err = buf.WriteRune('\\n')\n\t\t}\n\t\tbody = buf.Bytes()\n\n\tcase FormatText:\n\t\tvar buf bytes.Buffer\n\t\terr = proto.MarshalText(&buf, msg)\n\t\tbody = buf.Bytes()\n\n\tdefault:\n\t\tpanic(fmt.Errorf(\"impossible: invalid format %d\", format))\n\n\t}\n\tif err != nil {\n\t\twriteError(c, w, withCode(err, codes.Internal))\n\t\treturn\n\t}\n\n\tw.Header().Set(HeaderGRPCCode, strconv.Itoa(int(codes.OK)))\n\tw.Header().Set(headerContentType, format.MediaType())\n\tw.Header().Set(\"X-Content-Type-Options\", \"nosniff\")\n\tif _, err := w.Write(body); err != nil {\n\t\tlogging.WithError(err).Errorf(c, \"prpc: failed to write response body\")\n\t}\n}" ]
[ "0.57942605", "0.57841104", "0.560688", "0.5452389", "0.53893256", "0.5363706", "0.5342951", "0.5277034", "0.5256287", "0.52318096", "0.51665723", "0.5165779", "0.5154253", "0.51304793", "0.51088965", "0.5095732", "0.5083538", "0.5067103", "0.5031061", "0.50231755", "0.50188905", "0.49862653", "0.4946602", "0.4938537", "0.4929172", "0.4929172", "0.4926693", "0.49231842", "0.4906604", "0.49006668", "0.48987037", "0.48973972", "0.48777103", "0.4856129", "0.48519105", "0.48413497", "0.48309523", "0.4820574", "0.4813683", "0.48072103", "0.48062208", "0.4805842", "0.48021844", "0.47793132", "0.47641587", "0.47640267", "0.47621354", "0.47473118", "0.4746337", "0.47448662", "0.47425583", "0.47424015", "0.4739464", "0.47361887", "0.47314686", "0.47307923", "0.47296464", "0.47273803", "0.47267678", "0.47263142", "0.4718333", "0.47179988", "0.47151068", "0.47094822", "0.4706422", "0.47009587", "0.46758363", "0.46664524", "0.46660373", "0.4665518", "0.46643674", "0.46641484", "0.46632415", "0.4658846", "0.46573475", "0.46564767", "0.46548617", "0.46532086", "0.46528742", "0.46513256", "0.46492937", "0.46492767", "0.46487927", "0.46480253", "0.46452633", "0.4643253", "0.4640693", "0.46274272", "0.4626623", "0.46241805", "0.4623751", "0.4623374", "0.462313", "0.46204287", "0.46160188", "0.46148717", "0.4613282", "0.46131918", "0.4608712", "0.46054602" ]
0.5044047
18
Incoming attaches any additional metadata to a message
func (e *Extension) Incoming(ms *bayeux.Message) { switch ms.Channel.Type() { case bayeux.MetaChannel: switch ms.Channel { case bayeux.MetaHandshake: ext := ms.GetExt(false) if ext != nil { isSupported, ok := ext[ExtensionName].(bool) if ok && isSupported { atomic.CompareAndSwapInt32(e.supportedByServer, unsupported, supported) } } return case bayeux.MetaUnsubscribe: for _, channel := range ms.Subscription { e.replayStore.Delete(string(channel)) } return case bayeux.MetaConnect, bayeux.MetaSubscribe: return } case bayeux.BroadcastChannel: e.updateReplayID(ms) case bayeux.ServiceChannel: return } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (basic *BasicMessage) AddMetadata(key string, value string) {\n\tmetadata := NewMetadata(key, value)\n\tbasic.Metadata = append(basic.Metadata, metadata)\n}", "func (h Dnstap) TapMessageWithMetadata(ctx context.Context, m *tap.Message, state request.Request) {\n\tif h.ExtraFormat == \"\" {\n\t\th.tapWithExtra(m, nil)\n\t\treturn\n\t}\n\textraStr := h.repl.Replace(ctx, state, nil, h.ExtraFormat)\n\th.tapWithExtra(m, []byte(extraStr))\n}", "func (h Handler) Add(ctx context.Context, request *proto.Identifier) (*proto.Message, error) {\n\terr := h.meta.Create(ctx, request.UserID)\n\terr = errors.Wrap(err, \"Error while creating and adding new meta data\")\n\treturn &proto.Message{}, err\n}", "func (e *GameEngine) IncomingTalk(from ID, text string) {\n\te.talk <- struct {\n\t\tID\n\t\tstring\n\t}{from, text}\n}", "func (c *jsiiProxy_CfnDeliveryStream) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (m *IncomingContext) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func handleInteractionAddedMessage(s *common.Service, m *sarama.ConsumerMessage) error {\n\tvar ole kafka.ObjectLogEntry\n\tjson.Unmarshal(m.Value, &ole)\n\tvar i messages.ProximityInteraction\n\tjson.Unmarshal(ole.Object, &i)\n\n\tlog.Printf(\"umarshalled interaction: %+v\\n\", i)\n\n\t// Create new event score\n\tmodels.CreateEventScore(s.DB, i.UID, i.ID, models.ProximityInteraction, i.Timestamp, config.ProximityInteractionPoints)\n\n\t// TODO: Produce message for proximity service to consume\n\n\treturn nil\n}", "func (m *EventMessageDetail) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func Incoming(url string, msg *Message) (err error) {\n\tb, err := json.Marshal(msg)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treq, err := http.NewRequest(\"POST\", url, bytes.NewBuffer(b))\n\tif err != nil {\n\t\treturn\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tclient := &http.Client{\n\t\tTimeout: time.Second * 30,\n\t}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdefer resp.Body.Close()\n\treturn\n}", "func persist_inbound(s Store, m proto.Message) {\n\tswitch m := m.(type) {\n\tcase *proto.Publish:\n\t\tif m.Header.QosLevel == proto.QosAtLeastOnce {\n\t\t\t// store it in ibound until PubAck sent\n\t\t\ts.Put(ibound_mid2key(m.MessageId), m)\n\t\t}\n\n\tcase *proto.PubAck:\n\t\ts.Del(obound_mid2key(m.MessageId))\n\n\t}\n\n}", "func (dataChannel *DataChannel) AddDataToIncomingMessageBuffer(streamMessage StreamingMessage) {\n\tif len(dataChannel.IncomingMessageBuffer.Messages) == dataChannel.IncomingMessageBuffer.Capacity {\n\t\treturn\n\t}\n\tdataChannel.IncomingMessageBuffer.Mutex.Lock()\n\tdataChannel.IncomingMessageBuffer.Messages[streamMessage.SequenceNumber] = streamMessage\n\tdataChannel.IncomingMessageBuffer.Mutex.Unlock()\n}", "func FromIncomingContext(ctx context.Context) MetaData {\n\tm := make(MetaData)\n\tdata, ok := md.FromIncomingContext(ctx)\n\tif !ok {\n\t\treturn m\n\t}\n\tfor k, v := range data {\n\t\tif len(v) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tk = strings.ReplaceAll(strings.ToLower(k), \"_\", \"-\")\n\t\tm[k] = v[0]\n\t}\n\n\t// get remote address from peer\n\tpr, ok := peer.FromContext(ctx)\n\tif ok && pr.Addr != nil {\n\t\tss := strings.Split(pr.Addr.String(), \":\")\n\t\tm[remoteIPFromPeer] = ss[0]\n\t}\n\n\treturn m\n}", "func (iso *Message) AddMTI(data string) error {\n\tif err := ValidateMti(data); err != nil {\n\t\treturn err\n\t}\n\n\tiso.mti = data\n\t//fmt.Printf(\"MTI: %s\\n\", iso.Mti)\n\treturn nil\n}", "func (m *ChannelIdentity) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func InterceptMessage(conn *postmaster.Connection, msg postmaster.PublishMsg)(bool){\n\t//Filter out base url and split into components\n\turi := strings.Replace(msg.TopicURI,baseURL,\"\",1)\n\targs := strings.Split(uri,\"/\")\n\t\n\tusername := args[0]\n\t\n\tdata,ok := msg.Event.(map[string]interface{}) //cast\n\t\n\tif !ok{\n\t\tlog.Print(\"Message doesn't follow correct format: ignoring\")\n\t\treturn false\n\t}\n\t\n\t//Switch through command types\n\tswitch data[\"command\"]{\n\tcase \"addTrack\":\t\t\n\t\t//Parse recieved tracks\n\t\ttracks := data[\"data\"].([]interface{})\n\t\t\n\t\tfor _,m := range tracks{\n\t\t\ttrack := m.(map[string]interface{})\n\t\t\t\n\t\t\tt := TrackItem{\n\t\t\t\t\tProviderID:track[\"ProviderID\"].(string),\n\t\t\t\t\tTitle:track[\"Title\"].(string),\n\t\t\t\t\tArtistName:track[\"ArtistName\"].(string),\n\t\t\t\t\tAlbumName:track[\"AlbumName\"].(string),\n\t\t\t\t\tArtworkURL:track[\"ArtworkURL\"].(string),\n\t\t\t\t\tLength:track[\"Length\"].(float64),\n\t\t\t}\n\t\t\t\n\t\t\taddTrackToQueue(args[1],t)\n\t\t}\n\t\t\n\tcase \"removeTrack\":\n\tcase \"playTrack\":\n\t\tif len(args) > 1{\n\t\t\tsetMusicBoxPlaying(args[1],PLAYING) //Set Playing = true\n\t\t}\n\tcase \"pauseTrack\":\n\t\tfallthrough //Same as stop\n\tcase \"stopTrack\":\n\t\tif len(args) > 1{\n\t\t\tsetMusicBoxPlaying(args[1],PAUSED) //Set Playing = false\n\t\t}\n\tcase \"nextTrack\":\n\tcase \"startedTrack\":\n\t\tlog.Print(\"Track Started\")\n\t\t//Parse recieved track\n\t\td := data[\"data\"].(map[string]interface{})\n\t\t\n\t\ttrack := d[\"track\"].(map[string]interface{})\n\t\tt := TrackItem{\n\t\t\t\tProviderID:track[\"ProviderID\"].(string),\n\t\t\t\tTitle:track[\"Title\"].(string),\n\t\t\t\tArtistName:track[\"ArtistName\"].(string),\n\t\t\t\tAlbumName:track[\"AlbumName\"].(string),\n\t\t\t\tArtworkURL:track[\"ArtworkURL\"].(string),\n\t\t\t\tLength:track[\"Length\"].(float64),\n\t\t}\n\t\tfmt.Println(track)\n\t\tdeviceID := d[\"deviceID\"].(string)\n\t\tfmt.Println(deviceID)\n\t\t\n\t\t\n\t\t//\n\t\t// Remove from queue\n\t\t//\n\t\t\n\t\tpopTrackOffQueue(args[1])\n\t\t\n\t\t//\n\t\t// Save in track history\n\t\t//\n\t\t\n\t\t//Create aws item\n\t\tatts := []dynamodb.Attribute{\n\t\t\t*dynamodb.NewStringAttribute(\"Title\",t.Title),\n\t\t\t*dynamodb.NewStringAttribute(\"ArtistName\",t.ArtistName),\n\t\t\t*dynamodb.NewStringAttribute(\"ProviderID\",t.ProviderID),\n\t\t\t*dynamodb.NewStringAttribute(\"AlbumName\",t.AlbumName), //Moment.us doesn't always provide this\n\t\t\t*dynamodb.NewStringAttribute(\"ArtworkURL\",t.ArtworkURL), //Moment.us doesn't always provide this\n\t\t\t*dynamodb.NewNumericAttribute(\"Length\",strconv.FormatFloat(t.Length,'f',-1,32)), //Moment.us doesn't always provide this\n\t\t}\n\t\t\t\t\t\t\t\t\n\t\t//Add track to database for this user:musicbox\n\t\tif _,err := trackHistoryTable.PutItem(username+\":\"+deviceID,time.Now().UTC().Format(time.RFC3339),atts); err != nil{\n\t\t\tlog.Print(err.Error())\n\t\t}else{\n\t\t\tlog.Print(\"Put New track\")\n\t\t}\n\t\t\n\t\t//Set playing to true\n\t\tsetMusicBoxPlaying(deviceID,PLAYING) //Set Playing = true\n\t\t\n\tcase \"updateTheme\":\n\t\ttheme := themeItemFromMap(data[\"data\"].(map[string]interface{}))\n\t\tboxID := args[1]\n\t\t\n\t\t//Update box information with new theme\n\t\tthemeUpdate := []dynamodb.Attribute{*dynamodb.NewStringAttribute(\"ThemeID\",theme.ThemeID)}\n\t\t\n\t\t_, err := musicBoxesTable.UpdateAttributes(&dynamodb.Key{HashKey: boxID},themeUpdate)\n\t\tif err != nil{\n\t\t\tlog.Print(err.Error())\n\t\t}\n\t\t\n\t\t\n\tdefault:\n\t\t\tlog.Print(\"Unknown Command:\",data[\"command\"])\n\t}\n\t\t\t\t\n\treturn true\n}", "func (m *ChatMessageAttachment) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func NewIncomingContext()(*IncomingContext) {\n m := &IncomingContext{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func (manager *Manager) receiveMetadata() {\n\tfor t := range manager.metadata.timings {\n\t\tswitch ti := t.(type) {\n\t\tcase CloneTime:\n\t\t\tmanager.metadata.cloneTime += int64(ti)\n\t\tcase ScanTime:\n\t\t\tmanager.metadata.ScanTime += int64(ti)\n\t\tcase PatchTime:\n\t\t\tmanager.metadata.patchTime += int64(ti)\n\t\tcase RegexTime:\n\t\t\tmanager.metadata.RegexTime[ti.Regex] = manager.metadata.RegexTime[ti.Regex] + ti.Time\n\t\t}\n\t\tmanager.metaWG.Done()\n\t}\n}", "func (this *IoHandlerImp) MessageReceived(filter *IoFilter, obj BaseObject) {\n}", "func (h *PingHandler) HandleIncoming(conn mproto.Connection, p *proto.Packet) error {\n\tif p.Type != proto.PingEventType {\n\t\treturn nil\n\t}\n\tpayload, err := p.Payload()\n\tif err != nil {\n\t\treturn err\n\t}\n\tpingEvent, ok := payload.(*proto.PingEvent)\n\tif !ok {\n\t\treturn fmt.Errorf(\"Error asserting ping-event as such\")\n\t}\n\tpingReply := &proto.PingReply{\n\t\tUnixTime: pingEvent.UnixTime,\n\t}\n\tmarshalled, err := json.Marshal(pingReply)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn conn.Send(&proto.Packet{\n\t\tType: proto.PingReplyType,\n\t\tData: marshalled,\n\t})\n}", "func handleNewMessage(msg *arbor.ProtocolMessage, recents *RecentList, store *arbor.Store, broadcaster *Broadcaster) {\n\terr := msg.ChatMessage.AssignID()\n\tif err != nil {\n\t\tlog.Println(\"Error creating new message\", err)\n\t}\n\trecents.Add(msg.ChatMessage)\n\tstore.Add(msg.ChatMessage)\n\tbroadcaster.Send(msg)\n}", "func (r *AWSMediaLiveChannel_InputAttachment) SetMetadata(metadata map[string]interface{}) {\n\tr._metadata = metadata\n}", "func (c *jsiiProxy_CfnElasticLoadBalancerAttachment) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (m *InvitedUserMessageInfo) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *UserSimulationEventInfo) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (rb *routerBase) fwdMetadata(meta *execinfrapb.ProducerMetadata) {\n\tif meta == nil {\n\t\tlog.Fatalf(context.TODO(), \"asked to fwd empty metadata\")\n\t\treturn\n\t}\n\n\trb.semaphore <- struct{}{}\n\tdefer func() {\n\t\t<-rb.semaphore\n\t}()\n\tif metaErr := meta.Err; metaErr != nil {\n\t\t// Forward the error to all non-closed streams.\n\t\tif rb.fwdErrMetadata(metaErr) {\n\t\t\treturn\n\t\t}\n\t} else {\n\t\t// Forward the metadata to the first non-closed stream.\n\t\tfor i := range rb.outputs {\n\t\t\tro := &rb.outputs[i]\n\t\t\tro.mu.Lock()\n\t\t\tif ro.mu.streamStatus != execinfra.ConsumerClosed {\n\t\t\t\tro.addMetadataLocked(meta)\n\t\t\t\tro.mu.Unlock()\n\t\t\t\tro.mu.cond.Signal()\n\t\t\t\treturn\n\t\t\t}\n\t\t\tro.mu.Unlock()\n\t\t}\n\t}\n\t// If we got here it means that we couldn't even forward metadata anywhere;\n\t// all streams are closed.\n\tatomic.StoreUint32(&rb.aggregatedStatus, uint32(execinfra.ConsumerClosed))\n}", "func (s *BaseRFC5424Listener) EnterMsg(ctx *MsgContext) {}", "func (m *PresenceStatusMessage) SetAdditionalData(value map[string]any)() {\n err := m.GetBackingStore().Set(\"additionalData\", value)\n if err != nil {\n panic(err)\n }\n}", "func (nc *NetClient) HandleMeta(meta map[string]string) {\n\tfor key, value := range meta {\n\t\tswitch key {\n\t\tcase \"presence/who\":\n\t\t\tnc.Composer.AnnounceHere(nc.Session.ID)\n\t\tcase \"presence/here\":\n\t\t\tusername, sessionID, timestamp, err := parsePresence(value)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"error parsing presence/here message\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif username == nc.username && sessionID == nc.Session.ID {\n\t\t\t\t// don't track our own session\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\terr = nc.List.Track(username, session.Session{ID: sessionID, LastSeen: timestamp})\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"Error updating session\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tlog.Printf(\"Tracking session (id=%s) for user %s\\n\", sessionID, username)\n\t\tcase \"presence/leave\":\n\t\t\tusername, sessionID, _, err := parsePresence(value)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"error parsing presence/leave message\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif username == nc.username && sessionID == nc.Session.ID {\n\t\t\t\t// don't remove our own session\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\terr = nc.List.Remove(username, sessionID)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"Error removing session\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tlog.Printf(\"Removed session (id=%s) for user %s\\n\", sessionID, username)\n\t\tdefault:\n\t\t\tlog.Println(\"Unknown meta key:\", key)\n\t\t}\n\t}\n}", "func (s *FriendshipService) Incoming(params *FriendshipPendingParams) (*FriendIDs, *http.Response, error) {\n\tids := new(FriendIDs)\n\tapiError := new(APIError)\n\tresp, err := s.sling.New().Get(\"incoming.json\").QueryStruct(params).Receive(ids, apiError)\n\treturn ids, resp, relevantError(err, *apiError)\n}", "func (object *MQMessageHandler) OnMQMessage(raw []byte, offset int64) {\n}", "func (app *builder) WithIncoming(incoming []hash.Hash) Builder {\n\tapp.incoming = incoming\n\treturn app\n}", "func (s *seatPlugin) OnMessage(ctx *muxer.Context) {\n\tparts := ctx.Fields\n\tif len(parts) <= 1 {\n\t\tctx.SendPrivately(how2use)\n\t\treturn\n\t}\n\n\t// Do we have the user id?\n\tgroupID, ok := s.mapping[ctx.FromUserID]\n\tif !ok {\n\t\t// Update the cache...\n\t\tcacheWasLastUpdated := time.Now().Sub(s.lastUpdated)\n\t\tif cacheWasLastUpdated < cacheCanBeUpdatedEvery {\n\t\t\ttryAgain := cacheCanBeUpdatedEvery - cacheWasLastUpdated\n\t\t\tctx.SendPrivately(fmt.Sprintf(\"Please wait %s before trying again\", tryAgain.String()))\n\t\t\treturn\n\t\t}\n\n\t\tif err := s.updateCache(); err != nil {\n\t\t\tctx.SendPrivately(\"Sorry m8 try again later\")\n\t\t\treturn\n\t\t}\n\n\t\t// Try again after the cache update\n\t\tgroupID, ok = s.mapping[ctx.FromUserID]\n\t\tif !ok {\n\t\t\tctx.SendPrivately(fmt.Sprintf(registerMessage, s.server))\n\t\t\treturn\n\t\t}\n\t}\n\n\tswitch strings.ToLower(parts[1]) {\n\tcase \"characters\", \"chars\":\n\t\ts.sendCharacterInfo(groupID, ctx)\n\tdefault:\n\t\tctx.SendPrivately(how2use)\n\t}\n}", "func (m *Media) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (s *Switch) processMessage(ctx context.Context, ime net.IncomingMessageEvent) {\n\t// Extract request context and add to log\n\tif ime.RequestID != \"\" {\n\t\tctx = log.WithRequestID(ctx, ime.RequestID)\n\t} else {\n\t\tctx = log.WithNewRequestID(ctx)\n\t\ts.logger.WithContext(ctx).Warning(\"got incoming message event with no requestID, setting new id\")\n\t}\n\n\tif s.config.MsgSizeLimit != config.UnlimitedMsgSize && len(ime.Message) > s.config.MsgSizeLimit {\n\t\ts.logger.WithContext(ctx).With().Error(\"message is too big to process\",\n\t\t\tlog.Int(\"limit\", s.config.MsgSizeLimit),\n\t\t\tlog.Int(\"actual\", len(ime.Message)))\n\t\treturn\n\t}\n\n\tif err := s.onRemoteClientMessage(ctx, ime); err != nil {\n\t\t// TODO: differentiate action on errors\n\t\ts.logger.WithContext(ctx).With().Error(\"err reading incoming message, closing connection\",\n\t\t\tlog.FieldNamed(\"sender_id\", ime.Conn.RemotePublicKey()),\n\t\t\tlog.Err(err))\n\t\tif err := ime.Conn.Close(); err == nil {\n\t\t\ts.cPool.CloseConnection(ime.Conn.RemotePublicKey())\n\t\t\ts.Disconnect(ime.Conn.RemotePublicKey())\n\t\t}\n\t}\n}", "func (client *S3Upload) AddMetadata(key, value string) {\n\tclient.UploadInput.Metadata[key] = &value\n}", "func InMessage(messengerID, msg, stringBuffer string) (outServerMsg string, err error) {\n\tif msg == \"info\" {\n\t\toutServerMsg, err = controlsystemhome.GetInfoControlSystemHomeInterfaces()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tstr, errr := controlled.GetInfoControlledsString()\n\t\tif errr != nil {\n\t\t\terr = errr\n\t\t\treturn\n\t\t}\n\t\toutServerMsg += \"\\n\" + str\n\t\treturn\n\t}\n\n\toutServerMsg, err = commandrecord.UsedTextCommand(msg, stringBuffer)\n\treturn\n}", "func (m *CommsNotification) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (msg *stunMessage) addMessageIntegrity(password string) {\n\t// Use the password to make a new HMAC hash, which has sig.Size() == 20\n\tsig := hmac.New(sha1.New, []byte(password))\n\n\t// Add a dummy MESSAGE-INTEGRITY attribute, such that it is included in msg.length.\n\tattr := msg.addAttribute(stunAttrMessageIntegrity, zeros[0:20])\n\n\t// Compute hash of the message contents up to *just before* the MESSAGE-INTEGRITY.\n\tb := msg.Bytes()\n\tbeforeMessageIntegrity := len(b) - attr.numBytes()\n\tsig.Write(b[0:beforeMessageIntegrity])\n\n\tcopy(attr.Value, sig.Sum(nil))\n}", "func (m *IncomingContext) GetAdditionalData()(map[string]interface{}) {\n return m.additionalData\n}", "func newReceivedEventData(amqpMsg *amqp.Message) (*ReceivedEventData, error) {\n\tre := &ReceivedEventData{\n\t\tRawAMQPMessage: newAMQPAnnotatedMessage(amqpMsg),\n\t}\n\n\tif len(amqpMsg.Data) == 1 {\n\t\tre.Body = amqpMsg.Data[0]\n\t}\n\n\tif amqpMsg.Properties != nil {\n\t\tif id, ok := amqpMsg.Properties.MessageID.(string); ok {\n\t\t\tre.MessageID = &id\n\t\t}\n\n\t\tre.ContentType = amqpMsg.Properties.ContentType\n\t\tre.CorrelationID = amqpMsg.Properties.CorrelationID\n\t}\n\n\tif amqpMsg.ApplicationProperties != nil {\n\t\tre.Properties = make(map[string]any, len(amqpMsg.ApplicationProperties))\n\t\tfor key, value := range amqpMsg.ApplicationProperties {\n\t\t\tre.Properties[key] = value\n\t\t}\n\t}\n\n\tif err := updateFromAMQPAnnotations(amqpMsg, re); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn re, nil\n}", "func (c *jsiiProxy_CfnContactChannel) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (ms *sender) prep(ctx context.Context) error {\n\tif ms.invalid {\n\t\treturn fmt.Errorf(\"message sender has been invalidated\")\n\t}\n\tif ms.stream != nil {\n\t\treturn nil\n\t}\n\tnstr, err := ms.service.Node.PeerHost.NewStream(ctx, ms.pid, ms.protocol)\n\tif err != nil {\n\t\treturn err\n\t}\n\tms.reader = ggio.NewDelimitedReader(nstr, inet.MessageSizeMax)\n\tms.writer = ggio.NewDelimitedWriter(nstr)\n\tms.stream = nstr\n\treturn nil\n}", "func (h *clientHandler) handleMsg(p *peer) error {\n\t// Read the next message from the remote peer, and ensure it's fully consumed\n\tmsg, err := p.rw.ReadMsg()\n\tif err != nil {\n\t\treturn err\n\t}\n\tp.Log().Trace(\"Light Ethereum message arrived\", \"code\", msg.Code, \"bytes\", msg.Size)\n\n\tif msg.Size > ProtocolMaxMsgSize {\n\t\treturn errResp(ErrMsgTooLarge, \"%v > %v\", msg.Size, ProtocolMaxMsgSize)\n\t}\n\tdefer msg.Discard()\n\n\tvar deliverMsg *Msg\n\n\t// Handle the message depending on its contents\n\tswitch msg.Code {\n\tcase AnnounceMsg:\n\t\tp.Log().Trace(\"Received announce message\")\n\t\tvar req announceData\n\t\tif err := msg.Decode(&req); err != nil {\n\t\t\treturn errResp(ErrDecode, \"%v: %v\", msg, err)\n\t\t}\n\t\tif err := req.sanityCheck(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tupdate, size := req.Update.decode()\n\t\tif p.rejectUpdate(size) {\n\t\t\treturn errResp(ErrRequestRejected, \"\")\n\t\t}\n\t\tp.updateFlowControl(update)\n\n\t\tif req.Hash != (common.Hash{}) {\n\t\t\tif p.announceType == announceTypeNone {\n\t\t\t\treturn errResp(ErrUnexpectedResponse, \"\")\n\t\t\t}\n\t\t\tif p.announceType == announceTypeSigned {\n\t\t\t\tif err := req.checkSignature(p.ID(), update); err != nil {\n\t\t\t\t\tp.Log().Trace(\"Invalid announcement signature\", \"err\", err)\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tp.Log().Trace(\"Valid announcement signature\")\n\t\t\t}\n\t\t\tp.Log().Trace(\"Announce message content\", \"number\", req.Number, \"hash\", req.Hash, \"td\", req.Td, \"reorg\", req.ReorgDepth)\n\t\t\th.fetcher.announce(p, &req)\n\t\t}\n\tcase BlockHeadersMsg:\n\t\tp.Log().Trace(\"Received block header response message\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tHeaders []*types.Header\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tif h.fetcher.requestedID(resp.ReqID) {\n\t\t\th.fetcher.deliverHeaders(p, resp.ReqID, resp.Headers)\n\t\t} else {\n\t\t\tif err := h.downloader.DeliverHeaders(p.id, resp.Headers); err != nil {\n\t\t\t\tlog.Debug(\"Failed to deliver headers\", \"err\", err)\n\t\t\t}\n\t\t}\n\tcase BlockBodiesMsg:\n\t\tp.Log().Trace(\"Received block bodies response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tData []*types.Body\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgBlockBodies,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Data,\n\t\t}\n\tcase CodeMsg:\n\t\tp.Log().Trace(\"Received code response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tData [][]byte\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgCode,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Data,\n\t\t}\n\tcase ReceiptsMsg:\n\t\tp.Log().Trace(\"Received receipts response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tReceipts []types.Receipts\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgReceipts,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Receipts,\n\t\t}\n\tcase ProofsV2Msg:\n\t\tp.Log().Trace(\"Received les/2 proofs response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tData light.NodeList\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgProofsV2,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Data,\n\t\t}\n\tcase HelperTrieProofsMsg:\n\t\tp.Log().Trace(\"Received helper trie proof response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tData HelperTrieResps\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgHelperTrieProofs,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Data,\n\t\t}\n\tcase TxStatusMsg:\n\t\tp.Log().Trace(\"Received tx status response\")\n\t\tvar resp struct {\n\t\t\tReqID, BV uint64\n\t\t\tStatus []light.TxStatus\n\t\t}\n\t\tif err := msg.Decode(&resp); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ReceivedReply(resp.ReqID, resp.BV)\n\t\tdeliverMsg = &Msg{\n\t\t\tMsgType: MsgTxStatus,\n\t\t\tReqID: resp.ReqID,\n\t\t\tObj: resp.Status,\n\t\t}\n\tcase StopMsg:\n\t\tp.freezeServer(true)\n\t\th.backend.retriever.frozen(p)\n\t\tp.Log().Debug(\"Service stopped\")\n\tcase ResumeMsg:\n\t\tvar bv uint64\n\t\tif err := msg.Decode(&bv); err != nil {\n\t\t\treturn errResp(ErrDecode, \"msg %v: %v\", msg, err)\n\t\t}\n\t\tp.fcServer.ResumeFreeze(bv)\n\t\tp.freezeServer(false)\n\t\tp.Log().Debug(\"Service resumed\")\n\tdefault:\n\t\tp.Log().Trace(\"Received invalid message\", \"code\", msg.Code)\n\t\treturn errResp(ErrInvalidMsgCode, \"%v\", msg.Code)\n\t}\n\t// Deliver the received response to retriever.\n\tif deliverMsg != nil {\n\t\tif err := h.backend.retriever.deliver(p, deliverMsg); err != nil {\n\t\t\tp.responseErrors++\n\t\t\tif p.responseErrors > maxResponseErrors {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (l *OnionMessage) Encapsulate(to peer.ID, publicKey *[32]byte) (*OnionMessage, error) {\n\t// TODO: should use a better encoding than JSON.\n\tb, err := json.Marshal(l)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, ErrMarshal)\n\t}\n\n\tciphertext, err := seal(publicKey, b)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &OnionMessage{\n\t\tTo: []byte(to),\n\t\tContent: ciphertext,\n\t}, nil\n}", "func (m *SharePostRequestBody) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func receiveIncoming(reader *bufio.Reader, message *[]byte) (int, error) {\n numRead, err := reader.Read(*message)\n\n // let the caller choose how to deal with an EOF\n if err != nil && err != io.EOF {\n fmt.Println(\"Error reading incoming message: \", err.Error())\n }\n\n return numRead, err\n}", "func (p *InvoiceUpcomingInvoiceItemParams) AddMetadata(key string, value string) {\n\tif p.Metadata == nil {\n\t\tp.Metadata = make(map[string]string)\n\t}\n\n\tp.Metadata[key] = value\n}", "func handleIncomingMessage(session *discordgo.Session, message *discordgo.MessageCreate) {\n\t// Filters out messages from other bots, but also our own!\n\tif message.Author.Bot {\n\t\treturn\n\t}\n\n\t// Sneaky way to handle message components\n\tsendInteractionComponentIfNeeded(message)\n\n\t// Filter out non Direct Message messages\n\tchannel, channelErr := session.Channel(message.ChannelID)\n\tif channelErr != nil {\n\t\treturn\n\t}\n\n\tif channel.Type != discordgo.ChannelTypeDM {\n\t\treturn\n\t}\n\n\t// Check if this person already is in an ongoing conversation with the bot\n\tcurrentReportsMutex.RLock()\n\tif report, ok := currentOngoingReports[message.Author.ID]; ok {\n\t\tcurrentReportsMutex.RUnlock()\n\t\treport.lock.Lock()\n\t\tdefer report.lock.Unlock()\n\n\t\t// The user is already in an ongoing conversation, continue it\n\t\tcontinueOngoingReport(report, message.Content, message.Author.ID, message)\n\t} else {\n\t\t// The user is not in an ongoing conversation, make sure to start a new one\n\t\tcurrentReportsMutex.RUnlock()\n\t\tstartNewReportConversation(message.Author.ID, \"\")\n\t}\n}", "func (s *Switch) onRemoteClientMessage(ctx context.Context, msg net.IncomingMessageEvent) error {\n\tif msg.Message == nil || msg.Conn == nil {\n\t\treturn ErrBadFormat1\n\t}\n\n\t// protocol messages are encrypted in payload\n\t// Locate the session\n\tsession := msg.Conn.Session()\n\n\tif session == nil {\n\t\treturn ErrNoSession\n\t}\n\n\tdecPayload, err := session.OpenMessage(msg.Message)\n\tif err != nil {\n\t\treturn ErrFailDecrypt\n\t}\n\n\tpm := &ProtocolMessage{}\n\tif err = types.BytesToInterface(decPayload, pm); err != nil {\n\t\ts.logger.With().Error(\"error deserializing message\", log.Err(err))\n\t\treturn ErrBadFormat2\n\t}\n\n\t// check that the message was sent within a reasonable time\n\tif ok := timesync.CheckMessageDrift(pm.Metadata.Timestamp); !ok {\n\t\t// TODO: consider kill connection with this node and maybe blacklist\n\t\t// TODO : Also consider moving send timestamp into metadata(encrypted).\n\t\treturn ErrOutOfSync\n\t}\n\n\tdata, err := ExtractData(pm.Payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Add metadata collected from p2p message (todo: maybe pass sender and protocol inside metadata)\n\tp2pmeta := service.P2PMetadata{FromAddress: msg.Conn.RemoteAddr()}\n\n\t_, ok := s.gossipProtocolHandlers[pm.Metadata.NextProtocol]\n\n\ts.logger.WithContext(ctx).With().Debug(\"handle incoming message\",\n\t\tlog.String(\"protocol\", pm.Metadata.NextProtocol),\n\t\tlog.FieldNamed(\"peer_id\", msg.Conn.RemotePublicKey()),\n\t\tlog.Bool(\"is_gossip\", ok))\n\n\tif ok {\n\t\t// if this message is tagged with a gossip protocol, relay it.\n\t\treturn s.gossip.Relay(ctx, msg.Conn.RemotePublicKey(), pm.Metadata.NextProtocol, data)\n\t}\n\n\t// route authenticated message to the registered protocol\n\t// messages handled here are always processed by direct based protocols, only the gossip protocol calls ProcessGossipProtocolMessage\n\treturn s.ProcessDirectProtocolMessage(ctx, msg.Conn.RemotePublicKey(), pm.Metadata.NextProtocol, data, p2pmeta)\n}", "func IngameMessageCreate(lh *loghandler.LogHandler, server *servers.Server, event *loghandler.SayEvent) {\n\tlog.Println(fmt.Sprintf(\"Received command from '%s' on server '%s': %s\", event.Username, server.Name, event.Message))\n\tIngameCommand.Handle(ingame.CommandInfo{SayEvent: *event, Server: server}, event.Message, 0)\n}", "func (m *Monitor) handle(msg ndp.Message, host string) {\n\tm.debugf(\"monitor received %q from %s\", msg.Type(), host)\n\n\tm.cctx.mm.MonMessagesReceivedTotal(1.0, m.iface, host, msg.Type().String())\n\n\t// TODO(mdlayher): expand type switch.\n\tswitch msg := msg.(type) {\n\tcase *ndp.RouterAdvertisement:\n\t\tnow := m.now()\n\n\t\tm.cctx.mm.MonFlagManaged(\n\t\t\tboolFloat(msg.ManagedConfiguration),\n\t\t\tm.iface, host,\n\t\t)\n\n\t\tm.cctx.mm.MonFlagOther(\n\t\t\tboolFloat(msg.OtherConfiguration),\n\t\t\tm.iface, host,\n\t\t)\n\n\t\tif msg.RouterLifetime != 0 {\n\t\t\t// This is an advertisement from a default router. Calculate the\n\t\t\t// UNIX timestamp of when the default route will expire.\n\t\t\tm.cctx.mm.MonDefaultRouteExpirationTime(\n\t\t\t\tfloat64(now.Add(msg.RouterLifetime).Unix()),\n\t\t\t\tm.iface, host,\n\t\t\t)\n\t\t}\n\n\t\t// Export metrics for each prefix option.\n\t\tfor _, p := range pick[*ndp.PrefixInformation](msg.Options) {\n\t\t\tstr := cidrStr(p.Prefix, p.PrefixLength)\n\n\t\t\tm.cctx.mm.MonPrefixAutonomous(\n\t\t\t\tboolFloat(p.AutonomousAddressConfiguration),\n\t\t\t\tm.iface, str, host,\n\t\t\t)\n\n\t\t\tm.cctx.mm.MonPrefixOnLink(\n\t\t\t\tboolFloat(p.OnLink),\n\t\t\t\tm.iface, str, host,\n\t\t\t)\n\n\t\t\tm.cctx.mm.MonPrefixPreferredLifetimeExpirationTime(\n\t\t\t\tfloat64(now.Add(p.PreferredLifetime).Unix()),\n\t\t\t\tm.iface, str, host,\n\t\t\t)\n\n\t\t\tm.cctx.mm.MonPrefixValidLifetimeExpirationTime(\n\t\t\t\tfloat64(now.Add(p.ValidLifetime).Unix()),\n\t\t\t\tm.iface, str, host,\n\t\t\t)\n\t\t}\n\t}\n}", "func (i *invoker) appendMetadata(ctx context.Context) context.Context {\n\treturn metadata.AppendToOutgoingContext(ctx, \"Authorization\", \"Bearer token\")\n}", "func (joinSession *JoinSession) pushMaliciousInfo(missedPeers []uint32) {\n\n\tjoinSession.mu.Lock()\n\tdefer joinSession.mu.Unlock()\n\tlog.Debug(\"Number malicious\", len(missedPeers))\n\tmalicious := &pb.MaliciousPeers{}\n\tfor _, Id := range missedPeers {\n\t\tjoinSession.removePeer(Id)\n\t}\n\tmalicious.PeerIds = missedPeers\n\n\tif len(joinSession.Peers) == 0 {\n\t\tlog.Info(\"All peers not sending data in time, join session terminates\")\n\t\treturn\n\t}\n\n\t// Re-generate the session id and peer id\n\tmalicious.SessionId = GenId()\n\tnewPeers := make(map[uint32]*PeerInfo, 0)\n\tlog.Debug(\"Remaining peers in join session: \", joinSession.Peers)\n\tif len(joinSession.Peers) > 0 {\n\t\tfor _, peer := range joinSession.Peers {\n\t\t\tmalicious.PeerId = GenId()\n\t\t\t// Update new id generated.\n\t\t\tpeer.ResetData(malicious.PeerId, malicious.SessionId)\n\t\t\tdata, _ := proto.Marshal(malicious)\n\t\t\tpeer.TmpData = data\n\t\t\tnewPeers[peer.Id] = peer\n\t\t}\n\t}\n\tjoinSession.Peers = newPeers\n\tjoinSession.JoinedTx = nil\n\tjoinSession.PeersMsgInfo = []*pb.PeerInfo{}\n\tjoinSession.Id = malicious.SessionId\n\tjoinSession.State = StateKeyExchange\n\tjoinSession.TotalMsg = 0\n\n\t// After all change updated, inform clients for malicious information.\n\tlog.Debug(\"len of joinSession.Peers to push malicious \", len(joinSession.Peers))\n\tfor _, peer := range joinSession.Peers {\n\t\tmsg := messages.NewMessage(messages.S_MALICIOUS_PEERS, peer.TmpData).ToBytes()\n\t\tpeer.writeChan <- msg\n\t}\n\n\tlog.Debug(\"Remaining peers in join session after updated: \", joinSession.Peers)\n}", "func (r *AWSMediaLiveChannel_InputAttachment) Metadata() map[string]interface{} {\n\treturn r._metadata\n}", "func CloneServerIncomingData(ctx context.Context) metadata.MD {\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif !ok {\n\t\treturn metadata.MD{}\n\t}\n\n\t//return a copy\n\treturn md.Copy()\n}", "func (r Report) addIncoming(incomingChannel,\n\toutgoingChannel string, amount, fees lnwire.MilliSatoshi) {\n\trevenue := r.getRevenue(incomingChannel, outgoingChannel)\n\n\t// Add the fees and revenue that have been earned to the existing revenue\n\t// record.\n\trevenue.AmountIncoming += amount\n\trevenue.FeesIncoming += fees\n\n\t// Set the new revenue record in the revenue report.\n\tr.setRevenue(incomingChannel, outgoingChannel, revenue)\n}", "func (m *EventMessageDetail) GetAdditionalData()(map[string]interface{}) {\n return m.additionalData\n}", "func (o *handler) handle(client mqtt.Client, msg mqtt.Message) {\r\n\t// We extract the count and write that out first to simplify checking for missing values\r\n\tvar m Message\r\n\tvar resp Session\r\n\tif err := json.Unmarshal(msg.Payload(), &resp); err != nil {\r\n\t\tfmt.Printf(\"Message could not be parsed (%s): %s\", msg.Payload(), err)\r\n\t\treturn\r\n\t}\r\n\tfmt.Println(resp)\r\n\tswitch resp.Type {\r\n\tcase CMDMSG_OFFER:\r\n\t\tenc.Decode(resp.Data, &m)\r\n\t\tNotice(m)\r\n\tcase CMDMSG_DISC:\r\n\t\tvar devcmd DiscoveryCmd\r\n\t\tenc.Decode(resp.Data, &devcmd)\r\n\t\tDiscoveryDev(&devcmd)\r\n\tcase CMDMSG_WAKE:\r\n\t\tvar fing Fing\r\n\t\tenc.Decode(resp.Data, &fing)\r\n\t\twakemac(fing)\r\n\tcase CMDMSG_UPDATE:\r\n\t\tvar newver *versionUpdate\r\n\t\tGetUpdateMyself(newver)\r\n\tcase CMDMSG_MR2:\r\n\t\tvar mr2info Mr2Msg\r\n\t\tenc.Decode(resp.Data, &mr2info)\r\n\t\tMr2HostPort(&mr2info)\r\n\t}\r\n}", "func (reader *Reader) AppendMetadata(mdi MetadataInterface) {\n\tmd := mdi.(*Metadata)\n\treader.InputMetadata = append(reader.InputMetadata, *md)\n}", "func writeStatMessage(payload string, currentMessage *bytes.Buffer) *bytes.Buffer {\n\t// 11 -event type - 7 - 5 \"Stat\" 20\n\t// 13 -content-type -7 -8 \"text/xml\" 25\n\t// 13 -message-type -7 5 \"event\" 22\n\t// This is predefined from AMZ protocol found here:\n\t// https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectSELECTContent.html\n\theaderLen := len(statHeaders)\n\n\tcurrentMessage.Write(writePayloadSize(len(payload), headerLen))\n\n\tcurrentMessage.Write(writeHeaderSize(headerLen))\n\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\n\tcurrentMessage.Write(statHeaders)\n\n\t// This part is where the payload is written, this will be only one row, since\n\t// we're sending one message at a types\n\tcurrentMessage.Write(writePayload(payload))\n\n\t// Now we do a CRC check on the entire messages\n\tcurrentMessage.Write(writeCRC(currentMessage.Bytes()))\n\treturn currentMessage\n\n}", "func (ipam *IPAMServer) IngestCustomData(devicesWithAddr []custom.DeviceData, devicesWithoutAddr []custom.UnknownDeviceData, preferredHeaderOrder ...string) {\n\tipam.custom.RecreateDatastore(devicesWithAddr, devicesWithoutAddr, preferredHeaderOrder)\n}", "func (m *Message) Add(key string, value interface{}) {\n\tif m.Data == nil {\n\t\tm.Data = make(map[string]interface{})\n\t}\n\tm.Data[key] = value\n}", "func (c *Context) I(message string, fields ...interface{}) {\n\tc.Emitter.Emit(c.Tag, Info, message, fields...)\n}", "func (c Client) ProcessMessage(message []byte) error {\n\tvar (\n\t\tmsg ws.BaseDataResponse\n\t)\n\n\terr := json.Unmarshal(message, &msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Order book data\n\tif strings.HasPrefix(msg.Topic, OrderBook) {\n\t\tvar symbol string\n\n\t\t// Order book with a depth of 25 orders per side\n\t\tif strings.HasPrefix(msg.Topic, OrderBook25) {\n\t\t\tsymbol = strings.TrimLeft(msg.Topic, OrderBook25+\".\")\n\t\t}\n\t\t// Order book with a depth of 200 orders per side\n\t\tif strings.HasPrefix(msg.Topic, OrderBook200) {\n\t\t\tsymbol = strings.TrimLeft(msg.Topic, OrderBook200+\".\")\n\t\t}\n\n\t\tswitch msg.Type {\n\t\tcase Snapshot:\n\t\t\tif c.OnOrderBookSnapshot != nil{\n\t\t\t\tvar data *ws.OrderBookSnapshot\n\t\t\t\terr = json.Unmarshal(msg.Data.([]byte), &data)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.OnOrderBookSnapshot(symbol, data.Standard())\n\t\t\t}\n\t\t\tbreak\n\t\tcase Delta:\n\t\t\tif c.OnOrderBookDelta != nil {\n\t\t\t\tvar data *ws.OrderBookDelta\n\t\t\t\terr = json.Unmarshal(msg.Data.([]byte), &data)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.OnOrderBookDelta(symbol, data.Standard())\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif strings.HasPrefix(msg.Topic, Instrument) {\n\t\tsymbol := strings.TrimLeft(msg.Topic, Instrument)\n\n\t\tswitch msg.Type {\n\t\tcase Snapshot:\n\t\t\tif c.OnTicker != nil {\n\t\t\t\tvar data *ws.InstrumentSnapshot\n\t\t\t\terr = json.Unmarshal(msg.Data.([]byte), &data)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.instruments[symbol] = data.Instrument()\n\t\t\t\tc.OnTicker(symbol, data.Standard())\n\t\t\t}\n\t\t\tbreak\n\t\tcase Delta:\n\t\t\tif c.OnTicker != nil {\n\t\t\t\tvar data *ws.InstrumentDelta\n\t\t\t\terr = json.Unmarshal(msg.Data.([]byte), &data)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.instruments[symbol].Update(data)\n\t\t\t\tc.OnTicker(symbol, c.instruments[symbol].Standard())\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif strings.HasPrefix(msg.Topic, Trades) {\n\t\tif c.OnTrades != nil {\n\t\t\tsymbol := strings.TrimLeft(msg.Topic, Trades)\n\t\t\tvar data *ws.Trades\n\t\t\terr = json.Unmarshal(msg.Data.([]byte), &data)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tc.OnTrades(symbol, data.Standard())\n\t\t}\n\t}\n\n\treturn nil\n}", "func (x *internalXMPP) dispatchIncoming(dying chan<- struct{}) {\n\tfor {\n\t\t// The xml.StartElement tells us what is coming up.\n\t\tstartElement, err := readStartElement(x.xmlDecoder)\n\t\tif err != nil {\n\t\t\tif isXMLErrorClosedConnection(err) {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tlog.Errorf(\"Failed to read the next start element: %s\", err)\n\t\t\tbreak\n\t\t}\n\n\t\t// Parse the message.\n\t\tif startElement.Name.Local == \"message\" {\n\t\t\tvar message struct {\n\t\t\t\tXMLName xml.Name `xml:\"message\"`\n\t\t\t\tData string `xml:\"push>data\"`\n\t\t\t}\n\n\t\t\tif err := x.xmlDecoder.DecodeElement(&message, startElement); err != nil {\n\t\t\t\tif isXMLErrorClosedConnection(err) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tlog.Warningf(\"Error while parsing print jobs notification via XMPP: %s\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tmessageData, err := base64.StdEncoding.DecodeString(message.Data)\n\t\t\tif err != nil {\n\t\t\t\tlog.Warningf(\"Failed to convert XMPP message data from base64: %s\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tmessageDataString := string(messageData)\n\t\t\tif strings.ContainsRune(messageDataString, '/') {\n\t\t\t\tif strings.HasSuffix(messageDataString, \"/delete\") {\n\t\t\t\t\tgcpID := strings.TrimSuffix(messageDataString, \"/delete\")\n\t\t\t\t\tx.notifications <- PrinterNotification{gcpID, PrinterDelete}\n\t\t\t\t}\n\t\t\t\t// Ignore other suffixes, like /update_settings.\n\t\t\t} else {\n\t\t\t\tx.notifications <- PrinterNotification{messageDataString, PrinterNewJobs}\n\t\t\t}\n\n\t\t} else if startElement.Name.Local == \"iq\" {\n\t\t\tvar message struct {\n\t\t\t\tXMLName xml.Name `xml:\"iq\"`\n\t\t\t\tID string `xml:\"id,attr\"`\n\t\t\t\tType string `xml:\"type,attr\"`\n\t\t\t}\n\n\t\t\tif err := x.xmlDecoder.DecodeElement(&message, startElement); err != nil {\n\t\t\t\tif isXMLErrorClosedConnection(err) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tlog.Warningf(\"Error while parsing XMPP pong: %s\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tpingID, err := strconv.ParseUint(message.ID, 10, 8)\n\t\t\tif err != nil {\n\t\t\t\tlog.Warningf(\"Failed to convert XMPP ping ID: %s\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tx.pongs <- uint8(pingID)\n\n\t\t} else {\n\t\t\tlog.Warningf(\"Unexpected element while waiting for print message: %+v\", startElement)\n\t\t}\n\t}\n\n\tdying <- struct{}{}\n}", "func messageCreate(s *discordgo.Session, m *discordgo.MessageCreate) {\n\n\t// Ignore all messages created by the bot itself\n\t// This isn't required in this specific example but it's a good practice.\n\tif m.Author.ID == s.State.User.ID {\n\t\treturn\n\t}\n\n\tmsg := strings.ToLower(m.Content)\n\n\tfmt.Printf(\"%s: %s\\n\", m.Author.ID, msg)\n\n\tif strings.Contains(msg, \"alice\") || rand.Intn(1000) == 173 {\n\t\tmsg = strings.TrimPrefix(msg, \"alice\")\n\t\tmsg = strings.TrimLeft(msg, \":, \")\n\n\t\tfd, err := net.Dial(\"unix\", \"/tmp/alice\")\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"unix socket connect failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tbuf := m.Author.ID + \"\\007\" + msg\n\t\t_, err = fd.Write([]byte(buf))\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"write failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tout := make([]byte, 8192)\n\t\tn, err := fd.Read(out)\n\t\tif err != nil {\n\t\t\tfmt.Printf(\"read failed\\n\")\n\t\t\treturn\n\t\t}\n\n\t\tfmt.Printf(\"me: %s\\n\", string(out[:n]))\n\n\t\ts.ChannelMessageSend(m.ChannelID, string(out[:n]))\n\n\t\tfd.Close()\n\t}\n}", "func (s *Sourcer) SetMeta(message *rtmp.Message) error {\n\tdecoder := amf.NewDecoder()\n\treader := bytes.NewReader(message.Payload)\n\tl := reader.Len()\n\tfor {\n\t\tv, err := decoder.DecodeAmf0(reader)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif str := v.(string); str != \"@setDataFrame\" {\n\t\t\tmeta := message.Payload[int(message.Header.PayloadLength)-l:]\n\t\t\ts.metaHead = newMsg(message)\n\t\t\ts.metaHead.Payload = meta\n\t\t\ts.metaHead.Header.PayloadLength = uint32(len(meta))\n\t\t\treturn nil\n\t\t}\n\t\tl = reader.Len()\n\t}\n}", "func (m *AttachmentItem) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func messageFromEvent(event accounts.MessageEvent) *Message {\n\tp := getMessage(event.Post.MessageID)\n\tif p == nil {\n\t\tp = NewMessage(nil)\n\t}\n\n\tp.Forward = event.Forward\n\tp.Mention = event.Mention\n\tp.Like = event.Like\n\tp.Followed = event.Followed\n\tp.Reply = event.Reply\n\n\tif event.Post.MessageID != \"\" {\n\t\tp.MessageID = event.Post.MessageID\n\t\tp.PostID = event.Post.PostID\n\t\tp.PostURL = event.Post.URL\n\t\tp.Name = event.Post.AuthorName\n\t\tp.Author = event.Post.Author\n\t\tp.AuthorURL = event.Post.AuthorURL\n\t\tp.AuthorID = event.Post.AuthorID\n\t\tp.Avatar = event.Post.Avatar\n\t\tp.Body = strings.TrimSpace(event.Post.Body)\n\t\tp.Sensitive = event.Post.Sensitive\n\t\tp.Warning = event.Post.Warning\n\t\tp.CreatedAt = event.Post.CreatedAt\n\t\tp.ReplyToID = event.Post.ReplyToID\n\t\tp.ReplyToAuthor = event.Post.ReplyToAuthor\n\t\tp.Actor = event.Post.Actor\n\t\tp.ActorName = event.Post.ActorName\n\t\tp.ActorID = event.Post.ActorID\n\t\tp.Liked = event.Post.Liked\n\t\tp.Shared = event.Post.Shared\n\t\tp.RepliesCount = event.Post.RepliesCount\n\t\tp.SharesCount = event.Post.SharesCount\n\t\tp.LikesCount = event.Post.LikesCount\n\t\tp.Visibility = event.Post.Visibility\n\n\t\t// parse attachments\n\t\tp.MediaPreview = []string{}\n\t\tp.MediaURL = []string{}\n\t\tfor _, v := range event.Media {\n\t\t\tp.MediaPreview = append(p.MediaPreview, v.Preview)\n\t\t\tp.MediaURL = append(p.MediaURL, v.URL)\n\t\t}\n\n\t\tp.MentionIDs = []string{}\n\t\tp.MentionNames = []string{}\n\t\tfor _, v := range event.Post.Mentions {\n\t\t\tp.MentionIDs = append(p.MentionIDs, v.ID)\n\t\t\tp.MentionNames = append(p.MentionNames, v.Name)\n\t\t}\n\t}\n\n\tif event.Followed {\n\t\tp.MessageID = event.Follow.Account\n\t\tp.Actor = event.Follow.Account\n\t\tp.ActorName = event.Follow.Name\n\t\tp.Avatar = event.Follow.Avatar\n\t\tp.AuthorURL = event.Follow.ProfileURL\n\t\tp.AuthorID = event.Follow.ProfileID\n\t\tp.Following = event.Follow.Following\n\t\tp.FollowedBy = event.Follow.FollowedBy\n\t}\n\n\tif p.MessageID == \"\" {\n\t\tspw := &spew.ConfigState{Indent: \" \", DisableCapacities: true, DisablePointerAddresses: true}\n\t\tlog.Println(\"Invalid message received:\", spw.Sdump(event))\n\t}\n\treturn p\n}", "func WrapWithMetadataPIID(msgMap ...service.DIDCommMsg) {\n\tpiID := uuid.New().String()\n\n\tfor _, msg := range msgMap {\n\t\tmsg.Metadata()[metaPIID] = piID\n\t}\n}", "func (mh *MessageHandler) OnInv(p *peer.Peer, msg *protocol.MsgInv) {\n\tlogx.DevPrintf(\"messageHandler OnInv peer:%v from:%v invs:%+v\", p.GetListenAddr(), msg.AddrFrom, msg.InvList[0])\n\t//check existing\n}", "func (h Response) Tell(message string) {\n\th.emit(\":tell\", strings.Replace(strings.Replace(message, \"\\n\", \" \", -1), \"\\t\", \"\", -1))\n}", "func (c *jsiiProxy_CfnConnection) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (hub *Hub) ConsumeMessage(msgType string, bz []byte) {\n\thub.preHandleNewHeightInfo(msgType, bz)\n\tif hub.skipHeight {\n\t\treturn\n\t}\n\thub.recordMsg(msgType, bz)\n\tif !hub.isTimeToHandleMsg(msgType) {\n\t\treturn\n\t}\n\tif hub.skipToOldChain(msgType) {\n\t\treturn\n\t}\n\thub.handleMsg()\n}", "func (*GenericFramework) NewMessage(ctx *MessageContext) {}", "func (m *Manager) incrementReceived() {\n\tm.received.ch <- m.received.count.Inc()\n}", "func (config *MessageConfiguration) AsIncoming() *IncomingMessageConfiguration {\n\treturn config.endpoint.createOrGetIncomingMessageConfig(config)\n}", "func (p *InvoiceUpcomingLinesInvoiceItemParams) AddMetadata(key string, value string) {\n\tif p.Metadata == nil {\n\t\tp.Metadata = make(map[string]string)\n\t}\n\n\tp.Metadata[key] = value\n}", "func (s *BasemumpsListener) EnterRead_(ctx *Read_Context) {}", "func (s *Server) OnMessage(msg *Message, sess *Session) {\n\tmsg.SetHeader(proto.Sender, sess.ID)\n\tmsg.SetHeader(proto.Host, s.ServerAddress.Address)\n\tif msg.Id() == \"\" {\n\t\tmsg.SetId(uuid())\n\t}\n\tif msg.Cmd() != proto.Heartbeat {\n\t\t//log.Printf(msg.String())\n\t}\n\n\thandleUrlMessage(msg)\n\n\tcmd := msg.Cmd()\n\thandler, ok := s.handlerTable[cmd]\n\tif ok {\n\t\thandler(s, msg, sess)\n\t\treturn\n\t}\n\tres := NewMessageStatus(400, \"Bad format: command(%s) not support\", cmd)\n\tsess.WriteMessage(res)\n}", "func (m *PlayPromptPostRequestBody) SetAdditionalData(value map[string]interface{})() {\n m.additionalData = value\n}", "func (s *DnsServer) HandleIncoming(writer dns.ResponseWriter, incoming *dns.Msg) {\n\tm := &dns.Msg{}\n\tm.SetReply(incoming)\n\tm.Compress = false\n\tm.RecursionAvailable = true\n\tswitch incoming.Opcode {\n\tcase dns.OpcodeQuery:\n\t\ts.ParseDNS(m)\n\t}\n\terr := writer.WriteMsg(m)\n\tif err != nil {\n\t\ts.Logger.Errorf(\"%s\", err)\n\t}\n}", "func consumeUnicast(payload []byte) {\n\n\tmessage := networking.CalculationMessage{}\n\n\terr := json.Unmarshal(payload, &message)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif message.IsProbe {\n\t\tprobeEcho.Probe <- probeEcho.ProbeMessage{\n\t\t\tCalculationId: message.CalculationId,\n\t\t\tParent: message.Emitter,\n\t\t\tCandidate: message.Candidate,\n\t\t}\n\t} else {\n\t\tprobeEcho.Echo <- probeEcho.EchoMessage{\n\t\t\tCalculationId: message.CalculationId,\n\t\t\tMayBePrime: message.MayBePrime,\n\t\t}\n\t}\n}", "func addFields(ctx context.Context, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler, span *trace.Span) {\n\thandlerName := runtime.FuncForPC(reflect.ValueOf(handler).Pointer()).Name()\n\n\tspan.AddField(\"name\", handlerName)\n\tspan.AddField(\"meta.type\", \"grpc_request\")\n\tspan.AddField(\"handler.name\", handlerName)\n\tspan.AddField(\"handler.method\", info.FullMethod)\n\n\tpr, ok := peer.FromContext(ctx)\n\tif ok {\n\t\t// if we have an address, put it on the span\n\t\tif pr.Addr != net.Addr(nil) {\n\t\t\tspan.AddField(\"request.remote_addr\", pr.Addr.String())\n\t\t}\n\t}\n\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif ok {\n\t\tfor headerName, fieldName := range headersToFields {\n\t\t\tif val, ok := md[headerName]; ok {\n\t\t\t\tspan.AddField(fieldName, val[0])\n\t\t\t}\n\t\t}\n\t}\n}", "func prepareMessage(headerType uint16, cmd uint8, attributes []byte) (msg netlinkMessage) {\n\tmsg.Header.Type = headerType\n\tmsg.Header.Flags = syscall.NLM_F_REQUEST\n\tmsg.GenHeader.Command = cmd\n\tmsg.GenHeader.Version = 0x1\n\tmsg.Data = attributes\n\treturn msg\n}", "func (c *jsiiProxy_CfnApplication) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (c *jsiiProxy_CfnApplication) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (f *framer) payload() {\n\tf.flags |= flagCustomPayload\n}", "func (cr *CommandResponse) AddMeta(name, value string) {\n\tcr.Metadata = append(cr.Metadata, CommandResponseMetadata{\n\t\tName: name,\n\t\tValue: value,\n\t})\n}", "func mergeOutgoingMetadata(ctx context.Context, mds ...metadata.MD) context.Context {\n\t// There may not be metadata in the context, only insert the existing\n\t// metadata if it exists (ok).\n\tctxMD, ok := metadata.FromOutgoingContext(ctx)\n\tif ok {\n\t\t// The ordering matters, hence why ctxMD is added to the front.\n\t\tmds = append([]metadata.MD{ctxMD}, mds...)\n\t}\n\n\treturn metadata.NewOutgoingContext(ctx, metadata.Join(mds...))\n}", "func incoming(w http.ResponseWriter, r *http.Request) {\n\tbody := r.PostFormValue(\"Body\")\n\tfrom := r.PostFormValue(\"From\")\n\n\tif body == \"\" || from == \"\" {\n\t\thttp.Error(w, \"Missing required key\", http.StatusInternalServerError)\n\t\tdata, _ := ioutil.ReadAll(r.Body)\n\t\tlog.Printf(\"Unable to read expected fields from %v\", string(data))\n\t\treturn\n\t}\n\n\tdb := context.Get(r, \"db\").(*mgo.Session)\n\ttodo, err := todoListFor(from, db)\n\n\tif err != nil {\n\t\tlog.Printf(\"Error getting todolist: %v\", err)\n\t}\n\n\tif todo == nil {\n\t\tlog.Printf(\"Didn't pull up a todo list, creating a new one...\")\n\t\ttodo, err = createBlankTodoList(from, \"Basic Todo\", db)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Creating a new blank todo didn't work right: %v\", err)\n\t\t}\n\t}\n\n\tlog.Printf(\"Working with %v\", todo)\n\ttodo.TodoItems = append(todo.TodoItems, TodoItem{Title: body})\n\tlog.Printf(\"TODO: need to save %v\", todo)\n\t// c.UpdateId(todo.Id, bson.M{\"$set\": bson.M{\"name\": \"updated name\"}})\n\n\tmessage := SmsResponse{XMLName: xml.Name{Local: \"Response\"}, Message: \"Thank you, I got it.\"}\n\tx, err := xml.MarshalIndent(message, \"\", \" \")\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/xml\")\n\tw.Write(x)\n}", "func (b *Buffer) updateMeta() {\n\t// First 8 bytes have the first frame offset,\n\t// next 8 bytes have the last frame offset,\n\t// next 8 bytes are the next sequence number,\n\t// next 4 bytes are the biggest data record we've seen,\n\t// next 8 bytes are the total data in the buffer.\n\toff := int(b.capacity)\n\tbinary.PutLittleEndianUint64(b.data, off, b.first)\n\tbinary.PutLittleEndianUint64(b.data, off+8, b.last)\n\tbinary.PutLittleEndianUint64(b.data, off+16, b.nextSeq)\n\tbinary.PutLittleEndianUint32(b.data, off+24, b.biggest)\n\tbinary.PutLittleEndianUint64(b.data, off+28, b.length)\n}", "func (m *ChatMessageAttachment) GetAdditionalData()(map[string]any) {\n val , err := m.backingStore.Get(\"additionalData\")\n if err != nil {\n panic(err)\n }\n if val == nil {\n var value = make(map[string]any);\n m.SetAdditionalData(value);\n }\n return val.(map[string]any)\n}", "func (c *jsiiProxy_CfnContact) AddMetadata(key *string, value interface{}) {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"addMetadata\",\n\t\t[]interface{}{key, value},\n\t)\n}", "func (s *Stream) Provide(data []byte) {\n\ts.recv <- data\n}", "func (env *Envelope) AddReceivedLine(peer Peer) {\n\n\ttlsDetails := \"\"\n\n\ttlsVersions := map[uint16]string{\n\t\ttls.VersionSSL30: \"SSL3.0\",\n\t\ttls.VersionTLS10: \"TLS1.0\",\n\t\ttls.VersionTLS11: \"TLS1.1\",\n\t\ttls.VersionTLS12: \"TLS1.2\",\n\t\ttls.VersionTLS13: \"TLS1.3\",\n\t}\n\n\tif peer.TLS != nil {\n\t\tversion := \"unknown\"\n\n\t\tif val, ok := tlsVersions[peer.TLS.Version]; ok {\n\t\t\tversion = val\n\t\t}\n\n\t\tcipher := tls.CipherSuiteName(peer.TLS.CipherSuite)\n\n\t\ttlsDetails = fmt.Sprintf(\n\t\t\t\"\\r\\n\\t(version=%s cipher=%s);\",\n\t\t\tversion,\n\t\t\tcipher,\n\t\t)\n\t}\n\n\tpeerIP := \"\"\n\tif addr, ok := peer.Addr.(*net.TCPAddr); ok {\n\t\tpeerIP = addr.IP.String()\n\t}\n\n\tline := wrap([]byte(fmt.Sprintf(\n\t\t\"Received: from %s ([%s]) by %s with %s;%s\\r\\n\\t%s\\r\\n\",\n\t\tpeer.HeloName,\n\t\tpeerIP,\n\t\tpeer.ServerName,\n\t\tpeer.Protocol,\n\t\ttlsDetails,\n\t\ttime.Now().Format(\"Mon, 02 Jan 2006 15:04:05 -0700 (MST)\"),\n\t)))\n\n\tenv.Data = append(env.Data, line...)\n\n\t// Move the new Received line up front\n\n\tcopy(env.Data[len(line):], env.Data[0:len(env.Data)-len(line)])\n\tcopy(env.Data, line)\n\n}", "func (l DefaultBlazeListener) OnMessage(ctx context.Context, msg MessageView, userId string) error {\n\tlog.Println(\"I got a message: \", msg)\n\tdata, err := base64.StdEncoding.DecodeString(msg.Data)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif msg.Category == \"SYSTEM_ACCOUNT_SNAPSHOT\" {\n\t\tvar transfer TransferView\n\t\tif err := json.Unmarshal(data, &transfer); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tlog.Println(\"I got a snapshot: \", transfer)\n\t\treturn nil\n\t} else {\n\t\tlog.Printf(\"I got a message, it said: %s\", string(data))\n\t\treturn nil\n\t}\n}", "func (s *Server) Forward(id int64, event api.Event) {\n\tif event.Type == \"logging\" {\n\t\t// Parse the message\n\t\tlogEntry := api.EventLogging{}\n\t\terr := json.Unmarshal(event.Metadata, &logEntry)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tif !s.debug && logEntry.Level == \"dbug\" {\n\t\t\treturn\n\t\t}\n\n\t\tif !s.debug && !s.verbose && logEntry.Level == \"info\" {\n\t\t\treturn\n\t\t}\n\t}\n\n\terr := s.broadcast(event, true)\n\tif err != nil {\n\t\tlogger.Warnf(\"Failed to forward event from member %d: %v\", id, err)\n\t}\n}", "func (q *queue) persistMetaOfMessage(dataPageIndex int64, dataLen, messageOffset int) error {\n\tq.rwMutex.Lock()\n\tdefer q.rwMutex.Unlock()\n\n\tseq := q.appendedSeq.Load() + 1 // append sequence\n\tindexPageIndex := seq / indexItemsPerPage\n\tif indexPageIndex != q.indexPageIndex {\n\t\t// check size limit before index page acquire\n\t\tif err := q.checkDataSize(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// sync previous data page\n\t\tif err := q.indexPage.Sync(); err != nil {\n\t\t\tqueueLogger.Error(\"sync index page err when alloc\",\n\t\t\t\tlogger.String(\"queue\", q.dirPath), logger.Error(err))\n\t\t}\n\t\tindexPage, err := q.indexPageFct.AcquirePage(indexPageIndex)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tq.indexPage = indexPage\n\t\tq.indexPageIndex = indexPageIndex\n\t}\n\n\t// save index data\n\tindexOffset := int((seq % indexItemsPerPage) * indexItemLength)\n\tq.indexPage.PutUint64(uint64(dataPageIndex), indexOffset+queueDataPageIndexOffset)\n\tq.indexPage.PutUint32(uint32(messageOffset), indexOffset+messageOffsetOffset)\n\tq.indexPage.PutUint32(uint32(dataLen), indexOffset+messageLengthOffset)\n\n\t// save metadata\n\tq.metaPage.PutUint64(uint64(seq), queueAppendedSeqOffset)\n\tq.appendedSeq.Store(seq)\n\n\t// new data written, notify all waiting consumer groups can consume data\n\tq.notEmpty.Broadcast()\n\treturn nil\n}" ]
[ "0.5766339", "0.56079274", "0.5553875", "0.5425873", "0.534128", "0.5265968", "0.52381384", "0.52301913", "0.5211246", "0.5147059", "0.5129772", "0.512165", "0.5110792", "0.5100262", "0.5079197", "0.50734866", "0.500004", "0.4982307", "0.49607491", "0.49395782", "0.49265683", "0.491974", "0.4910204", "0.49096584", "0.49091044", "0.48830926", "0.48778355", "0.48763633", "0.48624334", "0.48485085", "0.4848206", "0.4836118", "0.48216495", "0.48126385", "0.4809134", "0.48048103", "0.4804378", "0.47969985", "0.47884434", "0.47834158", "0.47830978", "0.47713885", "0.4761541", "0.4756256", "0.47541687", "0.47472104", "0.4739717", "0.47119442", "0.4710531", "0.47076857", "0.4694536", "0.46835667", "0.46787682", "0.46675798", "0.46671197", "0.46660185", "0.46636283", "0.4663403", "0.46622542", "0.4647299", "0.4641702", "0.46359405", "0.4634692", "0.463121", "0.46304494", "0.46246886", "0.46212402", "0.46103916", "0.46083802", "0.4597141", "0.4596271", "0.45930567", "0.4590613", "0.458475", "0.45797905", "0.45764846", "0.4574528", "0.4560397", "0.45544422", "0.45541292", "0.45450562", "0.45445287", "0.45431718", "0.45400798", "0.45393312", "0.45382097", "0.45334974", "0.45334974", "0.45213112", "0.45204243", "0.45182502", "0.45170733", "0.45119333", "0.45047185", "0.4504148", "0.45032418", "0.45023623", "0.4497804", "0.4489987", "0.4486618" ]
0.5511433
3
Registered is called after an extension has been successfully registered
func (e *Extension) Registered(extensionName string, client *bayeux.BayeuxClient) { }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *Resolver) Register(name string, e Extension) bool {\n\tr.Execers[name] = e\n\treturn true\n}", "func (e *extensionRegistry) registerExtension(xd protoreflect.ExtensionDescriptor) error {\n\tif _, err := e.FindExtensionByName(xd.FullName()); err != protoregistry.NotFound {\n\t\t// Either the extension already exists or there was an error, either way we're done.\n\t\treturn err\n\t}\n\treturn e.local.RegisterExtension(dynamicpb.NewExtensionType(xd))\n}", "func (e *Extension) Unregistered() {\n}", "func (e *Client) Register(ctx context.Context, filename string) (*RegisterResponse, error) {\n\tconst action = \"/register\"\n\turl := e.baseURL + action\n\n\treqBody, err := json.Marshal(map[string]interface{}{\n\t\t\"events\": []EventType{Invoke, Shutdown},\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\thttpReq, err := http.NewRequestWithContext(ctx, \"POST\", url, bytes.NewBuffer(reqBody))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\thttpReq.Header.Set(extensionNameHeader, filename)\n\thttpRes, err := e.httpClient.Do(httpReq)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif httpRes.StatusCode != 200 {\n\t\treturn nil, fmt.Errorf(\"request failed with status %s\", httpRes.Status)\n\t}\n\tdefer httpRes.Body.Close()\n\tbody, err := ioutil.ReadAll(httpRes.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres := RegisterResponse{}\n\terr = json.Unmarshal(body, &res)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\te.ExtensionID = httpRes.Header.Get(extensionIdentiferHeader)\n\tfmt.Println(\"Extension id:\", e.ExtensionID)\n\treturn &res, nil\n}", "func (l *PluginLayer) Register(mw *layer.Layer) {\n\tmw.Use(\"error\", l.Run)\n\tmw.Use(\"request\", l.Run)\n}", "func RVExtensionRegisterCallback(cbptr unsafe.Pointer) {\n\tcb = C.callbackProc(cbptr)\n\n\tlog.Println(\"Calling callback function ……\")\n\tfunction := C.CString(\"registered\")\n\tdefer C.free(unsafe.Pointer(function))\n\tC.bridge_cb(cb, name, function, function)\n}", "func RegisterExtension(name string, plugin Plugin) {\n\tpluginRegistryMu.Lock()\n\tdefer pluginRegistryMu.Unlock()\n\tif plugin == nil {\n\t\tpanic(\"nil plugin passed to RegisterExtension\")\n\t}\n\tif _, ok := pluginRegistry[name]; ok {\n\t\tpanic(\"duplicate RegisterExtension name: \" + name)\n\t}\n\tpluginRegistry[name] = plugin\n}", "func (b *bot) Register(p Plugin, kind Kind, cb Callback) {\n\tr := regexp.MustCompile(`.*`)\n\tresp := func(r Request) bool {\n\t\treturn cb(r.Conn, r.Kind, r.Msg, r.Args...)\n\t}\n\tb.RegisterRegex(p, kind, r, resp)\n}", "func (c *Compiler) RegisterExtension(name string, meta *Schema, ext ExtCompiler) {\n\tc.extensions[name] = extension{meta, ext}\n}", "func Register(ext *sqlite.ExtensionApi, _ *options.Options) (_ sqlite.ErrorCode, err error) {\n\tvar fns = map[string]sqlite.Function{\n\t\t\"str_split\": &StringSplit{},\n\t\t\"toml_to_json\": &TomlToJson{},\n\t\t\"yaml_to_json\": &YamlToJson{},\n\t\t\"xml_to_json\": &XmlToJson{},\n\t}\n\n\t// alias yaml_to_json => yml_to_json\n\tfns[\"yml_to_json\"] = fns[\"yaml_to_json\"]\n\n\tfor name, fn := range fns {\n\t\tif err = ext.CreateFunction(name, fn); err != nil {\n\t\t\treturn sqlite.SQLITE_ERROR, errors.Wrapf(err, \"failed to register %q function\", name)\n\t\t}\n\t}\n\n\treturn sqlite.SQLITE_OK, nil\n}", "func (c *Component) Register() {}", "func (ec *ExtensionClient) RegisterExtension(ext *extension.Extension) (res *extension.Extension, err error) {\n\n\tresponse, err := ec.c.RegisterExtension()(context.Background(), ext)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn response.(*extension.Extension), nil\n}", "func (e *Client) Register(ctx context.Context, filename string) error {\n\tconst action = \"/register\"\n\turl := e.baseURL + action\n\n\treqBody, err := json.Marshal(map[string]interface{}{\n\t\t\"events\": []EventType{Shutdown},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\thttpReq, err := http.NewRequestWithContext(ctx, \"POST\", url, bytes.NewBuffer(reqBody))\n\tif err != nil {\n\t\treturn err\n\t}\n\thttpReq.Header.Set(extensionNameHeader, filename)\n\thttpRes, err := e.httpClient.Do(httpReq)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif httpRes.StatusCode != 200 {\n\t\treturn fmt.Errorf(\"request failed with status %s\", httpRes.Status)\n\t}\n\tdefer httpRes.Body.Close()\n\te.extensionID = httpRes.Header.Get(extensionIdentiferHeader)\n\treturn nil\n}", "func Register() map[string]string {\n\treturn map[string]string{\n\t\t\"name\": \"Beubo Example Plugin\",\n\t\t// identifier should be a unique identifier used to differentiate this plugin from other plugins\n\t\t\"identifier\": \"beubo_example_plugin\",\n\t}\n}", "func (p *Processor) RegisterExtension(id string, extension ProcessorExtension) {\n\tp.extensions[id] = extension\n}", "func register(name string, p Plugin) {\n\tdirectory[name] = p\n}", "func Register(ext *sqlite.ExtensionApi, opt *options.Options) (_ sqlite.ErrorCode, err error) {\n\t// register virtual table modules\n\tvar modules = map[string]sqlite.Module{\n\t\t\"commits\": &LogModule{Locator: opt.Locator, Context: opt.Context},\n\t\t\"refs\": &RefModule{Locator: opt.Locator, Context: opt.Context},\n\t\t\"stats\": native.NewStatsModule(opt.Locator, opt.Context),\n\t\t\"files\": native.NewFilesModule(opt.Locator, opt.Context),\n\t\t\"blame\": native.NewBlameModule(opt.Locator, opt.Context),\n\t}\n\n\tfor name, mod := range modules {\n\t\tif err = ext.CreateModule(name, mod); err != nil {\n\t\t\treturn sqlite.SQLITE_ERROR, errors.Wrapf(err, \"failed to register %q module\", name)\n\t\t}\n\t}\n\n\tvar fns = map[string]sqlite.Function{\n\t\t\"commit_from_tag\": &CommitFromTagFn{},\n\t}\n\n\tfor name, fn := range fns {\n\t\tif err = ext.CreateFunction(name, fn); err != nil {\n\t\t\treturn sqlite.SQLITE_ERROR, errors.Wrapf(err, \"failed to register %q function\", name)\n\t\t}\n\t}\n\n\treturn sqlite.SQLITE_OK, nil\n}", "func (proxy *OAuthProxy) registerExtensions(openshiftCAs StringArray) {\n\t//fill in\n\tproxy.RegisterRequestHandlers(cl.NewHandlers(openshiftCAs))\n}", "func (s *eremeticScheduler) Registered(driver sched.SchedulerDriver, frameworkID *mesos.FrameworkID, masterInfo *mesos.MasterInfo) {\n\tlog.Debugf(\"Framework %s registered with master %s\", frameworkID.GetValue(), masterInfo.GetHostname())\n\tif !s.initialised {\n\t\tdriver.ReconcileTasks([]*mesos.TaskStatus{})\n\t\ts.initialised = true\n\t} else {\n\t\ts.Reconcile(driver)\n\t}\n}", "func Register(plugin Plugin) {\n\tplugins = append(plugins, plugin)\n}", "func (sched *Scheduler) Registered(driver sched.SchedulerDriver, frameworkId *mesos.FrameworkID, masterInfo *mesos.MasterInfo) {\n\tsched.master = MasterConnStr(masterInfo)\n\tlog.Println(\"Taurus Framework Registered with Master\", sched.master)\n\t// Start the scheduler worker\n\tgo func() {\n\t\tlog.Printf(\"Starting %s framework scheduler worker\", FrameworkName)\n\t\tsched.errChan <- sched.Worker.Start(driver, masterInfo)\n\t}()\n}", "func Register() map[string]string {\n\treturn map[string]string{\n\t\t\"name\": \"Beubo gRPC\",\n\t\t// identifier should be a unique identifier used to differentiate this plugin from other plugins\n\t\t\"identifier\": \"beubo_grpc\",\n\t}\n}", "func registerAddon(addon kubernetes.Addon, addonType AddOnType, addonTemplater addonTemplater, preflightAddonTemplater preflightAddonTemplater, callbacks addonCallbacks, addonPriority addonPriority, getImageCallbacks []getImageCallback) {\n\tAddons[addon] = Addon{\n\t\tAddon: addon,\n\t\ttemplater: addonTemplater,\n\t\tpreflightTemplater: preflightAddonTemplater,\n\t\tcallbacks: callbacks,\n\t\taddonPriority: addonPriority,\n\t\tgetImageCallbacks: getImageCallbacks,\n\t\tAddOnType: addonType,\n\t}\n}", "func (self *Mediator) OnRegister() {\n\n}", "func (m *Manager) Register(name string, plugin Plugin) {\n\tm.mtx.Lock()\n\tdefer m.mtx.Unlock()\n\tm.plugins = append(m.plugins, namedplugin{\n\t\tname: name,\n\t\tplugin: plugin,\n\t})\n}", "func RegisterExtensions(sys *gohome.System) error {\n\tlog.V(\"registering extensions\")\n\n\tlog.V(\"register extension - belkin\")\n\tsys.Extensions.Register(belkin.NewExtension())\n\n\tlog.V(\"register extension - connectedbytcp\")\n\tsys.Extensions.Register(connectedbytcp.NewExtension())\n\n\tlog.V(\"register extension - fluxwifi\")\n\tsys.Extensions.Register(fluxwifi.NewExtension())\n\n\tlog.V(\"register extension - honeywell\")\n\tsys.Extensions.Register(honeywell.NewExtension())\n\n\tlog.V(\"register extension - lutron\")\n\tsys.Extensions.Register(lutron.NewExtension())\n\n\t/*\n\t\t// An example piece of hardware\n\t\tlog.V(\"register extension - example\")\n\t\tsys.Extensions.Register(example.NewExtension())\n\t*/\n\n\t//Uncomment for testing\n\tlog.V(\"register extension - testing\")\n\tsys.Extensions.Register(testing.NewExtension())\n\n\treturn nil\n}", "func RegisterExtensionModule(nsURL string, m Module) {\n\tns := strings.TrimPrefix(strings.TrimPrefix(nsURL, \"http://\"), \"https://\")\n\tb := modules[:0]\n\tfor _, v := range modules {\n\t\tif lookupModule(ns) == nil {\n\t\t\tb = append(b, v)\n\t\t}\n\t}\n\tb = append(b, module{ns, m})\n\tmodules = b\n}", "func RegisterExt(ext *ExtFormat) error {\n\tif ext == nil {\n\t\treturn fmt.Errorf(\"ext pointer is nil\")\n\t} else if ext.DecodeFunc == nil {\n\t\treturn fmt.Errorf(\"DecodeFunc is nil\")\n\t} else if !IsExt(ext.FirstByte) {\n\t\treturn fmt.Errorf(\"0x%02x is not ext type\", ext.FirstByte)\n\t}\n\n\textFormats[ext.FirstByte] = append(extFormats[ext.FirstByte], ext)\n\treturn nil\n}", "func Register(scheme string, b Broker) {\n\tbrokerRegistery[scheme] = b\n}", "func Register(ctx context.Context, init runtime.Initializer) error {\n\tw := &Walker{}\n\n\tif err := init.RegisterMatch(ModuleName, w.registerMatch); err != nil {\n\t\treturn err\n\t}\n\n\tif err := init.RegisterMatchmakerMatched(w.registerMatchmakerMatched); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Register() {\n\tglobalLock.Lock()\n\tdefer globalLock.Unlock()\n\tregister()\n}", "func RegisterExt(id int8, value interface{}) {\n\ttyp := reflect.TypeOf(value)\n\tif typ.Kind() == reflect.Ptr {\n\t\ttyp = typ.Elem()\n\t}\n\tptr := reflect.PtrTo(typ)\n\n\tif _, ok := extTypes[id]; ok {\n\t\tpanic(fmt.Errorf(\"msgpack: ext with id=%d is already registered\", id))\n\t}\n\n\tregisterExt(id, ptr, getEncoder(ptr), getDecoder(ptr))\n\tregisterExt(id, typ, getEncoder(typ), getDecoder(typ))\n}", "func (k *KubernetesExecutor) Registered(driver bindings.ExecutorDriver,\n\texecutorInfo *mesos.ExecutorInfo, frameworkInfo *mesos.FrameworkInfo, slaveInfo *mesos.SlaveInfo) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Executor %v of framework %v registered with slave %v\\n\",\n\t\texecutorInfo, frameworkInfo, slaveInfo)\n\tif !k.swapState(disconnectedState, connectedState) {\n\t\t//programming error?\n\t\tpanic(\"already connected?!\")\n\t}\n}", "func (p *JSONProtocol) Register(msg interface{}) {\n\tt := reflect.TypeOf(msg)\n\tif t.Kind() == reflect.Ptr {\n\t\tt = t.Elem()\n\t}\n\tname := t.PkgPath() + \"/\" + t.Name()\n\tp.types[name] = t\n\tp.names[t] = name\n}", "func (s *serverRegistry) Register(*FContext, FAsyncCallback) error {\n\treturn nil\n}", "func (k *KubernetesExecutor) Registered(driver bindings.ExecutorDriver,\n\texecutorInfo *mesos.ExecutorInfo, frameworkInfo *mesos.FrameworkInfo, slaveInfo *mesos.SlaveInfo) {\n\tif k.isDone() {\n\t\treturn\n\t}\n\tlog.Infof(\"Executor %v of framework %v registered with slave %v\\n\",\n\t\texecutorInfo, frameworkInfo, slaveInfo)\n\tif !(&k.state).transition(disconnectedState, connectedState) {\n\t\tlog.Errorf(\"failed to register/transition to a connected state\")\n\t}\n\n\tif executorInfo != nil && executorInfo.Data != nil {\n\t\tk.staticPodsConfig = executorInfo.Data\n\t}\n\n\tif slaveInfo != nil {\n\t\t_, err := node.CreateOrUpdate(k.client, slaveInfo.GetHostname(), node.SlaveAttributesToLabels(slaveInfo.Attributes))\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"cannot update node labels: %v\", err)\n\t\t}\n\t}\n\n\tk.initialRegistration.Do(k.onInitialRegistration)\n}", "func Register() {\n\tregisterMetrics.Do(func() {\n\t\tlegacyregistry.MustRegister(CronJobCreationSkew)\n\t})\n}", "func Register(proto interface{}, tag ...string) {\n\tRegisterName(\"\", proto, tag...)\n}", "func (f *FsWatcher) RegisterFileExtension(extension string, handler FsEventHandler) {\r\n\tf.handlers[extension] = handler\r\n}", "func (k *KubernetesScheduler) Registered(driver mesos.SchedulerDriver,\n\tframeworkId *mesos.FrameworkID, masterInfo *mesos.MasterInfo) {\n\tk.frameworkId = frameworkId\n\tk.masterInfo = masterInfo\n\tk.registered = true\n\tlog.Infof(\"Scheduler registered with the master: %v with frameworkId: %v\\n\", masterInfo, frameworkId)\n}", "func (reg *registrar) Register(example interface{}) error {\n\treg.lock.Lock()\n\tdefer reg.lock.Unlock()\n\treturn reg.Registry.Register(example)\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, initializer func() any)\n}) {\n\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, initializer func() any)\n}) {\n\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, initializer func() any)\n}) {\n\n}", "func Register(srv *grpc.Server) {\n\tglobal.Register(srv)\n}", "func RegisterHook(componentName string, hook func(settingName string, settingValue string)) {\n\tlog.Info().Msgf(\"Adding new hook for %s\", componentName)\n\thookList.mutex.Lock()\n\tdefer hookList.mutex.Unlock()\n\thookList.list[componentName] = append(hookList.list[componentName], hook)\n\thookList.count++\n}", "func (ct *ContentTypes) RegisterType(extension string, contentType ml.ContentType) {\n\t//check if there is type with such extension already, and if it's here then ignore\n\tfor _, def := range ct.ml.Defaults {\n\t\tif def.Extension == extension {\n\t\t\treturn\n\t\t}\n\t}\n\n\tct.ml.Defaults = append(ct.ml.Defaults, &ml.TypeDefault{\n\t\tExtension: extension,\n\t\tContentType: contentType,\n\t})\n\n\tct.file.MarkAsUpdated()\n}", "func (m *SecuritySchemeMutator) Extension(name string, value interface{}) *SecuritySchemeMutator {\n\tif m.proxy.extensions == nil {\n\t\tm.proxy.extensions = Extensions{}\n\t}\n\tm.proxy.extensions[name] = value\n\treturn m\n}", "func (wfw *impl) Register(p ComponentInterface) {\n\tif p == nil {\n\t\tif singleton.debug {\n\t\t\tlog.Debugf(\"Register workflow application nil component. Action missed\")\n\t\t}\n\t\treturn\n\t}\n\tif singleton.debug {\n\t\tlog.Debugf(\"Register workflow application component %q\", packageName(p))\n\t}\n\twfw.Components = append(wfw.Components, p)\n}", "func (h *Hook) Register() (*Hook, error) {\n\tvar err error\n\n\th.Registered = time.Now()\n\n\th.file, err = h.file.Set(h)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn h, nil\n}", "func (s *Server) Register(rcvr interface{}, metadata string) error {\n\tsname, err := s.register(rcvr, \"\", false)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn s.Plugins.DoRegister(sname, rcvr, metadata)\n}", "func (cli *CLI) Register(fun Function, name string, params []string, help string) {\n\tcli.commands = append(cli.commands, Command{fun, name, params, help})\n}", "func (a *Application) RegisterProvider() {\n\n}", "func Register(name string, plugin Plugin) {\n\n\tif Plugins[name].Exec != nil {\n\t\tlog.Fatalf(\"plugin \\\"%s\\\" already exists\", name)\n\t\treturn\n\t}\n\n\tPlugins[name] = plugin\n}", "func (inst *Instance) Register() error {\n\tpr := PluginRegistry()\n\tpr.mut.Lock()\n\tdefer pr.mut.Unlock()\n\n\t// default to the plugin's default if no RE was provided\n\tif inst.Regex == \"\" {\n\t\tinst.Regex = inst.Plugin.Regex\n\t}\n\t// TODO: the default regex still doesn't always show up\n\n\t// TODO: manually check/return the error so the bot doesn't crash\n\tinst.regex = regexp.MustCompile(inst.Regex)\n\n\t// call the instance init handler\n\tif inst.Plugin.Init != nil {\n\t\tinst.Plugin.Init(inst)\n\t}\n\n\t// once an instance is registered, the router will automatically\n\t// pick it up on the next message it processes\n\tpr.instances = append(pr.instances, inst)\n\n\tlog.Debugf(\"Registered plugin %q in room id %q on broker %q with RE match %q\",\n\t\tinst.Name, inst.RoomId, inst.Broker.Name(), inst.regex)\n\n\treturn nil\n}", "func Register(funcName string, backend TemplateFunc, buildFlags FlagsFunc) {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tbackends[funcName] = backend\n\tflags[funcName] = buildFlags\n}", "func (l *lifecycle) Register(h Hook) {\n\tl.hooks = append(l.hooks, h)\n}", "func (tqsc *Controller) Register() {\n}", "func RegisterPackage(pkgName string) {\n\tCorePackages = append(CorePackages, pkgName)\n}", "func init() {\n\tregister(\"jemoji\", jemojiPlugin{})\n\tregister(\"jekyll-mentions\", jekyllMentionsPlugin{})\n\tregister(\"jekyll-optional-front-matter\", jekyllOptionalFrontMatterPlugin{})\n\n\t// Gojekyll behaves as though the following plugins are always loaded.\n\t// Define them here so we don't see warnings that they aren't defined.\n\tregister(\"jekyll-live-reload\", plugin{})\n\tregister(\"jekyll-sass-converter\", plugin{})\n}", "func Register() {\n\n\t// Add module sub commands\n\tmoduleCmd.AddCommand(initCmd)\n\tmoduleCmd.AddCommand(commitCmd)\n\tmoduleCmd.AddCommand(prepareCmd)\n\tmoduleCmd.AddCommand(listModulesCmd)\n\tmoduleCmd.AddCommand(selectCmd)\n\n\t// Add event sub commands\n\teventCmd.AddCommand(listEventsCmd)\n\n\t// Add event sub commands\n\tdevCmd.AddCommand(moduleCmd)\n\tdevCmd.AddCommand(eventCmd)\n\n\t// Add event command to root\n\troot.RootCmd.AddCommand(devCmd)\n}", "func (e *EnterpriseEndpoints) Register(s *rpc.Server) {}", "func (r *Registrar) RegisterPlugin(info Info, cookie *string) error {\n\tregLog.Info(\"Received registration request. Signalling registrar.\")\n\n\treplyCh := make(chan regReply, 1)\n\tdefer close(replyCh)\n\tr.ch <- regRequest{Info: info, replyCh: replyCh}\n\n\tregLog.Info(\"Waiting for response from registrar...\")\n\treply := <-replyCh\n\n\tregLog.Info(\"Plugin \\\"%s\\\" registered with cookie \\\"%s\\\"\",\n\t\tinfo.Name, reply.cookie)\n\t(*cookie) = reply.cookie\n\treturn reply.err\n}", "func (c Class) Register() {\n\tobjc_registerClassPair(c)\n}", "func Register(cb func(string)) {\n\tcallbacks = append(callbacks, cb)\n}", "func Register(p Protocol, n NewFunc) {\n\treglock.Lock()\n\tdefer reglock.Unlock()\n\tregistry[p] = n\n}", "func (p *Plugin) Register() error {\n\tpr := PluginRegistry()\n\tpr.mut.Lock()\n\tdefer pr.mut.Unlock()\n\n\tfor _, plugin := range pr.plugins {\n\t\tif plugin.Name == p.Name {\n\t\t\tlog.Printf(\"Ignoring multiple calls to Register() for plugin '%s'\", p.Name)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tpr.plugins = append(pr.plugins, p)\n\n\treturn nil\n}", "func Register(u CliPipable) {\n\tallCliUtils[u.Name()] = u\n}", "func Register(name string, driver NewInstanceCreatorFunc) {\n\tdefaultRegistry.Register(name, driver)\n}", "func registerPlugin(name string, init constructor) {\n\tif Pipes == nil {\n\t\tPipes = make(map[string]constructor)\n\t}\n\tPipes[name] = init\n}", "func (s *Service) Register(name string, mp Provider) error {\n\ts.providers[name] = mp\n\treturn nil\n}", "func Register() error {\r\n\treturn DefaultServer.Register()\r\n}", "func (s SwxProxy) Register(_ context.Context, _ *protos.RegistrationRequest) (*protos.RegistrationAnswer, error) {\n\treturn &protos.RegistrationAnswer{}, nil\n}", "func (runner *runnerImpl) Register(g process.GetPID) error {\n\treturn runner.manager.Register(func() (int, error) {\n\t\tpid, err := g()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tfor _, hook := range runner.hooks {\n\t\t\tgo hook.PID(pid)\n\t\t}\n\n\t\treturn pid, err\n\t})\n}", "func (provider WechatWorkProvider) Register(context *auth.Context) {\n\tprovider.Login(context)\n}", "func Register(e interface{}, v...IAppEventHandler){\n\tinternalEventsHandlers.Register(e, v...)\n}", "func RegisterPlugin(p Plugin) {\n\tplugins = append(plugins, p)\n}", "func Register(fn interface{}) error {\n\t// Validate that its a function\n\tfnType := reflect.TypeOf(fn)\n\tif err := validateFnFormat(fnType); err != nil {\n\t\treturn err\n\t}\n\t// Check if already registered\n\tfnName := getFunctionName(fn)\n\t_, ok := fnLookup.getFn(fnName)\n\tif ok {\n\t\treturn nil\n\t}\n\tfor i := 0; i < fnType.NumIn(); i++ {\n\t\targType := fnType.In(i)\n\t\t// Interfaces cannot be registered, their implementations should be\n\t\t// https://golang.org/pkg/encoding/gob/#Register\n\t\tif argType.Kind() != reflect.Interface {\n\t\t\targ := reflect.Zero(argType).Interface()\n\t\t\tif err := GlobalBackend().Encoder().Register(arg); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"unable to register the message for encoding\")\n\t\t\t}\n\t\t}\n\t}\n\tfnLookup.addFn(fnName, fn)\n\treturn nil\n}", "func (_DelegateProfile *DelegateProfileCaller) Registered(opts *bind.CallOpts, _addr common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _DelegateProfile.contract.Call(opts, out, \"registered\", _addr)\n\treturn *ret0, err\n}", "func Register(id loader.PluginID, config *loader.InternalPluginConfig) {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tcatalog[id] = &Registration{\n\t\tConfig: config,\n\t}\n}", "func registerModules(){\n\tweb.Register()\t\t//http server.\n\tdevicetwin.Register()\t\n\teventhub.Register()\t\n}", "func Register(ctx context.Context, wContext *wrangler.Context, registryOverride string) error {\n\th := &handler{\n\t\tmanager: wContext.SystemChartsManager,\n\t\tnamespaces: wContext.Core.Namespace(),\n\t\tchartsConfig: chart.RancherConfigGetter{ConfigCache: wContext.Core.ConfigMap().Cache()},\n\t\tregistryOverride: registryOverride,\n\t}\n\n\twContext.Catalog.ClusterRepo().OnChange(ctx, \"bootstrap-charts\", h.onRepo)\n\trelatedresource.WatchClusterScoped(ctx, \"bootstrap-charts\", relatedFeatures, wContext.Catalog.ClusterRepo(), wContext.Mgmt.Feature())\n\n\trelatedresource.WatchClusterScoped(ctx, \"bootstrap-settings-charts\", relatedSettings, wContext.Catalog.ClusterRepo(), wContext.Mgmt.Setting())\n\n\t// ensure the system charts are installed with the correct values when there are changes to the rancher config map\n\trelatedresource.WatchClusterScoped(ctx, \"bootstrap-configmap-charts\", relatedConfigMaps, wContext.Catalog.ClusterRepo(), wContext.Core.ConfigMap())\n\treturn nil\n}", "func (e *runtimeInfoService) Register(rtime runtime.RuntimeId, enricherBuilder func(socket string) (runtime.ContainerEnricher, error)) error {\n\tif !e.sockets.Supports(rtime) {\n\t\treturn errfmt.Errorf(\"error registering enricher: unsupported runtime %s\", rtime.String())\n\t}\n\tsocket := e.sockets.Socket(rtime)\n\tenricher, err := enricherBuilder(socket)\n\tif err != nil {\n\t\treturn errfmt.WrapError(err)\n\t}\n\te.enrichers[rtime] = enricher\n\treturn nil\n}", "func register(fileExtensions []string, mimeTypes []string, factory LexerFactory) {\n\tfor _, extension := range fileExtensions {\n\t\textensionsMap[extension] = factory\n\t}\n\tfor _, mimeType := range mimeTypes {\n\t\tmimeTypesMap[mimeType] = factory\n\t}\n}", "func RegisterTypes(registry interface {\n\t RegisterType(name string, obj any)\n}) {\n\n}", "func (_m *IService) RegisterMix(info models.MixRegistrationInfo) {\n\t_m.Called(info)\n}", "func Register(params ...interface{}) {\n\t// appendParams will append the object that annotated with at.AutoConfiguration\n\tcomponentContainer, _ = appendParams(componentContainer, params...)\n\treturn\n}", "func AddCommandExtension(ext CommandRegistration, local bool) {\n\textensions = append(extensions, commandHook{ext, \"\", local, false})\n}", "func (s *GRPCServer) Register(ctx context.Context, registerRequest *dashboard.RegisterRequest) (*dashboard.RegisterResponse, error) {\n\tm, err := s.Impl.Register(ctx, registerRequest.DashboardAPIAddress)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcapabilities := convertFromCapabilities(m.Capabilities)\n\n\treturn &dashboard.RegisterResponse{\n\t\tPluginName: m.Name,\n\t\tDescription: m.Description,\n\t\tCapabilities: &capabilities,\n\t}, nil\n}", "func (this *ReceiverHolder) registerShutdownHook() {\n\tquit := make(chan os.Signal)\n\tsignal.Notify(quit, syscall.SIGINT, syscall.SIGTERM, syscall.SIGQUIT)\n\t//fine, wait for signal\n\tgo func() {\n\t\t<-quit\n\t\tfmt.Println(\"got shutdown call\")\n\t\tthis.Shutdown()\n\t\tos.Exit(0)\n\t}()\n}", "func Register(pf Provider) {\n\tlogrus.Infof(\"Registering log provider '%s'\", pf)\n\tproviders = append(providers, pf)\n}", "func (r *registry) Extensions() (*Extensions, error) {\n\tr.RLock()\n\tif r.setup {\n\t\textensions := r.extensions\n\t\tr.RUnlock()\n\t\treturn extensions, nil\n\t}\n\tr.RUnlock()\n\n\tr.Lock()\n\tdefer r.Unlock()\n\treturn r.doSetup()\n}", "func (command *Command) Register(shell sfinterfaces.IShell) {\n\n}", "func (b *bot) RegisterRegex(p Plugin, kind Kind, r *regexp.Regexp, resp ResponseHandler) {\n\tt := PluginName(p)\n\tif _, ok := b.callbacks[t]; !ok {\n\t\tb.callbacks[t] = make(map[Kind][]HandlerSpec)\n\t}\n\tif _, ok := b.callbacks[t][kind]; !ok {\n\t\tb.callbacks[t][kind] = []HandlerSpec{}\n\t}\n\tspec := HandlerSpec{\n\t\tKind: kind,\n\t\tRegex: r,\n\t\tHandler: resp,\n\t}\n\tb.callbacks[t][kind] = append(b.callbacks[t][kind], spec)\n}", "func (t *ZMQTransport) RegisterHook(h TransportHook) {\n\tt.hooks = append(t.hooks, h)\n}", "func (f *ExtensionStoreGetFeaturedExtensionsFunc) PushHook(hook func(context.Context) ([]*stores.Extension, error)) {\n\tf.mutex.Lock()\n\tf.hooks = append(f.hooks, hook)\n\tf.mutex.Unlock()\n}", "func Register(controller controller.Controller) {\n\tregistry.Register(controller)\n}", "func Register(m ...initializer.Simple) {\n\tall = append(all, m...)\n}", "func Register(plugins *admission.Plugins) {\n\tplugins.Register(\"FooBar\", func(config io.Reader) (admission.Interface, error) {\n\t\treturn New()\n\t})\n}", "func (p *Protocol) Register(r *protocol.Registry) error {\n\treturn r.Register(protocolID, p)\n}" ]
[ "0.64641726", "0.64082414", "0.6366064", "0.6070011", "0.60490626", "0.6028064", "0.59680957", "0.59625554", "0.5932356", "0.5923758", "0.58906144", "0.58837605", "0.5864451", "0.58362824", "0.57506686", "0.57484215", "0.5740233", "0.5724657", "0.5723639", "0.5720206", "0.57059675", "0.5685743", "0.5681493", "0.566724", "0.5666011", "0.5626879", "0.5623347", "0.5621214", "0.5597674", "0.55934674", "0.5556715", "0.55470115", "0.5542716", "0.5537134", "0.55370677", "0.55184954", "0.5513207", "0.5510334", "0.550198", "0.5501429", "0.54939723", "0.54799336", "0.54799336", "0.54799336", "0.5467946", "0.5465097", "0.5456116", "0.5445259", "0.54381615", "0.54337317", "0.54246145", "0.5423891", "0.5422987", "0.54205436", "0.54076725", "0.5399696", "0.53989565", "0.5396613", "0.5379931", "0.53687805", "0.5367583", "0.5361498", "0.53599143", "0.5355677", "0.535362", "0.5349311", "0.53472066", "0.53404367", "0.5336346", "0.53231174", "0.5315429", "0.53090215", "0.5302517", "0.5297121", "0.52961034", "0.5281029", "0.52791643", "0.52769667", "0.52742016", "0.5271441", "0.52696395", "0.526919", "0.5267896", "0.5266295", "0.52639645", "0.52600104", "0.5255078", "0.5231274", "0.5230562", "0.522807", "0.52266765", "0.5224628", "0.5216824", "0.5214363", "0.5211551", "0.5208046", "0.5203794", "0.5203497", "0.5201316", "0.5196125" ]
0.7803478
0
Unregistered is called when an extension is unregistered
func (e *Extension) Unregistered() { }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func unregisterAllExtensions() {\n\tpluginRegistryMu.Lock()\n\tdefer pluginRegistryMu.Unlock()\n\tpluginRegistry = make(map[string]Plugin)\n}", "func Unregister(name string) {\n\ti.Unregister(name)\n}", "func (s *serverRegistry) Unregister(*FContext) {}", "func (manager *Manager) Unregister(typ, username, password string, vnfmEndpoint *catalogue.Endpoint) {\n\tif vnfmEndpoint == nil {\n\t\tmanager.unregisterPlugin(typ, username, password)\n\t\treturn\n\t} else {\n\t\tmsg := catalogue.VnfmManagerUnregisterMessage{\n\t\t\tType: typ,\n\t\t\tAction: \"unregister\",\n\t\t\tUsername: username,\n\t\t\tPassword: password,\n\t\t\tEndpoint: vnfmEndpoint,\n\t\t}\n\t\tmanager.unregister(msg)\n\t}\n}", "func (inst *Instance) Unregister() error {\n\tpr := PluginRegistry()\n\tpr.mut.Lock()\n\tdefer pr.mut.Unlock()\n\n\tvar idx int\n\tfor j, i := range pr.instances {\n\t\t// TODO: verify if pointer equality is sufficient\n\t\tif i == inst {\n\t\t\tidx = j\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// delete the instance from the list\n\tpr.instances = append(pr.instances[:idx], pr.instances[idx+1:]...)\n\n\tlog.Printf(\"Unregistered plugin '%s' from room id '%s'\", inst.Name, inst.RoomId)\n\n\treturn nil\n}", "func (manager *Manager) unregisterPlugin(typ, username, password string) {\n\tmsg := catalogue.ManagerUnregisterMessage{\n\t\tType: typ,\n\t\tAction: \"unregister\",\n\t\tUsername: username,\n\t\tPassword: password,\n\t}\n\tmanager.unregister(msg)\n}", "func (w *WidgetList) UnloadExtensionWidget(widgetName string) error {\n\tfor i, widget := range w.List {\n\t\tif widget.Name == widgetName {\n\t\t\tw.List = append(w.List[:i], w.List[i+1:]...)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (communication *Wrapper) Unregister() common.SyncServiceError {\n\tif common.Configuration.CommunicationProtocol == \"http\" {\n\t\tcomm, err := communication.selectCommunicator(common.Configuration.CommunicationProtocol, \"\", \"\", \"\")\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn comm.Unregister()\n\t}\n\n\treturn &Error{\"ESS unregister only support in http communication protocol\\n\"}\n}", "func Unregister(prefix string) {\n\tpluginLock.Lock()\n\tdefer pluginLock.Unlock()\n\tdelete(plugins, prefix)\n}", "func Deregister() error {\r\n\treturn DefaultServer.Deregister()\r\n}", "func (rs *ResourceSubscription) unregister() {\n\tif rs.query == \"\" {\n\t\trs.e.base = nil\n\t} else {\n\t\tdelete(rs.e.queries, rs.query)\n\t}\n\tfor _, q := range rs.links {\n\t\tif q == \"\" {\n\t\t\trs.e.base = nil\n\t\t} else {\n\t\t\tdelete(rs.e.links, q)\n\t\t}\n\t}\n\trs.links = nil\n}", "func (innerExecutorImpl) unregister(w http.ResponseWriter, req *http.Request) {\n\tlogger.Logging(logger.DEBUG)\n\tdefer logger.Logging(logger.DEBUG, \"OUT\")\n\n\tif !common.CheckSupportedMethod(w, req.Method, POST) {\n\t\treturn\n\t}\n\n\te := healthExecutor.Unregister()\n\tif e != nil {\n\t\tcommon.MakeErrorResponse(w, e)\n\t\treturn\n\t}\n\n\tresponse := make(map[string]interface{})\n\tresponse[\"result\"] = \"success\"\n\tcommon.MakeResponse(w, common.ChangeToJson(response))\n}", "func (m *manager) Unregister(name string) {\n\tm.tagsMutex.Lock()\n\tdefer m.tagsMutex.Unlock()\n\n\tm.registry.Unregister(name)\n\tm.metricTags[name] = nil\n}", "func (h *NotificationHub) Unregister(ctx context.Context, registration RegisteredDevice) (err error) {\n\tvar (\n\t\tregURL = h.generateAPIURL(path.Join(\"registrations\", registration.RegistrationID))\n\t\theaders = map[string]string{\n\t\t\t\"Content-Type\": \"application/atom+xml;type=entry;charset=utf-8\",\n\t\t\t\"If-Match\": registration.ETag,\n\t\t}\n\t)\n\n\t_, _, err = h.exec(ctx, deleteMethod, regURL, headers, nil)\n\treturn\n}", "func Unregister(name string) {\n\tif Registered(name) {\n\t\tdelete(registeredCalendars, canonCalendarName(name))\n\t}\n}", "func (ds *DataSubject) unregisterObserver(o DataListener) {\n\n}", "func (s SwxProxy) Deregister(_ context.Context, _ *protos.RegistrationRequest) (*protos.RegistrationAnswer, error) {\n\treturn &protos.RegistrationAnswer{}, nil\n}", "func (h *Healthz) Unregister(name string) {\n\th.Lock()\n\tdefer h.Unlock()\n\tdelete(h.checks, name)\n}", "func Unsubscribe(tag string, handlers ...Handler) {\n\tdefaultDispatcher.Unsubscribe(tag, handlers...)\n}", "func Unregister(handle FnHandle) {\n\tif running() {\n\t\treturn\n\t}\n\tfnsMutex.Lock()\n\tdefer fnsMutex.Unlock()\n\tdelete(fns, handle)\n}", "func (manager *Manager) unregister(msg interface{}) {\n\tmsgBytes, err := json.Marshal(msg)\n\tif err != nil {\n\t\tmanager.logger.Errorf(\"Error while marshalling unregister message: %v\", err)\n\t\treturn\n\t}\n\terr = SendMsg(nfvoManagerHandling, msgBytes, manager.Channel, manager.logger)\n\tif err != nil {\n\t\tmanager.logger.Errorf(\"Error unregistering: %v\", err)\n\t\treturn\n\t}\n}", "func (s *Service) unregister() error {\n\tif _, ok := s.cfg.Discovery.(discovery.Discovery); ok {\n\t\tif err := s.cfg.Discovery.Unregister(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func UnregisterProviders() {\n\t_setupMux.Lock()\n\tdefer _setupMux.Unlock()\n\t_staticProviderFuncs = nil\n\t_dynamicProviderFuncs = nil\n}", "func (p *plugin) Unload(prof irc.Profile) error {\n\tp.profile = nil\n\treturn nil\n}", "func TestRemoveReferencedExtension(t *testing.T) {\n\t// TDB after resource APIs are implemented\n}", "func DllUnregisterServer() { main() }", "func UnRegisterListener(listenerObj core.EventListener, key ...string) error {\n\treturn DefaultConf.ConfigFactory.UnRegisterListener(listenerObj, key...)\n}", "func Unregister(name string) {\n\tif _, ok := factories[name]; !ok {\n\t\tlog.Warningf(\"workflow %v doesn't exist, cannot remove it\", name)\n\t} else {\n\t\tdelete(factories, name)\n\t}\n}", "func (p *plugin) Unload(prof irc.Profile) error {\n\tp.quitOnce.Do(func() {\n\t\tclose(p.quit)\n\t\tp.saveFile()\n\t})\n\treturn nil\n}", "func (f *Factory) Unregister(componentName string) error {\n\tif _, exists := f.registry[componentName]; exists {\n\t\tdelete(f.registry, componentName)\n\t\treturn nil\n\t}\n\treturn fmt.Errorf(\"Registry error: component '%s' is not registered\", componentName)\n}", "func (h *Hook) Unregister() error {\n\tsp, err := h.GetSnapshot().FastForward()\n\tif err != nil {\n\t\treturn err\n\t}\n\texists, _, err := sp.Exists(h.file.Path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !exists {\n\t\treturn errorf(ErrNotFound, `hook \"%s\" not found`, h.Name)\n\t}\n\treturn h.file.Del()\n}", "func (addon DockerregistryAddon) Uninstall() {\n\tnode := *addon.masterNode\n\t_, err := addon.communicator.RunCmd(node, \"helm delete --purge `helm list | grep docker-registry | awk '{print $1;}'`\")\n\tFatalOnError(err)\n\tlog.Println(\"docker-registry uninstalled\")\n}", "func UnregisterEncoder(encType uint32) error {\n\tif _, ok := encoders[encType]; !ok {\n\t\treturn fmt.Errorf(\"encoder type %d doesn't exist\", encType)\n\t}\n\n\tdelete(encoders, encType)\n\n\treturn nil\n}", "func (r *Router) UnregisterHandler(string) {\n\n}", "func (l *Loader) UnregisterProviders() {\n\tl.lock.Lock()\n\tdefer l.lock.Unlock()\n\n\tl.staticProviderFuncs = nil\n\tl.dynamicProviderFuncs = nil\n}", "func UnregisterUnloadable(unloadable Unloadable) {\n\tif !unloadingAll {\n\t\tfor i, u := range unloadables {\n\t\t\tif u == unloadable {\n\t\t\t\tunloadables[i] = unloadables[len(unloadables)-1]\n\t\t\t\tunloadables[len(unloadables)-1] = nil\n\t\t\t\tunloadables = unloadables[:len(unloadables)-1]\n\t\t\t\tTraceLog(LogTrace, \"[UNLOAD] Removed Unloadable\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n}", "func Unregister(name string) bool {\n\tbackendDrivers.Lock()\n\tdefer backendDrivers.Unlock()\n\n\t_, exist := backendDrivers.drivers[name]\n\tif !exist {\n\t\treturn false\n\t}\n\n\tdelete(backendDrivers.drivers, name)\n\n\treturn true\n}", "func (c *Command) Unregister() {\n\tc.cobraCmd = nil\n}", "func (a *PushKeyAPI) unregister(params interface{}) (resp *rpc.Response) {\n\treturn rpc.Success(a.mods.PushKey.Unregister(cast.ToStringMap(params)))\n}", "func Unregister(path string, methods HttpMethod) {\n\treg, regExists := registries[methods]\n\tif regExists {\n\t\t_, handlerExists := reg[path]\n\t\tif handlerExists {\n\t\t\tdelete(reg, path)\n\t\t}\n\t}\n}", "func deregister(machinePath string) {\n\tif isActive(VBoxName) {\n\t\tm, err := virtualbox.GetMachine(VBoxName)\n\t\tif err != nil {\n\t\t\thelp.ExitOnError(err)\n\t\t}\n\t\tif m.State == virtualbox.Running {\n\t\t\terr = m.Poweroff()\n\t\t\tif err != nil {\n\t\t\t\thelp.ExitOnError(err)\n\t\t\t}\n\t\t}\n\t\thelp.ExecCmd(\"VBoxManage\",\n\t\t\t[]string{\n\t\t\t\t\"unregistervm\",\n\t\t\t\tfmt.Sprintf(\"%s\", machinePath),\n\t\t\t})\n\t\tfmt.Println(\"[+] Done\")\n\t}\n}", "func unRegisterTrackerInstance(connStr string) {\n\tmanagedLock.Lock()\n\tdefer managedLock.Unlock()\n\tif managedTrackerInstance[connStr] == nil {\n\t\tlogger.Info(\"tracker instance not exists\")\n\t} else {\n\t\tdelete(managedTrackerInstance, connStr)\n\t}\n}", "func UnRegister(accountID string, tx *sql.Tx) error {\n\tmapper := rlt.NewAccountMapper(tx) //关联至数据库相连结构体\n\treturn mapper.DeleteAccountByID(accountID)\n}", "func unregisterCallback(cbArg *callbackArg) {\n\tcallbacksLock.Lock()\n\tdefer callbacksLock.Unlock()\n\tif _, ok := callbacks[cbArg.id]; !ok {\n\t\tpanic(fmt.Sprintf(\"Callback ID %d not registered\", cbArg.id))\n\t}\n\tdelete(callbacks, cbArg.id)\n}", "func Uninstall() error { return mageextras.Uninstall(\"factorio\") }", "func (p *Plugin) Unregister() error {\n\tpr := PluginRegistry()\n\tpr.mut.Lock()\n\tdefer pr.mut.Unlock()\n\n\tplugins := make([]*Plugin, len(pr.plugins)-1)\n\tvar i int\n\tfor _, plugin := range pr.plugins {\n\t\tif plugin.Name == p.Name {\n\t\t\tcontinue\n\t\t} else {\n\t\t\t// TODO: this might segfault if this is called on an unregistered or never-registered plugin\n\t\t\tplugins[i] = plugin\n\t\t\ti++\n\t\t}\n\t}\n\n\tpr.plugins = plugins\n\n\treturn nil\n}", "func (p *DynamicComputationPolicy) UnregisterAll(path string) {\n\tdelete(p.capabilities, path)\n}", "func UnregisterPath(path string) {\n\troot.unregisterPath(newPathSpec(path))\n}", "func (m MockEventPubSub) Unsubscribe(name string, hdl events.Handler) {\n\tif m.UnsubscribeFunc == nil {\n\t\tpanic(\"Unsubscribe called, but not set\")\n\t}\n\tm.UnsubscribeFunc(name, hdl)\n}", "func (w *deviceWrapper) Unload() {\n\tw.Ctor.DeviceInterface.(device.IDevice).Unload()\n\tif 0 != w.jobID {\n\t\tw.Ctor.Cron.RemoveFunc(w.jobID)\n\t}\n\n\tclose(w.Ctor.LoadData.DeviceStateUpdateChan)\n\n\tif w.Ctor.IsRootDevice {\n\t\tclose(w.Ctor.LoadData.DeviceDiscoveredChan)\n\t}\n}", "func Unsubscribe(name string, handler func(Addresses)) {\n\tUnsubscribeByTag(name, \"\", handler)\n}", "func DeRegister() {\n\tdefer trace()()\n\tif !conf.UCMConfig.ServiceDiscovery.Enabled {\n\t\treturn\n\t}\n\t//func (c *Catalog) Deregister(dereg *CatalogDeregistration, q *WriteOptions) (*WriteMeta, error)\n\tdereg := consul.CatalogDeregistration{\n\t\tNode: hostname,\n\t\tDatacenter: conf.UCMConfig.ServiceDiscovery.Datacenter,\n\t\tServiceID: conf.UCMConfig.ServiceDiscovery.ServiceID,\n\t}\n\t//Get the Consul client\n\tcconfig := consul.DefaultNonPooledConfig()\n\tcconfig.Address = conf.UCMConfig.ServiceDiscovery.GetAddress()\n\tclient, err := consul.NewClient(cconfig)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tcatalog := client.Catalog()\n\n\t//make the API call to register\n\tw, err := catalog.Deregister(&dereg, nil)\n\tif err != nil {\n\t\tlog.Error(err)\n\t} else {\n\t\tlog.Infof(\"OK: Consul deregistration succeeded after %f ns.\", w.RequestTime.Nanoseconds())\n\t}\n}", "func (p *ProcType) Unregister() (err error) {\n\treturn p.Dir.del(\"/\")\n}", "func uninstall() {\n\tC.al_uninstall_system()\n}", "func (r *GoMetricsRegistry) Unregister(name string) {\n\tr.mutex.Lock()\n\tdefer r.mutex.Unlock()\n\n\tst := r.rmState(name)\n\tr.reg.Remove(st.name)\n\tr.shadow.Unregister(name)\n}", "func (s NoUseSwxProxy) Deregister(\n\tctx context.Context,\n\treq *protos.RegistrationRequest,\n) (*protos.RegistrationAnswer, error) {\n\treturn &protos.RegistrationAnswer{}, fmt.Errorf(\"Deregister is NOT IMPLEMENTED\")\n}", "func (e *Engine) deregister() error {\n\t// remove rule associated with engine\n\terr := e.Rule.Remove()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't remove rule %s of engine %s: %v\", e.Rule.Name, e.Name, err)\n\t}\n\n\t// remove engine as export client\n\texportClient := ExportClient{}\n\terr = exportClient.Remove(*e)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't remove export client registration for engine %s: %v\", e.Name, err)\n\t}\n\n\t// instruct engine to deregister value descriptors\n\t//deregister := \"deregister\"\n\t//communication.Publish(\n\t//\tnaming.Topic(e.Index, naming.Command),\n\t//\tnaming.Publisher(e.Index, naming.Command),\n\t//\tderegister)\n\n\t// wait until value descriptors are deregistered\n\n\treturn nil\n}", "func (backend *Backend) Deregister(deviceID string) {\n\tif device, ok := backend.devices[deviceID]; ok {\n\t\tbackend.onDeviceUninit(deviceID)\n\t\tdelete(backend.devices, deviceID)\n\t\tbackend.DeregisterKeystore()\n\n\t\t// Old-school\n\t\tbackend.events <- backendEvent{Type: \"devices\", Data: \"registeredChanged\"}\n\t\t// New-school\n\t\tbackend.Notify(observable.Event{\n\t\t\tSubject: \"devices/registered\",\n\t\t\tAction: action.Reload,\n\t\t})\n\t\tswitch device.ProductName() {\n\t\tcase bitbox.ProductName:\n\t\t\tbackend.banners.Deactivate(banners.KeyBitBox01)\n\t\tcase bitbox02.ProductName:\n\t\t\tbackend.banners.Deactivate(banners.KeyBitBox02)\n\t\t}\n\n\t}\n}", "func UnregisterAll() {\n\tfactorysMu.Lock()\n\tdefer factorysMu.Unlock()\n\t// For tests.\n\tfactories = make(map[string]Factory)\n}", "func (r *rpcServerService) Unregister(name string) error {\n if _, found := r.serviceMap.Del(name); !found {\n return os.ErrNotExist\n }\n return nil\n}", "func (backend *Backend) OnDeviceUninit(f func(string)) {\n\tbackend.onDeviceUninit = f\n}", "func Unregister(level, name int) {\r\n\tparserMu.Lock()\r\n\tdelete(parsers, int64(level)<<32|int64(name))\r\n\tparserMu.Unlock()\r\n}", "func (backend *Backend) OnAccountUninit(f func(accounts.Interface)) {\n\tbackend.onAccountUninit = f\n}", "func (p *Proc) Unregister() error {\n\tsp, err := p.GetSnapshot().FastForward()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn p.dir.Join(sp).Del(\"/\")\n}", "func (bus *EventBus) Unregister(h Handler) {\n\tbus.unregister <- h\n}", "func (sn *ViewUpdateNotification) UnNotify(sub ViewUpdateSubscriber) {\n\tsn.do(func() {\n\t\tindex, ok := sn.register[sub]\n\t\tif !ok {\n\t\t\treturn\n\t\t}\n\n\t\tsn.subs = append(sn.subs[:index], sn.subs[index+1:]...)\n\t})\n}", "func (bt *Blobies) deregister(guid uuid.UUID) {\n\tdelete(bt.Objects, guid)\n}", "func (t *Server) UnregisterRPC(uri string) {\n\tdelete(t.rpcHandlers, uri)\n}", "func (r *DefaultTaggedRegistry) UnregisterAll() {\n\tr.mutex.Lock()\n\tdefer r.mutex.Unlock()\n\tfor name, _ := range r.metrics {\n\t\tdelete(r.metrics, name)\n\t}\n}", "func UnRegisterTaiDriverHandler(name string, handler DriverHandler) {\n\ttai.handlersMutex.Lock()\n\tdefer tai.handlersMutex.Unlock()\n\tdelete(tai.handlers, name)\n}", "func (r *Registry) Unregister(pdb *db.PostgresDB, info *rentities.RegisterInfo) error {\n\tsList, okName := r.ServicesMap[info.TName]\n\tif okName != true {\n\t\treturn fmt.Errorf(\"Unregister error: service not exist\")\n\t}\n\tfor i, ri := range sList {\n\t\tif ri.IP == info.IP { //delete elem in slice\n\t\t\tsList[i] = sList[len(sList)-1]\n\t\t\tsList[len(sList)-1] = nil\n\t\t\tsList = sList[:len(sList)-1]\n\t\t\t//delete from db\n\t\t\tpdb.DeleteReg(info.IID)\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn fmt.Errorf(\"Deregister error: service not exist\")\n}", "func unregisterViews(r *Reporter) error {\n\tif !r.initialized {\n\t\treturn errors.New(\"reporter is not initialized\")\n\t}\n\tmetricstest.Unregister(countName, latencyName, qdepthName)\n\tr.initialized = false\n\treturn nil\n}", "func (c *client) UnregisterCapability(ctx context.Context, params *UnregistrationParams) (err error) {\n\terr = c.Conn.Call(ctx, MethodClientUnregisterCapability, params, nil)\n\treturn\n}", "func Unsubscribe(subToken SubToken) error {\n\treturn getDefaultNats().Unsubscribe(subToken)\n}", "func (s *InputNotifier) unregisterAll() {\n\ts.observers = []InputDetector{}\n}", "func Unregister(manager *mgdb.MongoManager, userCertificateHash []byte) *api.ErrorCode {\n\tcount, err := manager.Get(\"users\").DeleteAll(bson.M{\n\t\t\"certHash\": userCertificateHash,\n\t})\n\tif err != nil || count == 0 {\n\t\treturn &api.ErrorCode{Code: api.ErrorCode_INVARG, Message: \"No user matching provided certificate\"}\n\t}\n\n\treturn &api.ErrorCode{Code: api.ErrorCode_SUCCESS}\n}", "func Unload(filename string) {\n\tloadedLock.Lock()\n\tdelete(loaded, filename)\n\tloadedLock.Unlock()\n}", "func Unregister(qi inomap.QIno) {\n\tt.Lock()\n\tdefer t.Unlock()\n\n\te := t.entries[qi]\n\te.refCount--\n\tif e.refCount == 0 {\n\t\tdelete(t.entries, qi)\n\t}\n}", "func (m *manager) Unsubscribe(nh datapath.NodeHandler) {\n\tm.nodeHandlersMu.Lock()\n\tdelete(m.nodeHandlers, nh)\n\tm.nodeHandlersMu.Unlock()\n}", "func (c *clientRegistry) Unregister(ctx *FContext) {\n\topID := ctx.opID()\n\tc.mu.Lock()\n\tdelete(c.handlers, opID)\n\tc.mu.Unlock()\n}", "func unregApp(cmd marcopolo.CmdMsg, srvConn *marcopolo.ServerConn) {\n\tappName := cmd.AppNameParam.AppName\n\tfmt.Println(\"unregister app \", appName)\n\n\t// lookup regd app w. this name\n\tprevApp, found := apps[appName]\n\t_ = prevApp\n\tif found {\n\t\tfmt.Printf(\"unregapp, removing @ %s\\n\", prevApp.appAddr)\n\n\t\t// remove app entry\n\t\tdelete(apps, appName)\n\n\t\t// send OK to caller\n\t\tsrvConn.SendRespUnregApp(cmd.UdpPacket.RemoteAddr)\n\t} else {\n\t\terr := fmt.Errorf(\"unregapp, not found : '%s'\", appName)\n\t\tfmt.Println(err)\n\n\t\t// send error back to app\n\t\tsrvConn.SendRespUnregAppErr(err, cmd.UdpPacket.RemoteAddr)\n\t}\n}", "func (n *Network) Unregister(registration fab.Registration) {\n\tn.event.Unregister(registration)\n}", "func (dr *DefaultCheckRegister) Unregister(addr, path, checkID string) error {\n\treturn unregister(addr, path, checkID)\n}", "func (s *SysContext) UnqualifiedRegistries(ctx context.Context) []string {\n\treturn s.unqualifiedRegistries\n}", "func UnregisterCommand(command string) {\n\tdelete(Commands, command)\n}", "func (w *worker) deregister() {\n\tw.Lock()\n\tid := w.id\n\teid := w.eid\n\tw.active = false\n\tw.Unlock()\n\tactivePipelines.Lock()\n\tdelete(activePipelines.i, id)\n\tdelete(activePipelines.eids, eid)\n\tactivePipelines.Unlock()\n}", "func (_m *Interface) Unwatch() {\n\t_m.Called()\n}", "func (asr *sessionRegistry) deregister(clt *Client) {\n\tasr.lock.Lock()\n\tdelete(asr.registry, clt.Session.Key)\n\tasr.lock.Unlock()\n}", "func Uninit() {\n\tC.notify_uninit()\n}", "func (s *RuntimeVersionSubscription) Unsubscribe() {\n\ts.sub.Unsubscribe()\n\ts.quitOnce.Do(func() {\n\t\tclose(s.channel)\n\t})\n}", "func adapterUnRegisterFromService(c *gin.Context) {\n\t// send post request to other endpoint to trigger connection cycle\n\tvar informationUserDTO id.InformationUserDTO\n\terr := c.BindJSON(&informationUserDTO)\n\tif err != nil {\n\t\tlogrus.Fatalf(\"[api.adapterUnRegisterFromService] Error marshal informationUserDTO with error %s\", err)\n\t}\n\tsuccess := id.UnregisterUserFromYourUserList(informationUserDTO)\n\tif success {\n\t\tc.String(200, DefaultSuccessMessage)\n\t} else {\n\t\tc.String(404, DefaultErrorMessage)\n\t}\n}", "func (ref *EventClientRef) Unregister(reg fab.Registration) {\n\tif service, err := ref.get(); err != nil {\n\t\tlogger.Warnf(\"Error unregistering event registration: %s\", err)\n\t} else {\n\t\tservice.Unregister(reg)\n\t}\n}", "func adapterSendUnRegisterToServices(c *gin.Context) {\n\t// trigger method to send all unregister messages to users\n\tsuccess := id.SendUnregisterUserFromYourUserList()\n\tif success {\n\t\tc.String(200, DefaultSuccessMessage)\n\t} else {\n\t\tc.String(404, DefaultErrorMessage)\n\t}\n\t// c.String(403, DefaultNotAvailableMessage)\n}", "func (acnl *Channel) unregister() error {\n\tconst Attempts = 2\n\n\tmsg, err := json.Marshal(acnl.endpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tacnl.l.WithFields(log.Fields{\n\t\t\"tag\": \"channel-amqp-unregister\",\n\t\t\"max-attempts\": Attempts,\n\t\t\"endpoint\": string(msg),\n\t}).Debug(\"sending an unregistering request\")\n\n\tfor i := 0; i < Attempts; i++ {\n\t\t// Try to use the current connection the first time.\n\t\t// Recreate it otherwise\n\t\tif i > 0 {\n\t\t\tacnl.l.WithFields(log.Fields{\n\t\t\t\t\"tag\": \"channel-amqp-unregister\",\n\t\t\t\t\"try\": i,\n\t\t\t}).Warn(\"attempting to re-initialize the connection\")\n\n\t\t\tif _, err = acnl.setup(); err != nil {\n\t\t\t\tacnl.l.WithFields(log.Fields{\n\t\t\t\t\t\"tag\": \"channel-amqp-unregister\",\n\t\t\t\t\t\"try\": i,\n\t\t\t\t\t\"err\": err,\n\t\t\t\t}).Warn(\"setup failed\")\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\tif err = acnl.publish(QueueVNFMUnregister, msg); err == nil {\n\t\t\tacnl.l.WithFields(log.Fields{\n\t\t\t\t\"tag\": \"channel-amqp-unregister\",\n\t\t\t\t\"try\": i,\n\t\t\t\t\"success\": true,\n\t\t\t}).Info(\"endpoint unregister request successfully sent\")\n\t\t\treturn nil\n\t\t}\n\n\t\tacnl.l.WithFields(log.Fields{\n\t\t\t\"tag\": \"channel-amqp-unregister\",\n\t\t\t\"try\": i,\n\t\t\t\"success\": false,\n\t\t}).Info(\"endpoint unregister failed to send\")\n\t}\n\n\treturn err\n}", "func (rs *RegistryService) Unregister(ctx context.Context, in *proto.UnregisterType) (*proto.EmptyResponse, error) {\n\trs.mu.Lock()\n\tdefer rs.mu.Unlock()\n\n\tif h, ok := rs.hosts[in.GetName()]; ok {\n\t\tdelete(rs.hosts, h)\n\t}\n\n\treturn &proto.EmptyResponse{}, nil\n}", "func (p *PRep) Deregister() {\n\tp.Cancel()\n\tpReps.Mutex.Lock()\n\tdelete(pReps.PReps, p.ID.String())\n\tpReps.Mutex.Unlock()\n\tif err := p.Websocket.Close(); err != nil {\n\t\tlog.Printf(\"Error closing prep: %s\", err.Error())\n\t}\n}", "func (m *Manager) Unregister(key string, reply *string) error {\n\tfmt.Println(\"Unregistering key\", key)\n\tdelete(m.validKeys, key)\n\t*reply = \"OK\"\n\treturn nil\n}", "func UninstallJoystick() {\n\tC.al_uninstall_joystick()\n}", "func UnDeviceReady(cb func()) {\n\tjs.Global.Get(\"document\").Call(\"removeEventListener\", \"deviceready\", cb, false)\n}", "func (self *DB) Unsubscribe(name string) {\n\tself.subscriptionsMutex.Lock()\n\tdefer self.subscriptionsMutex.Unlock()\n\tfor _, typeSubs := range self.subscriptions {\n\t\tdelete(typeSubs, name)\n\t}\n}" ]
[ "0.65315443", "0.64458025", "0.6411859", "0.6329457", "0.6286149", "0.62621164", "0.6252598", "0.6172278", "0.6136201", "0.60733706", "0.6065752", "0.60603946", "0.6053101", "0.6033575", "0.60280937", "0.6011342", "0.59920835", "0.5987336", "0.5959819", "0.5945642", "0.59260803", "0.59087497", "0.5887287", "0.5868968", "0.58669335", "0.5860071", "0.5854893", "0.5835114", "0.5823526", "0.5820768", "0.58083206", "0.5803676", "0.57985073", "0.57935244", "0.57670563", "0.57539153", "0.5733817", "0.5711685", "0.57068676", "0.5704264", "0.5703213", "0.5682805", "0.5679445", "0.567706", "0.5653257", "0.56358844", "0.5624968", "0.5612357", "0.56113905", "0.5608517", "0.5603175", "0.55978054", "0.55974644", "0.5594339", "0.5589666", "0.55875164", "0.5582621", "0.557837", "0.5576354", "0.5558796", "0.5551886", "0.5545016", "0.55368716", "0.5525948", "0.5516925", "0.5508094", "0.5502867", "0.5501273", "0.5497065", "0.54919416", "0.5487362", "0.54859465", "0.5485587", "0.5483724", "0.5474457", "0.5473205", "0.5450316", "0.54477054", "0.54381675", "0.5437708", "0.5436031", "0.5421298", "0.5417063", "0.54024476", "0.53987145", "0.5397039", "0.5391988", "0.5390748", "0.53651375", "0.53619856", "0.536185", "0.53592914", "0.5355892", "0.5354339", "0.5348763", "0.53471965", "0.5339685", "0.53387225", "0.53358835", "0.5327684" ]
0.89574146
0
NewMapStorage creates a new MapStorage instance
func NewMapStorage() *MapStorage { return &MapStorage{store: make(map[string]int)} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewMapStorage(name string) *MapStorage {\n\tvar storage MapStorage\n\tstorage.name = name\n\tstorage.Stations = make(map[string]Station)\n\tstorage.MonthlySeries = make(map[string]MonthlyMeasureSerie)\n\n\treturn &storage\n}", "func NewStorage() SafeMap {\n\tsm := make(safeMap)\n\tgo sm.run()\n\treturn sm\n}", "func newStorage() *storage {\n\tr := make(map[string][]byte)\n\treturn &storage{\n\t\trepository: r,\n\t}\n}", "func InitMapStorage() *MapURLStorage {\n\treturn &MapURLStorage{\n\t\tStorage: make(map[string]string),\n\t}\n}", "func (p *PersistentSyncMap) New() {\n\tp.storage = make(map[interface{}]interface{}, 100)\n}", "func newStorage() *storage {\n\treturn &storage{\n\t\tsto: make(map[uint16]mqtt.Message),\n\t\tmux: new(sync.RWMutex),\n\t}\n}", "func NewStorage(s map[string]interface{}) (Storage, error) {\n\tstype, ok := s[\"Type\"].(string)\n\tif !ok || stype == \"\" {\n\t\treturn nil, errors.New(\"Template do not have Storage type\")\n\t}\n\n\tswitch stype {\n\tcase \"Local\":\n\t\treturn newStorageLocal(s), nil\n\tcase \"S3\":\n\t\treturn newStorageS3(s)\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unexecepted Storage type: %v\", stype)\n\t}\n}", "func New() *StorageInMemory {\n\treturn &StorageInMemory{\n\t\tm: sync.Map{},\n\t}\n}", "func New() *Map { return new(Map).Init() }", "func NewMap(size int) *Map {\n\tif size <= 0 {\n\t\tsize = runtime.GOMAXPROCS(0)\n\t}\n\tsplits := make([]Split, size)\n\tfor i := range splits {\n\t\tsplits[i].Map = make(map[interface{}]interface{})\n\t}\n\treturn &Map{splits}\n}", "func NewStorage() *Storage {\n\treturn &Storage{\n\t\tgames: make(map[game.ID]*game.Game),\n\t}\n}", "func NewStorage(t mockConstructorTestingTNewStorage) *Storage {\n\tmock := &Storage{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewStorage(t mockConstructorTestingTNewStorage) *Storage {\n\tmock := &Storage{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewMapStore() RawStore {\n\treturn &mapStorage{\n\t\tinner: make(map[string][]byte),\n\t}\n}", "func NewMemoryStorage() *MemoryStorage {\n return &MemoryStorage{tickets: make(map[string]*types.Ticket)}\n}", "func newServiceStorage() *ServiceStorage {\n\ts := new(ServiceStorage)\n\ts.cache = cache.NewCache(24 * time.Hour)\n\treturn s\n}", "func NewMap(store map[string]*rsa.PrivateKey) *KeyStore {\n\treturn &KeyStore{\n\t\tstore: store,\n\t}\n}", "func New(_ string) (s *Store, err error) {\n\treturn &Store{xz.NewMap()}, nil\n}", "func NewMemoryStorage() (*MemoryStorage, error) {\n\treturn &MemoryStorage{\n\t\tdata: new(sync.Map),\n\t}, nil\n}", "func NewStorage(lgr *log.Logger) *Storage {\n\treturn &Storage{\n\t\tdataStack: graph.NewStack(),\n\t\tdataStorage: NewKVStorage(),\n\t\tlgr: lgr,\n\t}\n}", "func New(items map[storage.Key]*storage.Value, opts ...func(*Storage)) *Storage {\n\tstrg := &Storage{\n\t\tmeta: make(map[storage.MetaKey]storage.MetaValue),\n\t}\n\tstrg.setItems(items)\n\n\tfor _, f := range opts {\n\t\tf(strg)\n\t}\n\n\tif strg.clck == nil {\n\t\tstrg.clck = clock.New()\n\t}\n\n\treturn strg\n}", "func NewStorage() *InMemoryHashStorage {\n\treturn &InMemoryHashStorage{\n\t\tmake(map[int64]string),\n\t\t0,\n\t\t&sync.RWMutex{},\n\t}\n}", "func NewMap(r *goja.Runtime) *Map {\n\treturn &Map{runtime: r}\n}", "func newStorageObject(URL string, source interface{}, fileInfo os.FileInfo) storage.Object {\n\tabstract := storage.NewAbstractStorageObject(URL, source, fileInfo)\n\tresult := &object{\n\t\tAbstractObject: abstract,\n\t}\n\tresult.AbstractObject.Object = result\n\treturn result\n}", "func NewStorage(mds MetadataStore, tr track.Tracker, chunks *chunk.Storage, opts ...StorageOption) *Storage {\n\ts := &Storage{\n\t\tstore: mds,\n\t\ttracker: tr,\n\t\tchunks: chunks,\n\t\tidxCache: index.NewCache(chunks, DefaultIndexCacheSize),\n\t\tmemThreshold: DefaultMemoryThreshold,\n\t\tshardConfig: &index.ShardConfig{\n\t\t\tNumFiles: index.DefaultShardNumThreshold,\n\t\t\tSizeBytes: index.DefaultShardSizeThreshold,\n\t\t},\n\t\tcompactionConfig: &CompactionConfig{\n\t\t\tLevelFactor: DefaultCompactionLevelFactor,\n\t\t},\n\t\tfilesetSem: semaphore.NewWeighted(math.MaxInt64),\n\t\tprefetchLimit: DefaultPrefetchLimit,\n\t}\n\tfor _, opt := range opts {\n\t\topt(s)\n\t}\n\tif s.compactionConfig.LevelFactor < 1 {\n\t\tpanic(\"level factor cannot be < 1\")\n\t}\n\treturn s\n}", "func New() *Storage {\n\treturn &Storage{\n\t\tstat: &statApp{},\n\t}\n}", "func New(timeout time.Duration) (*Storage, error) {\n\tif timeout <= 0 {\n\t\treturn nil, errors.New(\"timeout must be positive\")\n\t}\n\treturn &Storage{\n\t\tdata: make(map[string]interface{}),\n\t\tcancelFuncs: make(map[string]context.CancelFunc),\n\t\ttimeout: timeout,\n\t}, nil\n}", "func NewStorage() *Storage {\r\n\treturn new(Storage)\r\n}", "func NewStorage() *Storage {\n\treturn &Storage{\n\t\tkv: btree.NewKVTree(),\n\t}\n}", "func New() (*g3storage.Storage, error) {\n\treturn NewWithHostname(makoAppHostname)\n}", "func New() *memoryMap {\n\treturn &memoryMap{}\n}", "func NewStorage(l log.Logger, reg prometheus.Registerer, stCallback startTimeCallback, walDir string, flushDeadline time.Duration, sm ReadyScrapeManager) *Storage {\n\tif l == nil {\n\t\tl = log.NewNopLogger()\n\t}\n\tlogger := logging.Dedupe(l, 1*time.Minute)\n\n\ts := &Storage{\n\t\tlogger: logger,\n\t\tlocalStartTimeCallback: stCallback,\n\t}\n\ts.rws = NewWriteStorage(s.logger, reg, walDir, flushDeadline, sm)\n\treturn s\n}", "func NewStorage() *Storage {\n\treturn &Storage{}\n}", "func New(config *config.ConfYaml) *Storage {\n\treturn &Storage{\n\t\tconfig: config,\n\t}\n}", "func New() kv.Store {\n\treturn newStore(newMapStore())\n}", "func NewStorageServer(funcs map[string]*wasmtime.Func, memory *wasmtime.Memory) *StorageServer {\n\treturn &StorageServer{\n\t\tfuncs: funcs,\n\t\tmemory: memory,\n\t\tport: IP,\n\t}\n}", "func newOfferStorage() *offerStorage {\n\treturn &offerStorage{\n\t\tcidMap: make(map[string]*digestOffer),\n\t\tlock: sync.RWMutex{},\n\t}\n}", "func New(opts *Options) *Map {\n\tif opts == nil {\n\t\topts = &Options{}\n\t}\n\tstore := newStore(opts)\n\tm := &Map{\n\t\tstore: store,\n\t\tkeeper: newKeeper(store),\n\t}\n\tgo m.keeper.run()\n\treturn m\n}", "func NewStorage(size int, hash HashFunc) *Storage {\n\ts := &Storage{n: size, h: hash}\n\ts.Reset()\n\treturn s\n}", "func New(dir, name string) (mp *MapDB, err error) {\n\tvar m MapDB\n\t// Initialize map\n\tm.m = make(map[string]string)\n\n\t// Encryption middleware\n\tcmw := middleware.NewCryptyMW([]byte(\" encryption key \"), make([]byte, 16))\n\tif cmw == nil {\n\n\t}\n\n\t// Create a new instance of mrT\n\tif m.mrT, err = mrT.New(dir, name); err != nil {\n\t\treturn\n\t}\n\n\tif err = m.mrT.ForEach(m.load); err != nil {\n\t\treturn\n\t}\n\n\t// Assign pointer to our MapDB\n\tmp = &m\n\treturn\n}", "func NewStorage(typ string) (*Storage, error) {\n\tswitch typ {\n\tcase \"\", \"postgres\":\n\t\tpostgresStore, err := NewPostgresStorage()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tinst := Storage(postgresStore)\n\t\treturn &inst, nil\n\tcase \"inmem\":\n\t\tinst := Storage(new(InMemStorage))\n\t\treturn &inst, nil\n\tdefault:\n\t\treturn nil, errors.New(\"Invalid storage type provided.\")\n\t}\n}", "func makeStorage(name string) ds.Storage {\n\tswitch name {\n\tcase \"skiplist\":\n\t\treturn ds.NewSkipList()\n\tcase \"dict\":\n\t\treturn ds.NewDict()\n\tcase \"b-tree\":\n\t\treturn ds.InitBTree(10)\n\t}\n\treturn ds.NewDict()\n}", "func (m *Manager) MapClientStorage(stor oauth2.ClientStore) error {\n\tif stor == nil {\n\t\treturn errors.ErrNilValue\n\t}\n\tm.injector.Map(stor)\n\treturn nil\n}", "func FakeNewStorage() *fakeStorage {\n\treturn &fakeStorage{}\n}", "func NewStorage() Storage {\n\treturn &storage{}\n}", "func NewStorage() (*Storage, error) {\n\tvar err error\n\ts := new(Storage)\n\tcwd, _ := os.Getwd()\n\ts.db, err = scribble.New(cwd+viper.GetString(\"storage.basedir\"), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn s, nil\n}", "func newStorageImpl(db orm.DB) *storageImpl {\n\treturn &storageImpl{\n\t\tdb: db,\n\t}\n}", "func NewStorage(opts ...StorageOption) *Storage {\n\ts := &Storage{\n\t\tcr: config.DefaultManager,\n\t\tmu: sync.RWMutex{},\n\t}\n\tfor _, opt := range opts {\n\t\tif opt != nil {\n\t\t\topt(s)\n\t\t}\n\t}\n\treturn s\n}", "func NewStorage() (*Storage, error) {\n\tvar err error\n\n\ts := new(Storage)\n\n\t_, filename, _, _ := runtime.Caller(0)\n\tp := path.Dir(filename)\n\n\ts.db, err = scribble.New(p+dir, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s, nil\n}", "func newStorage(\n\tmachineName,\n\ttablename,\n\tdatabase,\n\tretentionPolicy,\n\tusername,\n\tpassword,\n\tinfluxdbHost string,\n\tisSecure bool,\n\tbufferDuration time.Duration,\n) (*influxdbStorage, error) {\n\turl := &url.URL{\n\t\tScheme: \"http\",\n\t\tHost: influxdbHost,\n\t}\n\tif isSecure {\n\t\turl.Scheme = \"https\"\n\t}\n\n\tconfig := &influxdb.Config{\n\t\tURL: *url,\n\t\tUsername: username,\n\t\tPassword: password,\n\t\tUserAgent: fmt.Sprintf(\"%v/%v\", \"cAdvisor\", version.Info[\"version\"]),\n\t}\n\tclient, err := influxdb.NewClient(*config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tret := &influxdbStorage{\n\t\tclient: client,\n\t\tmachineName: machineName,\n\t\tdatabase: database,\n\t\tretentionPolicy: retentionPolicy,\n\t\tbufferDuration: bufferDuration,\n\t\tlastWrite: time.Now(),\n\t\tpoints: make([]*influxdb.Point, 0),\n\t}\n\tret.readyToFlush = ret.defaultReadyToFlush\n\treturn ret, nil\n}", "func newClusterStorage() *ClusterStorage {\n\ts := new(ClusterStorage)\n\treturn s\n}", "func New() storage.Storage {\n\treturn &memStorage{\n\t\tclients: make(map[string]storage.Client),\n\t\tauthCodes: make(map[string]storage.AuthCode),\n\t\trefreshTokens: make(map[string]storage.Refresh),\n\t\tauthReqs: make(map[string]storage.AuthRequest),\n\t}\n}", "func newPodStorage() *PodStorage {\n\ts := new(PodStorage)\n\treturn s\n}", "func NewStorage(cfg *Config) *Storage {\n\tif cfg.Engine == nil {\n\t\tlog.Fatalln(\"Cannot create a ops proxy without an engine\")\n\t}\n\tif cfg.App == nil {\n\t\tnrConfig := newrelic.NewConfig(\"widget\", \"\")\n\t\tnrConfig.Enabled = false\n\t\tapp, err := newrelic.NewApplication(nrConfig)\n\t\tif err != nil {\n\t\t\tlogrus.WithField(\"error\", err).Fatalln(\"could not create dummy new relic app\")\n\t\t}\n\t\tcfg.App = app\n\t}\n\treturn &Storage{engine: cfg.Engine, newrelic: cfg.App}\n}", "func NewStorage(\n\tlogger *zerolog.Logger,\n\tcfg *config.TSDBStorageConfig,\n\tmetadataStorage metadata.Storage,\n) (data.Storage, error) {\n\tgoKitWrapper, err := NewGoKitLogWrapper(logger)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// create data directory if not exists\n\tif _, err = os.Stat(cfg.DataDir); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\tif err = os.MkdirAll(cfg.DataDir, 0750); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// create storage\n\tstor, err := prometheus.OpenTSDB(cfg.DataDir, goKitWrapper)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx, cancel := context.WithCancel(context.Background())\n\n\ts := &storage{\n\t\tcodec: newB64Codec(),\n\t\tmetadataStorage: metadataStorage,\n\t\tcfg: cfg,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\twg: sync.WaitGroup{},\n\t\tlogger: logger,\n\t\ttsdbStorage: stor,\n\t}\n\treturn s, nil\n}", "func NewStorage(namespace, name string) (*Storage, error) {\n\tif err := k8sutil.CreateCRD(name); err != nil {\n\t\treturn nil, err\n\t}\n\tcli := k8sutil.NewRESTClient()\n\treturn &Storage{\n\t\tNamespace: namespace,\n\t\tName: strings.ToLower(name),\n\t\trestcli: cli,\n\t}, nil\n}", "func newStorage(account *account, prov provider.Account, cfg *config.Storage) (*storage, error) {\n\tlog.Debug(\"Initializing Storage\")\n\n\t// Validate the config.Storage object.\n\tif cfg.Buckets == nil {\n\t\treturn nil, fmt.Errorf(\"The buckets element is missing from the storage configuration\")\n\t}\n\n\ts := &storage{\n\t\tResources: resource.NewResources(),\n\t\tStorage: cfg,\n\t\taccount: account,\n\t}\n\n\tvar err error\n\ts.providerStorage, err = prov.NewStorage(cfg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ts.buckets, err = newBuckets(s, prov, cfg.Buckets)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ts.Append(s.buckets)\n\treturn s, nil\n}", "func NewMap() *Map {\n\tm := &Map{}\n\n\tm.Map = make(EnvMap)\n\treturn m\n}", "func NewSyncMap() (store *SyncMap, err error) {\n\tstore = new(SyncMap)\n\tstore.data = new(sync.Map)\n\treturn store, nil\n}", "func NewStorage() (Storage, error) {\n\tstorage, err := newBoltStorage()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn store{storage}, nil\n}", "func NewStorageMock(t minimock.Tester) *StorageMock {\n\tm := &StorageMock{t: t}\n\tif controller, ok := t.(minimock.MockController); ok {\n\t\tcontroller.RegisterMocker(m)\n\t}\n\n\tm.GetUserMock = mStorageMockGetUser{mock: m}\n\tm.GetUserMock.callArgs = []*StorageMockGetUserParams{}\n\n\tm.GetUserLocationMock = mStorageMockGetUserLocation{mock: m}\n\tm.GetUserLocationMock.callArgs = []*StorageMockGetUserLocationParams{}\n\n\tm.SaveUserMock = mStorageMockSaveUser{mock: m}\n\tm.SaveUserMock.callArgs = []*StorageMockSaveUserParams{}\n\n\tm.SaveUserLocationMock = mStorageMockSaveUserLocation{mock: m}\n\tm.SaveUserLocationMock.callArgs = []*StorageMockSaveUserLocationParams{}\n\n\treturn m\n}", "func (s *ImagesByRepositoryRegistryStorage) New() interface{} {\n\treturn &imageapi.ImageRepositoryMapping{}\n}", "func New() Map {\n\treturn empty\n}", "func NewStorageMock(t minimock.Tester) *StorageMock {\n\tm := &StorageMock{t: t}\n\tif controller, ok := t.(minimock.MockController); ok {\n\t\tcontroller.RegisterMocker(m)\n\t}\n\n\tm.AllMock = mStorageMockAll{mock: m}\n\tm.AllMock.callArgs = []*StorageMockAllParams{}\n\n\tm.CloneMock = mStorageMockClone{mock: m}\n\tm.CloneMock.callArgs = []*StorageMockCloneParams{}\n\n\tm.ForIDMock = mStorageMockForID{mock: m}\n\tm.ForIDMock.callArgs = []*StorageMockForIDParams{}\n\n\tm.SplitMock = mStorageMockSplit{mock: m}\n\tm.SplitMock.callArgs = []*StorageMockSplitParams{}\n\n\tm.UpdateMock = mStorageMockUpdate{mock: m}\n\tm.UpdateMock.callArgs = []*StorageMockUpdateParams{}\n\n\treturn m\n}", "func newTestModel() *Storage {\n\treturn &Storage{}\n}", "func NewMap() Map {\n\treturn Map{NewSet()}\n}", "func NewMapDisk(ctx context.Context, wg *sync.WaitGroup) *MapDisk {\n\tmd := &MapDisk{\n\t\tstore: make(map[string][]byte),\n\t\tinChan: make(chan Request),\n\t}\n\twg.Add(1)\n\tgo md.serve(ctx, wg)\n\treturn md\n}", "func (m *Manager) MapTokenStorage(stor oauth2.TokenStore) error {\n\tif stor == nil {\n\t\treturn errors.ErrNilValue\n\t}\n\tm.injector.Map(stor)\n\treturn nil\n}", "func New(path string) (Storage, error) {\n\tif path == \":memory:\" {\n\t\treturn newMemoryStorage(), nil\n\t}\n\n\treturn newFileStorage(path)\n}", "func NewStorage(vol string) *Storage {\n\tloc := vol\n\n\tif vol[len(vol)-1:] != \"/\" {\n\t\tloc = fmt.Sprintf(\"%s/\", vol)\n\t}\n\n\treturn &Storage{\n\t\tloc,\n\t}\n}", "func NewMapEngine(storageLevel int, degree int) *MapEngine {\n\tindex := NewS2Storage(storageLevel, degree)\n\treturn &MapEngine{\n\t\tedges: make(map[int64]map[int64]*Edge),\n\t\ts2Storage: index,\n\t}\n}", "func newStore() storage.Store {\n\tdata := `{\n \"management_chain\": {\n \"bob\": [\n \"ken\",\n \"janet\"\n ],\n \"alice\": [\n \"janet\"\n ]\n }\n}`\n\n\tvar json map[string]interface{}\n\n\terr := util.UnmarshalJSON([]byte(data), &json)\n\tif err != nil {\n\t\t// Handle error.\n\t}\n\n\t// Manually create the storage layer. inmem.NewFromObject returns an\n\t// in-memory store containing the supplied data.\n\tstore := inmem.NewFromObject(json)\n\treturn store\n}", "func NewStorage(db *sql.DB) *Storage {\n\treturn &Storage{db}\n}", "func NewStorage(first, second storage.Storage) (*Storage, error) {\n\treturn &Storage{first, second}, nil\n}", "func NewStorage(client *clientv3.Client, codec codec.Codec) storage.Store {\n\treturn &Storage{\n\t\tclient: client,\n\t\tcodec: codec,\n\t}\n}", "func NewStorage(db *gorm.DB) *Storage {\n\treturn &Storage{\n\t\tdb: db,\n\t}\n}", "func New(o *Options) *Storage {\n\ts := &Storage{}\n\tif o.GraphiteAddress != \"\" {\n\t\tc := graphite.NewClient(\n\t\t\to.GraphiteAddress, o.GraphiteTransport,\n\t\t\to.StorageTimeout, o.GraphitePrefix)\n\t\ts.queues = append(s.queues, NewStorageQueueManager(c, defaultConfig))\n\t}\n\tif o.OpentsdbURL != \"\" {\n\t\tc := opentsdb.NewClient(o.OpentsdbURL, o.StorageTimeout)\n\t\ts.queues = append(s.queues, NewStorageQueueManager(c, defaultConfig))\n\t}\n\tif o.InfluxdbURL != nil {\n\t\tconf := influx.Config{\n\t\t\tURL: *o.InfluxdbURL,\n\t\t\tUsername: o.InfluxdbUsername,\n\t\t\tPassword: o.InfluxdbPassword,\n\t\t\tTimeout: o.StorageTimeout,\n\t\t}\n\t\tc := influxdb.NewClient(conf, o.InfluxdbDatabase, o.InfluxdbRetentionPolicy)\n\t\tprometheus.MustRegister(c)\n\t\ts.queues = append(s.queues, NewStorageQueueManager(c, defaultConfig))\n\t}\n\tif o.GenericURL != \"\" {\n\t\theaders := http.Header{}\n\t\tif o.GenericHeaderName != \"\" {\n\t\t\theaders.Add(o.GenericHeaderName, o.GenericHeaderValue)\n\t\t}\n\t\tc := generic.NewClient(o.GenericURL, headers, o.StorageTimeout)\n\t\ts.queues = append(s.queues, NewStorageQueueManager(c, defaultConfig))\n\t}\n\tif len(s.queues) == 0 {\n\t\treturn nil\n\t}\n\treturn s\n}", "func New() *OMap {\n\treturn &OMap{\n\t\tkeys: make([]string, 0),\n\t\tbaseMap: make(map[string]interface{}, 0),\n\t}\n}", "func newStorage(\n\tmachineName,\n\ttablename,\n\tdatabase,\n\tusername,\n\tpassword,\n\tinfluxdbHost string,\n\tisSecure bool,\n\tbufferDuration time.Duration,\n) (*influxdbStorage, error) {\n\tconfig := &influxdb.ClientConfig{\n\t\tHost: influxdbHost,\n\t\tUsername: username,\n\t\tPassword: password,\n\t\tDatabase: database,\n\t\tIsSecure: isSecure,\n\t}\n\tclient, err := influxdb.NewClient(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// TODO(monnand): With go 1.3, we cannot compress data now.\n\tclient.DisableCompression()\n\n\tret := &influxdbStorage{\n\t\tclient: client,\n\t\tmachineName: machineName,\n\t\ttableName: tablename,\n\t\tbufferDuration: bufferDuration,\n\t\tlastWrite: time.Now(),\n\t\tseries: make([]*influxdb.Series, 0),\n\t}\n\tret.readyToFlush = ret.defaultReadyToFlush\n\treturn ret, nil\n}", "func NewStorage() Storage {\n\tmemtable := memtable.NewRollingMemtable()\n\n\tstorage := Storage{memtable: memtable, wal: wal.Wal{Memtable: &memtable}}\n\tstorage.wal.Start()\n\treturn storage\n}", "func NewStorage() *Storage {\n\tstorageHandler = new(Storage)\n\n\treturn storageHandler\n}", "func NewStorage() *Storage {\n\treturn &Storage{\n\t\tReferenceStorage: make(ReferenceStorage),\n\t\tConfigStorage: ConfigStorage{},\n\t\tShallowStorage: ShallowStorage{},\n\t\tObjectStorage: ObjectStorage{\n\t\t\tObjects: make(map[plumbing.Hash]plumbing.EncodedObject),\n\t\t\tCommits: make(map[plumbing.Hash]plumbing.EncodedObject),\n\t\t\tTrees: make(map[plumbing.Hash]plumbing.EncodedObject),\n\t\t\tBlobs: make(map[plumbing.Hash]plumbing.EncodedObject),\n\t\t\tTags: make(map[plumbing.Hash]plumbing.EncodedObject),\n\t\t},\n\t\tModuleStorage: make(ModuleStorage),\n\t}\n}", "func NewStorage() *Storage {\n\tConfig := util.NewConfig()\n\tses, err := mgo.Dial(string(Config.Mongo.Addr))\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn &Storage{database: Config.Mongo.DB, table: Config.Mongo.Table, session: ses}\n}", "func NewStorage(path string) (*Storage, error) {\n\tdb, err := bolt.Open(path, 0600, nil)\n\tif err != nil {\n\t\treturn nil, common.Error(common.InitFailed, err)\n\t}\n\tstore := &Storage{db}\n\treturn store, nil\n}", "func NewMapFunc(t mockConstructorTestingTNewMapFunc) *MapFunc {\n\tmock := &MapFunc{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func NewMap(client *clientv3.Client, path string, value interface{}, codec Codec, handlers ...EventHandler) (Map, error) {\n\t// Check value type.\n\ttyp := reflect.TypeOf(value)\n\tif typ.Kind() != reflect.Ptr {\n\t\treturn nil, ErrValueTypeNotPointer\n\t}\n\t// Create cache.\n\treturn newCache(client, path, typ.Elem(), codec, handlers...)\n}", "func newStorageLayer(disk string) (storage StorageAPI, err error) {\n\tif !strings.ContainsRune(disk, ':') || filepath.VolumeName(disk) != \"\" {\n\t\t// Initialize filesystem storage API.\n\t\treturn newPosix(disk)\n\t}\n\t// Initialize rpc client storage API.\n\treturn newRPCClient(disk)\n}", "func New(db *mgo.Database, collectionPrefix string) *Storage {\n\ts := &Storage{\n\t\tfs: db.GridFS(collectionPrefix),\n\t}\n\t// TODO check error\n\ts.fs.Files.EnsureIndex(mgo.Index{\n\t\tKey: []string{\"filename\"},\n\t\tUnique: true,\n\t})\n\treturn s\n}", "func New() SharedMap {\n\tsm := sharedMap{\n\t\tm: make(map[string]interface{}),\n\t\tc: make(chan command),\n\t}\n\tgo sm.run()\n\treturn sm\n}", "func New() hctx.Map {\n\treturn hctx.Map{\n\t\tPathForKey: PathFor,\n\t}\n}", "func (m *LocalManager) New(ctx context.Context, id string) (linker.Storage, error) {\n\tdb, err := NewLocalStorage(ctx, fmt.Sprintf(\"%s/db-%s\", m.path, id))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func NewStorage(conf *ConfService) (*goejdb.Ejdb, error) {\n\tvar storageConf StorageConfiguration\n\tconf.Get(&storageConf)\n\n\t// Create a new database file and open it\n\tdb, err := goejdb.Open(storageConf.DBName, goejdb.JBOWRITER|goejdb.JBOCREAT)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func NewMap(div js.Value, options map[string]interface{}) *Map {\n\tif div == js.Null() {\n\t\tpanic(\"leaflet: cannot use null map div\")\n\t}\n\tif nodeName := div.Get(\"nodeName\").String(); nodeName != \"DIV\" {\n\t\tpanic(\"leaflet: map div nodeName should be DIV but is \" + nodeName)\n\t}\n\tInitialize()\n\treturn &Map{\n\t\tValue: L.Call(\"map\", div, options),\n\t}\n}", "func NewStorage(cat *repository.MongoCatalog, cache *rediscache.Redis) *Storage {\n\treturn &Storage{\n\t\tcat,\n\t\tcache,\n\t}\n}", "func NewMap(ctx *pulumi.Context,\n\tname string, args *MapArgs, opts ...pulumi.ResourceOption) (*Map, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Configuration == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Configuration'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Map\n\terr := ctx.RegisterResource(\"aws-native:location:Map\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func New(configs ...Configurator) (*Storage, error) {\n\tinstance := &Storage{}\n\tfor _, configure := range configs {\n\t\tif err := configure(instance); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn instance, nil\n}", "func NewMemoryStorage(features []Feature) *MemoryStorage {\n\tm := &MemoryStorage{\n\t\tfeatures: map[string]Feature{},\n\t}\n\tfor _, f := range features {\n\t\tm.features[f.Meta.Name] = f\n\t\tm.featureNames = append(m.featureNames, f.Meta.Name)\n\t}\n\n\tsort.Strings(m.featureNames)\n\treturn m\n}", "func (s *StorageBase) SetMap(ctx context.Context, sessionId string, mapData map[string]interface{}, ttl time.Duration) error {\n\treturn ErrorDisabled\n}", "func NewMapDB() kvstore.KVStore {\n\treturn &mapDB{\n\t\tm: &syncedKVMap{m: make(map[string][]byte)},\n\t}\n}", "func New(ctx context.Context, logger logger.Logger) (storage.Storage, error) {\n\tm := &memstore{\n\t\tteams: make(map[string]storage.Team),\n\t\tusersByTeam: make(map[string][]string),\n\t\tlogger: logger,\n\t}\n\n\tif err := m.EnsureAdminsTeam(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}" ]
[ "0.7882739", "0.7772161", "0.7381058", "0.72726333", "0.69259906", "0.688915", "0.67103523", "0.66192317", "0.6364578", "0.63280195", "0.6277338", "0.6212047", "0.6212047", "0.61600333", "0.6122099", "0.6101804", "0.6101486", "0.60964805", "0.6095136", "0.6051582", "0.60463166", "0.60343254", "0.5983521", "0.5977621", "0.597373", "0.5961138", "0.5949365", "0.5943946", "0.5920769", "0.591759", "0.5906262", "0.5898058", "0.5871379", "0.5870103", "0.586992", "0.58696127", "0.5860414", "0.5852803", "0.58518696", "0.58358794", "0.5833818", "0.58255434", "0.5804471", "0.5802752", "0.57993686", "0.57912534", "0.5789915", "0.57862353", "0.577605", "0.57750696", "0.5771188", "0.5759687", "0.5759605", "0.5754094", "0.57513595", "0.57341623", "0.572618", "0.5717729", "0.571627", "0.57012", "0.5698819", "0.56929827", "0.5691437", "0.56904465", "0.5685685", "0.56725204", "0.565768", "0.5655121", "0.5654221", "0.5654143", "0.5651839", "0.56501794", "0.5648154", "0.5646532", "0.5641895", "0.56411314", "0.5636455", "0.56291497", "0.562829", "0.5628236", "0.56188846", "0.5612625", "0.5611575", "0.56028193", "0.5586778", "0.55801934", "0.55700845", "0.55499566", "0.5541206", "0.5540922", "0.5536134", "0.5533469", "0.5533118", "0.55279475", "0.552539", "0.5518358", "0.55165744", "0.5499879", "0.54964244", "0.54950005" ]
0.76322776
2
Set implements the IDStore interface
func (s *MapStorage) Set(channel string, replayID int) { s.lock.Lock() defer s.lock.Unlock() s.store[channel] = replayID }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (DummyStore) Set(key string, value interface{}) error {\n\treturn nil\n}", "func (s Store) Set(ctx context.Context, key int64, value string) error {\n\tconn := s.Pool.Get()\n\tdefer conn.Close()\n\n\t_, err := conn.Do(\"SET\", key, value)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *CustomizeRdsStore) Set(id string, value string)(err error) {\n\terr = r.Context.CacheClient.Set(context.Background(), id, value, time.Minute*10).Err()\n\tif err != nil {\n\t\tr.Context.Error(map[string]interface{}{\n\t\t\t\"message\": \"auth.AuthLogin\",\n\t\t\t\"error\": err,\n\t\t})\n\t}\n\treturn\n}", "func (s *customizeRdsStore) Set(id string, value string) {\n\terr := s.redisClient.Set(id, value, time.Minute*10).Err()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (store *EntryStore) Set(id []byte, entry *hexalog.Entry) error {\n\tval, err := proto.Marshal(entry)\n\tif err == nil {\n\t\terr = store.db.Put(store.wo, id, val)\n\t}\n\treturn err\n}", "func (rs *Store) Set(ctx context.Context, key, value interface{}) error {\n\trs.lock.Lock()\n\tdefer rs.lock.Unlock()\n\trs.values[key] = value\n\treturn nil\n}", "func (cs *ClientStore) Set(id string, cli oauth2.Client) (err error) {\n\tcs.Lock()\n\tdefer cs.Unlock()\n\tcs.data[id] = cli\n\treturn\n}", "func (s *Store) Set(entity workloadmeta.Entity) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\tentityID := entity.GetID()\n\n\tif _, ok := s.store[entityID.Kind]; !ok {\n\t\ts.store[entityID.Kind] = make(map[string]workloadmeta.Entity)\n\t}\n\n\ts.store[entityID.Kind][entityID.ID] = entity\n}", "func (cs *ClientStore) Set(id string, cli oauth2.ClientInfo) (err error) {\n\tcs.Lock()\n\tdefer cs.Unlock()\n\tcs.data[id] = cli\n\treturn\n}", "func (db *DB) Set(id int, data *[]byte) {\n\tdb.Lock()\n\tdefer db.Unlock()\n\tdb.docs[id] = *data\n\t// Make sure ID is in the set. Needed when a DB is loaded from file.\n\tdb.all.Add(id)\n\t// Always update db.idMax to the highest Id number\n\tif id > db.idMax {\n\t\tdb.idMax = id\n\t}\n}", "func (mavls *Store) Set(datas *types.StoreSet, sync bool) ([]byte, error) {\r\n\treturn mavl.SetKVPair(mavls.GetDB(), datas, sync)\r\n}", "func (s *Store) Set(k string, v interface{}) error {\n\treturn s.SetEx(k, v, 0)\n}", "func (s *Store) Set(k string, v interface{}) error {\n\treturn s.SetEx(k, v, 0)\n}", "func (i *InmemStore) Set(key []byte, val []byte) error {\n\ti.l.Lock()\n\tdefer i.l.Unlock()\n\ti.kv[string(key)] = val\n\treturn nil\n}", "func (l *LocalStore) Set(ctx context.Context, key, data string) error {\n\tl.lock.Lock()\n\tl.store[key] = data\n\tl.lock.Unlock()\n\n\treturn nil\n}", "func (s *Store) Set(id string, url string, meta *store.Meta) error {\n\tmetaJSON, err := json.Marshal(meta)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't unmarshal meta: %v\", err)\n\t}\n\tconn := s.pool.Get()\n\tdefer conn.Close()\n\tconn.Send(\"MULTI\")\n\t// Store full url and meta in hashmap.\n\tconn.Send(\"HMSET\", s.idKey(id), \"url\", url, \"meta\", metaJSON)\n\t// Store a reverse map of full url and id.\n\tconn.Send(\"SET\", s.urlKey(url), id)\n\trep, err := redis.Values(conn.Do(\"EXEC\"))\n\t// Check if there are any errors.\n\tfor _, r := range rep {\n\t\tif _, ok := r.(redis.Error); ok {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (store *RedisStore) Set(key string, contents []byte) error {\n\terr := store.Client.Set(key, contents, 0).Err()\n\treturn err\n}", "func (s *mem) Set(sid string, key string, value interface{}, _ time.Duration, immutable bool) error {\n\ts.mu.RLock()\n\tstore, ok := s.values[sid]\n\ts.mu.RUnlock()\n\tif ok {\n\t\tstore.Save(key, value, immutable)\n\t}\n\n\treturn nil\n}", "func (s *MongodbStore) Set(key, val interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.data[key] = val\n\treturn nil\n}", "func (s *Store) Set(key string, value interface{}) {\n\ts.access.Lock()\n\tdefer s.access.Unlock()\n\n\ts.data[key] = value\n\ts.sendDataChanged()\n\ts.doKeyChanged(key)\n}", "func (client *MemcachedClient4T) Set(e *common.Element) error {\n\treturn client.store(\"set\", e)\n}", "func (me *TartIdTypeInt) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (s *Store) Set(session *entities.Session) error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\ts.sess[session.ClientID] = session\n\treturn nil\n}", "func (s *LevelDBStore) Set(key string, value []byte) {\n\t_ = s.db.Put([]byte(key), value, nil)\n}", "func (me *TArtIdTypeUnion4) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (me *TArtIdTypeUnion) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (r *RedisStore) Set(key, value string) error {\n\tif _, err := r.conn.Do(\"SET\", key, value); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (me *TartIdType) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (me *TArtIdTypeUnion2) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (me *TSAFPTGLAccountID) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (s *Store) Set(key string, value string) {\r\n\ts.store[key] = value\r\n}", "func (me *TArtIdTypeUnion1) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (st *SessionStoreMySQL) Set(key, value interface{}) error {\n\tst.lock.Lock()\n\tdefer st.lock.Unlock()\n\tst.values[key] = value\n\treturn nil\n}", "func (st *Store) Set(key, value []byte) error {\n\tst.lock.Lock()\n\tdefer st.lock.Unlock()\n\tif err := st.beforeSet(key, value); err != nil {\n\t\treturn err\n\t}\n\tif st.compression {\n\t\tvalue = compress(value)\n\t}\n\terr := st.store.Put(key, value, ds.ItemOptions{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn st.writeToLogFile(key, value)\n}", "func (s *VirtualStore) Set(key, val any) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.data[key] = val\n\treturn nil\n}", "func (m *RedisStore) Set(key interface{}, value interface{}) error {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\tm.sessionData[key] = value\n\treturn nil\n}", "func (mvccs *KVMVCCStore) Set(datas *types.StoreSet, hash []byte, sync bool) ([]byte, error) {\n\tif hash == nil {\n\t\thash = calcHash(datas)\n\t}\n\tkvlist, err := mvccs.mvcc.AddMVCC(datas.KV, hash, datas.StateHash, datas.Height)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmvccs.saveKVSets(kvlist, sync)\n\treturn hash, nil\n}", "func Set(c *template.Context, value interface{}) error {\n\tid := c.ID()\n\n\t// Get the store.\n\ts, err := getStore(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// The context has to be locked to perform any changes.\n\tif !IsLocked(c) {\n\t\treturn fmt.Errorf(\"store.Set: can't set store data if the context is not locked by the session!\")\n\t}\n\n\t// Lock the mutex.\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\t// Create the map if it doesn't exists.\n\ts.data.createMapIfNil()\n\n\t// Set the value. The key is the context's ID.\n\ts.data.Values[id] = newDBStoreData(value)\n\n\t// Update data to the database.\n\terr = flushUpdatesToDB(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Broadcast changes to other sessions in edit mode.\n\tbroadcastChangedContext(id, c.Session())\n\n\treturn nil\n}", "func (s *StoreDb) Set(key, v string) error {\n\treturn s.db.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Set([]byte(key), []byte(v))\n\t})\n}", "func (b *BadgerStore) Set(key []byte, val []byte) error {\n\treturn b.conn.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Set(key, val)\n\t})\n}", "func (b *BadgerStore) Set(key []byte, val []byte) error {\n\treturn b.conn.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Set(key, val)\n\t})\n}", "func (s *shardAccessor) Set(id node.ID, hash []byte) {\n\ts.writes = append(s.writes, Node{ID: id, Hash: hash})\n}", "func (s *StorageBase) Set(ctx context.Context, sessionId string, key string, value interface{}, ttl time.Duration) error {\n\treturn ErrorDisabled\n}", "func (s *Store) Set(ctx context.Context, key interface{}, v json.Marshaler) error {\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tdefault:\n\t}\n\n\tb, err := v.MarshalJSON()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tdefault:\n\t}\n\n\ts.m[key] = entry{data: b}\n\treturn nil\n}", "func (ms *memoryStore) Set(ns *NameSpace) error {\n\terr := ms.Delete(ns)\n\tms.Lock()\n\tdefer ms.Unlock()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, prefix := range ns.Prefixes() {\n\t\tms.prefix2base[prefix] = ns\n\t}\n\tfor _, base := range ns.BaseURIs() {\n\t\tms.base2prefix[base] = ns\n\t}\n\tid := ns.GetID()\n\tms.namespaces[id] = ns\n\treturn nil\n}", "func (me *TArtIdTypeUnion3) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (d DBase) Set(table, key string, value interface{}) {\n\tCritical(d.Store.Set(table, key, value))\n}", "func (db *inMemoryDB) Set(key, val string) error {\n\tdb.lock.Lock()\n\tdefer db.lock.Unlock()\n\tdb.m[key] = val\n\treturn nil\n}", "func (l *ShardedNoTS) Set(tenantID, key string, val interface{}) error {\n\t_, ok := l.cache[tenantID]\n\tif !ok {\n\t\tl.cache[tenantID] = l.constructor()\n\t\tl.itemCount[tenantID] = 0\n\t}\n\n\tl.itemCount[tenantID]++\n\treturn l.cache[tenantID].Set(key, val)\n}", "func set(s kvs.Store) {\n\tc := context.Background()\n\tdb := &Data{}\n\n\t// Creating the first node object\n\tn := Node{\n\t\tID: \"100\",\n\t\tName: \"foobar\",\n\t\tDescription: \"This is a nice node\",\n\t}\n\tstore.Set(s, c, db, \"/db/stored/here/\", n, \"Nodes\", \"100\")\n\n\t// Creating the second node object\n\tn = Node{\n\t\tID: \"101\",\n\t\tName: \"barfoo\",\n\t\tDescription: \"This is another nice node\",\n\t}\n\tstore.Set(s, c, db, \"/db/stored/here/\", n, \"Nodes\", \"101\")\n\n\t// Creating an edge\n\tstore.Set(s, c, db, \"/db/stored/here/\", Edge{\n\t\tNodeID1: \"100\",\n\t\tNodeID2: \"101\",\n\t}, \"Edges\", \"10\")\n\n\t// Setting the QuitDemo boolean\n\tstore.Set(s, c, db, \"/db/stored/here/\", true, \"QuitDemo\")\n}", "func (l *localStore) Set(key string, b []byte) error {\n\t_, ok := l.m[key]\n\tl.m[key] = b\n\tif ok {\n\t\treturn ErrKeyExist\n\t}\n\treturn nil\n}", "func (s *MemStore) Set(key, val interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.data[key] = val\n\treturn nil\n}", "func (s *RedisClusterStore) Set(ctx context.Context, key interface{}, value interface{}, options *Options) error {\n\tif options == nil {\n\t\toptions = s.options\n\t}\n\n\terr := s.clusclient.Set(ctx, key.(string), value, options.ExpirationValue()).Err()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif tags := options.TagsValue(); len(tags) > 0 {\n\t\ts.setTags(ctx, key, tags)\n\t}\n\n\treturn nil\n}", "func (me *TArtIdTypeUnion6) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func Set(redisdb *redis.Client, sessionID, key string, value interface{}) error {\n\n\tif _, err := redisdb.HSet(sessionID, key, value).Result(); err != nil {\n\t\treturn fmt.Errorf(\"failed to write %s to session: %s\", key, err.Error())\n\t}\n\n\treturn nil\n}", "func TestGetSetID(t *testing.T) {\n\tid := ID(\"someid\")\n\tvar r Record\n\tr.Set(&id)\n\n\tvar id2 ID\n\trequire.NoError(t, r.Load(&id2))\n\tassert.Equal(t, id, id2)\n}", "func Set(key string, value interface{}) {\n\tglobalStore.Lock()\n\tdefer globalStore.Unlock()\n\n\tglobalStore.store[key] = value\n}", "func (s *PostgresStore) Set(key, value interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.data[key] = value\n\treturn nil\n}", "func (m *Memory) Set(key, value uint64) error {\n\tm.kvstore[key] = value\n\treturn nil\n}", "func (me *TArtIdTypeUnion5) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (st *MemSessionStore) Set(key, value interface{}) error {\n\tst.lock.Lock()\n\tdefer st.lock.Unlock()\n\tst.value[key] = value\n\treturn nil\n}", "func (storage *remoteStorage) Set(key *data.ID, value []byte) error {\n\tnode := storage.node\n\n\turl := fmt.Sprintf(\"http://%s/storage/%s\", node.TCPAddr().String(), key.String())\n\n\t// Base64 encoding, RFC 4648.\n\t// str := base64.StdEncoding.EncodeToString(value)\n\treq, err := http.NewRequest(http.MethodPut, url, bytes.NewReader(value))\n\tif req.Body != nil {\n\t\tdefer req.Body.Close()\n\t}\n\tif err != nil {\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\tres, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\tif res.Body != nil {\n\t\tdefer res.Body.Close()\n\t}\n\tif res.StatusCode < 200 || res.StatusCode > 299 {\n\t\terr := fmt.Errorf(\"HTTP storage Put %s -> %d %s\", url, res.StatusCode, res.Status)\n\t\tnode.disconnect(err)\n\t\treturn err\n\t}\n\treturn nil\n}", "func (m *IdentitySetCollectionResponse) SetValue(value []IdentitySetable)() {\n err := m.GetBackingStore().Set(\"value\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *IndexedMap[PrimaryKey, Value, Idx]) Set(ctx context.Context, pk PrimaryKey, value Value) error {\n\terr := m.ref(ctx, pk, value)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn m.m.Set(ctx, pk, value)\n}", "func (ms *mongoStore) Set(key []byte, data []byte, ttl time.Duration) (err error) {\n\tctx, cancel := context.WithTimeout(context.Background(), ms.timeout)\n\tdefer cancel()\n\tupsert := true\n\t_, err = ms.collection().UpdateOne(ctx, &mongoCache{\n\t\tKey: string(key),\n\t}, bson.M{\n\t\t\"$set\": &mongoCache{\n\t\t\tKey: string(key),\n\t\t\tData: data,\n\t\t\tExpiredAt: time.Now().Add(ttl),\n\t\t},\n\t}, &options.UpdateOptions{\n\t\tUpsert: &upsert,\n\t})\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func (tracer *Instance) Set(id string) {\n\ttracer.collections.Set(tracer.key(), id)\n}", "func (kv KeyValueStore) Set(key, value string) error {\n\treturn set(kv.DB, key, value)\n}", "func (bm tsidbitmap) Set(id core.TractserverID) tsidbitmap {\n\tidx := uint64(id) / 64\n\toff := uint64(id) % 64\n\tfor uint64(len(bm)) <= idx {\n\t\tbm = append(bm, 0)\n\t}\n\tbm[idx] |= uint64(1) << off\n\treturn bm\n}", "func (cs *ClientStore) Set(info oauth2.ClientInfo) (err error) {\n\tcs.cHandler(cs.ccfg.ClientsCName, func(c *mongo.Collection) {\n\t\tentity := &client{\n\t\t\tID: info.GetID(),\n\t\t\tSecret: info.GetSecret(),\n\t\t\tDomain: info.GetDomain(),\n\t\t\tUserID: info.GetUserID(),\n\t\t}\n\n\t\tif _, cerr := c.InsertOne(context.TODO(), entity); cerr != nil {\n\t\t\terr = cerr\n\t\t\treturn\n\t\t}\n\t})\n\n\treturn\n}", "func (c *StateStore) Set(req *state.SetRequest) error {\n\terr := state.CheckSetRequestOptions(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\toptions := []documentdb.CallOption{documentdb.PartitionKey(req.Key)}\n\n\tif req.ETag != \"\" {\n\t\toptions = append(options, documentdb.IfMatch((req.ETag)))\n\t}\n\tif req.Options.Consistency == state.Strong {\n\t\toptions = append(options, documentdb.ConsistencyLevel(documentdb.Strong))\n\t}\n\tif req.Options.Consistency == state.Eventual {\n\t\toptions = append(options, documentdb.ConsistencyLevel(documentdb.Eventual))\n\t}\n\n\t_, err = c.client.UpsertDocument(c.collection.Self, CosmosItem{ID: req.Key, Value: req.Value}, options...)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (t *Tkeyid) Set(key string, id uint) {\n\tt.idtokey[id] = key\n\tt.keytoid[key] = id\n}", "func (bs *badgerStore) Set(key []byte, data []byte, ttl time.Duration) (err error) {\n\treturn bs.db.Update(func(txn *badger.Txn) error {\n\t\te := badger.NewEntry(key, data).\n\t\t\tWithTTL(ttl)\n\t\treturn txn.SetEntry(e)\n\t})\n}", "func (s *RedisDeviceStore) Set(new *Device, properties ...string) (err error) {\n\tnow := time.Now()\n\tnew.UpdatedAt = now\n\tkey := fmt.Sprintf(\"%s:%s\", new.AppID, new.DevID)\n\tif new.old == nil {\n\t\tnew.CreatedAt = now\n\t}\n\terr = s.store.Set(key, *new, properties...)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn nil\n}", "func (m *PieceInMemoryStorage) Set(p *Piece) {\n\tm.storageMux.Lock()\n\tdefer m.storageMux.Unlock()\n\tm.storage[p.pieceID] = p\n}", "func (ms *MemStore) Set(key string, w io.WriterTo) error {\n\td := make(memio.Buffer, 0)\n\tif _, err := w.WriteTo(&d); err != nil && err != io.EOF {\n\t\treturn err\n\t}\n\tms.set(key, d)\n\treturn nil\n}", "func (db *TriasDB) Set(key []byte, value []byte) {\n\tdb.mtx.Lock()\n\tdefer db.mtx.Unlock()\n\n\tdb.SetNoLock(key, value)\n}", "func (i *Int64) Set(v int64) {\n\t*i = Int64(v)\n}", "func (i Index) Set(sourceID string, id int, table string, record *Record) {\n\tinfo := RecordInfo{\n\t\tID: id,\n\t\tTable: table,\n\t\tRecord: record,\n\t}\n\n\tsourceID = strings.ToLower(sourceID)\n\tsourceID = strings.TrimSpace(sourceID)\n\ti[sourceID] = info\n}", "func Set(key string, val interface{}) error {\n\treturn DB.Set(key, val)\n}", "func (self *Store) Set(k string) []byte {\n\tself.mu.Lock()\n\tself.mu.Unlock()\n\treturn nil\n}", "func (s *session) Set(key string, value interface{}) {\n\ts.mu.Lock()\n\ts.values[key] = value\n\ts.mu.Unlock()\n\ts.provider.update(s.sid)\n}", "func (d *DI) Set(name string, v interface{}) {\n\td.mutex.Lock()\n\td.store[name] = v\n\td.mutex.Unlock()\n}", "func (tx *Tx) Set(key string, value string) error {\n\te := newRecord([]byte(key), []byte(value), StringRecord, StringSet)\n\ttx.addRecord(e)\n\n\treturn nil\n}", "func (h *Int64GaugeHandle) Set(ctx context.Context, value int64) {\n\th.RecordOne(ctx, NewInt64MeasurementValue(value))\n}", "func (s *DiffStore) Set(key []byte, hash uint64, value interface{}) {\n\titem := s.tree.Get(&storeKV{key: key})\n\n\tif item == nil {\n\t\ts.tree.ReplaceOrInsert(&storeKV{\n\t\t\tkey: key,\n\t\t\thash: hash,\n\t\t\tvalue: value,\n\t\t\tstate: ItemChanged,\n\t\t})\n\t\treturn\n\t}\n\n\tv := item.(*storeKV)\n\n\tif v.hash == hash {\n\t\tif v.state == ItemDeleted {\n\t\t\tv.value = value\n\t\t\tv.state = ItemUnchanged\n\t\t}\n\t\treturn\n\t}\n\n\tv.hash = hash\n\tv.value = value\n\tv.state = ItemChanged\n}", "func (t *Tenants) Set(tenant *Tenant) error {\n\tw := &bytes.Buffer{}\n\tif err := tenant.Encode(w); err != nil {\n\t\treturn err\n\t}\n\treturn t.store.Set(tenant.ID, w.Bytes())\n}", "func (r userQueryIDString) Set(value string) userSetParams {\n\n\treturn userSetParams{\n\t\tdata: builder.Field{\n\t\t\tName: \"id\",\n\t\t\tValue: value,\n\t\t},\n\t}\n\n}", "func (kvs *MemoryKeyValueStore) Set(key string, value []byte) error {\n\tkvs.mu.Lock()\n\tdefer kvs.mu.Unlock()\n\tkvs.m[key] = value\n\treturn nil\n}", "func (r *Redis) Set(key cipher.SHA256, val []byte) (*data.Object, error) {\n\treturn r.SetIncr(key, val, 1)\n}", "func (r *Store) Set(name string, content []byte, reason string) error {\n\tstore := r.getStore(name)\n\treturn store.Set(strings.TrimPrefix(name, store.Alias()), content, reason)\n}", "func (i *UInt64) Set(v uint64) {\n\t*i = UInt64(v)\n}", "func (s *InMemoryHashStorage) Set(key int64, value string) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\n\ts.hashes[key] = value\n}", "func (s Storage) set(k string, v *metadata) {\n\ts.store[k] = v\n}", "func (session KeyValueSession) Set(key DatabaseKey, value interface{}) bool {\n\tbuffer := convertData(value)\n\tsession.store.tree.Set(key, buffer)\n\n\treturn true\n}", "func (m *MongoDB) Set(key string, val Model) error {\n\tsess := m.sess.Copy()\n\tdefer sess.Close()\n\tcoll := sess.DB(dbName).C(candiesColl)\n\treturn coll.Update(bson.M{\"key\": key}, objWrapper{Key: key, Data: val})\n}", "func (m *MemoryStorage) Set(e Entry) {\n\tm.mux.Lock()\n\tdefer m.mux.Unlock()\n\tm.entries[e.Key] = e\n}", "func (me *TSAFPTProductID) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (fs *FakeSession) Set(oid string, typ gosnmp.Asn1BER, val any) *FakeSession {\n\tif _, ok := fs.data[oid]; !ok {\n\t\t// new OID, need to rebuild oids\n\t\tfs.dirty = true\n\t}\n\tfs.data[oid] = gosnmp.SnmpPDU{\n\t\tName: oid,\n\t\tType: typ,\n\t\tValue: val,\n\t}\n\treturn fs\n}", "func (s *DynamoDB) Set(ctx context.Context, identifier string, findings []domain.Finding) error {\n\tb, _ := json.Marshal(fromDomain(findings))\n\t_, err := s.Client.PutItemWithContext(ctx, &dynamodb.PutItemInput{\n\t\tTableName: aws.String(s.TableName),\n\t\tItem: map[string]*dynamodb.AttributeValue{\n\t\t\ts.PartitionKeyName: {\n\t\t\t\tS: aws.String(identifier),\n\t\t\t},\n\t\t\tdataKey: {\n\t\t\t\tS: aws.String(string(b)),\n\t\t\t},\n\t\t\ts.TTLKeyName: {\n\t\t\t\tN: aws.String(fmt.Sprintf(\"%d\", time.Now().Add(time.Hour).Unix())),\n\t\t\t},\n\t\t},\n\t})\n\treturn err\n}", "func (m *MemAdmin) Set(siteID string, arec AdminRec) {\n\tm.data[siteID] = arec\n}" ]
[ "0.73111916", "0.72890264", "0.72428745", "0.7087768", "0.7085901", "0.7049794", "0.69764024", "0.6957838", "0.69094276", "0.6905412", "0.69002354", "0.688193", "0.688193", "0.68666136", "0.68094313", "0.68043345", "0.67774326", "0.6775505", "0.6751434", "0.6716437", "0.6695529", "0.66830635", "0.667903", "0.6672411", "0.6672111", "0.6659058", "0.6655904", "0.6655896", "0.66472435", "0.66396767", "0.6637304", "0.6616971", "0.66123646", "0.660833", "0.65827274", "0.65662134", "0.656002", "0.65343696", "0.6516802", "0.6507084", "0.6507084", "0.65059114", "0.6485762", "0.6477832", "0.64776635", "0.64761156", "0.6466963", "0.646395", "0.6463094", "0.64535356", "0.64422625", "0.6431719", "0.6428306", "0.6419551", "0.6410105", "0.63897246", "0.63862115", "0.6379655", "0.6362725", "0.6361269", "0.63474846", "0.6346299", "0.63329536", "0.6331613", "0.6327607", "0.6317851", "0.6310656", "0.62923956", "0.62850446", "0.627728", "0.6274693", "0.627133", "0.6248005", "0.62437385", "0.62371874", "0.62360704", "0.62331563", "0.6226376", "0.62244904", "0.62229705", "0.6222706", "0.6220877", "0.62187284", "0.6197957", "0.61944723", "0.6192833", "0.6185559", "0.617883", "0.61775154", "0.61701196", "0.6168059", "0.61585927", "0.6149529", "0.61461145", "0.6145643", "0.6144494", "0.6127836", "0.61276466", "0.61168796", "0.6115494" ]
0.6730087
19
Get implements the IDStore interface
func (s *MapStorage) Get(channel string) (replayID int, ok bool) { s.lock.RLock() defer s.lock.RUnlock() replayID, ok = s.store[channel] return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *baseStore[T, E, TPtr, EPtr]) Get(ctx context.Context, id int64) (T, error) {\n\treturn s.FindOne(ctx, db.FindQuery{Where: gosql.Column(\"id\").Equal(id)})\n}", "func (s Store) Get(id []byte) (perm.AccessControl, error) {\n\treturn accessControl{}, nil\n}", "func (s *Store) Get(rid ffs.RetrievalID) (ffs.RetrievalInfo, error) {\n\tvar ri ffs.RetrievalInfo\n\tbuf, err := s.ds.Get(makeKey(rid))\n\tif err == datastore.ErrNotFound {\n\t\treturn ri, ErrNotFound\n\t}\n\tif err != nil {\n\t\treturn ri, fmt.Errorf(\"getting retrieval info from datastore: %s\", err)\n\t}\n\tif err := json.Unmarshal(buf, &ri); err != nil {\n\t\treturn ri, fmt.Errorf(\"unmarshaling retrieval info from datastore: %s\", err)\n\t}\n\treturn ri, nil\n}", "func (mp *Provider) Get(sessionID []byte) (session.Storer, error) {\n\tcurrentStore := mp.memoryDB.GetBytes(sessionID)\n\tif currentStore != nil {\n\t\treturn currentStore.(*Store), nil\n\t}\n\n\tnewStore := mp.acquireStore(sessionID, mp.expiration)\n\tmp.memoryDB.SetBytes(sessionID, newStore)\n\n\treturn newStore, nil\n}", "func (m *MemoryStore) Get(id string) (string, error) {\n\treturn m.URLs[id], nil\n}", "func (v *Venom) Get(key string) interface{} {\n\tval, _ := v.Store.Find(key)\n\treturn val\n}", "func (s *Store) Get(id string) (e Entry, exists bool) {\n\te, exists = (*s)[id]\n\treturn\n}", "func (s *Store) Get(id packet.ID) *Future {\n\t// acquire mutex\n\ts.mutex.RLock()\n\tdefer s.mutex.RUnlock()\n\n\treturn s.store[id]\n}", "func (self *CassandraMetaStore) Get(v *meta.RequestVars) (*meta.Object, error) {\n\treturn self.findOid(v.Oid, false)\n}", "func (r *CustomizeRdsStore) Get(id string, clear bool) (value string) {\n\tctx := context.Background()\n\tval, err := r.Context.CacheClient.Get(ctx, id).Result()\n\tif err != nil {\n\t\tr.Context.Error(map[string]interface{}{\n\t\t\t\"message\": \"auth.AuthLogin\",\n\t\t\t\"error\": err,\n\t\t})\n\t\treturn\n\t}\n\tif !clear {\n\t\treturn val\n\t}\n\terr = r.Context.CacheClient.Del(ctx, id).Err()\n\tif err != nil {\n\t\tr.Context.Error(map[string]interface{}{\n\t\t\t\"message\": \"auth.AuthLogin\",\n\t\t\t\"error\": err,\n\t\t})\n\t\treturn\n\t}\n\treturn val\n}", "func (c *dynamicClient) Get(uid meta.UID) (meta.Object, error) {\n\treturn c.storage.GetByID(c.kind, uid)\n}", "func (ms *mysqlstore) Get(id string) (record store.Record, err error) {\n\tdefer func(t0 time.Time) {\n\t\terrStr := \"\"\n\t\tif err != nil {\n\t\t\terrStr = err.Error()\n\t\t}\n\t\tms.logger.Printf(\"%v\",\n\t\t\tlogrec{\n\t\t\t\tService: \"mysql\",\n\t\t\t\tOperation: \"get\",\n\t\t\t\tID: id,\n\t\t\t\tError: errStr,\n\t\t\t\tDuration: fmt.Sprintf(\"%v\", time.Since(t0)),\n\t\t\t},\n\t\t)\n\t}(time.Now())\n\n\t// get by id\n\n\treturn\n}", "func (s *Store) Get(_ context.Context, id string) (*rpc.Meta, error) {\n\tresponse, ok := s.metadata.Load(id)\n\tif !ok {\n\t\treturn nil, &v1.NotFoundError{Link: &v1.Link{Hash: id}, Err: fmt.Errorf(\"Not found id: %s\", id)}\n\t}\n\n\tv, ok := response.(*rpc.Meta)\n\tif !ok {\n\t\treturn nil, &v1.NotFoundError{Link: &v1.Link{Hash: id}, Err: fmt.Errorf(\"Not found id: %s\", id)}\n\t}\n\n\treturn v, nil\n}", "func (s *customizeRdsStore) Get(id string, clear bool) (value string) {\n\tval, err := s.redisClient.Get(id).Result()\n\tif err != nil {\n\t\t// TODO 这里区分不存在还是出错\n\t\treturn \"\"\n\t}\n\tif clear {\n\t\terr := s.redisClient.Del(id).Err()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\treturn val\n}", "func (s *Store) Get(key string) interface{} {\n\treturn s.data.Get(key)\n}", "func (store *payloadStore) Get(id string) *Payload {\n\tstore.RLock()\n\tdefer store.RUnlock()\n\treturn store.m[id]\n}", "func (s *Store) Get(argID string) (value string, keyExists bool) {\r\n\tvalue, keyExists = s.store[argID]\r\n\tif keyExists {\r\n\t\treturn value, true\r\n\t}\r\n\tif s.parent != nil {\r\n\t\treturn s.parent.Get(argID)\r\n\t}\r\n\treturn value, false\r\n}", "func (c Client) Get(name string) (string, error) {\n\treturn c.Store.Get(c.ctx, name)\n}", "func (c *crdBackend) Get(ctx context.Context, key allocator.AllocatorKey) (idpool.ID, error) {\n\tidentity := c.get(ctx, key)\n\tif identity == nil {\n\t\treturn idpool.NoID, nil\n\t}\n\n\tid, err := strconv.ParseUint(identity.Name, 10, 64)\n\tif err != nil {\n\t\treturn idpool.NoID, fmt.Errorf(\"unable to parse value '%s': %s\", identity.Name, err)\n\t}\n\n\treturn idpool.ID(id), nil\n}", "func (s *DIDStore) Get(id string) (*did.Doc, error) {\n\tif id == \"\" {\n\t\treturn nil, errors.New(\"ID is mandatory\")\n\t}\n\n\tdeltas, err := s.getDeltas(id)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"delta data fetch from store failed: %w\", err)\n\t}\n\n\t// TODO construct document from all the deltas (https://github.com/hyperledger/aries-framework-go/issues/54)\n\t// For now, assume storage contains only one delta(genesis document)\n\tdelta := deltas[0]\n\n\tdoc, err := base64.URLEncoding.DecodeString(delta.Change)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"decoding of document delta failed: %w\", err)\n\t}\n\n\tdocument, err := did.ParseDocument(doc)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"document ParseDocument() failed: %w\", err)\n\t}\n\n\treturn document, nil\n}", "func (s *DatastoreStore) Get(r *http.Request, name string) (*sessions.Session,\n\terror) {\n\treturn sessions.GetRegistry(r).Get(s, name)\n}", "func Get() *Store {\n\treturn coreStore\n}", "func (store *KapacitorStore) Get(ctx context.Context, id int) (chronograf.Server, error) {\n\tif store.Kapacitor == nil || store.Kapacitor.ID != id {\n\t\treturn chronograf.Server{}, fmt.Errorf(\"unable to find Kapacitor with id %d\", id)\n\t}\n\treturn *store.Kapacitor, nil\n}", "func (DummyStore) Get(key string) (interface{}, error) {\n\treturn nil, nil\n}", "func (s *inMemorySessionStore) Get(id string) *USSDSession {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\te := s.store[id]\n\tif e != nil {\n\t\ts.gcList.MoveToFront(e)\n\t\tif session, ok := e.Value.(*USSDSession); ok {\n\t\t\treturn session\n\t\t} else {\n\t\t\tpanic(\"Data Store corrupted: non-string key value in garbage collector\")\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn nil\n}", "func (s *Store) Get(clientID string) (*entities.Session, error) {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\treturn s.sess[clientID], nil\n}", "func (m *MongoStore) Get(r *http.Request, id string) (map[string]interface{}, error) {\n\t// get data in db\n\tms := m.Ms.Clone()\n\tdefer ms.Close()\n\tmc := ms.DB(\"\").C(m.C)\n\n\tmp := make(map[string]interface{})\n\tif err := mc.FindId(id).One(&mp); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn mp, nil\n}", "func (s *Store) Get(string) ([]byte, error) {\n\treturn s.GetReturn, s.ErrGet\n}", "func (s *Store) Get(key string) (v interface{}, ok bool) {\n\tv, ok = s.store.GetPtrOk(key)\n\treturn\n}", "func (d *DI) Get(name string) interface{} {\n\td.mutex.RLock()\n\tv := d.store[name]\n\td.mutex.RUnlock()\n\treturn v\n}", "func (idGen *IdGenerator) Get() uint64 {\n\tbuf, err := getIdBuffer(idGen)\n\tif err != nil {\n\t\treturn 0\n\t}\n\treturn uint64(binary.LittleEndian.Uint64(buf))\n}", "func (d *Datastore) Get() Object {\n\td.localLock.RLock()\n\tdefer d.localLock.RUnlock()\n\treturn d.meta.object\n}", "func (m *MemEDVStore) Get(k string) ([]byte, error) {\n\treturn m.coreStore.Get(k)\n}", "func (s MockStore) Get(id int) (u User, err error) {\n\tu, ok := s.id[id]\n\tif !ok {\n\t\terr = errors.New(\"User not found in memory store.\")\n\t}\n\n\treturn u, err\n}", "func (s Store) Get(ctx context.Context, key int64) (string, error) {\n\tconn := s.Pool.Get()\n\tdefer conn.Close()\n\n\treply, err := redis.String(conn.Do(\"GET\", key))\n\treturn reply, err\n}", "func (s *Store) Get(id string) (string, *store.Meta, error) {\n\tconn := s.pool.Get()\n\tdefer conn.Close()\n\tvals, err := redis.Values(conn.Do(\"HMGET\", s.idKey(id), \"url\", \"meta\"))\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tvar (\n\t\turl string\n\t\tmetaJSON []byte\n\t)\n\tif _, err := redis.Scan(vals, &url, &metaJSON); err != nil {\n\t\treturn \"\", nil, err\n\t}\n\tif url == \"\" {\n\t\treturn \"\", nil, store.ErrNotFound\n\t}\n\tmeta := &store.Meta{}\n\tif err := json.Unmarshal(metaJSON, meta); err != nil {\n\t\treturn \"\", nil, fmt.Errorf(\"couldn't unmarshal meta: %v\", err)\n\t}\n\treturn url, meta, nil\n}", "func (store TodoStore) Get(id string) (gtimer.Todo, error) {\n\tif todo, ok := store[id]; ok {\n\t\treturn todo, nil\n\t}\n\treturn gtimer.Todo{}, gtimer.ErrNotFound\n}", "func Get(id string, m Model) error {\n\tmodel, err := Store.Get(id, m)\n\tm.Convert(model)\n\treturn err\n}", "func (w *Wallet) Get(label string) (Identity, error) {\n\tcontent, err := w.store.Get(label)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar data map[string]interface{}\n\tif err := json.Unmarshal(content, &data); err != nil {\n\t\treturn nil, errors.Wrap(err, \"Invalid identity format\")\n\t}\n\n\tidType, ok := data[\"type\"].(string)\n\n\tif !ok {\n\t\treturn nil, errors.New(\"Invalid identity format: missing type property\")\n\t}\n\n\tvar id Identity\n\n\tswitch idType {\n\tcase x509Type:\n\t\tid = &X509Identity{}\n\tdefault:\n\t\treturn nil, errors.New(\"Invalid identity format: unsupported identity type: \" + idType)\n\t}\n\n\treturn id.fromJSON(content)\n}", "func (m *MemoryCache) Get(id string) (*Session, error) {\n\tm.mx.RLock()\n\ts, ok := m.store[id]\n\tif !ok {\n\t\tm.mx.RUnlock()\n\t\treturn nil, ErrNotFound\n\t}\n\tif !s.Valid() {\n\t\t// We have to upgrade the lock. There's no harm in a yield between.\n\t\tm.mx.RUnlock()\n\t\tm.mx.Lock()\n\t\tdelete(m.store, id)\n\t\tm.mx.Unlock()\n\t\treturn nil, ErrExpired\n\t}\n\tm.mx.RUnlock()\n\treturn s, nil\n}", "func (m *CacheStore) Get(id *fields.QualifiedHash) (forest.Node, bool, error) {\n\treturn m.getUsingFuncs(id, m.Cache.Get, m.Back.Get)\n}", "func (s *MongodbStore) Get(key interface{}) interface{} {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\n\treturn s.data[key]\n}", "func (s *ServersStore) Get(ctx context.Context, id int) (chronograf.Server, error) {\n\terr := validOrganization(ctx)\n\tif err != nil {\n\t\treturn chronograf.Server{}, err\n\t}\n\n\td, err := s.store.Get(ctx, id)\n\tif err != nil {\n\t\treturn chronograf.Server{}, err\n\t}\n\n\tif d.Organization != s.organization {\n\t\treturn chronograf.Server{}, chronograf.ErrServerNotFound\n\t}\n\n\treturn d, nil\n}", "func (db *DB) Get(ctx context.Context, id string) (rec *databroker.Record, err error) {\n\tc := db.pool.Get()\n\t_, span := trace.StartSpan(ctx, \"databroker.redis.Get\")\n\tdefer span.End()\n\tdefer recordOperation(ctx, time.Now(), \"get\", err)\n\tdefer c.Close()\n\n\tb, err := redis.Bytes(c.Do(\"HGET\", db.recordType, id))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db.toPbRecord(b)\n}", "func (store *EntryStore) Get(id []byte) (*hexalog.Entry, error) {\n\tsl, err := store.db.Get(store.ro, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer sl.Free()\n\n\tdata := sl.Data()\n\tif data == nil {\n\t\treturn nil, hexatype.ErrEntryNotFound\n\t}\n\n\tvar entry hexalog.Entry\n\terr = proto.Unmarshal(data, &entry)\n\treturn &entry, err\n}", "func (tracer *Instance) Get() (id string) {\n\tid, ok := tracer.collections.Get(tracer.key())\n\tif ok {\n\t\treturn id\n\t}\n\tid = uuid.NewV4().String()\n\ttracer.Set(id)\n\treturn\n}", "func (so *SQLOrderItemStore) Get(id string) (*model.OrderItem, error) {\n\toi, err := so.SQLStore.Tx.Get(model.OrderItem{}, id)\n\treturn oi.(*model.OrderItem), err\n}", "func (ds *gcdatastore) Get(c context.Context, k *datastore.Key, dest interface{}) (err error) {\n\treturn ds.Client.Get(c, k, dest)\n}", "func (rs *Store) Get(ctx context.Context, key interface{}) interface{} {\n\trs.lock.RLock()\n\tdefer rs.lock.RUnlock()\n\tif v, ok := rs.values[key]; ok {\n\t\treturn v\n\t}\n\treturn nil\n}", "func (a *Adapter) Get(key uint64) ([]byte, bool) {\n\ta.mutex.RLock()\n\tresponse, ok := a.store[key]\n\ta.mutex.RUnlock()\n\n\tif ok {\n\t\treturn response, true\n\t}\n\n\treturn nil, false\n}", "func (z *subscriptionStore) Get(tx Transaction, id ...string) *Subscription {\n\tcompoundID := \"\"\n\tswitch len(id) {\n\tcase 1:\n\t\tcompoundID = id[0]\n\tcase 2:\n\t\tcompoundID = keyEncode(id...)\n\tdefault:\n\t\treturn nil\n\t}\n\tbData := tx.Bucket(bucketData, entitySubscription)\n\tif data := bData.Get([]byte(compoundID)); data != nil {\n\t\tsubscription := &Subscription{}\n\t\tif err := subscription.decode(data); err == nil {\n\t\t\treturn subscription\n\t\t}\n\t}\n\treturn nil\n}", "func (e *Store) Get(id string) *Config {\n\te.RLock()\n\tres := e.commands[id]\n\te.RUnlock()\n\treturn res\n}", "func (db *DB) Get(_ context.Context, id string) (*databroker.Record, error) {\n\tc := db.pool.Get()\n\tdefer c.Close()\n\n\tb, err := redis.Bytes(c.Do(\"HGET\", db.recordType, id))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db.toPbRecord(b)\n}", "func (s *LDBStore) Get(_ context.Context, addr Address) (chunk Chunk, err error) {\n\tmetrics.GetOrRegisterCounter(\"ldbstore.get\", nil).Inc(1)\n\tlog.Trace(\"ldbstore.get\", \"key\", addr)\n\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\treturn s.get(addr)\n}", "func (rc *Store) Get(key string) []byte {\n\tif rc.conn == nil {\n\t\tif err := rc.connectInit(); err != nil {\n\t\t\treturn nil\n\t\t\t//return err\n\t\t}\n\t}\n\tif item, err := rc.conn.Get(key); err == nil {\n\t\treturn item.Value\n\t}\n\treturn nil\n}", "func (s *Store) Get(key string) []byte {\n\n\tval, _ := s.objects.Load(key)\n\n\treturn val.([]byte)\n}", "func (s *dbstore) Get(key string) (*api.Value, error) {\n\tdata, err := fetchData(s.db, string(encodeKey(key, time.Time{})))\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Failed to load key [%s] from db\", key)\n\t}\n\n\tif data != nil {\n\t\tval := &api.Value{Value: []byte(data.Data), TxID: data.TxnID, ExpiryTime: time.Unix(0, data.Expiry)}\n\t\treturn val, nil\n\t}\n\n\treturn nil, nil\n}", "func (s *boltStore) Get(sourceID string) (map[string]interface{}, error) {\n\titems := map[string]interface{}{}\n\terr := s.db.View(func(tx *bolt.Tx) error {\n\t\tbucket := tx.Bucket([]byte(boltBucket))\n\t\tvalue := bucket.Get([]byte(sourceID))\n\n\t\tif value != nil {\n\t\t\tif err := json.Unmarshal(value, &items); err != nil {\n\t\t\t\treturn fmt.Errorf(\"error unmarshaling json: %s\", err)\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"tx error: %s\", err)\n\t}\n\n\treturn items, nil\n}", "func (s *RedisDeviceStore) Get(appID, devID string) (*Device, error) {\n\tdeviceI, err := s.store.Get(fmt.Sprintf(\"%s:%s\", appID, devID))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif device, ok := deviceI.(Device); ok {\n\t\treturn &device, nil\n\t}\n\treturn nil, errors.New(\"Database did not return a Device\")\n}", "func (self *LevelDBStore) Get(key []byte) ([]byte, error) {\n\tvalue, err := self.db.Get(key, nil)\n\tif err != nil {\n\t\tif err == leveldb.ErrNotFound {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, err\n\t}\n\treturn value, nil\n}", "func (s MyEntityManager) Get(id uint64) ecs.Entity {\n\treturn *s.items[id].entity\n}", "func (store *DynamoDBFeatureStore) Get(kind ld.VersionedDataKind, key string) (ld.VersionedData, error) {\n\tresult, err := store.Client.GetItem(&dynamodb.GetItemInput{\n\t\tTableName: aws.String(store.Table),\n\t\tConsistentRead: aws.Bool(true),\n\t\tKey: map[string]*dynamodb.AttributeValue{\n\t\t\ttablePartitionKey: {S: aws.String(kind.GetNamespace())},\n\t\t\ttableSortKey: {S: aws.String(key)},\n\t\t},\n\t})\n\tif err != nil {\n\t\tstore.Logger.Printf(\"ERROR: Failed to get item (key=%s): %s\", key, err)\n\t\treturn nil, err\n\t}\n\n\tif len(result.Item) == 0 {\n\t\tstore.Logger.Printf(\"DEBUG: Item not found (key=%s)\", key)\n\t\treturn nil, nil\n\t}\n\n\titem, err := unmarshalItem(kind, result.Item)\n\tif err != nil {\n\t\tstore.Logger.Printf(\"ERROR: Failed to unmarshal item (key=%s): %s\", key, err)\n\t\treturn nil, err\n\t}\n\n\tif item.IsDeleted() {\n\t\tstore.Logger.Printf(\"DEBUG: Attempted to get deleted item (key=%s)\", key)\n\t\treturn nil, nil\n\t}\n\n\treturn item, nil\n}", "func (i *InmemStore) Get(key []byte) ([]byte, error) {\n\ti.l.RLock()\n\tdefer i.l.RUnlock()\n\treturn i.kv[string(key)], nil\n}", "func (st *MemSessionStore) Get(key interface{}) interface{} {\n\treturn (*session.MemSessionStore)(st).Get(context.Background(), key)\n}", "func (s *store) Get(key string) (*api.Value, error) {\n\tlogger.Debugf(\"load key [%s] from db\", key)\n\tvalue, err := s.db.Get(encodeKey(key, time.Time{}))\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to load key [%s] from db\", key)\n\t}\n\tif value != nil {\n\t\tval, err := decodeVal(value)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to decode value [%s] for key [%s]\", value, key)\n\t\t}\n\t\treturn val, nil\n\t}\n\treturn nil, nil\n}", "func (t *Table) Get(key []byte) ([]byte, error) {\n\treturn t.Store.Get(key)\n}", "func (m *MemoryStorage) Get(key string) (interface{}, error) {\n\tm.lock.RLock()\n\tdefer m.lock.RUnlock()\n\tvalue, ok := m.store[key]\n\tif !ok {\n\t\treturn nil, ErrNotFound\n\t}\n\treturn value, nil\n}", "func Get(c *template.Context, vars ...func() interface{}) (interface{}, bool, error) {\n\t// Get the store.\n\ts, err := getStore(c)\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\n\t// The key is the context's ID.\n\tid := c.ID()\n\n\t// Lock the mutex.\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\t// Create the map if it doesn't exists.\n\ts.data.createMapIfNil()\n\n\t// Get the value.\n\tdata, ok := s.data.Values[id]\n\tif ok {\n\t\treturn data.Data, true, nil\n\t}\n\n\t// If no data is found and the create function variable\n\t// is set, then call the function and set the new value.\n\tif len(vars) > 0 {\n\t\t// The context has to be locked to perform any changes.\n\t\tif !IsLocked(c) {\n\t\t\treturn nil, false, fmt.Errorf(\"store.Get: create function: can't set store data if the context is not locked by the session!\")\n\t\t}\n\n\t\tvalue := vars[0]()\n\t\ts.data.Values[id] = newDBStoreData(value)\n\n\t\t// Update data to the database.\n\t\terr = flushUpdatesToDB(s)\n\t\tif err != nil {\n\t\t\treturn nil, false, err\n\t\t}\n\n\t\t// Broadcast changes to other sessions in edit mode.\n\t\tbroadcastChangedContext(id, c.Session())\n\n\t\treturn value, true, nil\n\t}\n\n\treturn nil, false, nil\n}", "func (ss *redisStore) Get(key string) (*storage.Secret, error) {\n\tbb, err := ss.rdb.Get(ctx, key).Bytes()\n\tif err == redis.Nil {\n\t\treturn nil, storage.ErrNoRecord\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar s storage.Secret\n\terr = json.Unmarshal(bb, &s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ts.ID = key\n\n\treturn &s, nil\n}", "func (db *SQLStore) Get(id string) (*CertRecord, error) {\n\tif err := db.conn.Ping(); err != nil {\n\t\treturn nil, err\n\t}\n\tr := &CertRecord{}\n\treturn r, db.get.Get(r, id)\n}", "func (storage *Storage) Get(key []byte) ([]byte, error) {\n\tvalue, ok := storage.Db[string(key)]\n\tvar err error\n\tif !ok {\n\t\terr = errors.E(errors.Op(\"Get value from memory database\"))\n\t}\n\treturn []byte(value), err\n}", "func (s *LevelDBStore) Get(key string) []byte {\n\tv, err := s.db.Get([]byte(key), nil)\n\tif err != nil {\n\t\treturn []byte{}\n\t}\n\treturn v\n}", "func (s *shardAccessor) Get(id node.ID) ([]byte, error) {\n\tif hash, ok := s.reads[id]; ok && hash != nil {\n\t\treturn hash, nil\n\t}\n\treturn s.w.h.hashEmpty(id), nil\n}", "func (st *Store) Get(key []byte) []byte {\n\treturn st.get(key, \"\")\n}", "func (store *Storage) Get(key []byte) ([]byte, error) {\n\tslice, err := store.db.Get(store.readOptions, key)\n\tif err != nil {\n\t\tfmt.Println(\"Get data from RocksDB failed!\")\n\t\treturn nil, err\n\t}\n\treturn slice.Data(), nil\n}", "func (a HassEntity) GetID() string { return a.ID }", "func (imd *InMemoryDb) Retrieve(id string) (interface{}, bool) {\n\tif !imd.connected {\n\t\treturn nil, false\n\t}\n\telem, ok := imd.store[id]\n\treturn elem, ok\n}", "func (p *idElementPool) get() *idElement {\n\treturn p.pool.Get().(*idElement)\n}", "func (d *Datastore) Get(key []byte) (interface{}, error) {\n\td.mutex.Lock()\n\tdefer d.mutex.Unlock()\n\n\tsKey := convertKey(key)\n\titem, ok := d.items[sKey]\n\tif !ok {\n\t\treturn nil, errors.New(\"not found\")\n\t} else {\n\t\treturn item.data, nil\n\t}\n}", "func (s *GenericStorage) Get(gvk schema.GroupVersionKind, uid runtime.UID) (runtime.Object, error) {\n\tstorageKey := KeyForUID(gvk, uid)\n\tcontent, err := s.raw.Read(storageKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.decode(content, gvk)\n}", "func (n *NodeStore) Get(key []byte) (data []byte, err error) {\n\treturn n.store.Get(n.addr, key)\n}", "func (l *LocalStore) Get(ctx context.Context, key string) (string, error) {\n\tl.lock.RLock()\n\tdata, ok := l.store[key]\n\tl.lock.RUnlock()\n\n\tif !ok {\n\t\treturn \"\", httperrs.NotFound(fmt.Errorf(\"no item found at %s\", key), \"\")\n\t}\n\treturn data, nil\n}", "func (r *Registry) get(name string) (v []byte, err error) {\n\tpair, err := r.store.Get(r.getModelName(name))\n\tif err != nil {\n\t\treturn\n\t}\n\treturn pair.Value, nil\n}", "func (s *Store) Get(item storage.Item) error {\n\tval, err := s.db.Get(key(item), nil)\n\n\tif errors.Is(err, leveldb.ErrNotFound) {\n\t\treturn storage.ErrNotFound\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = item.Unmarshal(val); err != nil {\n\t\treturn fmt.Errorf(\"failed decoding value %w\", err)\n\t}\n\n\treturn nil\n}", "func (s storeServer) Get(ref upspin.Reference) ([]byte, *upspin.Refdata, []upspin.Location, error) {\n\tif ref == accessRef {\n\t\treturn s.accessBytes, &accessRefdata, nil, nil\n\t}\n\tkey, err := refToIssueKey(ref)\n\tif err != nil {\n\t\treturn nil, nil, nil, errors.E(errors.NotExist, err)\n\t}\n\ts.mu.Lock()\n\tissue, ok := s.issue[key]\n\ts.mu.Unlock()\n\tif !ok {\n\t\treturn nil, nil, nil, errors.E(errors.NotExist)\n\t}\n\treturn issue.data, &upspin.Refdata{Reference: ref}, nil, nil\n}", "func (s *inMemoryStore) Get(ctx context.Context, h types.Metric, resetTime time.Time, fieldVals []any) any {\n\tif resetTime.IsZero() {\n\t\tresetTime = clock.Now(ctx)\n\t}\n\n\tm := s.getOrCreateData(h)\n\tt := s.findTarget(ctx, m)\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\treturn m.get(fieldVals, t, resetTime).Value\n}", "func (c *KeyStoreClient) Get(ctx context.Context, id int32) (*KeyStore, error) {\n\treturn c.Query().Where(keystore.ID(id)).Only(ctx)\n}", "func (store *GSStore) Get(key string, form GSType) (interface{}, error) {\n\n}", "func (m *Memory) Get(key uint64) (uint64, error) {\n\treturn m.kvstore[key], nil\n}", "func (ds *MySQLDatastore) Get(ctx context.Context, key []byte) ([]byte, error) {\n\trow := ds.db.QueryRow(\"SELECT value FROM extras WHERE id=?\", key)\n\n\tvar value string\n\terr := row.Scan(&value)\n\tif err == sql.ErrNoRows {\n\t\treturn nil, nil\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn []byte(value), nil\n}", "func (t *Track) Get() error {\n\n\tif err := db.Where(&t).First(&t).Error; err != nil {\n\t\tlogrus.Warn(err)\n\t\treturn err\n\t}\n\n\tlogrus.Infof(\"icii retrieved information for track #%x\", t.ID)\n\n\treturn nil\n\n}", "func (s *Store) Get(key []byte) ([]byte, error) {\n\ts.mu.RLock()\n\tdefer s.mu.RUnlock()\n\n\tif rec := s.recByKey(key, maxTX); rec != nil {\n\t\treturn rec.value()\n\t}\n\n\treturn nil, fmt.Errorf(\"key not found\")\n}", "func (t *Triplestore) getID(txn *badger.Txn, value []byte, create bool) ([]byte, error) {\n\tvar v, seek []byte\n\tvar rawid *badger.Item\n\tvar err error\n\tvar hash []byte\n\n\t// Prepend the value with our value identifier\n\tv = append(dbValue, value...)\n\tif len(value) > 39 {\n\t\t// If the value is 40 bytes or more (including the identifier) compute the hash to use as a lookup\n\t\th := sha256.Sum256(v)\n\t\thash = h[:]\n\t\tseek = append(dbHash, hash...)\n\t} else {\n\t\t// If the value is shorter the 40 bytes just store the value\n\t\tseek = v\n\t}\n\tret := make([]byte, 9)\n\n\t// Try and get the key for the value from the database\n\trawid, err = txn.Get(seek)\n\n\tif err != badger.ErrKeyNotFound {\n\t\tif err != nil {\n\t\t\t// Error for any error apart from key not found\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// We have a key to return\n\t\terr2 := rawid.Value(func(val []byte) error {\n\t\t\tret = val\n\t\t\treturn nil\n\t\t})\n\t\tif err2 != nil {\n\t\t\treturn nil, err2\n\t\t}\n\t\treturn ret, nil\n\t}\n\n\t// Error if the key is not found, and we don't have permission to create one.\n\tif !create {\n\t\treturn nil, err\n\t}\n\n\t// Create a new sequence\n\tseq, seqerr := t.db.GetSequence(dbValue, 10)\n\tif seqerr != nil {\n\t\treturn nil, seqerr\n\t}\n\tdefer seq.Release()\n\t// Get the next available ID in sequence\n\tkey, err := seq.Next()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Store the new key and value\n\treturn store(txn, key, v, hash)\n}", "func (r *RedisStore) Get(key string) (string, error) {\n\treturn redis.String(r.conn.Do(\"GET\", key))\n}", "func (rs *RedisStore) Get(sid SessionID, state interface{}) error {\n\tpipe := rs.Client.Pipeline()\n\tsc := pipe.Get(sid.getRedisKey())\n\tpipe.Expire(sid.getRedisKey(), rs.SessionDuration)\n\n\t_, err := pipe.Exec()\n\n\tif err == redis.Nil {\n\t\treturn ErrStateNotFound\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\tdata, err := sc.Bytes()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal(data, state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (f *Int64Slice) Get() interface{} {\n\treturn *f\n}", "func (s *StateStorerAdapter) Get(key string, obj interface{}) (err error) {\n\treturn s.storage.Get(newProxyItem(key, obj))\n}", "func (store *RedisStore) Get(key string) ([]byte, error) {\n\tcontent, err := store.Client.Get(key).Bytes()\n\treturn content, err\n}", "func (cs *contentStore) Get(ct ContentType, key string) ([]byte, error) {\n\treturn cs.store.Get(getContentKeyPrefix(ct, key))\n}", "func (cs *contentStore) Get(ct ContentType, key string) ([]byte, error) {\n\treturn cs.store.Get(getContentKeyPrefix(ct, key))\n}" ]
[ "0.7367177", "0.7091891", "0.70566446", "0.7002503", "0.68959856", "0.68576235", "0.6857348", "0.6855728", "0.6788413", "0.6777845", "0.67493325", "0.6737099", "0.67070603", "0.6689448", "0.66649306", "0.6660002", "0.66537136", "0.6643339", "0.66232944", "0.6616294", "0.66112965", "0.6610023", "0.6585567", "0.6574946", "0.6573787", "0.6570362", "0.65677965", "0.6560091", "0.65538585", "0.65466565", "0.65418", "0.65366477", "0.65165687", "0.650461", "0.64783424", "0.6456067", "0.6414545", "0.6404649", "0.63441783", "0.63317794", "0.63296825", "0.6326759", "0.63229525", "0.6317949", "0.6303594", "0.6302101", "0.6294134", "0.6282694", "0.6279808", "0.6274017", "0.62730324", "0.62664795", "0.62602544", "0.6258166", "0.6246301", "0.6241007", "0.62329435", "0.6231813", "0.62316567", "0.6230451", "0.62301975", "0.62217027", "0.6218411", "0.6216321", "0.6211296", "0.6210798", "0.61989194", "0.6195823", "0.6192638", "0.6191968", "0.61892", "0.6172751", "0.6158161", "0.61548924", "0.61508846", "0.61431295", "0.61310357", "0.61270505", "0.6107457", "0.6096222", "0.60938954", "0.6089953", "0.6088018", "0.60875404", "0.6083735", "0.60829234", "0.6076243", "0.6073057", "0.60670274", "0.6064558", "0.6063838", "0.6050892", "0.6049618", "0.60496116", "0.604893", "0.60488516", "0.6045699", "0.6044012", "0.6042878", "0.6042878" ]
0.62448335
55
Delete implements the IDStore interface
func (s *MapStorage) Delete(channel string) { s.lock.Lock() defer s.lock.Unlock() delete(s.store, channel) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (store *EntryStore) Delete(id []byte) error {\n\treturn store.db.Delete(store.wo, id)\n}", "func (s *baseStore[T, E, TPtr, EPtr]) Delete(ctx context.Context, id int64) error {\n\teventPtr := s.newObjectEvent(ctx, DeleteEvent)\n\teventPtr.SetObjectID(id)\n\treturn s.createObjectEvent(ctx, eventPtr)\n}", "func (s *Store) Delete(string) error {\n\tpanic(\"implement me\")\n}", "func (ms *mysqlstore) Delete(id string) (err error) {\n\tdefer func(t0 time.Time) {\n\t\terrStr := \"\"\n\t\tif err != nil {\n\t\t\terrStr = err.Error()\n\t\t}\n\t\tms.logger.Printf(\"%v\",\n\t\t\tlogrec{\n\t\t\t\tService: \"mysql\",\n\t\t\t\tOperation: \"delete\",\n\t\t\t\tID: id,\n\t\t\t\tError: errStr,\n\t\t\t\tDuration: fmt.Sprintf(\"%v\", time.Since(t0)),\n\t\t\t},\n\t\t)\n\t}(time.Now())\n\n\t// Check if object exists first\n\tif !ms.isPresent(id) {\n\t\terr = ErrNotFound\n\t\treturn\n\t}\n\n\t// Update state\n\treturn\n}", "func (s *CqlStore) Delete(id ID) error {\n\tif s.session == nil {\n\t\treturn NewStoreClosedError()\n\t}\n\terr := s.session.Query(\"insert into items (id, updated, status) values(?,now(),?)\",\n\t\tIDToString(id), \"DELETED\").Exec()\n\tif err != nil {\n\t\treturn errors.Wrap(err, 0)\n\t}\n\treturn nil\n}", "func (s *Store) Delete(id packet.ID) {\n\t// acquire mutex\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\t// delete future\n\tdelete(s.store, id)\n}", "func (store TodoStore) Delete(_ sqlx.Ext, id string) error {\n\tif _, err := store.Get(id); err != nil {\n\t\treturn err\n\t}\n\tdelete(store, id)\n\treturn nil\n}", "func (DummyStore) Delete(key string) error {\n\treturn nil\n}", "func (m *MemEDVStore) Delete(docID string) error {\n\treturn m.coreStore.Delete(docID)\n}", "func (ps *Store) Delete(ctx context.Context, entity datastore.Entity) error {\n\tc := GetCon(ctx)\n\temd := entity.GetEntityMetadata()\n\terr := c.Query(getDeleteStmt(emd.GetName(), emd.GetIDColumnName()), entity.GetKey()).Exec()\n\treturn err\n}", "func (s *Store) Delete(c *gin.Context) {\n\n}", "func (store *ProviderBackedStore) Del(id Identifier) {\n\tkey := id.Value()\n\tif _, existing := store.retriever[key]; existing {\n\t\tdelete(store.retriever, key)\n\t\tnewIDs := make([]Identifier, 0, len(store.ids)-1)\n\t\tfor _, oldID := range store.ids {\n\t\t\tif oldID.Value() != key {\n\t\t\t\tnewIDs = append(newIDs, oldID)\n\t\t\t}\n\t\t}\n\t\tstore.ids = newIDs\n\t}\n}", "func (st *WriteOnlyStore) Delete(gun data.GUN) error {\n\treturn ErrWriteOnly{}\n}", "func (e *Store) Delete(id string) {\n\te.Lock()\n\tdelete(e.commands, id)\n\te.Unlock()\n}", "func (o *Store) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Store provided for delete\")\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), storePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"stores\\\" WHERE \\\"id\\\"=$1\"\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, sql)\n\t\tfmt.Fprintln(writer, args...)\n\t}\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from stores\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for stores\")\n\t}\n\n\treturn rowsAff, nil\n}", "func (imd *InMemoryDb) Delete(id string) bool {\n\tif !imd.connected {\n\t\treturn false\n\t}\n\tif _, ok := imd.store[id]; !ok {\n\t\treturn false\n\t}\n\timd.mux.Lock()\n\tdelete(imd.store, id)\n\timd.mux.Unlock()\n\treturn true\n}", "func (mysql *MySqlStore) Delete(id int64) error {\n\ttx, err := mysql.db.Begin() // begin transaction\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := tx.Exec(del, id); err != nil {\n\t\ttx.Rollback() // rollback anything thats been done\n\t\treturn ErrUserNotFound\n\t}\n\ttx.Commit() // commit transaction\n\treturn nil\n}", "func (datastore *Datastore) Delete() error {\n\t_, err := client.Call(\"one.datastore.delete\", datastore.ID)\n\treturn err\n}", "func (s Store) Delete(name string) error {\n\treturn s.backingStore.Delete(ItemType, name)\n}", "func (s Store) Delete(name string) error {\n\treturn s.backingStore.Delete(ItemType, name)\n}", "func (db *FlatDatabase) Delete(key []byte) error { panic(\"not supported\") }", "func (ds *KVStorage) Delete(ID string) {\n\tdelete(ds.data, ID)\n}", "func (s *Store) Delete(ID string) error {\n\tcount, err := s.client.Del(ID).Result()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif count < 1 {\n\t\treturn ErrSessionNotFound\n\t}\n\n\treturn nil\n}", "func (rs *RedisStore) Delete(sid SessionID) error {\n\terr := rs.Client.Del(sid.getRedisKey()).Err()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *LiftingStorage) Delete(id int) error {\n\ttx, err := s.db.Beginx()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tworkout := lifting.WorkoutRow{ID: &id}\n\n\t_, err = tx.NamedExec(\n\t\tnamedDelete,\n\t\tworkout,\n\t)\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\n\treturn tx.Commit()\n}", "func (s MockStore) Del(id int) error {\n\tu, err := s.Get(id)\n\tdelete(s.id, id)\n\tdelete(s.name, u.Name)\n\tdelete(s.email, u.Email)\n\n\t// delete tokens associated with uid\n\treturn err\n}", "func (db *DB) Delete(ctx context.Context, id string) error {\n\tc := db.pool.Get()\n\tdefer c.Close()\n\n\tr, err := db.Get(ctx, id)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get record: %w\", err)\n\t}\n\n\tlastVersion, err := redis.Int64(c.Do(\"INCR\", db.lastVersionKey))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.DeletedAt = ptypes.TimestampNow()\n\tr.Version = fmt.Sprintf(\"%012X\", lastVersion)\n\tb, err := proto.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmds := []map[string][]interface{}{\n\t\t{\"MULTI\": nil},\n\t\t{\"HSET\": {db.recordType, id, string(b)}},\n\t\t{\"SADD\": {db.deletedSet, id}},\n\t\t{\"ZADD\": {db.versionSet, lastVersion, id}},\n\t}\n\tif err := db.tx(c, cmds); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (rs *RedisStore) Delete(sid SessionID) error {\n\trs.Client.Del(sid.getRedisKey())\n\treturn nil\n}", "func Delete(c *template.Context) error {\n\tid := c.ID()\n\n\t// Get the store.\n\ts, err := getStore(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// The context has to be locked to perform any changes.\n\tif !IsLocked(c) {\n\t\treturn fmt.Errorf(\"store.Delete: can't remove store data if the context is not locked by the session!\")\n\t}\n\n\t// Lock the mutex.\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\tif s.data.Values == nil {\n\t\t// Nothing to do.\n\t\treturn nil\n\t}\n\n\t// Remove the value. The key is the context's ID.\n\tdelete(s.data.Values, id)\n\n\t// Update data to the database.\n\terr = flushUpdatesToDB(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Broadcast changes to other sessions in edit mode.\n\tbroadcastChangedContext(id, c.Session())\n\n\treturn nil\n}", "func (m *SQLIgnoreStore) Delete(id int, userId string) (int, error) {\n\tstmt := \"DELETE FROM ignorerule WHERE id=?\"\n\tret, err := m.vdb.DB.Exec(stmt, id)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\trowsAffected, err := ret.RowsAffected()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif rowsAffected > 0 {\n\t\tm.inc()\n\t}\n\treturn int(rowsAffected), nil\n}", "func (o *Kvstore) Delete(ctx context.Context, exec boil.ContextExecutor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"models: no Kvstore provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), kvstorePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"kvstore\\\" WHERE \\\"member\\\"=$1 AND \\\"key\\\"=$2\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.ExecContext(ctx, sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: unable to delete from kvstore\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"models: failed to get rows affected by delete for kvstore\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(ctx, exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (self *LevelDBStore) Delete(key []byte) error {\n\treturn self.db.Delete(key, nil)\n}", "func (store *DynamoDBFeatureStore) Delete(kind ld.VersionedDataKind, key string, version int) error {\n\tdeletedItem := kind.MakeDeletedItem(key, version)\n\treturn store.updateWithVersioning(kind, deletedItem)\n}", "func Del(id string) error {\n\treturn getC().Del(id)\n}", "func Del(id string) error {\n\treturn getC().Del(id)\n}", "func Del(id string) error {\n\treturn getC().Del(id)\n}", "func (s *MySQLStore) Delete(id string) error {\n\n\ts.mu.Lock()\n\n\tif len(id) > 64 {\n\t\treturn fmt.Errorf(\"failed to delete ticket; ID not formatted properly\")\n\t}\n\n\tsql := fmt.Sprintf(`DELETE FROM tickets WHERE t_id = '%s';`, id)\n\t_, err := s.DB.ExecSingle(sql)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.mu.Unlock()\n\treturn nil\n}", "func (bs *BoltStore) Delete(token string) error {\n\treturn bs.db.Update(func(tx *bolt.Tx) error {\n\t\ttokenBytes := []byte(token)\n\t\treturn txDelete(tx, tokenBytes)\n\t})\n}", "func (t *TestStore) Delete(data interface{}) error {\n\treturn t.delete(data)\n}", "func (me imageStore) Delete(imageID uint) (bool, error) {\n\tif err := me.Db.Unscoped().Exec(\"DELETE FROM images WHERE id = ? \", imageID).Error; err != nil {\n\t\treturn false, err\n\t}\n\n\treturn true, nil\n}", "func (i *Identity) Delete() error {\n\treturn errors.New(\"NotImplemented\")\n}", "func (s *RedisStore) Delete(key interface{}) error {\n\t_, err := s.client.Del(key.(string)).Result()\n\treturn err\n}", "func (store *MySQLStore) Delete(id int64) error {\n\tquery := \"DELETE FROM users WHERE id = ?\"\n\t_, err := store.db.Exec(query, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (db *DB) Delete(ctx context.Context, id string) (err error) {\n\tc := db.pool.Get()\n\t_, span := trace.StartSpan(ctx, \"databroker.redis.Delete\")\n\tdefer span.End()\n\tdefer recordOperation(ctx, time.Now(), \"delete\", err)\n\tdefer c.Close()\n\n\tr, err := db.Get(ctx, id)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get record: %w\", err)\n\t}\n\n\tlastVersion, err := redis.Int64(c.Do(\"INCR\", db.lastVersionKey))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.DeletedAt = ptypes.TimestampNow()\n\tr.Version = fmt.Sprintf(\"%012X\", lastVersion)\n\tb, err := proto.Marshal(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcmds := []map[string][]interface{}{\n\t\t{\"MULTI\": nil},\n\t\t{\"HSET\": {db.recordType, id, string(b)}},\n\t\t{\"SADD\": {db.deletedSet, id}},\n\t\t{\"ZADD\": {db.versionSet, lastVersion, id}},\n\t}\n\tif err := db.tx(c, cmds); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *Store) Delete(key string) {\n\ts.data.Del(key)\n}", "func DeleteStore(id int, db *gorm.DB) error {\n\tstore := Store{}\n\n\tif err := db.Debug().Table(\"stores\").Where(\"id = ?\", id).Delete(store).Error; err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n\n}", "func (s *MongodbStore) Delete(key interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tdelete(s.data, key)\n\treturn nil\n}", "func (b *BadgerDBStore) Delete(key []byte) error {\n\treturn b.db.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Delete(key)\n\t})\n}", "func (s *Store) Delete(key string) error {\n\treturn s.Store.Delete(key)\n}", "func (s *PostgresStore) Delete(key interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tdelete(s.data, key)\n\treturn nil\n}", "func (s *Store) Delete(id string) error {\n\t// We're going to be modifying our items slice - lock for writing.\n\ts.mutex.Lock()\n\n\t// Unlock once we're done.\n\tdefer s.mutex.Unlock()\n\n\tremainingItems := make([]item, 0, len(s.items))\n\n\t// Only keep items which do not match the ID provided.\n\tfor _, item := range s.items {\n\t\tif item.id != id {\n\t\t\tremainingItems = append(remainingItems, item)\n\t\t}\n\t}\n\n\t// If the number of items is the same then we haven't found anything to delete.\n\tif len(s.items) == len(remainingItems) {\n\t\treturn moodboard.ErrNoSuchItem\n\t}\n\n\ts.items = remainingItems\n\n\treturn nil\n}", "func (s *Store) Del(id string) error {\n\turl, _, err := s.Get(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\tconn := s.pool.Get()\n\tdefer conn.Close()\n\tconn.Send(\"MULTI\")\n\tconn.Send(\"DEL\", s.idKey(id))\n\tconn.Send(\"DEL\", s.urlKey(url))\n\trep, err := redis.Values(conn.Do(\"EXEC\"))\n\t// Check if there are any errors.\n\tfor _, r := range rep {\n\t\tif _, ok := r.(redis.Error); ok {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s *PersonStore) Delete(record *Person) error {\n\treturn s.Store.Delete(Schema.Person.BaseSchema, record)\n}", "func (s *State) Delete(db XODB) error {\n\tvar err error\n\n\t// if doesn't exist, bail\n\tif !s._exists {\n\t\treturn nil\n\t}\n\n\t// if deleted, bail\n\tif s._deleted {\n\t\treturn nil\n\t}\n\n\t// sql query\n\tconst sqlstr = `DELETE FROM public.state WHERE \"id\" = $1`\n\n\t// run query\n\tXOLog(sqlstr, s.ID)\n\t_, err = db.Exec(sqlstr, s.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// set deleted\n\ts._deleted = true\n\n\treturn nil\n}", "func (m *DBMem) Delete(idToDelete int) {\n m.Lock()\n defer m.Unlock()\n\n\tif len(m.data) <= idToDelete {\n\t\tfmt.Println(\"ID\", idToDelete,\"is out of range\")\n\t\treturn\n\t}\n entryToDelete := m.data[idToDelete]\n delete(m.data, idToDelete)\n\n // Reindex database where ID > deleted\n // I think with using .Lock() m.data wasn't updating immediatly so I ignore the key\n tempMap := make(map[int]Person)\n for k, v := range m.data {\n switch{\n case k > idToDelete:\n tempMap[k-1] = v\n case k == idToDelete:\n default:\n tempMap[k] = v\n }\n }\n\n m.data = tempMap\n m.history.Append(\"DELETE\", idToDelete, entryToDelete)\n}", "func (kv *KVStore) Delete(key string) error {\n\tpayload := kvPayload{\n\t\tKey: key,\n\t\tOp: OpDel,\n\t}\n\n\t_, err := kv.db.Add(&payload)\n\treturn err\n}", "func (rs *Store) Delete(ctx context.Context, key interface{}) error {\n\trs.lock.Lock()\n\tdefer rs.lock.Unlock()\n\tdelete(rs.values, key)\n\treturn nil\n}", "func (ms memstore) Delete(key string) {\n\tms.lock.Lock()\n\tdelete(ms.store, key)\n\tms.lock.Unlock()\n}", "func (o *Storestate) Delete(exec boil.Executor) (int64, error) {\n\tif o == nil {\n\t\treturn 0, errors.New(\"stellarcore: no Storestate provided for delete\")\n\t}\n\n\tif err := o.doBeforeDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\targs := queries.ValuesFromMapping(reflect.Indirect(reflect.ValueOf(o)), storestatePrimaryKeyMapping)\n\tsql := \"DELETE FROM \\\"storestate\\\" WHERE \\\"statename\\\"=$1\"\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, sql)\n\t\tfmt.Fprintln(boil.DebugWriter, args...)\n\t}\n\n\tresult, err := exec.Exec(sql, args...)\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: unable to delete from storestate\")\n\t}\n\n\trowsAff, err := result.RowsAffected()\n\tif err != nil {\n\t\treturn 0, errors.Wrap(err, \"stellarcore: failed to get rows affected by delete for storestate\")\n\t}\n\n\tif err := o.doAfterDeleteHooks(exec); err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rowsAff, nil\n}", "func (s *StoreService) Delete(hash string) {\n\ts.store.Delete(hash)\n}", "func (s *Store) Delete(key string) {\n\ts.store.Del(key)\n}", "func (s *Store) Delete(item storage.Item) error {\n\t// this is a small hack to make the deletion of old entries work. As they\n\t// don't have a namespace, we need to check for that and use the ID as key without\n\t// the separator.\n\tvar k []byte\n\tif item.Namespace() == \"\" {\n\t\tk = []byte(item.ID())\n\t} else {\n\t\tk = key(item)\n\t}\n\n\treturn s.db.Delete(k, nil)\n}", "func (rc *Store) Delete(key string) error {\n\tif rc.conn == nil {\n\t\tif err := rc.connectInit(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn rc.conn.Delete(key)\n}", "func (d *DB) Delete(a *Article) {\n\tst, err := d.db.Prepare(\"update articles set deleted = true where id = ?\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\tdefer st.Close()\n\n\tif _, err := st.Exec(a.id); err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func (r *repository) Delete(id int) error {\n\tq := \"DELETE FROM records WHERE id = $1;\"\n\t_, err := r.db.Exec(q, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *LDBStore) Delete(addr Address) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tikey := getIndexKey(addr)\n\n\tidata, err := s.db.Get(ikey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar idx dpaDBIndex\n\tdecodeIndex(idata, &idx)\n\tproximity := s.po(addr)\n\treturn s.deleteNow(&idx, ikey, proximity)\n}", "func (store *RedisStore) Delete(key string) error {\n\terr := store.Client.Del(key).Err()\n\treturn err\n}", "func (mss *MySQLStore) Delete(id int64) error {\n\tinsq := \"delete from users where user_id=?\"\n\t_, err := mss.Client.Exec(insq, id)\n\tif err != nil {\n\t\treturn ErrUserNotFound\n\t}\n\treturn nil\n}", "func (ms Memorystore) Delete(key string) error {\n\treturn ms.DeleteMulti([]string{key})\n}", "func (ss *SQLStore) Delete(id int64) error {\n\tdesq := \"delete from USERS where id = ?\"\n\t_, err := ss.db.Exec(desq, id)\n\tif err != nil {\n\t\treturn errors.New(\"Error deleting row, given ID might be invalid.\")\n\t}\n\treturn nil\n}", "func (s *MemStore) Delete(key interface{}) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tdelete(s.data, key)\n\treturn nil\n}", "func (ms *MySQLStore) Delete(id int64) error {\n\tdel := string(\"DELETE FROM TblSurvey WHERE SurveyID = ?\")\n\tres, err := ms.Database.Exec(del, id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trowsAffected, rowsAffectedErr := res.RowsAffected()\n\tif rowsAffectedErr != nil {\n\t\treturn rowsAffectedErr\n\t}\n\n\tif rowsAffected != 1 {\n\t\treturn ErrSurveyNotFound\n\t}\n\n\treturn nil\n}", "func (s *DiffStore) Delete(key []byte) {\n\titem := s.tree.Get(&storeKV{key: key})\n\titem.(*storeKV).state = ItemDeleted\n}", "func (t *Tenants) Del(id string) error {\n\treturn t.store.Del(id)\n}", "func (db *memorydb) Del(key []byte) error {\n\n\tif db.enableBatch {\n\t\tdb.batch.Del(key)\n\t} else {\n\t\tdb.writeLock <- struct{}{}\n\t\tdefer func() {\n\t\t\t<-db.writeLock\n\t\t}()\n\n\t\tdb.sm.Lock()\n\t\tdefer db.sm.Unlock()\n\n\t\tdelete(db.db, string(key))\n\t}\n\n\treturn nil\n}", "func (bs *badgerStore) Delete(key []byte) (err error) {\n\treturn bs.db.Update(func(txn *badger.Txn) error {\n\t\treturn txn.Delete(key)\n\t})\n}", "func (ss *storageServer) Delete(args *storagerpc.DeleteArgs, reply *storagerpc.DeleteReply) error {\n\n\thash := libstore.StoreHash(args.Key)\n\tif rangeOK := ss.rangeChecker(hash); !rangeOK {\n\t\t*reply = storagerpc.DeleteReply{\n\t\t\tStatus: storagerpc.WrongServer,\n\t\t}\n\t} else {\n\t\tss.valLock.Lock()\n\t\tv, ok := ss.valTable[args.Key]\n\t\tif !ok {\n\t\t\t*reply = storagerpc.DeleteReply{\n\t\t\t\tStatus: storagerpc.KeyNotFound,\n\t\t\t}\n\t\t} else {\n\t\t\tdelete(ss.valTable, args.Key)\n\t\t\t*reply = storagerpc.DeleteReply{\n\t\t\t\tStatus: storagerpc.OK,\n\t\t\t}\n\t\t}\n\n\t\tss.valLock.Unlock()\n\n\t\tv.writeLock.Lock()\n\t\tv.readLock.Lock()\n\t\tss.Revoke(args.Key, v.leasePool.Freeze())\n\t\tv.readLock.Unlock()\n\t\tv.writeLock.Unlock()\n\n\t}\n\treturn nil\n}", "func (db *database) Delete(id int64) error {\n\tif id == 0 || id == defaultID {\n\t\treturn requestError{fmt.Errorf(\"cannot delete snippet (ID: %d)\", id)}\n\t}\n\terr := db.db.Update(func(tx *bolt.Tx) error {\n\t\t// Locate and delete key from bucketsByID.\n\t\tbktByID := tx.Bucket([]byte(bucketByID))\n\t\tv := bktByID.Get(idKey(id))\n\t\tif v == nil {\n\t\t\treturn errNotFound\n\t\t}\n\t\tif err := bktByID.Delete(idKey(id)); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Delete key from bucketsByDate.\n\t\tvar s snippet\n\t\tif err := s.UnmarshalBinary(v); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tk := dualKey(s.ID, s.Modified)\n\t\treturn tx.Bucket([]byte(bucketByDate)).Delete(k)\n\t})\n\tif err == nil {\n\t\tdb.mu.Lock()\n\t\tdelete(db.names, id)\n\t\tdb.mu.Unlock()\n\t}\n\treturn err\n}", "func (s *Store) Delete(key string) error {\n\treturn s.Remove(key)\n}", "func (s *ServersStore) Delete(ctx context.Context, d chronograf.Server) error {\n\terr := validOrganization(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\td, err = s.store.Get(ctx, d.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn s.store.Delete(ctx, d)\n}", "func (dao ArtistDAO) Delete(id string) error {\n\treturn dao.getCollection().Remove(bson.M{\"_id\": bson.ObjectIdHex(id)})\n}", "func (s *Storage) Del(key []byte) error {\n\topts := gorocksdb.NewDefaultWriteOptions()\n\n\tdefer opts.Destroy()\n\n\treturn s.db.Delete(opts, key)\n}", "func (s *VirtualStore) Delete(key any) error {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\tdelete(s.data, key)\n\treturn nil\n}", "func (i *API) Delete(kind, key string) error {\n\tgo func() {\n\t\tif i.backupStore != nil {\n\t\t\tif err := i.backupStore.RawDelete(kind, key); err != nil {\n\t\t\t\tlog.Error(errors.Wrapf(types.ErrBackupStoreDelete, \"Infra API: Err: %v\", err))\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn i.primaryStore.RawDelete(kind, key)\n}", "func (els *EventLocalStore) Delete(id string) error {\n\tels.mutex.Lock()\n\tdefer els.mutex.Unlock()\n\n\tif _, ok := els.events[id]; ok {\n\t\tdelete(els.events, id)\n\t\treturn nil\n\t}\n\n\treturn event.ErrEventNotFound\n}", "func (sS StoreService) Delete(id int64) (message string, err error) {\n\tvar order models.Order\n\terr = sS.DB.First(&order, id).Error\n\tif err != nil {\n\t\treturn \"not found!\", err\n\t}\n\n\tsS.DB.Delete(&order)\n\treturn \"success\", err\n}", "func (nr *NoteRepository) Delete(ID string) error {\n\tif ID == \"\" {\n\t\treturn errors.New(\"ID should NOT be empty\")\n\t}\n\n\tif (nr.Notes[ID] == model.Note{}) {\n\t\treturn errors.New(\"note NOT found\")\n\t}\n\tdelete(nr.Notes, ID)\n\tgo storeNotes(nr.DataFileName, nr.Notes)\n\treturn nil\n}", "func (ss *redisStore) Delete(key string) error {\n\ti, err := ss.rdb.Del(ctx, key).Result()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif i == 0 {\n\t\treturn storage.ErrNoRecord\n\t}\n\treturn nil\n}", "func (r *RedisStore) Delete(token string) error {\n\tconn := r.pool.Get()\n\tdefer conn.Close()\n\n\t_, err := conn.Do(\"DEL\", r.prefix+token)\n\treturn err\n}", "func (s *RedisStore) delete(ctx context.Context, session *sessions.Session) error {\n\tif err := s.Cmd.Del(ctx, s.keyPrefix+session.ID).Err(); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c Controller) Delete(ctx context.Context, tableID, id string) (string, error) {\n\tdeletedID, err := c.Storage.Delete(ctx, tableID, id)\n\tif err != nil {\n\t\tc.Logger.WithFields(logrus.Fields{\n\t\t\t\"error\": err.Error(),\n\t\t}).Error(ErrDelete.Error())\n\n\t\treturn \"\", ErrDelete\n\t}\n\n\treturn deletedID, nil\n}", "func (m *RedisStore) Delete(key interface{}) error {\n\tm.lock.Lock()\n\tdefer m.lock.Unlock()\n\tdelete(m.sessionData, key)\n\treturn nil\n}", "func (s *PaymentStorage) Delete(ctx context.Context, id aggregate.ID) error {\n\tlogger := log.FromContext(ctx)\n\n\tquery := `DELETE FROM %[1]s WHERE id = $1`\n\tquery = fmt.Sprintf(query, s.table)\n\n\tif logger != nil {\n\t\tlogger.Debugf(\"exec in transaction sql %s, values %+v\", query, []interface{}{\n\t\t\tid,\n\t\t})\n\t}\n\n\treturn execInTransaction(s.db, func(tx *sqlx.Tx) error {\n\t\t_, err := tx.ExecContext(ctx, query, id)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func Delete(txn *badger.Txn, entType schema.EntityType, ID []byte) error {\n\treturn item.Delete(txn, entType.EntityID(ID))\n}", "func (stdnt *Student) Delete(ctx context.Context) error {\n\tkey, err := datastore.DecodeKey(stdnt.ID)\n\tif err != nil {\n\t\treturn errors.New(err.Error())\n\t}\n\terr = datastore.Delete(ctx, key)\n\tif err != nil {\n\t\treturn errors.New(err.Error())\n\t}\n\treturn nil\n}", "func (r *RedisStore) Delete(token string) error {\n\tr.RWMutex.Lock()\n\tdefer r.RWMutex.Unlock()\n\t_, err := r.redis.DelWithErr(r.prefix+token)\n\treturn err\n}", "func (m *MySQLStore) Delete(token string) error {\n\t_, err := m.DB.Exec(\"DELETE FROM sessions WHERE token = ?\", token)\n\treturn err\n}", "func (d *Dosen) Delete(db *sql.DB) error {\n\tif d.ID == \"\" {\n\t\treturn fmt.Errorf(\"id tidak bolehh kosong\")\n\t}\n\tquery := \"DELETE FROM dosen where id = $1\"\n\t_, err := db.Exec(query, &d.ID)\n\treturn err\n}", "func (p *Store) Del(ctx context.Context, round uint64) error {\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn ctx.Err()\n\tdefault:\n\t}\n\n\tconst query = `\n\tDELETE FROM\n\t\tbeacon_details\n\tWHERE\n\t\tbeacon_id = :id AND\n\t\tround = :round`\n\n\tdata := struct {\n\t\tID int `db:\"id\"`\n\t\tRound uint64 `db:\"round\"`\n\t}{\n\t\tID: p.beaconID,\n\t\tRound: round,\n\t}\n\n\t_, err := p.db.NamedExecContext(ctx, query, data)\n\treturn err\n}", "func (s *LDBStore) delete(batch *leveldb.Batch, idx *dpaDBIndex, idxKey []byte, po uint8) {\n\tmetrics.GetOrRegisterCounter(\"ldbstore.delete\", nil).Inc(1)\n\n\tgcIdxKey := getGCIdxKey(idx)\n\tbatch.Delete(gcIdxKey)\n\tdataKey := getDataKey(idx.Idx, po)\n\tbatch.Delete(dataKey)\n\tbatch.Delete(idxKey)\n\ts.entryCnt--\n\tdbEntryCount.Dec(1)\n\tcntKey := make([]byte, 2)\n\tcntKey[0] = keyDistanceCnt\n\tcntKey[1] = po\n\tbatch.Put(keyEntryCnt, U64ToBytes(s.entryCnt))\n\tbatch.Put(cntKey, U64ToBytes(s.bucketCnt[po]))\n}", "func (es *etcdStore) Delete(imageID string) error {\n\tkey := path.Join(es.prefix, imageID)\n\tif _, err := es.client.Delete(key, true); err != nil {\n\t\tetcdErr := err.(*etcd.EtcdError)\n\t\tif etcdErr.ErrorCode != etcderr.EcodeKeyNotFound {\n\t\t\tlog.WithFields(etcdLogFields).WithFields(log.Fields{\n\t\t\t\t\"error\": err,\n\t\t\t\t\"key\": key,\n\t\t\t}).Error(\"failed to delete image\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}" ]
[ "0.7466687", "0.7384", "0.7331129", "0.73217475", "0.7315649", "0.7236049", "0.71782357", "0.7138321", "0.7090377", "0.70798945", "0.7011291", "0.7008103", "0.69747865", "0.69713014", "0.69643414", "0.6916275", "0.6909399", "0.6876967", "0.68627083", "0.68627083", "0.6860765", "0.68514293", "0.6848135", "0.6842922", "0.68373513", "0.6822284", "0.6814298", "0.6797039", "0.67904574", "0.678423", "0.67817134", "0.67583895", "0.6755021", "0.67217344", "0.67217344", "0.67217344", "0.6668264", "0.66625077", "0.66380554", "0.6621779", "0.6610711", "0.6604493", "0.6602755", "0.6593974", "0.65905124", "0.6575306", "0.65566504", "0.6553863", "0.65508693", "0.6548076", "0.6546956", "0.6545551", "0.6531891", "0.65282077", "0.6520226", "0.6519604", "0.6517326", "0.65164566", "0.6514928", "0.65009576", "0.649329", "0.6481568", "0.647946", "0.64728034", "0.6469162", "0.64522165", "0.6451719", "0.6445438", "0.642825", "0.64258224", "0.64228964", "0.6417058", "0.64168596", "0.63926256", "0.6385431", "0.63818514", "0.6377233", "0.6373097", "0.6370068", "0.6364679", "0.6362574", "0.6354634", "0.6346832", "0.6341975", "0.6341036", "0.633971", "0.6335507", "0.63338214", "0.63329643", "0.63314605", "0.63245726", "0.6323936", "0.6310138", "0.6308033", "0.6307468", "0.63074225", "0.6306797", "0.6304854", "0.6304652", "0.6300999", "0.6297802" ]
0.0
-1
AsMap implements the IDStore interface
func (s *MapStorage) AsMap() map[string]int { s.lock.RLock() defer s.lock.RUnlock() replay := make(map[string]int) for k, v := range s.store { replay[k] = v } return replay }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *InMemoryHashStorage) Map() map[int64]string {\n\treturn s.hashes\n}", "func (cs *CStore) ToMap() map[string]string {\n\tnewMap := make(map[string]string, len(cs.store))\n\n\tfor k, v := range cs.store {\n\t\tnewMap[k] = v\n\t}\n\n\treturn newMap\n}", "func toMap(s *DefaultIDSetMap) map[int]*IDSet {\n\tif s.key != 0 {\n\t\treturn map[int]*IDSet{\n\t\t\ts.key: s.value,\n\t\t}\n\t}\n\n\tif s.m != nil {\n\t\tm := map[int]*IDSet{}\n\t\tfor k, v := range s.m {\n\t\t\tm[k] = v\n\t\t}\n\t\treturn m\n\t}\n\n\treturn nil\n}", "func (store *Store) Map(item interface{}) *Map {\n\treturn store.MapWithName(\"\", item)\n}", "func (vind *Varbinary) Map(_ VCursor, ids []interface{}) ([][]byte, error) {\n\tout := make([][]byte, 0, len(ids))\n\tfor _, id := range ids {\n\t\tdata, err := getVarbinaryHash(id)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"VarBinary_hash.Map :%v\", err)\n\t\t}\n\t\tout = append(out, data)\n\t}\n\treturn out, nil\n}", "func (s SliceOfBar) AsMap() map[string]*Bar {\n\tresult := make(map[string]*Bar)\n\tfor _, value := range s {\n\t\tresult[value.ID] = value\n\t}\n\treturn result\n}", "func mockMapStore(storage map[string]interface{}) ResultStoreFn {\n\treturn func(id string, key string, value interface{}) {\n\t\tutil.SetNestedField(storage, value, id, key)\n\t}\n}", "func NewMapStore() RawStore {\n\treturn &mapStorage{\n\t\tinner: make(map[string][]byte),\n\t}\n}", "func (DummyStore) SetMap(key string, value map[string]interface{}) error {\n\treturn nil\n}", "func HStoreFromStringPtrMap(val map[string]*string) driver.Valuer {\n\treturn hstoreFromStringPtrMap{val: val}\n}", "func SliceToIDMap(slice interface{}, idField string) map[primitive.ObjectID]interface{} {\n\tsliceV := reflect.ValueOf(slice)\n\tm := make(map[primitive.ObjectID]interface{}, sliceV.Len())\n\tif sliceV.Len() == 0 {\n\t\treturn m\n\t}\n\n\tfor i := 0; i < sliceV.Len(); i++ {\n\t\tidV := reflect.\n\t\t\tIndirect(sliceV.Index(i)).\n\t\t\tFieldByName(idField)\n\t\tid := idV.Interface()\n\t\tvar oid primitive.ObjectID\n\t\tif idV.Kind() == reflect.Ptr {\n\t\t\toid = *(id.(*primitive.ObjectID))\n\t\t} else {\n\t\t\toid = id.(primitive.ObjectID)\n\t\t}\n\t\tif idV.IsValid() && id != nil && !oid.IsZero() {\n\t\t\tm[oid] = sliceV.Index(i).Interface()\n\t\t}\n\t}\n\n\treturn m\n}", "func (vind *UTF8cihash) Map(_ VCursor, ids []interface{}) ([][]byte, error) {\n\tout := make([][]byte, 0, len(ids))\n\tfor _, id := range ids {\n\t\tdata, err := getutf8cihash(id)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"utf8cihash.Map :%v\", err)\n\t\t}\n\t\tout = append(out, data)\n\t}\n\treturn out, nil\n}", "func (d docs) mapByID(key ...string) (result map[bson.ObjectId]map[string]interface{}) {\n\tresult = map[bson.ObjectId]map[string]interface{}{}\n\tif len(key) == 0 {\n\t\tkey = []string{\"_id\"}\n\t}\n\tfor _, doc := range d {\n\t\tif _, ok := doc[key[0]]; !ok {\n\t\t\tcontinue\n\t\t}\n\t\tresult[doc[key[0]].(bson.ObjectId)] = doc\n\t}\n\treturn\n}", "func ToMap(i interface{}) (map[string]interface{}, error) {\n\tvar stringInterfaceMap map[string]interface{}\n\titr, _ := bson.Marshal(i)\n\terr := bson.Unmarshal(itr, &stringInterfaceMap)\n\treturn stringInterfaceMap, err\n}", "func (c *Configmap) AsMapSI() map[string]interface{} {\n\treturn *c\n}", "func (d Document) Map() map[string]interface{} {\n\treturn d.m\n}", "func HStoreFromStringMap(val map[string]string) driver.Valuer {\n\treturn hstoreFromStringMap{val: val}\n}", "func StoreCAUniqueIDToCNMap(c context.Context, mapping map[int64]string) error {\n\tbuf := bytes.Buffer{}\n\tenc := gob.NewEncoder(&buf)\n\tif err := enc.Encode(mapping); err != nil {\n\t\treturn err\n\t}\n\t// Note that in practice 'mapping' is usually very small, so we are not\n\t// concerned about 1MB entity size limit.\n\treturn errors.WrapTransient(datastore.Get(c).Put(&CAUniqueIDToCNMap{\n\t\tGobEncodedMap: buf.Bytes(),\n\t}))\n}", "func (s *Int64) Map() map[int64]bool {\n\tres := make(map[int64]bool, len(s.m))\n\n\tfor val := range s.m {\n\t\tres[val] = true\n\t}\n\treturn res\n}", "func (ud *UserData) ToMap() UserDataMap {\n\tif ud == nil {\n\t\treturn nil\n\t}\n\tm := UserDataMap{}\n\tud.Range(func(key, value interface{}) bool {\n\t\tm[key.(string)] = value.(string)\n\t\treturn true\n\t})\n\treturn m\n}", "func (DummyStore) GetMap(key string) (map[string]interface{}, error) {\n\treturn nil, nil\n}", "func AsMap(dt DataType) *Map {\n\tswitch t := dt.(type) {\n\tcase *UserTypeExpr:\n\t\treturn AsMap(t.Type)\n\tcase *ResultTypeExpr:\n\t\treturn AsMap(t.Type)\n\tcase *Map:\n\t\treturn t\n\tdefault:\n\t\treturn nil\n\t}\n}", "func (s *Superpower) ToMap() map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"id\": s.ID,\n\t\t\"name\": s.Name,\n\t\t\"damage\": s.Damage,\n\t\t\"energy\": s.Energy,\n\t\t\"passive\": s.Passive,\n\t}\n}", "func HStoreToStringMap(val *map[string]string) sql.Scanner {\n\treturn hstoreToStringMap{val: val}\n}", "func HStoreToStringPtrMap(val *map[string]*string) sql.Scanner {\n\treturn hstoreToStringPtrMap{val: val}\n}", "func (image *Image) ToMap() map[string]interface{} {\n\tdata := make(map[string]interface{})\n\timage.UpdateMap(data)\n\n\treturn data\n}", "func (o *Echo) ToMap() map[string]interface{} {\n\to.setDefaults(false)\n\treturn map[string]interface{}{\n\t\t\"id\": toEchoObject(o.ID, false),\n\t\t\"message\": toEchoObject(o.Message, true),\n\t\t\"updated_ts\": toEchoObject(o.UpdatedAt, false),\n\t}\n}", "func (s *Store) Data() map[string]interface{} {\n\ts.access.RLock()\n\tdefer s.access.RUnlock()\n\n\tcpy := make(map[string]interface{}, len(s.data))\n\tfor key, value := range s.data {\n\t\tcpy[key] = value\n\t}\n\treturn cpy\n}", "func (this *DmnList) Map() (DmnMap, error) {\n\n\tdm := make(DmnMap)\n\n\tfor _, di := range *this {\n\n\t\tif dm[di.Key] == nil {\n\t\t\tdm[di.Key] = make(map[int]*DmnInfo)\n\t\t}\n\n\t\tdm[di.Key][di.Version] = di\n\t}\n\n\treturn dm, nil\n}", "func (s *MockStore) Write(collection, id, value string) error {\n\tc, ok := s.Data[collection]\n\tif ok {\n\t\tc[id] = value\n\t} else {\n\t\tc = map[string]string{id: value}\n\t\tif s.Data == nil {\n\t\t\ts.Data = map[string]map[string]string{collection: c}\n\t\t} else {\n\t\t\ts.Data[collection] = c\n\t\t}\n\t}\n\treturn nil\n}", "func (ts Uint64Uint64Tuples) ToMap() Uint64Uint64Map {\n\treturn NewUint64Uint64Map(ts...)\n}", "func (p *Client) ConvertMap() db.StData {\n\treturn db.StData{\n\t\t\"name\": p.Name,\n\t\t\"segment1\": p.Segment1,\n\t\t\"segment2\": p.Segment2,\n\t\t\"segment3\": p.Segment3,\n\t\t\"segment4\": p.Segment4,\n\t\t\"platformId\": p.PlatformID,\n\t\t\"clientId\": p.ClientID,\n\t}\n}", "func (qq Qualifiers) Map() map[string]string {\n\tm := make(map[string]string)\n\n\tfor i := 0; i < len(qq); i++ {\n\t\tk := qq[i].Key\n\t\tv := qq[i].Value\n\t\tm[k] = v\n\t}\n\n\treturn m\n}", "func (o *OutputState) ApplyIDArrayMap(applier interface{}) IDArrayMapOutput {\n\treturn o.ApplyT(applier).(IDArrayMapOutput)\n}", "func (DummyStore) DeleteMap(key string, fields ...string) error { return nil }", "func Map(args ...interface{}) dgo.MapType {\n\treturn internal.MapType(args...)\n}", "func (s InputSecureFileUploadedArray) ToMap() map[int64]InputSecureFileUploaded {\n\tr := make(map[int64]InputSecureFileUploaded, len(s))\n\ts.FillMap(r)\n\treturn r\n}", "func (h *sizeHistogram) toMap() map[string]uint64 {\n\tres := make(map[string]uint64, dataUsageBucketLen)\n\tfor i, count := range h {\n\t\tres[ObjectsHistogramIntervals[i].name] = count\n\t}\n\treturn res\n}", "func (o *OutputState) ApplyIDMap(applier interface{}) IDMapOutput {\n\treturn o.ApplyT(applier).(IDMapOutput)\n}", "func (s *Uint64) Map() map[uint64]bool {\n\tres := make(map[uint64]bool, len(s.m))\n\n\tfor val := range s.m {\n\t\tres[val] = true\n\t}\n\treturn res\n}", "func MarshalMap(in interface{}) (map[string]*dynamodb.AttributeValue, error) {\n\tav, err := dynamodbattribute.MarshalMap(in)\n\tif nil != err {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"cause\": err,\n\t\t\t\"data\": fmt.Sprintf(\"%+v\", in),\n\t\t}).Error(\"Problem marshaling User in GetUserByID\")\n\t}\n\treturn av, err\n}", "func flattenImageShieldedInstanceInitialStateDbMap(c *Client, i interface{}) map[string]ImageShieldedInstanceInitialStateDb {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageShieldedInstanceInitialStateDb{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageShieldedInstanceInitialStateDb{}\n\t}\n\n\titems := make(map[string]ImageShieldedInstanceInitialStateDb)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageShieldedInstanceInitialStateDb(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (repo *Repository) AsMap() map[Name]worker.ArtifactSource {\n\tresult := make(map[Name]worker.ArtifactSource)\n\n\trepo.repoL.RLock()\n\tfor name, source := range repo.repo {\n\t\tresult[name] = source\n\t}\n\trepo.repoL.RUnlock()\n\n\treturn result\n}", "func (o *ExportData) ToMap() map[string]interface{} {\n\to.setDefaults(false)\n\treturn map[string]interface{}{\n\t\t\"customer_id\": toExportDataObject(o.CustomerID, false),\n\t\t\"id\": toExportDataObject(o.ID, false),\n\t\t\"integration_instance_id\": toExportDataObject(o.IntegrationInstanceID, false),\n\t\t\"job_id\": toExportDataObject(o.JobID, false),\n\t\t\"objects\": toExportDataObject(o.Objects, false),\n\t\t\"ref_id\": toExportDataObject(o.RefID, false),\n\t\t\"ref_type\": toExportDataObject(o.RefType, false),\n\t\t\"hashcode\": toExportDataObject(o.Hashcode, false),\n\t}\n}", "func (r Room) ToMap() *map[string]interface{} {\n\treturn &map[string]interface{}{\n\t\t\"UId\": r.UId,\n\t\t\"title\": r.title,\n\t\t\"contentMd\": r.contentMd,\n\t\t\"imageUId\": r.imageUId,\n\t}\n}", "func (d BsonD) Map() (m bson.M) {\n\tm = make(bson.M, len(d))\n\tfor _, item := range d {\n\t\tswitch v := item.Value.(type) {\n\t\tcase BsonD:\n\t\t\tm[item.Key] = v.Map()\n\t\tcase []BsonD:\n\t\t\tel := []bson.M{}\n\t\t\tfor i := range v {\n\t\t\t\tel = append(el, v[i].Map())\n\t\t\t}\n\t\t\tm[item.Key] = el\n\t\tcase []interface{}:\n\t\t\t// mgo/bson doesn't expose UnmarshalBSON interface\n\t\t\t// so we can't create custom bson.Unmarshal()\n\t\t\tel := []bson.M{}\n\t\t\tfor i := range v {\n\t\t\t\tif b, ok := v[i].(BsonD); ok {\n\t\t\t\t\tel = append(el, b.Map())\n\t\t\t\t}\n\t\t\t}\n\t\t\tm[item.Key] = el\n\t\tdefault:\n\t\t\tm[item.Key] = item.Value\n\t\t}\n\t}\n\treturn m\n}", "func (pdu *Pdu) AsMapKey() string {\n // return fmt.Sprintf(\"%d[%x]\", pdu.MessageID, pdu.Token)\n return fmt.Sprintf(\"%x\", pdu.Token)\n // return fmt.Sprintf(\"%d\", pdu.MessageID)\n}", "func (d *Document) ToMap() map[string]interface{} {\n\tl := getLn()\n\ttags := make(map[string]string, l)\n\tout := make(map[string]interface{}, l)\n\n\tappendDocumentKeys(tags, d)\n\tappendBatchKeys(tags)\n\n\tif len(d.Title) > 0 {\n\t\tout[tags[\"Title\"]] = d.Title\n\t}\n\tout[tags[\"Url\"]] = d.Url\n\tout[tags[\"BatchType\"]] = d.BatchType\n\tout[tags[\"BatchName\"]] = d.BatchName\n\tout[tags[\"BatchTime\"]] = d.BatchTime\n\tout[tags[\"BatchId\"]] = d.BatchId\n\tout[tags[\"Body\"]] = d.Body\n\tout[tags[\"Type\"]] = d.Type\n\tout[tags[\"Id\"]] = d.Id\n\tif len(d.Project) > 0 {\n\t\tout[tags[\"Project\"]] = d.Project\n\t}\n\n\treturn out\n}", "func (f *FileList) Map() FileMap {\n\treturn f.store\n}", "func (store *HexastoreDB) MapStringsToIds(subject, property, object string) (int, int, int) {\n\tvar ok bool\n\tvar objID, subjID, propID int\n\tsubjID, ok = store.entities.GetKey(subject)\n\tif !ok {\n\t\tsubjID = store.entities.Put(subject)\n\t}\n\tpropID, ok = store.props.GetKey(property)\n\tif !ok {\n\t\tpropID = store.props.Put(property)\n\t}\n\tobjID, ok = store.entities.GetKey(object)\n\tif !ok {\n\t\tobjID = store.entities.Put(object)\n\t}\n\n\treturn subjID, propID, objID\n}", "func (daemon *Daemon) IdentityMapping() idtools.IdentityMapping {\n\treturn daemon.idMapping\n}", "func newMapCache(dss map[string]rrd.DataSourcer) *mapCache {\n\tmc := &mapCache{make(map[string]int64), make(map[int64]rrd.DataSourcer)}\n\tvar n int64\n\tfor name, ds := range dss {\n\t\tmc.byName[name] = n\n\t\tmc.byId[n] = ds\n\t\tn++\n\t}\n\treturn mc\n}", "func Map() map[string]interface{} {\n\treturn conf.Map()\n}", "func (c Certificates) AsMap() CertificateMap {\n\tcertMap := make(map[string]*CmdAdminCert)\n\tfor _, cert := range c {\n\t\tcertMap[cert.Name] = cert\n\t}\n\treturn certMap\n}", "func stripID(Map map[string]interface{}) map[string]interface{} {\n\tdelete(Map, \"_id\")\n\treturn Map\n}", "func (mpi *mempoolImpl) pubKeyAsNodeIDMap(nodePubKey *cryptolib.PublicKey, _ int) gpa.NodeID {\n\treturn mpi.pubKeyAsNodeID(nodePubKey)\n}", "func (archiver *Archiver) IdentityMapping() idtools.IdentityMapping {\n\treturn archiver.IDMapping\n}", "func IdentityMapper(input Keyer, outputs chan<- Keyer) {\n\tdefer close(outputs)\n\toutputs <- input\n}", "func (store *ProviderBackedStore) IDs() []Identifier {\n\treturn cloneIDs(store.ids)\n}", "func (g *UsersGetStoriesMaxIDsRequest) MapID() (value InputUserClassArray) {\n\treturn InputUserClassArray(g.ID)\n}", "func (c CPUMap) Map() map[string]int {\n\treturn map[string]int(c)\n}", "func (d *adapterMemoryData) SetMap(data map[interface{}]interface{}, expireTime int64) error {\n\td.mu.Lock()\n\tfor k, v := range data {\n\t\td.data[k] = adapterMemoryItem{\n\t\t\tv: v,\n\t\t\te: expireTime,\n\t\t}\n\t}\n\td.mu.Unlock()\n\treturn nil\n}", "func ToMap(object *js.Object) map[string]interface{} {\n\treturn object.Interface().(map[string]interface{})\n}", "func (v *Value) Map() map[string]interface{} {\n\tval := make(map[string]interface{})\n\tkeys := &Value{value: C.value_array_keys(v.value)}\n\n\tfor _, k := range keys.Slice() {\n\t\tswitch key := k.(type) {\n\t\tcase int64:\n\t\t\tt := &Value{value: C.value_array_index_get(v.value, C.ulong(key))}\n\t\t\tsk := strconv.Itoa((int)(key))\n\n\t\t\tval[sk] = t.Interface()\n\t\t\tt.Destroy()\n\t\tcase string:\n\t\t\tstr := C.CString(key)\n\t\t\tt := &Value{value: C.value_array_key_get(v.value, str)}\n\t\t\tC.free(unsafe.Pointer(str))\n\n\t\t\tval[key] = t.Interface()\n\t\t\tt.Destroy()\n\t\t}\n\t}\n\n\tkeys.Destroy()\n\treturn val\n}", "func (c *MapCache) ToMap() map[interface{}]interface{} {\n\tret := make(map[interface{}]interface{})\n\tfor k, v := range c.entries {\n\t\tret[k] = v\n\t}\n\treturn ret\n}", "func NewMapStore() *MapStore {\n\treturn &MapStore { store: make(map[string]domain.Customer)}\n}", "func (p Post) ToMap() map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"id\": p.ID,\n\t\t\"title\": p.Title,\n\t\t\"body\": p.Body,\n\t\t\"private\": p.Body,\n\t\t\"author\": p.Author,\n\t\t\"created_at\": p.CreatedAt,\n\t\t\"updated_at\": p.UpdatedAt,\n\t}\n}", "func (this *ExDomain) Values_asMap() map[int]bool {\n\treturn this.GetValues()\n}", "func (m *docsMap) Store(key string, value *Doc) {\n\tm.Data().Store(key, value)\n}", "func (m OrderMap) insertOrderIntoMap(order *Order) {\n m[order.IdNumber] = order\n}", "func (c *counters) toMap() (export map[string]int64) {\n\tres := make(map[string]int64)\n\n\tfor _, t := range c.subscriptions.keys() {\n\t\tc := c.subscriptions.load(t)\n\t\tmt := strings.ToLower(ptp.MessageType(t).String())\n\t\tres[fmt.Sprintf(\"subscriptions.%s\", mt)] = c\n\t}\n\n\tfor _, t := range c.rx.keys() {\n\t\tc := c.rx.load(t)\n\t\tmt := strings.ToLower(ptp.MessageType(t).String())\n\t\tres[fmt.Sprintf(\"rx.%s\", mt)] = c\n\t}\n\n\tfor _, t := range c.tx.keys() {\n\t\tc := c.tx.load(t)\n\t\tmt := strings.ToLower(ptp.MessageType(t).String())\n\t\tres[fmt.Sprintf(\"tx.%s\", mt)] = c\n\t}\n\n\tfor _, t := range c.rxSignaling.keys() {\n\t\tc := c.rxSignaling.load(t)\n\t\tmt := strings.ToLower(ptp.MessageType(t).String())\n\t\tres[fmt.Sprintf(\"rx.signaling.%s\", mt)] = c\n\t}\n\n\tfor _, t := range c.txSignaling.keys() {\n\t\tc := c.txSignaling.load(t)\n\t\tmt := strings.ToLower(ptp.MessageType(t).String())\n\t\tres[fmt.Sprintf(\"tx.signaling.%s\", mt)] = c\n\t}\n\n\tfor _, t := range c.workerQueue.keys() {\n\t\tc := c.workerQueue.load(t)\n\t\tres[fmt.Sprintf(\"worker.%d.queue\", t)] = c\n\t}\n\n\tfor _, t := range c.workerSubs.keys() {\n\t\tc := c.workerSubs.load(t)\n\t\tres[fmt.Sprintf(\"worker.%d.subscriptions\", t)] = c\n\t}\n\n\tfor _, t := range c.txtsattempts.keys() {\n\t\tc := c.txtsattempts.load(t)\n\t\tres[fmt.Sprintf(\"worker.%d.txtsattempts\", t)] = c\n\t}\n\n\tres[\"utcoffset\"] = c.utcoffset\n\n\treturn res\n}", "func (mt MapType) canStoreMap() bool {\n\treturn mt == ArrayOfMaps || mt == HashOfMaps\n}", "func (s InputSecureFileArray) ToMap() map[int64]InputSecureFile {\n\tr := make(map[int64]InputSecureFile, len(s))\n\ts.FillMap(r)\n\treturn r\n}", "func (dc *DockerClient) ListContainersAsMap() (map[string]docker.APIContainers, error) {\n\tcontainers := make(map[string]docker.APIContainers)\n\n\tcontainerList, err := dc.cl.ListContainers(docker.ListContainersOptions{})\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, cont := range containerList {\n\t\tshortID, err := GetShortID(cont.ID)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tcontainers[shortID] = cont\n\t}\n\n\tcontainers[\"root\"] = docker.APIContainers{ID: \"/\"}\n\n\tif len(containers) == 0 {\n\t\treturn nil, errors.New(\"No docker container found\")\n\t}\n\n\treturn containers, nil\n}", "func (m *Manager) pluginMap(id string) map[string]plugin.Plugin {\n\tpmap := map[string]plugin.Plugin{}\n\n\t// for _, pinfo := range m.Plugins {\n\t// \tpmap[pinfo.ID] = m.pluginImpl\n\t// }\n\n\tpmap[id] = m.pluginImpl\n\n\treturn pmap\n}", "func (store *HexastoreDB) MapIdsToStrings(subjID, propID, objectID int) (string, string, string) {\n\tsubject, _ := store.entities.Get(subjID)\n\tobject, _ := store.entities.Get(objectID)\n\tprop, _ := store.props.Get(propID)\n\n\treturn subject, prop, object\n}", "func (vectorizer *Vectorizer) indexMap(src Map) {\n\tfor _, key := range src.orderedKeys() {\n\t\telem := src[key]\n\t\tif vectorizer.indexes[key] == nil {\n\t\t\tvectorizer.indexes[key] = index{}\n\t\t}\n\t\tif vectorizer.indexes[key][elem] == 0 {\n\t\t\tvectorizer.indexes[key][elem] = vectorizer.nextID(key)\n\t\t}\n\t}\n}", "func (o *ExportData) FromMap(kv map[string]interface{}) {\n\n\to.ID = \"\"\n\n\t// if coming from db\n\tif id, ok := kv[\"_id\"]; ok && id != \"\" {\n\t\tkv[\"id\"] = id\n\t}\n\tif val, ok := kv[\"customer_id\"].(string); ok {\n\t\to.CustomerID = val\n\t} else {\n\t\tif val, ok := kv[\"customer_id\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.CustomerID = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.CustomerID = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"id\"].(string); ok {\n\t\to.ID = val\n\t} else {\n\t\tif val, ok := kv[\"id\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.ID = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.ID = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"integration_instance_id\"].(string); ok {\n\t\to.IntegrationInstanceID = val\n\t} else {\n\t\tif val, ok := kv[\"integration_instance_id\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.IntegrationInstanceID = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.IntegrationInstanceID = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"job_id\"].(string); ok {\n\t\to.JobID = val\n\t} else {\n\t\tif val, ok := kv[\"job_id\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.JobID = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.JobID = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"objects\"].(string); ok {\n\t\to.Objects = val\n\t} else {\n\t\tif val, ok := kv[\"objects\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.Objects = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.Objects = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"ref_id\"].(string); ok {\n\t\to.RefID = val\n\t} else {\n\t\tif val, ok := kv[\"ref_id\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.RefID = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.RefID = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\tif val, ok := kv[\"ref_type\"].(string); ok {\n\t\to.RefType = val\n\t} else {\n\t\tif val, ok := kv[\"ref_type\"]; ok {\n\t\t\tif val == nil {\n\t\t\t\to.RefType = \"\"\n\t\t\t} else {\n\t\t\t\tv := pstrings.Value(val)\n\t\t\t\tif v != \"\" {\n\t\t\t\t\tif m, ok := val.(map[string]interface{}); ok && m != nil {\n\t\t\t\t\t\tval = pjson.Stringify(m)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tval = v\n\t\t\t\t}\n\t\t\t\to.RefType = fmt.Sprintf(\"%v\", val)\n\t\t\t}\n\t\t}\n\t}\n\to.setDefaults(false)\n}", "func (p OrderedMap) RawMap() map[interface{}]interface{} {\n\treturn p.m\n}", "func (m *infoMap) Store(key string, value Info) {\n\tm.data.Store(key, value)\n}", "func NewMapStorage() *MapStorage {\n\treturn &MapStorage{store: make(map[string]int)}\n}", "func (cosi *cosiAggregate) StoreIdentities(idents map[string]proto.Message) {\n\tfor k, v := range idents {\n\t\tpoint := suite.G2().Point()\n\t\terr := point.UnmarshalBinary(v.(*BdnIdentity).PublicKey)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tcosi.skipchain.identities[k] = point\n\t}\n}", "func (s *syncMapInt64) store(key int, value int64) {\n\ts.Lock()\n\ts.m[key] = value\n\ts.Unlock()\n}", "func (node *Mapping) Map() (m map[interface{}]interface{}) {\n\tm = make(map[interface{}]interface{}, len(node.Pairs))\n\tfor _, pair := range node.Pairs {\n\t\tm[pair.Key.Data()] = pair.Value.Data()\n\t}\n\treturn\n}", "func Int64Map(src map[string]*int64) map[string]int64 {\n\tdst := make(map[string]int64)\n\tfor k, val := range src {\n\t\tif val != nil {\n\t\t\tdst[k] = *val\n\t\t}\n\t}\n\treturn dst\n}", "func (arc *AppRoleCredentials) Map() map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"role_id\": arc.RoleID,\n\t\t\"secret_id\": arc.SecretID,\n\t}\n}", "func (reply Reply) Map() (map[string]string, error) {\n\tresult, err := redis.StringMap(reply.data, reply.err)\n\tif err != nil {\n\t\treturn nil, redisError(err)\n\t}\n\n\treturn result, nil\n}", "func (object MQObject) InqMap(goSelectors []int32) (map[int32]interface{}, error) {\n\treturn object.Inq(goSelectors)\n}", "func ToPersistenceInfoMap(t map[string]*admin.PersistenceInfo) map[string]*types.PersistenceInfo {\n\tif t == nil {\n\t\treturn nil\n\t}\n\tv := make(map[string]*types.PersistenceInfo, len(t))\n\tfor key := range t {\n\t\tv[key] = ToPersistenceInfo(t[key])\n\t}\n\treturn v\n}", "func (m *privateSetMap) ToNativeMap() map[Value]struct{} {\n\tresult := make(map[Value]struct{})\n\tm.Range(func(key Value, value struct{}) bool {\n\t\tresult[key] = value\n\t\treturn true\n\t})\n\n\treturn result\n}", "func toMap(i interface{}) (iMap map[string]interface{}, err error) {\n\tbytes, err := json.Marshal(i)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = json.Unmarshal(bytes, &iMap)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func (iis *IIndexStore) PutIMap(key string, imap *trinity.InvertedMap) error {\n\tdata, err := imap.Marshal()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar compressed []byte\n\tcompressed, err = zstd.Compress(compressed, data)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = iis.store.Put(ds.NewKey(key), compressed)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (bm ByteMap) AsMap() map[string]interface{} {\n\tresult := make(map[string]interface{}, 10)\n\tbm.IterateValues(func(key string, value interface{}) bool {\n\t\tresult[key] = value\n\t\treturn true\n\t})\n\treturn result\n}", "func (this *DynMap) ToMap() map[string]interface{} {\n\tmp := make(map[string]interface{})\n\tfor k, v := range(this.Map) {\n\t\tsubmp, ok := ToDynMap(this.Map[k])\n\t\tif ok {\n\t\t\tv = submp.ToMap()\n\t\t}\n\t\tmp[k] = v\n\t}\n\treturn mp\n}", "func NewInMemoryMap() Map {\n\treturn &inMemoryMap{\n\t\tstore: make(map[string]StartStopper),\n\t}\n}", "func (n nexter) Map(id string, x Applicative, options ...*Option) Builder {\n\topt := &Option{\n\t\tBufferSize: intP(0),\n\t}\n\n\tif len(options) > 0 {\n\t\topt = opt.merge(options...)\n\t}\n\n\tnext := &node{}\n\tedge := newEdge(opt.BufferSize)\n\n\tnext.vertex = vertex{\n\t\tid: id,\n\t\tvertexType: \"map\",\n\t\tmetrics: createMetrics(id, \"map\"),\n\t\toption: opt,\n\t\thandler: func(payload []*Packet) {\n\t\t\tfor _, packet := range payload {\n\t\t\t\tpacket.apply(id, x)\n\t\t\t}\n\n\t\t\tedge.channel <- payload\n\t\t},\n\t\tconnector: func(ctx context.Context, b *builder) error {\n\t\t\tif next.next == nil {\n\t\t\t\treturn fmt.Errorf(\"non-terminated map\")\n\t\t\t}\n\t\t\treturn next.next.cascade(ctx, b, edge)\n\t\t},\n\t}\n\n\tnext = n(next)\n\n\treturn nexter(func(n *node) *node {\n\t\tnext.next = n\n\t\treturn n\n\t})\n}", "func (indexer Indexer) GetIdfDict() *map[int]float64 {\n\treturn &indexer.idfDict\n}", "func newMetricIDMapping(metricID, sequence uint32) MetricIDMapping {\n\treturn &metricIDMapping{\n\t\tmetricID: metricID,\n\t\thash2SeriesID: make(map[uint64]uint32),\n\t\tidSequence: *atomic.NewUint32(sequence), // first value is 1\n\t\tmaxSeriesIDsLimit: *atomic.NewUint32(constants.DefaultMaxSeriesIDsCount),\n\t}\n}", "func (rm *FilteredResultMap) Store(key string, r *FilteredResult) {\n\trm.sm.Store(key, r)\n}", "func (grid *SquareGrid) Map() map[Loc]interface{} {\n\treturn grid.Data\n}" ]
[ "0.64095354", "0.6337629", "0.63284034", "0.6106668", "0.5993814", "0.5982267", "0.5865659", "0.5710766", "0.56578773", "0.5631537", "0.5630683", "0.5629808", "0.5601113", "0.55943334", "0.5587781", "0.55481577", "0.55180997", "0.5489575", "0.5474147", "0.5472326", "0.54506284", "0.54015446", "0.5401397", "0.5395907", "0.53795224", "0.5378234", "0.53625053", "0.535306", "0.5340042", "0.53166777", "0.53083575", "0.5306904", "0.53021264", "0.52998805", "0.529048", "0.52873635", "0.52795327", "0.5261939", "0.5256385", "0.52480745", "0.5232834", "0.5229738", "0.5224723", "0.5222585", "0.5212398", "0.5210287", "0.5208473", "0.52034336", "0.5198437", "0.51977766", "0.5184106", "0.51722884", "0.5163595", "0.5153813", "0.5147571", "0.5121037", "0.5119386", "0.5118036", "0.5099959", "0.5086099", "0.50832635", "0.508149", "0.50792617", "0.50726944", "0.5069731", "0.506948", "0.50632393", "0.5061601", "0.5059683", "0.50528634", "0.5051278", "0.50463355", "0.5043463", "0.5039952", "0.5038445", "0.503531", "0.5029825", "0.5023965", "0.5020675", "0.5018343", "0.50103915", "0.5007454", "0.5001169", "0.49985564", "0.49969706", "0.4996557", "0.49948707", "0.49944943", "0.49917108", "0.49810433", "0.49752373", "0.49623543", "0.49621558", "0.49612597", "0.49582738", "0.49527037", "0.4952515", "0.49524626", "0.49504432", "0.49490827" ]
0.7167768
0
Return a new interceptor function that authorizes RPCs using a password stored in the config.
func unaryAuthInterceptor(auth Authenticate, access Access) grpc.UnaryServerInterceptor { return func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) { //fmt.Printf("AuthInt: %#v\n", ctx) md, _ := metadata.FromIncomingContext(ctx) //fmt.Printf("Metadata: %#v\n", md) //omd, _ := metadata.FromOutgoingContext(ctx) //fmt.Printf("Raw: %#v\n", omd) metaData := MetaData{} for i := range md { metaData[i] = md[i] } user, err := auth.Validate(metaData) if err != nil { return nil, status.Error(codes.Unauthenticated, "PermissionDenied") } if op, ok := MethodMap[info.FullMethod]; ok { graph, err := getUnaryRequestGraph(req, info) if err != nil { return nil, status.Error(codes.Unknown, fmt.Sprintf("Unable to get graph: %s", err)) } err = access.Enforce(user, graph, op) if err != nil { return nil, status.Error(codes.PermissionDenied, "PermissionDenied") } return handler(ctx, req) } return nil, status.Error(codes.Unknown, "Unknown method") } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *server) authInterceptor(c context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\tif s.configs.Authorization {\n\t\tmethods := strings.Split(info.FullMethod, \"/\")\n\t\tif md, ok := metadata.FromIncomingContext(c); !ok {\n\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t} else {\n\t\t\tif methods[len(methods)-1] == \"UserLogin\" {\n\t\t\t\tr := req.(*pb.AuthInfo)\n\t\t\t\tif result, err := s.gdb.userLogin(authInfo{\n\t\t\t\t\tUserName: r.GetUserName(),\n\t\t\t\t\tPassWord: r.GetPassWord(),\n\t\t\t\t}); err != nil {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t} else {\n\t\t\t\t\treturn &pb.UserToken{Token: result.Token}, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tvar au string\n\t\t\t\tif d, ok := md[\"authorization\"]; ok {\n\t\t\t\t\tau = d[0]\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t}\n\t\t\t\tif userName, token, ok := parseBasicAuth(au); !ok {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t} else {\n\t\t\t\t\tif r, err := s.gdb.query(\"select token from user_cfg where userName='\" + userName + \"'\"); err != nil || len(r) == 0 {\n\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif token != r[0][\"token\"] {\n\t\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tsub, obj, act := userName, methods[len(methods)-1], \"POST\" // replace gRCP with POST\n\t\t\t\t\t\t\tif ok, _ := s.gdb.e.Enforce(sub, obj, act); !ok {\n\t\t\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn handler(c, req)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn handler(c, req)\n\t}\n}", "func New(c Config) lilty.ChainHandler {\n\treturn func(next lilty.HandlerFunc) lilty.HandlerFunc {\n\t\treturn func(ctxt *lilty.Context) {\n\t\t\tusername, password, ok := ctxt.Request.BasicAuth()\n\n\t\t\tmatch := c.Username == username && c.Password == password\n\n\t\t\tif !ok || !match {\n\t\t\t\tctxt.SetResponseHeader(lilty.WWWAuthenticate, `Basic realm=\"`+c.Realm+`\"`)\n\t\t\t\tctxt.SetStatusCode(401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext(ctxt)\n\t\t}\n\t}\n}", "func authWrapFunc(f func(http.ResponseWriter, *http.Request)) http.Handler {\n\treturn negroni.New(&AuthMiddleware{}, negroni.Wrap(http.HandlerFunc(f)))\n}", "func authWrapper(handler http.HandlerFunc, secrets auth.SecretProvider, host string) http.HandlerFunc {\n\tauthenticator := &auth.BasicAuth{Realm: host, Secrets: secrets}\n\treturn auth.JustCheck(authenticator, handler)\n}", "func authUnaryIntercept(\n\tctx context.Context,\n\treq interface{},\n\tinfo *grpc.UnaryServerInfo,\n\thandler grpc.UnaryHandler,\n) (resp interface{}, err error) {\n\t//bypass auth if method is /hahiye.AuthService/Login\n\tif info.FullMethod == \"/hahiye.AuthService/Login\" {\n\t\tfmt.Println(\"bypassing auth cz it's login action\")\n\t\treturn handler(ctx, req)\n\t}\n\tif err := auth(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(\"authorization OK\")\n\treturn handler(ctx, req)\n}", "func New(username, password string) *client {\n\tr := newRequestHandler()\n\tc := &client{\n\t\tusername: username,\n\t\tpassword: password,\n\t\toauth: AccessTokenStruct{},\n\t\thandler: r,\n\t\twsConn: nil,\n\t\treconnectToken: uuid.Nil,\n\t}\n\terr := c.authenticate()\n\tif err != nil {\n\t\tc.handler.override = responseOverride{override: true, data: *err}\n\t}\n\tgo c.authenticator() // Launch authenticator\n\treturn c\n}", "func interceptor(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\n\tif err := auth(ctx); err != nil {\n\t\tfmt.Println(\"111\")\n\t\treturn nil, err\n\t}\n\t//继续处理请求\n\treturn handler(ctx, req)\n\n}", "func authEndpoint(rw http.ResponseWriter, req *http.Request) {\n\n\t// request has to be POST\n\tif req.Method != \"POST\" {\n\t\thttp.Error(rw, \"bad method, only post allowed\", http.StatusBadRequest)\n\t}\n\n\t// has to be authenticated, in a real we would use soemthing more\n\t// secure like certificates etc.\n\tuser, _, ok := req.BasicAuth()\n\n\tif !ok {\n\t\thttp.Error(rw, \"authentication required\", http.StatusForbidden)\n\t}\n\n\tlog.Println(\"basic authentication successful for \", user)\n\n\t// now we issue token and return it\n\n\t// This context will be passed to all methods.\n\tctx := req.Context()\n\n\t// Create an empty session object which will be passed to the request handlers\n\tmySessionData := newSession(\"\")\n\n\t// This will create an access request object and iterate through the registered TokenEndpointHandlers to validate the request.\n\taccessRequest, err := fositeInstance.NewAccessRequest(ctx, req, mySessionData)\n\n\t// Catch any errors, e.g.:\n\t// * unknown client\n\t// * invalid redirect\n\t// * ...\n\tif err != nil {\n\t\tlog.Printf(\"Error occurred in NewAccessRequest: %+v\", err)\n\t\tfositeInstance.WriteAccessError(rw, accessRequest, err)\n\t\treturn\n\t}\n\n\t// If this is a client_credentials grant, grant all requested scopes\n\t// NewAccessRequest validated that all requested scopes the client is allowed to perform\n\t// based on configured scope matching strategy.\n\tif accessRequest.GetGrantTypes().ExactOne(\"client_credentials\") {\n\t\tfor _, scope := range accessRequest.GetRequestedScopes() {\n\t\t\taccessRequest.GrantScope(scope)\n\t\t}\n\t}\n\n\t// Next we create a response for the access request. Again, we iterate through the TokenEndpointHandlers\n\t// and aggregate the result in response.\n\tresponse, err := fositeInstance.NewAccessResponse(ctx, accessRequest)\n\tif err != nil {\n\t\tlog.Printf(\"Error occurred in NewAccessResponse: %+v\", err)\n\t\tfositeInstance.WriteAccessError(rw, accessRequest, err)\n\t\treturn\n\t}\n\n\t// All done, send the response.\n\tfositeInstance.WriteAccessResponse(rw, accessRequest, response)\n\n}", "func (fn AuthenticatorFunc) Authenticate(username, password, otp string) (*Session, error) {\n\treturn fn(username, password, otp)\n}", "func AuthInterceptor(authURL string, insecure bool) grpc.UnaryServerInterceptor {\n\tauth := newKeystoneAuth(authURL, insecure)\n\n\treturn func(ctx context.Context, req interface{},\n\t\tinfo *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\tmd, ok := metadata.FromIncomingContext(ctx)\n\t\tif !ok {\n\t\t\treturn nil, errutil.ErrorUnauthenticated\n\t\t}\n\t\ttoken := md[\"x-auth-token\"]\n\t\tif len(token) == 0 {\n\t\t\treturn nil, errutil.ErrorUnauthenticated\n\t\t}\n\t\tnewCtx, err := authenticate(ctx, auth, token[0])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn handler(newCtx, req)\n\t}\n}", "func New(authenticator auth.Authenticator) clevergo.MiddlewareFunc {\n\treturn func(next clevergo.Handle) clevergo.Handle {\n\t\treturn func(c *clevergo.Context) error {\n\t\t\tidentity, err := authenticator.Authenticate(c.Request, c.Response)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t\tauthenticator.Challenge(c.Request, c.Response)\n\t\t\t} else {\n\t\t\t\tc.WithValue(auth.IdentityKey, identity)\n\t\t\t}\n\t\t\treturn next(c)\n\t\t}\n\t}\n}", "func AuthWrapper(fn server.HandlerFunc) server.HandlerFunc {\n\treturn func(ctx context.Context, req server.Request, resp interface{}) error {\n\t\tmeta, ok := metadata.FromContext(ctx)\n\t\tif !ok {\n\t\t\treturn errors.New(\"no auth meta-data found in request\")\n\t\t}\n\n\t\t// Note this is now uppercase (not entirely sure why this is...)\n\t\ttoken := meta[\"token\"]\n\n\t\tif token == \"\" {\n\t\t\treturn errors.New(\"token is empty\")\n\t\t}\n\n\t\tlog.Println(\"Authenticating with token: \", token)\n\n\t\t// Auth here\n\t\tauthClient := userService.NewUserServiceClient(\"user\", k8s.NewService().Client())\n\t\t_, err := authClient.ValidateToken(context.Background(), &userService.Token{\n\t\t\tToken: token,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = fn(ctx, req, resp)\n\t\treturn err\n\t}\n}", "func streamAuthIntercept(\n\tserver interface{},\n\tstream grpc.ServerStream,\n\tinfo *grpc.StreamServerInfo,\n\thandler grpc.StreamHandler,\n) error {\n\t//bypass auth if method is /hahiye.AuthService/Login\n\tif info.FullMethod == \"/hahiye.AuthService/Login\" {\n\t\tfmt.Println(\"bypassing auth cz it's login action\")\n\t\treturn handler(server, stream)\n\t}\n\tif err := auth(stream.Context()); err != nil {\n\t\treturn err\n\t}\n\tlog.Println(\"authorization OK\")\n\treturn handler(server, stream)\n}", "func NewAuthInterceptor(\n\tjwtToken string,\n\tauthMethods map[string]bool,\n) (*AuthInterceptor, error) {\n\tinterceptor := &AuthInterceptor{\n\t\tauthMethods: authMethods,\n\t\taccessToken: jwtToken,\n\t}\n\tlog.Printf(\"Inside Client AuthIntercept\")\n\treturn interceptor, nil\n}", "func New(realm string, credentials map[string]string, protectedHTTPMethods []string, enabled bool) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tif !enabled {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tmethodIsProtected := false\n\t\t\tfor _, protectedHTTPMethod := range protectedHTTPMethods {\n\t\t\t\tif r.Method == protectedHTTPMethod {\n\t\t\t\t\tmethodIsProtected = true\n\t\t\t\t}\n\t\t\t}\n\t\t\tif !methodIsProtected {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tusername, password, ok := r.BasicAuth()\n\t\t\tif !ok {\n\t\t\t\tunauthorized(w, realm)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tvalidPassword, userFound := credentials[username]\n\t\t\tif userFound {\n\t\t\t\tif password == validPassword {\n\t\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tunauthorized(w, realm)\n\t\t})\n\t}\n}", "func AuthAndCallAPI(w http.ResponseWriter, r *http.Request, service string, method string, version string) {\n\te := Execution{name: \"AuthAndCallAPI \" + service}\n\te.Start()\n\n\tauthorization := r.Header.Get(\"authorization\")\n\n\ttoken := \"\"\n\ts := strings.Split(authorization, \" \")\n\tif len(s) >= 2 {\n\t\ttoken = s[1]\n\t}\n\n\tconfig := config.GetConfig()\n\tresp, _ := resty.R().\n\t\tSetFormData(map[string]string{\n\t\t\t\"token\": token,\n\t\t\t\"service\": service,\n\t\t}).\n\t\tSetResult(&Respon{}).\n\t\tPost(config.API.Auth + \"v100/auth/check_token\")\n\n\tvar respon Respon\n\t_ = json.Unmarshal(resp.Body(), &respon)\n\n\tif respon.Code != 200 {\n\t\trespond := Respon{\n\t\t\tStatus: respon.Status,\n\t\t\tCode: respon.Code,\n\t\t\tMessage: respon.Message,\n\t\t\tExeTime: respon.ExeTime,\n\t\t\tData: respon.Data,\n\t\t\tError: respon.Error,\n\t\t}\n\t\tRespondJson(w, resp.StatusCode(), respond)\n\t\treturn\n\t}\n\n\tCallAPI(w, r, service, method, version)\n}", "func (p *UserPool) AuthUser(name string, password string) {\n\n}", "func Authenticator(t *Token) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\tfn := func(w http.ResponseWriter, req *http.Request) {\n\t\t\ttokenString, err := t.GetToken(req)\n\t\t\tif err != nil {\n\t\t\t\thttp.Error(w, err.Error(), http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tdata, err := t.CheckToken(tokenString)\n\t\t\tif err != nil {\n\t\t\t\thttp.Error(w, err.Error(), http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// If we get here, everything worked and we can set the\n\t\t\t// user data in context\n\t\t\tnewReq := req.WithContext(context.WithValue(req.Context(), t.options.UserProperty, data))\n\t\t\t// update the current request with the new context information\n\t\t\t*req = *newReq\n\t\t\tnext.ServeHTTP(w, req)\n\t\t}\n\t\treturn http.HandlerFunc(fn)\n\t}\n}", "func (b *BasicAuthenticationBackend) Wrap(wrapped auth.AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := tokenFromRequest(r)\n\n\t\t// if not coming from cookie or X-Auth-Token\n\t\ts := strings.SplitN(token, \" \", 2)\n\t\tif len(s) == 2 || s[0] == \"Basic\" {\n\t\t\ttoken = s[1]\n\t\t}\n\n\t\t// add \"fake\" header to let the basic auth library do the authentication\n\t\tr.Header.Set(\"Authorization\", \"Basic \"+token)\n\n\t\tif username := b.CheckAuth(r); username == \"\" {\n\t\t\tUnauthorized(w, r, ErrWrongCredentials)\n\t\t} else {\n\t\t\tauthCallWrapped(w, r, username, wrapped)\n\t\t}\n\t}\n}", "func (*SpecialAuth) AuthFunc(_ context.Context, apiKey string, _ map[string]string) bool {\n\treturn apiKey == \"12345\"\n}", "func NewAuthInterceptor(\n\tauthn authn.AuthenticationServiceClient,\n\tauthz GRPCAuthorizationHandler,\n) AuthorizationInterceptor {\n\treturn &authInterceptor{authn: authn, authz: authz}\n}", "func CustomAuth(cfg *types.Config) gin.HandlerFunc {\n\tbasicAuthHandler := gin.BasicAuth(gin.Accounts{\n\t\t// Use the config's username and password for basic auth\n\t\tcfg.Username: cfg.Password,\n\t})\n\n\toidcHandler := getOIDCMiddleware(cfg.OIDCIssuer, cfg.OIDCSubject, cfg.OIDCGroups)\n\n\treturn func(c *gin.Context) {\n\t\tauthHeader := c.GetHeader(\"Authorization\")\n\t\tif strings.HasPrefix(authHeader, \"Bearer \") {\n\t\t\toidcHandler(c)\n\t\t} else {\n\t\t\tbasicAuthHandler(c)\n\t\t}\n\t}\n}", "func WrapAuthenticate(hfn http.Handler) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlVars := mux.Vars(r)\n\t\turlValues := r.URL.Query()\n\n\t\t// if the url parameter 'key' is empty or absent, end the request with an unauthorized response\n\t\tif urlValues.Get(\"key\") == \"\" {\n\t\t\terr := APIErrorUnauthorized()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\t\tserviceToken := gorillaContext.Get(r, \"auth_service_token\").(string)\n\n\t\tprojectName := urlVars[\"project\"]\n\t\tprojectUUID := projects.GetUUIDByName(urlVars[\"project\"], refStr)\n\n\t\t// In all cases instead of project create\n\t\tif \"projects:create\" != mux.CurrentRoute(r).GetName() {\n\t\t\t// Check if given a project name the project wasn't found\n\t\t\tif projectName != \"\" && projectUUID == \"\" {\n\t\t\t\tapiErr := APIErrorNotFound(\"project\")\n\t\t\t\trespondErr(w, apiErr)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t// Check first if service token is used\n\t\tif serviceToken != \"\" && serviceToken == urlValues.Get(\"key\") {\n\t\t\tgorillaContext.Set(r, \"auth_roles\", []string{\"service_admin\"})\n\t\t\tgorillaContext.Set(r, \"auth_user\", \"\")\n\t\t\tgorillaContext.Set(r, \"auth_user_uuid\", \"\")\n\t\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\t\thfn.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\n\t\troles, user := auth.Authenticate(projectUUID, urlValues.Get(\"key\"), refStr)\n\n\t\tif len(roles) > 0 {\n\t\t\tuserUUID := auth.GetUUIDByName(user, refStr)\n\t\t\tgorillaContext.Set(r, \"auth_roles\", roles)\n\t\t\tgorillaContext.Set(r, \"auth_user\", user)\n\t\t\tgorillaContext.Set(r, \"auth_user_uuid\", userUUID)\n\t\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\t\thfn.ServeHTTP(w, r)\n\t\t} else {\n\t\t\terr := APIErrorUnauthorized()\n\t\t\trespondErr(w, err)\n\t\t}\n\n\t})\n}", "func AuthFunc(ctx context.Context) (context.Context, error) {\n\treturn ctx, nil\n}", "func New(addr string, c auth.Cache, opts ...auth.Option) auth.Strategy {\n\tfn := GetAuthenticateFunc(addr, opts...)\n\treturn token.New(fn, c, opts...)\n}", "func New(addr string, c auth.Cache, opts ...auth.Option) auth.Strategy {\n\tfn := GetAuthenticateFunc(addr, opts...)\n\treturn token.New(fn, c, opts...)\n}", "func (b *KeystoneAuthenticationBackend) Wrap(wrapped auth.AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := tokenFromRequest(r)\n\n\t\tif username, err := b.CheckUser(token); username == \"\" {\n\t\t\tUnauthorized(w, r, err)\n\t\t} else {\n\t\t\tauthCallWrapped(w, r, username, wrapped)\n\t\t}\n\t}\n}", "func authWrap(h http.Handler) http.Handler {\n\treturn negroni.New(&AuthMiddleware{}, negroni.Wrap(h))\n}", "func AuthenticationInterceptor(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (out interface{}, err error) {\n\tmethodParts := strings.Split(info.FullMethod, \"/\")\n\texceptions := []string{\"Auth\", \"Register\", \"Items\", \"Item\", \"AddOrder\"}\n\t_, found := utils.Find(exceptions, methodParts[len(methodParts)-1])\n\tif found {\n\t\treturn handler(ctx, req)\n\t}\n\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif !ok {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\ttokenString, ok := md[\"authorization\"]\n\tif !ok || len(tokenString) < 1 || len(tokenString[0]) < 8 {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\tclaims := &auth.Claims{}\n\ttoken, err := jwt.ParseWithClaims(tokenString[0][7:], claims, func(token *jwt.Token) (interface{}, error) {\n\t\treturn jwtKey, nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tuserID := claims.UserID\n\n\t// load userID into context\n\tauthContext := auth.AuthContext{\n\t\tUserID: userID,\n\t}\n\n\tif token.Valid {\n\t\tctx = context.WithValue(ctx, auth.UserAuthKey, authContext)\n\t} else {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\treturn handler(ctx, req)\n}", "func login(service Service, logger log.Logger) routing.Handler {\n return func(c *routing.Context) error {\n var req struct {\n Username string `json:\"username\"`\n Password string `json:\"password\"`\n }\n\n if err := c.Read(&req); err != nil {\n logger.With(c.Request.Context()).Errorf(\"invalid request: %v\", err)\n return errors.BadRequest(\"\")\n }\n\n token, err := service.Login(c.Request.Context(), req.Username, req.Password)\n if err != nil {\n return err\n }\n\n return c.Write(struct {\n Token string `json:\"token\"`\n }{token})\n }\n}", "func (c *Commander) SignIn(w http.ResponseWriter, r *http.Request) {\n\tvar client model.Client\n\tvar err error\n\tvar error model.Error\n\tdb := database.DbConn()\n\tdefer db.Close()\n\t// create a new client\n\tif client, err = New(model.Config{\n\t\tBaseDN: \"DC=sls,DC=ads,DC=valuelabs,DC=net\",\n\t\t//BaseDN: \"cn=ltest,ou=SERVICE ACCOUNTS,ou=SLS,dc=SLS,dc=ads,dc=valuelabs,dc=net\",\n\t\tFilter: \"userPrincipalName\",\n\t\tROUser: model.User{Name: \"L test\", Password: \"Welcome@123\"},\n\t\tTitle: \"title\",\n\t\tHost: \"10.10.52.113:389\",\n\t}); err != nil {\n\t\tWriteLogFile(err)\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\tvar creds model.Credentials\n\t//\tvar pass string\n\tSetupResponse(&w, r)\n\tif (*r).Method == \"OPTIONS\" {\n\t\tw.Header().Set(\"Access-Control-Max-Age\", \"86400\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t// Get the JSON body and decode into credentials\n\terr = json.NewDecoder(r.Body).Decode(&creds)\n\tif err != nil {\n\t\t// If the structure of the body is wrong, return an HTTP error\n\t\tWriteLogFile(err)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\t// var usr = creds.Username\n\t// var bytePassword = []byte(creds.Password)\n\tusername := creds.Username\n\tpassword := creds.Password\n\tRole = creds.Role\n\tsplitUser := strings.Split(username, \"@\")\n\tprint := splitUser[0]\n\tuser1 := fmt.Sprintf(\"%[email protected]\", print)\n\tuser2 := fmt.Sprintf(\"%[email protected]\", print)\n\terr = client.Auth(user2, password)\n\tif err == nil {\n\t\tfmt.Println(\"Success!\")\n\t\ttoken(w, user1)\n\t} else if err.Error() == \"not found\" {\n\t\tfmt.Println(\"H2\")\n\t\tif errr := client.Auth(user1, password); errr != nil {\n\t\t\tfmt.Println(\"H3\")\n\t\t\tWriteLogFile(errr)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\terror.Code = \"401\"\n\t\t\terror.Message = \"Invalid Username or Password\"\n\t\t\tjson.NewEncoder(w).Encode(error)\n\t\t\treturn\n\t\t} //else {\n\t\tfmt.Println(\"Success!\")\n\t\ttoken(w, user1)\n\t\t//}\n\t} else {\n\t\tfmt.Println(\"H4\")\n\t\tWriteLogFile(err)\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\terror.Code = \"401\"\n\t\terror.Message = \"Invalid Username or Password\"\n\t\tjson.NewEncoder(w).Encode(error)\n\t\treturn\n\t}\n}", "func Credentials(addr, username, password string) InfluxReporterFunc {\n\treturn func(c *InfluxReporter) {\n\t\tc.addr = addr\n\t\tc.username = username\n\t\tc.password = password\n\t}\n}", "func (b *basicAuth) set(r *http.Request) { r.SetBasicAuth(b.username, b.password) }", "func makeAuthFunc(schemes map[string]bool) func(ctx context.Context, input *AuthenticationInput) error {\n\treturn func(ctx context.Context, input *AuthenticationInput) error {\n\t\t// If the scheme is valid and present in the schemes\n\t\tvalid, present := schemes[input.SecuritySchemeName]\n\t\tif valid && present {\n\t\t\treturn nil\n\t\t}\n\n\t\t// If the scheme is present in che schemes\n\t\tif present {\n\t\t\t// Return an unmet scheme error\n\t\t\treturn fmt.Errorf(\"security scheme for %q wasn't met\", input.SecuritySchemeName)\n\t\t}\n\t\t// Return an unknown scheme error\n\t\treturn fmt.Errorf(\"security scheme for %q is unknown\", input.SecuritySchemeName)\n\t}\n}", "func (m *BasicAuthentication) SetPassword(value *string)() {\n m.password = value\n}", "func setBasicAuth(u *vcs.URL, extSvcType, username, password string) error {\n\tswitch extSvcType {\n\tcase extsvc.TypeGitHub, extsvc.TypeGitLab:\n\t\treturn errors.New(\"need token to push commits to \" + extSvcType)\n\n\tcase extsvc.TypeBitbucketServer, extsvc.TypeBitbucketCloud:\n\t\tu.User = url.UserPassword(username, password)\n\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"setBasicAuth: invalid external service type %q\", extSvcType))\n\t}\n\treturn nil\n}", "func (s *Server) AuthFuncOverride(ctx context.Context, fullMethodName string) (context.Context, error) {\n\treturn ctx, nil\n}", "func NewSetAuth() context.Handler {\n\treturn func(ctx context.Context) {\n\t\taddr := ctx.RemoteAddr()\n\t\tif _, ok := tempForbidRemote[addr]; ok {\n\t\t\tctx.StatusCode(404)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\n\t\tkey := ctx.Params().Get(\"name\")\n\t\ttoken := ctx.GetHeader(\"token\")\n\t\tif token == \"\" || getServerToken(key) != token {\n\t\t\ttempForbidRemote[addr] = 1\n\t\t\tctx.StatusCode(404)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\t\tctx.Next()\n\t}\n}", "func newTokenInjector(meta *metadata.Client, acc string) func(*http.Request) error {\n\tif acc == \"\" {\n\t\tacc = \"default\"\n\t}\n\tacc = url.PathEscape(acc)\n\treturn func(req *http.Request) error {\n\t\taud := fmt.Sprintf(\"%s://%s\", req.URL.Scheme, req.URL.Host)\n\t\taud = url.QueryEscape(aud)\n\t\t// TODO(smut): Cache the token and reuse if not yet expired.\n\t\t// Currently the only user of this package only makes one\n\t\t// request per boot so caching isn't too important yet.\n\t\ttok, err := meta.Get(fmt.Sprintf(tokMetadata, acc, aud))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Header.Set(vmtoken.Header, tok)\n\t\treturn nil\n\t}\n}", "func ClientPassword(password string) func(*Client) {\n\treturn func(c *Client) { c.password = password }\n}", "func (ba BasicAuthenticator) Wrap(handler http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tusername, password, ok := r.BasicAuth()\n\n\t\tif !ok {\n\t\t\tvar strRequest string\n\t\t\tdumpRequest, err := httputil.DumpRequest(r, true)\n\t\t\tif err == nil {\n\t\t\t\tstrRequest = string(dumpRequest)\n\t\t\t}\n\t\t\terrStr := fmt.Sprint(\"credentials error\", \" error:\", ErrInvalidCredFormat, \" request:\", strRequest)\n\t\t\tw.Header().Set(\"WWW-Authenticate\", \"Basic realm=Authorization Required\")\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tdtlogger.Error(errStr)\n\t\t\t//http.Error(w, errStr, http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\tif username != ba.username || password != ba.password {\n\t\t\tvar strRequest string\n\t\t\tdumpRequest, err := httputil.DumpRequest(r, true)\n\t\t\tif err == nil {\n\t\t\t\tstrRequest = string(dumpRequest)\n\t\t\t}\n\t\t\terrStr := fmt.Sprint(\"credentials error\", \" error:\", ErrInvalidCreds, \" request:\", strRequest)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tdtlogger.Error(errStr)\n\t\t\t//http.Error(w, errStr, http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\thandler.ServeHTTP(w, r)\n\t})\n}", "func Authenticate(endpoint func(http.ResponseWriter, *http.Request)) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\tendpoint(w, r)\n\t})\n}", "func CredentialCtx(s *Server) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t// Get session cookie.\n\t\t\tsession := \"\"\n\t\t\tif cookie, err := r.Cookie(\"session\"); err == nil {\n\t\t\t\tsession = cookie.Value\n\t\t\t}\n\n\t\t\t// Get API key header.\n\t\t\tapiKey := \"\"\n\t\t\tif apiKeys, ok := r.Header[\"X-Api-Key\"]; ok {\n\t\t\t\tapiKey = apiKeys[0]\n\t\t\t}\n\t\t\tapiSecret := \"\"\n\t\t\tif apiSecrets, ok := r.Header[\"X-Api-Secret\"]; ok {\n\t\t\t\tapiSecret = apiSecrets[0]\n\t\t\t}\n\n\t\t\tif session == \"\" && apiKey == \"\" {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tauthInfo := chassis.AuthInfo{}\n\t\t\tif session != \"\" {\n\t\t\t\tif userID, _, isAdmin, err := s.db.LookupSession(session); err == nil {\n\t\t\t\t\tauthInfo.AuthMethod = chassis.SessionAuth\n\t\t\t\t\tauthInfo.UserID = userID\n\t\t\t\t\tauthInfo.UserIsAdmin = isAdmin\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t//TODO: CACHE MAY BE WORTH IT\n\t\t\t\tif user, err := s.userSvc.GetUserByApiKey(apiKey, apiSecret); err == nil {\n\t\t\t\t\tauthInfo.AuthMethod = chassis.SessionAuth\n\t\t\t\t\tauthInfo.UserID = user.ID\n\t\t\t\t\tauthInfo.UserIsAdmin = user.IsAdmin\n\t\t\t\t}\n\t\t\t}\n\t\t\tnext.ServeHTTP(w, r.WithContext(chassis.NewAuthContext(r.Context(), &authInfo)))\n\t\t})\n\t}\n}", "func (a *BasicAuth) Wrap(wrapped AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tif username := a.CheckAuth(r); username == \"\" {\n\t\t\ta.RequireAuth(w, r)\n\t\t} else {\n\t\t\tar := &AuthenticatedRequest{Request: *r, Username: username}\n\t\t\twrapped(w, ar)\n\t\t}\n\t}\n}", "func (c Clients) Auth(ctx context.Context, username, password string) error {\n\tvar req *request\n\tif username == \"\" {\n\t\treq = newRequest(\"*2\\r\\n$4\\r\\nAUTH\\r\\n$\")\n\t\treq.addString(password)\n\t} else {\n\t\treq = newRequest(\"*3\\r\\n$4\\r\\nAUTH\\r\\n$\")\n\t\treq.addString2(username, password)\n\t}\n\treturn c.c.cmdSimple(ctx, req)\n}", "func CustomAuth(handler http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := r.Header.Get(\"Authorization\")\n\t\ttokenPieces := strings.Split(token, \" \")\n\t\tif len(tokenPieces) == 2 && tokenPieces[0] == \"Basic\" {\n\t\t\t// TODO: do basic auth\n\t\t\thandler.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\t\t// TODO: do session auth\n\n\t\t// if both validations fails\n\t\thttp.Error(w, \"error while trying to authenticate\", http.StatusUnauthorized)\n\n\t})\n}", "func New(config Config) App {\n\treturn App{\n\t\treq: request.New().Post(strings.TrimSpace(*config.url)).BasicAuth(strings.TrimSpace(*config.user), *config.pass),\n\t}\n}", "func newProxy(config *Config) (*oauthProxy, error) {\n\t// create the service logger\n\tlog, err := createLogger(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.Info(\"starting the service\", zap.String(\"prog\", prog), zap.String(\"author\", author), zap.String(\"version\", version))\n\tsvc := &oauthProxy{\n\t\tconfig: config,\n\t\tlog: log,\n\t\tmetricsHandler: prometheus.Handler(),\n\t}\n\n\t// parse the upstream endpoint\n\tif svc.endpoint, err = url.Parse(config.Upstream); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// initialize the store if any\n\tif config.StoreURL != \"\" {\n\t\tif svc.store, err = createStorage(config.StoreURL); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// initialize the openid client\n\tif !config.SkipTokenVerification {\n\t\tif svc.client, svc.idp, svc.idpClient, err = svc.newOpenIDClient(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tlog.Warn(\"TESTING ONLY CONFIG - the verification of the token have been disabled\")\n\t}\n\n\tif config.ClientID == \"\" && config.ClientSecret == \"\" {\n\t\tlog.Warn(\"client credentials are not set, depending on provider (confidential|public) you might be unable to auth\")\n\t}\n\n\t// are we running in forwarding mode?\n\tif config.EnableForwarding {\n\t\tif err := svc.createForwardingProxy(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tif err := svc.createReverseProxy(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn svc, nil\n}", "func Auth(fn func(ctx *system.Context)) func(ctx *system.Context) {\n\treturn func(ctx *system.Context) {\n\t\tgconfig, err := ctx.System.DB.CreateGuildIfNotExists(ctx.Msg.GuildID)\n\t\tif err != nil {\n\t\t\tctx.ReplyError(\"Error getting guild configuration: \", err)\n\t\t\treturn\n\t\t}\n\n\t\tisAdmin, err := ctx.IsAdmin()\n\t\tif err != nil {\n\t\t\tctx.ReplyError(\"Error checking administrator status: \", err)\n\t\t\treturn\n\t\t}\n\n\t\tif !isAdmin {\n\t\t\tctx.ReplyError(\"You need to be an administrator or own the guild to configure guild settings\")\n\t\t\treturn\n\t\t}\n\n\t\tctx.Set(\"gconfig\", gconfig)\n\t\tfn(ctx)\n\t}\n}", "func (r *RPC) Auth(c context.Context, arg *rpc.Auth, res *struct{}) (err error) {\n\treturn\n}", "func WrapHandlerInBasicAuth(h http.Handler, b BasicAuth) http.HandlerFunc {\n\tif b.Username == \"\" || b.Password == \"\" {\n\t\tlogrus.Warn(\"Metrics are exposed without protection. Make sure you set up protection at proxy level.\")\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t// Serve without authorization if either Username or Password is unset\n\t\tif b.Username == \"\" || b.Password == \"\" {\n\t\t\th.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\t\tuser, pass, ok := r.BasicAuth()\n\n\t\tif !ok || user != b.Username || pass != b.Password {\n\t\t\thttp.Error(w, http.StatusText(http.StatusForbidden), http.StatusForbidden)\n\t\t\treturn\n\t\t}\n\t\th.ServeHTTP(w, r)\n\t}\n}", "func UnaryAuth() grpc.UnaryServerInterceptor {\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\t\t//Get auth data that's passed from client in context.\n\t\tmd, ok := metadata.FromContext(ctx)\n\t\tif !ok {\n\t\t\tfmt.Println(\"Could not get metadata\", err)\n\t\t}\n\n\t\t//Check it to make sure it's good\n\t\tcred := md[authKey]\n\t\tfmt.Println(\"cred\", cred)\n\t\tif cred == nil {\n\t\t\t//Reject call if not\n\t\t\treturn nil, errors.New(\"Not authorized to make this call!\")\n\t\t}\n\n\t\t//Add user data to ctx.\n\t\tctx = context.WithValue(ctx, \"user\", cred)\n\n\t\t//Pass to next handler\n\t\treturn handler(ctx, req)\n\t}\n}", "func New() iris2.HandlerFunc {\n\tl := &authMiddleware{}\n\treturn l.Serve\n}", "func JWTAuthenticationInterceptor(publicKey string, excludingPath []string) grpc.UnaryServerInterceptor {\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\treturn handler(ctx, req)\n\t}\n}", "func getToken(urlStr string, creds []byte)string{\n\n\tvar urlBuffer bytes.Buffer\n\n\tproxyStr := os.Getenv(\"HTTPS_PROXY\")\n\tproxyURL, err := url.Parse(proxyStr)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tfmt.Println(\"Authenticating with CloudBolt API....\")\n\turlBuffer.WriteString(urlStr)\n\tfmt.Println(urlStr)\n\turlBuffer.WriteString(\"/api/v2/api-token-auth/\")\n\treq, err := http.NewRequest(\"POST\", urlBuffer.String(), bytes.NewBuffer(creds))\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t\tProxy: http.ProxyURL(proxyURL),\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer resp.Body.Close()\n\tfmt.Println(resp.StatusCode)\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\ttoken := new(Token)\n\terr = json.Unmarshal(body, &token)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\treturn token.Token\n\n}", "func runAuthProxy(s *tsnet.Server, rt http.RoundTripper, logf logger.Logf) {\n\tln, err := s.Listen(\"tcp\", \":443\")\n\tif err != nil {\n\t\tlog.Fatalf(\"could not listen on :443: %v\", err)\n\t}\n\tu, err := url.Parse(fmt.Sprintf(\"https://%s:%s\", os.Getenv(\"KUBERNETES_SERVICE_HOST\"), os.Getenv(\"KUBERNETES_SERVICE_PORT_HTTPS\")))\n\tif err != nil {\n\t\tlog.Fatalf(\"runAuthProxy: failed to parse URL %v\", err)\n\t}\n\n\tlc, err := s.LocalClient()\n\tif err != nil {\n\t\tlog.Fatalf(\"could not get local client: %v\", err)\n\t}\n\tap := &authProxy{\n\t\tlogf: logf,\n\t\tlc: lc,\n\t\trp: &httputil.ReverseProxy{\n\t\t\tDirector: func(r *http.Request) {\n\t\t\t\t// We want to proxy to the Kubernetes API, but we want to use\n\t\t\t\t// the caller's identity to do so. We do this by impersonating\n\t\t\t\t// the caller using the Kubernetes User Impersonation feature:\n\t\t\t\t// https://kubernetes.io/docs/reference/access-authn-authz/authentication/#user-impersonation\n\n\t\t\t\t// Out of paranoia, remove all authentication headers that might\n\t\t\t\t// have been set by the client.\n\t\t\t\tr.Header.Del(\"Authorization\")\n\t\t\t\tr.Header.Del(\"Impersonate-Group\")\n\t\t\t\tr.Header.Del(\"Impersonate-User\")\n\t\t\t\tr.Header.Del(\"Impersonate-Uid\")\n\t\t\t\tfor k := range r.Header {\n\t\t\t\t\tif strings.HasPrefix(k, \"Impersonate-Extra-\") {\n\t\t\t\t\t\tr.Header.Del(k)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// Now add the impersonation headers that we want.\n\t\t\t\twho := r.Context().Value(whoIsKey{}).(*apitype.WhoIsResponse)\n\t\t\t\tif who.Node.IsTagged() {\n\t\t\t\t\t// Use the nodes FQDN as the username, and the nodes tags as the groups.\n\t\t\t\t\t// \"Impersonate-Group\" requires \"Impersonate-User\" to be set.\n\t\t\t\t\tr.Header.Set(\"Impersonate-User\", strings.TrimSuffix(who.Node.Name, \".\"))\n\t\t\t\t\tfor _, tag := range who.Node.Tags {\n\t\t\t\t\t\tr.Header.Add(\"Impersonate-Group\", tag)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tr.Header.Set(\"Impersonate-User\", who.UserProfile.LoginName)\n\t\t\t\t}\n\n\t\t\t\t// Replace the URL with the Kubernetes APIServer.\n\t\t\t\tr.URL.Scheme = u.Scheme\n\t\t\t\tr.URL.Host = u.Host\n\t\t\t},\n\t\t\tTransport: rt,\n\t\t},\n\t}\n\ths := &http.Server{\n\t\t// Kubernetes uses SPDY for exec and port-forward, however SPDY is\n\t\t// incompatible with HTTP/2; so disable HTTP/2 in the proxy.\n\t\tTLSConfig: &tls.Config{\n\t\t\tGetCertificate: lc.GetCertificate,\n\t\t\tNextProtos: []string{\"http/1.1\"},\n\t\t},\n\t\tTLSNextProto: make(map[string]func(*http.Server, *tls.Conn, http.Handler)),\n\t\tHandler: ap,\n\t}\n\tif err := hs.ServeTLS(ln, \"\", \"\"); err != nil {\n\t\tlog.Fatalf(\"runAuthProxy: failed to serve %v\", err)\n\t}\n}", "func basicAuth(password string) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tw.Header().Set(\"WWW-Authenticate\", `Basic realm=\"Restricted\"`)\n\n\t\t\ts := strings.SplitN(r.Header.Get(\"Authorization\"), \" \", 2)\n\t\t\tif len(s) != 2 {\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tb, err := base64.StdEncoding.DecodeString(s[1])\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"base64.StdEncoding.DecodeString() Error: %s\\n\", err)\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tpair := strings.SplitN(string(b), \":\", 2)\n\t\t\tif len(pair) != 2 {\n\t\t\t\tlog.Printf(\"strings.SplitN() Error: %s\\n\", err)\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif pair[0] != \"admin\" || pair[1] != password {\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func Authenticator(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tctx := r.Context()\n\n\t\tjwtToken, err := TokenFromContext(ctx)\n\t\tif err != nil {\n\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\treturn\n\t\t}\n\n\t\t_ = jwtToken\n\n\t\t// Token is authenticated, pass it through\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func RequestCredentials(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\tidentity AuthIdentity\n\t\tcredentials AuthCredentials\n\t)\n\n\teaaCtx := r.Context().Value(contextKey(\"appliance-ctx\")).(*Context)\n\n\tconst fName = \"/Auth RequestCredentials \"\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\n\terr := json.NewDecoder(r.Body).Decode(&identity)\n\tif err != nil {\n\t\tlog.Errf(fName+\"decode failed: %v\", err.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\thost, port, err := net.SplitHostPort(r.RemoteAddr)\n\tif err != nil {\n\t\tlog.Errf(fName+\"Cannot retrieve IP from RemoteAddr: %v [%v:%v] %v\",\n\t\t\tr.RemoteAddr, host, port, err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tcert, err := SignCSR(identity.Csr, eaaCtx)\n\tif err != nil {\n\t\tlog.Errf(fName+\"failed: %v\", err.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tsignedCertBlock := pem.EncodeToMemory(\n\t\t&pem.Block{Type: \"CERTIFICATE\", Bytes: cert.Raw})\n\tif signedCertBlock == nil {\n\t\tlog.Err(fName + \"/failed to enode signed cert\")\n\t\treturn\n\t}\n\trcaBlock := pem.EncodeToMemory(\n\t\t&pem.Block{Type: \"CERTIFICATE\",\n\t\t\tBytes: eaaCtx.certsEaaCa.rca.x509Cert.Raw})\n\tif rcaBlock == nil {\n\t\tlog.Err(fName + \"failed to enode rca cert\")\n\t\treturn\n\t}\n\n\tcredentials.ID = cert.Subject.CommonName\n\tcredentials.Certificate = string(signedCertBlock)\n\tcredentials.CaChain = []string{string(rcaBlock)}\n\tcredentials.CaPool = []string{string(rcaBlock)}\n\n\tencoder := json.NewEncoder(w)\n\terr = encoder.Encode(credentials)\n\tif err != nil {\n\t\tlog.Errf(fName+\"encoding output to JSON failed: %s\",\n\t\t\terr.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlog.Info(fName + \" request from CN: \" + credentials.ID + \", from IP: \" +\n\t\thost + \" properly handled\")\n}", "func redirectPolicyFunc(req *http.Request, via []*http.Request) error {\n\treq.Header.Add(\"Authorization\", \"Basic \"+globalBasicAuthString)\n\treturn nil\n}", "func WrapMockAuthConfig(hfn http.HandlerFunc, cfg *config.APICfg, brk brokers.Broker, str stores.Store, mgr *oldPush.Manager, c push.Client, roles ...string) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlVars := mux.Vars(r)\n\n\t\tuserRoles := []string{\"publisher\", \"consumer\"}\n\t\tif len(roles) > 0 {\n\t\t\tuserRoles = roles\n\t\t}\n\n\t\tnStr := str.Clone()\n\t\tdefer nStr.Close()\n\n\t\tprojectUUID := projects.GetUUIDByName(urlVars[\"project\"], nStr)\n\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\tgorillaContext.Set(r, \"brk\", brk)\n\t\tgorillaContext.Set(r, \"str\", nStr)\n\t\tgorillaContext.Set(r, \"mgr\", mgr)\n\t\tgorillaContext.Set(r, \"apsc\", c)\n\t\tgorillaContext.Set(r, \"auth_resource\", cfg.ResAuth)\n\t\tgorillaContext.Set(r, \"auth_user\", \"UserA\")\n\t\tgorillaContext.Set(r, \"auth_user_uuid\", \"uuid1\")\n\t\tgorillaContext.Set(r, \"auth_roles\", userRoles)\n\t\tgorillaContext.Set(r, \"push_worker_token\", cfg.PushWorkerToken)\n\t\tgorillaContext.Set(r, \"push_enabled\", cfg.PushEnabled)\n\t\thfn.ServeHTTP(w, r)\n\n\t})\n}", "func Auth() func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tgivenAPIKey := r.Header.Get(\"API-KEY\")\n\n\t\t\t// Pass through unauthenticated requests.\n\t\t\tif givenAPIKey == \"\" {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Inject admin user.\n\t\t\tif givenAPIKey == adminAPIKey {\n\t\t\t\tuser := &User{Name: \"Eric\", IsAdmin: true}\n\t\t\t\tctx := context.WithValue(r.Context(), userKey, user)\n\t\t\t\tr = r.WithContext(ctx)\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Inject non admin user.\n\t\t\tuser := &User{Name: \"Greg\", IsAdmin: false}\n\t\t\tctx := context.WithValue(r.Context(), userKey, user)\n\t\t\tr = r.WithContext(ctx)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func (c *Client) Auth() (string, error) {\n\t// First do an empty get to get the auth challenge\n\treq, err := http.NewRequest(http.MethodGet, c.BaseURL+\"/v2/\", nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\trsp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed sending auth request: %w\", err)\n\t}\n\tdefer rsp.Body.Close()\n\tio.Copy(io.Discard, rsp.Body)\n\n\tif rsp.StatusCode == http.StatusOK {\n\t\t// no auth needed\n\t\treturn \"\", nil\n\t}\n\n\tif rsp.StatusCode != http.StatusUnauthorized {\n\t\treturn \"\", fmt.Errorf(\"unexpected status %s\", rsp.Status)\n\t}\n\n\t// The Www-Authenticate header tells us where to go to get a token\n\tvals, err := parseWWWAuthenticate(rsp.Header.Get(\"Www-Authenticate\"))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tu, err := url.Parse(vals[\"realm\"])\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"could not parse authentication realm: %w\", err)\n\t}\n\tq := u.Query()\n\tq.Set(\"service\", vals[\"service\"])\n\tq.Set(\"scope\", \"repository:\"+c.Name+\":pull,push\")\n\tu.RawQuery = q.Encode()\n\n\tfmt.Printf(\"get %s\\n\", u)\n\n\treq, err = http.NewRequest(http.MethodGet, u.String(), nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treq.SetBasicAuth(c.User, c.Password)\n\n\trsp, err = http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed sending auth request: %w\", err)\n\t}\n\tdefer rsp.Body.Close()\n\tif rsp.StatusCode != http.StatusOK {\n\t\treturn \"\", fmt.Errorf(\"unexpected status %s\", rsp.Status)\n\t}\n\tbody, err := io.ReadAll(rsp.Body)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"could not read auth response body: %w\", err)\n\t}\n\n\ttype token struct {\n\t\tToken string `json:\"token\"`\n\t}\n\tvar tok token\n\tif err := json.Unmarshal(body, &tok); err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to unmarshal token: %w\", err)\n\t}\n\n\treturn tok.Token, nil\n}", "func Auth() gin.HandlerFunc {\r\n\tif gin.Mode() == \"debug\" {\r\n\t\treturn func(c *gin.Context) { c.Next() }\r\n\t}\r\n\treturn func(c *gin.Context) {\r\n\t\tAccessKey := c.GetHeader(\"AccessKey\")\r\n\t\tif c.GetHeader(\"AccessKey\") == \"\" {\r\n\t\t\tAccessKey = c.GetHeader(\"Token\")\r\n\t\t}\r\n\r\n\t\tsession := sessions.Default(c)\r\n\t\tLoginUserID := session.Get(\"UserID\")\r\n\t\tIsLeader := session.Get(\"IsLeader\")\r\n\r\n\t\tfmt.Println(\"AccessKey: \", AccessKey)\r\n\t\tswitch AccessKey {\r\n\t\tcase \"\":\r\n\t\t\tif LoginUserID != nil {\r\n\t\t\t\tc.Set(\"UserID\", LoginUserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\t\t\t} else {\r\n\t\t\t\tsession := sessions.Default(c)\r\n\t\t\t\tUserID := session.Get(\"UserID\")\r\n\t\t\t\tIsLeader := session.Get(\"IsLeader\")\r\n\r\n\t\t\t\tfmt.Println(\"UserID, IsLeader\", UserID, IsLeader)\r\n\t\t\t\tif UserID == nil {\r\n\t\t\t\t\tc.JSON(http.StatusUnauthorized, gin.H{\"message\": \"Empty AccessKey Please authorize before requesting\"})\r\n\t\t\t\t\tc.Abort()\r\n\t\t\t\t}\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\r\n\t\t\t}\r\n\r\n\t\tdefault:\r\n\t\t\tUserID, IsLeader, err := utils.LoadAccessKey(AccessKey)\r\n\r\n\t\t\tif LoginUserID != nil {\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\t\t\t} else {\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\tc.JSON(http.StatusUnauthorized, gin.H{\"message\": \"Please authorize before requesting\"})\r\n\t\t\t\t\tc.Abort()\r\n\t\t\t\t}\r\n\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\r\n\t\t\t\tc.Next()\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n}", "func NewServerInterceptor(authenticator Authenticator) rpc.ServerInterceptor {\n\treturn rpc.ServerInterceptorFunc(\n\t\tfunc(\n\t\t\tctx context.Context,\n\t\t\trequest interface{},\n\t\t\tserverInfo *rpc.ServerInfo,\n\t\t\tserverHandler rpc.ServerHandler,\n\t\t) (interface{}, error) {\n\t\t\tuser, ok := authenticate(ctx, authenticator)\n\t\t\tif ok {\n\t\t\t\tctx = context.WithValue(ctx, authContextKey{}, user)\n\t\t\t}\n\t\t\treturn serverHandler.Handle(ctx, request)\n\t\t},\n\t)\n}", "func clientInterceptor(ctx context.Context, method string, req, reply interface{}, cc *grpc.ClientConn, invoker grpc.UnaryInvoker, opts ...grpc.CallOption) error {\n log.Printf(\"Method Called: %s\", method)\n return invoker(ctx, method, req, reply, cc, opts...)\n}", "func (b *Basic) InjectAuth(req *http.Request) error {\n\tif _, _, ok := req.BasicAuth(); ok {\n\t\treturn errors.New(\"Basic Auth header already exists\")\n\t}\n\n\treq.SetBasicAuth(b.Username, b.Password)\n\n\treturn nil\n}", "func Auth(route string) gin.HandlerFunc {\n\n\treturn func(c *gin.Context) {\n\n\t\tvar auth JwtPayload\n\n\t\t//authenticate jwt\n\t\tsecret := config.Config().GetString(\"hmac_key\")\n\n\t\ttoken := strings.Split(c.GetHeader(\"Authorization\"), \" \")\n\t\tif (len(token) != 2) || (token[0] != \"JWT\") {\n\t\t\tabort(c, 401, msg.Get(\"7\")) //JWT authorization header malformed\n\t\t\treturn\n\t\t}\n\n\t\tvar e *msg.Message\n\t\tauth, e = JwtAuth(token[1], secret)\n\t\tif e != nil {\n\t\t\tabort(c, 401, *e)\n\t\t\treturn\n\t\t}\n\n\t\ta := ACL()\n\t\tg := Grant{RoleID: auth.RoleID, Route: route, Method: c.Request.Method}\n\t\tr, ok := a[g]\n\t\tif !ok {\n\t\t\tabort(c, 401, msg.Get(\"8\")) //Not enough permissions\n\t\t\treturn\n\t\t}\n\n\t\tnow := time.Now()\n\t\tif now.Before(r.From) || now.After(r.To) {\n\t\t\tabort(c, 401, msg.Get(\"9\")) //Role access expired or not yet valid\n\t\t\treturn\n\t\t}\n\n\t\tif tps.IsEnabled() && tps.Transaction(auth.UserID, auth.TPS) != nil {\n\t\t\tabort(c, 401, msg.Get(\"10\")) //TPS limit exceeded\n\t\t\treturn\n\t\t}\n\n\t\tif _, ok := DeletedUsersMap[auth.UserID]; ok {\n\t\t\tabort(c, 401, msg.Get(\"11\")) //Unauthorized\n\t\t\treturn\n\t\t}\n\n\t\tc.Set(\"Auth\", auth)\n\t\tc.Next()\n\t}\n}", "func New(config *Config, next http.RoundTripper) http.RoundTripper {\n\t// Need to delay fetching auth settings until env vars have had a chance to propagate\n\tif authSettings == nil {\n\t\tauthSettings = awsds.ReadAuthSettingsFromEnvironmentVariables()\n\t}\n\n\treturn RoundTripperFunc(func(r *http.Request) (*http.Response, error) {\n\t\tif next == nil {\n\t\t\tnext = http.DefaultTransport\n\t\t}\n\t\treturn (&middleware{\n\t\t\tconfig: config,\n\t\t\tnext: next,\n\t\t}).exec(r)\n\t})\n}", "func AuthorizedAPIClientFor(username string, password string) *baloo.Client {\n\tjwt, err := Authenticate(username, password)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\treturn baloo.New(constants.API_URL).\n\t\tUseRequest(AuthMiddleware(jwt))\n}", "func GetLoginFunc(db *sqlx.DB) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusername := \"\"\n\t\tpassword := \"\"\n\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error reading body: \", err.Error())\n\t\t\thttp.Error(w, \"Error reading body: \"+err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tvar lj loginJson\n\t\tlog.Println(body)\n\t\terr = json.Unmarshal(body, &lj)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error unmarshalling JSON: \", err.Error())\n\t\t\thttp.Error(w, \"Invalid JSON: \"+err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tusername = lj.U\n\t\tpassword = lj.P\n\t\tuserInterface, err := api.GetUser(username, db)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Invalid user: \"+err.Error(), http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\t\tu, ok := userInterface.(api.Users)\n\t\tif !ok {\n\t\t\thttp.Error(w, \"Error GetUser returned a non-user.\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tencBytes := sha1.Sum([]byte(password))\n\t\tencString := hex.EncodeToString(encBytes[:])\n\t\tif err != nil {\n\t\t\tctx.Set(r, \"user\", nil)\n\t\t\tlog.Println(\"Invalid password\")\n\t\t\thttp.Error(w, \"Invalid password: \"+err.Error(), http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\t\tif u.LocalPassword.String != encString {\n\t\t\tctx.Set(r, \"user\", nil)\n\t\t\tlog.Println(\"Invalid password\")\n\t\t\thttp.Error(w, \"Invalid password\", http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\t// Create the token\n\t\ttoken := jwt.New(jwt.SigningMethodHS256)\n\t\t// Set some claims\n\t\ttoken.Claims[\"userid\"] = u.Username\n\t\ttoken.Claims[\"role\"] = u.Links.RolesLink.ID\n\t\ttoken.Claims[\"exp\"] = time.Now().Add(time.Hour * 72).Unix()\n\t\t// Sign and get the complete encoded token as a string\n\t\ttokenString, err := token.SignedString([]byte(\"mySigningKey\")) // TODO JvD\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tjs, err := json.Marshal(TokenResponse{Token: tokenString})\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(js)\n\t}\n}", "func AuthFunc() types.AuthFunc {\n\treturn globalBusiness.AuthFunc\n}", "func Handler(handler http.HandlerFunc, security *Config) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusernameEntered, passwordEntered, ok := r.BasicAuth()\n\t\tif !ok || usernameEntered != security.Basic.Username || Sha512(passwordEntered) != strings.ToLower(security.Basic.PasswordSha512Hash) {\n\t\t\tw.Header().Set(\"WWW-Authenticate\", \"Basic\")\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\t_, _ = w.Write([]byte(\"Unauthorized\"))\n\t\t\treturn\n\t\t}\n\t\thandler(w, r)\n\t}\n}", "func makeAuthInfoWriterCustom(cmd *cobra.Command) (runtime.ClientAuthInfoWriter, error) {\n\tauths := []runtime.ClientAuthInfoWriter{}\n\t/*x-auth-token User authorization token*/\n\tif viper.IsSet(\"uisp.x-auth-token\") {\n\t\tXAuthTokenKey := viper.GetString(\"uisp.x-auth-token\")\n\t\tauths = append(auths, httptransport.APIKeyAuth(\"x-auth-token\", \"header\", XAuthTokenKey))\n\t}\n\tif len(auths) == 0 {\n\t\tlogDebugf(\"Warning: No auth params detected.\")\n\t\treturn nil, nil\n\t}\n\t// compose all auths together\n\treturn httptransport.Compose(auths...), nil\n}", "func basicAuthHandler(handler func(string, string) http.HandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusername, password, ok := r.BasicAuth()\n\n\t\tif !ok {\n\t\t\tw.Header().Add(\"WWW-Authenticate\", `Basic realm=\"cwmp-proxy\"`)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\thandler(username, password).ServeHTTP(w, r)\n\t}\n}", "func (proto *Protocol) AUTHENTICATE(command *Command) {\n\targs := strings.Split(command.args, \" \")\n\n\tif len(args) < 1 {\n\t\t// FIXME what error?\n\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\treturn\n\t}\n\n\tswitch strings.ToUpper(args[0]) {\n\tcase \"PLAIN\":\n\t\tif len(args) > 1 {\n\t\t\t// Do auth now\n\t\t\treturn\n\t\t}\n\t\tproto.readIntercept = func(l string) {\n\t\t\tproto.readIntercept = nil\n\t\t\tuser, pass, err := auth.DecodePLAIN(l)\n\t\t\tif err != nil {\n\t\t\t\t// FIXME what error?\n\t\t\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif ok := proto.ValidateAuthenticationHandler(\"PLAIN\", user, pass); !ok {\n\t\t\t\t// FIXME what error?\n\t\t\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// OK\n\t\t\tproto.State = AUTH\n\t\t\tproto.Responses <- &Response{command.tag, Status(ResponseOK), nil, \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tproto.Responses <- &Response{\"+\", nil, nil, \"\", nil}\n\t\treturn\n\t}\n\n\t// FIXME what error?\n\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n}", "func NewProvider(username string, password string) *Provider {\n transport := httptransport.New(\"api.autodns.com\", \"/v1\", []string{\"https\"})\n transport.DefaultAuthentication = httptransport.BasicAuth(username, password)\n\n formats := strfmt.Default\n\n return &Provider{\n username: username,\n password: password,\n client: zone_tasks.New(transport, formats),\n }\n}", "func WrapAuthorize(hfn http.Handler, routeName string) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlValues := r.URL.Query()\n\n\t\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\t\trefRoles := gorillaContext.Get(r, \"auth_roles\").([]string)\n\t\tserviceToken := gorillaContext.Get(r, \"auth_service_token\").(string)\n\n\t\t// Check first if service token is used\n\t\tif serviceToken != \"\" && serviceToken == urlValues.Get(\"key\") {\n\t\t\thfn.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\n\t\tif auth.Authorize(routeName, refRoles, refStr) {\n\t\t\thfn.ServeHTTP(w, r)\n\t\t} else {\n\t\t\terr := APIErrorForbidden()\n\t\t\trespondErr(w, err)\n\t\t}\n\t})\n}", "func StreamAuth() grpc.StreamServerInterceptor {\n\n\treturn func(srv interface{}, ss grpc.ServerStream, info *grpc.StreamServerInfo, handler grpc.StreamHandler) error {\n\t\tcred := ss.Context().Value(authKey)\n\t\tif cred == nil {\n\t\t\treturn errors.New(\"Not authorized to make this call!\")\n\t\t}\n\n\t\treturn handler(srv, ss)\n\t}\n}", "func (a *App) Auth() negroni.HandlerFunc {\n\treturn negroni.HandlerFunc(func(w http.ResponseWriter, req *http.Request, next http.HandlerFunc) {\n\t\tdb, ok := context.Get(req, \"db\").(*mgo.Database)\n\t\tif !ok {\n\t\t\ta.R.JSON(w, http.StatusInternalServerError, &Response{Status: \"Error\", Message: \"Internal server error\"})\n\t\t\treturn\n\t\t}\n\t\tauthHeader := req.Header.Get(\"Authorization\")\n\t\tif authHeader == \"\" {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tdata, err := base64.StdEncoding.DecodeString(strings.Replace(authHeader, \"Basic \", \"\", 1))\n\t\tif err != nil {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tuser := &User{}\n\t\tparts := strings.Split(string(data), \":\")\n\t\tif len(parts) < 2 {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tif parts[0] == parts[1] {\n\t\t\tshaHash := sha256.New()\n\t\t\tif _, err := shaHash.Write([]byte(parts[0])); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t\ttoken := base64.StdEncoding.EncodeToString(shaHash.Sum(nil))\n\t\t\tif err := db.C(\"users\").Find(bson.M{\n\t\t\t\t\"services.resume.loginTokens\": bson.M{\"$elemMatch\": bson.M{\"hashedToken\": token}},\n\t\t\t}).One(&user); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t} else {\n\t\t\tin := []bson.M{bson.M{\"address\": parts[0], \"verified\": false}}\n\t\t\tif err := db.C(\"users\").Find(bson.M{\"emails\": bson.M{\"$in\": in}}).One(&user); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t\tshaHash := sha256.New()\n\t\t\tif _, err := shaHash.Write([]byte(parts[1])); err != nil {\n\t\t\t\thttp.Error(w, \"Not Authorized\", http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\th := hex.EncodeToString(shaHash.Sum(nil))\n\t\t\tif err := bcrypt.CompareHashAndPassword([]byte(user.Services.Password.Bcrypt), []byte(h)); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tcontext.Set(req, \"user\", user)\n\t\tnext(w, req)\n\t})\n}", "func (srv *targetServiceHandler) auth(h http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tctx := httpbakery.ContextWithRequest(context.TODO(), req)\n\t\tops, err := opsForRequest(req)\n\t\tif err != nil {\n\t\t\tfail(w, http.StatusInternalServerError, \"%v\", err)\n\t\t\treturn\n\t\t}\n\t\tauthChecker := srv.checker.Auth(httpbakery.RequestMacaroons(req)...)\n\t\tif _, err = authChecker.Allow(ctx, ops...); err != nil {\n\t\t\thttpbakery.WriteError(ctx, w, srv.oven.Error(ctx, req, err))\n\t\t\treturn\n\t\t}\n\t\th.ServeHTTP(w, req)\n\t})\n}", "func auth(conn net.Conn, username string, password string) error {\n\tbufConn := bufio.NewReader(conn)\n\n\t// Read the version byte\n\tversion := []byte{0}\n\tif _, err := bufConn.Read(version); err != nil {\n\t\terr := fmt.Errorf(\"[AUTH] socks: Failed to get version byte: %v\", err)\n\t\treturn err\n\t}\n\n\t// Ensure we are compatible\n\tif version[0] != socks5Version {\n\t\terr := fmt.Errorf(\"[AUTH] Unsupported SOCKS version: %v\", version)\n\t\treturn err\n\t}\n\n\t// Get the methods\n\tmethods, err := readMethods(bufConn)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"[AUTH] Failed to get auth methods: %v\", err)\n\t}\n\n\t// Select a usable method (only auth for us here)\n\tfor _, method := range methods {\n\t\tif method == UserPassAuth {\n\t\t\t// Tell the client to use user/pass auth\n\t\t\tif _, err := conn.Write([]byte{socks5Version, UserPassAuth}); err != nil {\n\t\t\t\treturn fmt.Errorf(\"[AUTH] Can't write method reply: %v\", err)\n\t\t\t}\n\n\t\t\t//read username/password\n\t\t\tu, p, err := getauthdata(bufConn, conn)\n\t\t\tif (err != nil) || (u != username || p != password) {\n\t\t\t\tconn.Write([]byte{userAuthVersion, authFailure})\n\t\t\t\treturn fmt.Errorf(\"[AUTH] Username/password auth failed: %v\", err)\n\t\t\t}\n\n\t\t\t//correct auth\n\t\t\tconn.Write([]byte{userAuthVersion, authSuccess})\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// No usable method found\n\tconn.Write([]byte{socks5Version, noAcceptable})\n\treturn fmt.Errorf(\"[AUTH] No acceptable auth method\")\n}", "func (a *AuthCommand) Initialize(app *kingpin.Application, config *service.Config) {\n\ta.config = config\n\n\t// operations with authorities\n\tauth := app.Command(\"auth\", \"Operations with user and host certificate authorities (CAs)\").Hidden()\n\ta.authExport = auth.Command(\"export\", \"Export public cluster (CA) keys to stdout\")\n\ta.authExport.Flag(\"keys\", \"if set, will print private keys\").BoolVar(&a.exportPrivateKeys)\n\ta.authExport.Flag(\"fingerprint\", \"filter authority by fingerprint\").StringVar(&a.exportAuthorityFingerprint)\n\ta.authExport.Flag(\"compat\", \"export cerfiticates compatible with specific version of Teleport\").StringVar(&a.compatVersion)\n\ta.authExport.Flag(\"type\", \"certificate type: 'user', 'host' or 'tls'\").StringVar(&a.authType)\n\n\ta.authGenerate = auth.Command(\"gen\", \"Generate a new SSH keypair\").Hidden()\n\ta.authGenerate.Flag(\"pub-key\", \"path to the public key\").Required().StringVar(&a.genPubPath)\n\ta.authGenerate.Flag(\"priv-key\", \"path to the private key\").Required().StringVar(&a.genPrivPath)\n\n\ta.authSign = auth.Command(\"sign\", \"Create an identity file(s) for a given user\")\n\ta.authSign.Flag(\"user\", \"Teleport user name\").StringVar(&a.genUser)\n\ta.authSign.Flag(\"host\", \"Teleport host name\").StringVar(&a.genHost)\n\ta.authSign.Flag(\"out\", \"identity output\").Short('o').StringVar(&a.output)\n\ta.authSign.Flag(\"format\", \"identity format: 'file' (default) or 'dir'\").Default(string(client.DefaultIdentityFormat)).StringVar((*string)(&a.outputFormat))\n\ta.authSign.Flag(\"ttl\", \"TTL (time to live) for the generated certificate\").Default(fmt.Sprintf(\"%v\", defaults.CertDuration)).DurationVar(&a.genTTL)\n\ta.authSign.Flag(\"compat\", \"OpenSSH compatibility flag\").StringVar(&a.compatibility)\n}", "func AuthLoginWrapper(ctx *fasthttp.RequestCtx, mgoClient *mgo.Session, redisClient *redis.Client, cfg datastructures.Configuration) {\n\tlog.Info(\"AuthLoginWrapper | Starting authentication | Parsing authentication credentials\")\n\tctx.Response.Header.SetContentType(\"application/json; charset=utf-8\")\n\tusername, password := ParseAuthenticationCoreHTTP(ctx) // Retrieve the username and password encoded in the request from BasicAuth headers, GET & POST\n\tif authutils.ValidateCredentials(username, password) { // Verify if the input parameter respect the rules ...\n\t\tlog.Debug(\"AuthLoginWrapper | Input validated | User: \", username, \" | Pass: \", password, \" | Calling core functionalities ...\")\n\t\tcheck := authutils.LoginUserHTTPCore(username, password, mgoClient, cfg.Mongo.Users.DB, cfg.Mongo.Users.Collection) // Login phase\n\t\tif strings.Compare(check, \"OK\") == 0 { // Login Succeed\n\t\t\tlog.Debug(\"AuthLoginWrapper | Login succesfully! Generating token!\")\n\t\t\ttoken := basiccrypt.GenerateToken(username, password) // Generate a simple md5 hashed token\n\t\t\tlog.Info(\"AuthLoginWrapper | Inserting token into Redis \", token)\n\t\t\tbasicredis.InsertIntoClient(redisClient, username, token, cfg.Redis.Token.Expire) // insert the token into the DB\n\t\t\tlog.Info(\"AuthLoginWrapper | Token inserted! All operation finished correctly! | Setting token into response\")\n\t\t\tauthcookie := authutils.CreateCookie(\"GoLog-Token\", token, cfg.Redis.Token.Expire)\n\t\t\tctx.Response.Header.SetCookie(authcookie) // Set the token into the cookie headers\n\t\t\tctx.Response.Header.Set(\"GoLog-Token\", token) // Set the token into a custom headers for future security improvments\n\t\t\tlog.Warn(\"AuthLoginWrapper | Client logged in succesfully!! | \", username, \":\", password, \" | Token: \", token)\n\t\t\terr := json.NewEncoder(ctx).Encode(datastructures.Response{Status: true, Description: \"User logged in!\", ErrorCode: username + \":\" + password, Data: token})\n\t\t\tcommonutils.Check(err, \"AuthLoginWrapper\")\n\t\t} else {\n\t\t\tcommonutils.AuthLoginWrapperErrorHelper(ctx, check, username, password)\n\t\t}\n\t} else { // error parsing credential\n\t\tlog.Info(\"AuthLoginWrapper | Error parsing credential!! |\", username+\":\"+password)\n\t\tctx.Response.Header.DelCookie(\"GoLog-Token\")\n\t\tctx.Error(fasthttp.StatusMessage(fasthttp.StatusUnauthorized), fasthttp.StatusUnauthorized)\n\t\tctx.Response.Header.Set(\"WWW-Authenticate\", \"Basic realm=Restricted\")\n\t\t//err := json.NewEncoder(ctx).Encode(datastructures.Response{Status: false, Description: \"Error parsing credential\", ErrorCode: \"Missing or manipulated input\", Data: nil})\n\t\t//commonutils.Check(err, \"AuthLoginWrapper\")\n\t}\n}", "func Auth() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\trawCallerID := mlauth.GetCaller(c.Request)\n\t\tisAdmin := mlauth.IsCallerAdmin(c.Request)\n\n\t\tcallerID, err := strconv.ParseUint(rawCallerID, 10, 64)\n\n\t\t// If request is not from an admin, and we failed parsing caller ID, fail\n\t\tif !isAdmin && err != nil {\n\t\t\terrors.ReturnError(c, &errors.Error{\n\t\t\t\tCode: errors.BadRequestApiError,\n\t\t\t\tCause: \"parsing header value\",\n\t\t\t\tMessage: \"invalid caller.id\",\n\t\t\t\tValues: map[string]string{\n\t\t\t\t\t\"caller.id\": rawCallerID,\n\t\t\t\t},\n\t\t\t})\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tc.Set(\"callerID\", callerID)\n\t\tc.Set(\"isAdmin\", isAdmin)\n\t\tc.Next()\n\t}\n}", "func (a *AbstractNetworkConnectionHandler) OnAuthPassword(_ string, _ []byte) (response AuthResponse, reason error) {\n\treturn AuthResponseUnavailable, nil\n}", "func (s *HTTPServer) wrap(handler endpoint, methods []string) http.HandlerFunc {\n\treturn func(resp http.ResponseWriter, req *http.Request) {\n\t\tsetHeaders(resp, s.agent.config.HTTPResponseHeaders)\n\t\tsetTranslateAddr(resp, s.agent.config.TranslateWANAddrs)\n\n\t\t// Obfuscate any tokens from appearing in the logs\n\t\tformVals, err := url.ParseQuery(req.URL.RawQuery)\n\t\tif err != nil {\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Failed to decode query: %s from=%s\", err, req.RemoteAddr)\n\t\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tlogURL := req.URL.String()\n\t\tif tokens, ok := formVals[\"token\"]; ok {\n\t\t\tfor _, token := range tokens {\n\t\t\t\tif token == \"\" {\n\t\t\t\t\tlogURL += \"<hidden>\"\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tlogURL = strings.Replace(logURL, token, \"<hidden>\", -1)\n\t\t\t}\n\t\t}\n\t\tlogURL = aclEndpointRE.ReplaceAllString(logURL, \"$1<hidden>$4\")\n\n\t\tif s.blacklist.Block(req.URL.Path) {\n\t\t\terrMsg := \"Endpoint is blocked by agent configuration\"\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Request %s %v, error: %v from=%s\", req.Method, logURL, err, req.RemoteAddr)\n\t\t\tresp.WriteHeader(http.StatusForbidden)\n\t\t\tfmt.Fprint(resp, errMsg)\n\t\t\treturn\n\t\t}\n\n\t\tisForbidden := func(err error) bool {\n\t\t\tif acl.IsErrPermissionDenied(err) || acl.IsErrNotFound(err) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\t_, ok := err.(ForbiddenError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisMethodNotAllowed := func(err error) bool {\n\t\t\t_, ok := err.(MethodNotAllowedError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisBadRequest := func(err error) bool {\n\t\t\t_, ok := err.(BadRequestError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisTooManyRequests := func(err error) bool {\n\t\t\t// Sadness net/rpc can't do nice typed errors so this is all we got\n\t\t\treturn err.Error() == consul.ErrRateLimited.Error()\n\t\t}\n\n\t\taddAllowHeader := func(methods []string) {\n\t\t\tresp.Header().Add(\"Allow\", strings.Join(methods, \",\"))\n\t\t}\n\n\t\thandleErr := func(err error) {\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Request %s %v, error: %v from=%s\", req.Method, logURL, err, req.RemoteAddr)\n\t\t\tswitch {\n\t\t\tcase isForbidden(err):\n\t\t\t\tresp.WriteHeader(http.StatusForbidden)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase structs.IsErrRPCRateExceeded(err):\n\t\t\t\tresp.WriteHeader(http.StatusTooManyRequests)\n\t\t\tcase isMethodNotAllowed(err):\n\t\t\t\t// RFC2616 states that for 405 Method Not Allowed the response\n\t\t\t\t// MUST include an Allow header containing the list of valid\n\t\t\t\t// methods for the requested resource.\n\t\t\t\t// https://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html\n\t\t\t\taddAllowHeader(err.(MethodNotAllowedError).Allow)\n\t\t\t\tresp.WriteHeader(http.StatusMethodNotAllowed) // 405\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase isBadRequest(err):\n\t\t\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase isTooManyRequests(err):\n\t\t\t\tresp.WriteHeader(http.StatusTooManyRequests)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tdefault:\n\t\t\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\t}\n\t\t}\n\n\t\tstart := time.Now()\n\t\tdefer func() {\n\t\t\ts.agent.logger.Printf(\"[DEBUG] http: Request %s %v (%v) from=%s\", req.Method, logURL, time.Since(start), req.RemoteAddr)\n\t\t}()\n\n\t\tvar obj interface{}\n\n\t\t// if this endpoint has declared methods, respond appropriately to OPTIONS requests. Otherwise let the endpoint handle that.\n\t\tif req.Method == \"OPTIONS\" && len(methods) > 0 {\n\t\t\taddAllowHeader(append([]string{\"OPTIONS\"}, methods...))\n\t\t\treturn\n\t\t}\n\n\t\t// if this endpoint has declared methods, check the request method. Otherwise let the endpoint handle that.\n\t\tmethodFound := len(methods) == 0\n\t\tfor _, method := range methods {\n\t\t\tif method == req.Method {\n\t\t\t\tmethodFound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif !methodFound {\n\t\t\terr = MethodNotAllowedError{req.Method, append([]string{\"OPTIONS\"}, methods...)}\n\t\t} else {\n\t\t\terr = s.checkWriteAccess(req)\n\n\t\t\tif err == nil {\n\t\t\t\t// Invoke the handler\n\t\t\t\tobj, err = handler(resp, req)\n\t\t\t}\n\t\t}\n\t\tcontentType := \"application/json\"\n\t\thttpCode := http.StatusOK\n\t\tif err != nil {\n\t\t\tif errPayload, ok := err.(CodeWithPayloadError); ok {\n\t\t\t\thttpCode = errPayload.StatusCode\n\t\t\t\tif errPayload.ContentType != \"\" {\n\t\t\t\t\tcontentType = errPayload.ContentType\n\t\t\t\t}\n\t\t\t\tif errPayload.Reason != \"\" {\n\t\t\t\t\tresp.Header().Add(\"X-Consul-Reason\", errPayload.Reason)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\thandleErr(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif obj == nil {\n\t\t\treturn\n\t\t}\n\t\tvar buf []byte\n\t\tif contentType == \"application/json\" {\n\t\t\tbuf, err = s.marshalJSON(req, obj)\n\t\t\tif err != nil {\n\t\t\t\thandleErr(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t} else {\n\t\t\tif strings.HasPrefix(contentType, \"text/\") {\n\t\t\t\tif val, ok := obj.(string); ok {\n\t\t\t\t\tbuf = []byte(val)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tresp.Header().Set(\"Content-Type\", contentType)\n\t\tresp.WriteHeader(httpCode)\n\t\tresp.Write(buf)\n\t}\n}", "func BasicAuth(username, password string) Credentials {\n\treturn &basicAuth{username, password}\n}", "func AuthMiddlewareHandlerFunc(authenticator *Authenticator, handleFunc http.HandlerFunc) func(w http.ResponseWriter, r *http.Request) {\n\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\tuser := authenticator.AuthenticateHttpRequest(r)\n\t\tif user != nil {\n\t\t\thandleFunc(w, r.WithContext(context.WithValue(r.Context(), \"User\", user)))\n\t\t} else {\n\t\t\tw.Header().Add(\"WWW-Authenticate\", \"Basic realm=restricted\")\n\t\t\thttp.Error(w, http.StatusText(http.StatusUnauthorized),\n\t\t\t\thttp.StatusUnauthorized)\n\t\t}\n\n\t}\n\n}", "func (t *OpetCode) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n fmt.Printf(\"Invoke function\\n\")\n // Retrieve the requested Smart Contract function and arguments\n function, args := APIstub.GetFunctionAndParameters()\n // Route to the appropriate handler function to interact with the ledger appropriately\n if function == \"initLedger\" {\n return t.initLedger(APIstub)\n } else if function == \"createUser\" {\n return t.createUser(APIstub, args)\n } else if function == \"retrieveUser\" {\n return t.retrieveUser(APIstub, args)\n } else if function == \"createDocument\" {\n return t.createDocument(APIstub, args)\n } else if function == \"retrieveDocument\" {\n return t.retrieveDocument(APIstub, args)\n }\n return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func Auth() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tc.Writer.Write([]byte(\"auth required for endpoint\"))\n\t}\n}", "func NewGetAuth() context.Handler {\n\treturn func(ctx context.Context) {\n\t\tname := ctx.Params().Get(\"name\")\n\t\ttoken := ctx.GetHeader(\"token\")\n\t\tif token == \"\" || getClientToken(name) != token {\n\t\t\tctx.StatusCode(403)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\t\tctx.Next()\n\t}\n}", "func NewMiddleware(creds *secrets.BasicAuthCredentials) es.CommandHandlerMiddleware {\n\treturn func(handler es.CommandHandler) es.CommandHandler {\n\t\treturn es.CommandHandlerFunc(func(ctx context.Context, cmd es.Command) error {\n\t\t\tcur, err := secrets.AuthFromContext(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif !creds.Equals(cur) {\n\t\t\t\treturn ErrInvalidCredentials\n\t\t\t}\n\n\t\t\treturn handler.HandleCommand(ctx, cmd)\n\t\t})\n\t}\n}", "func AuthFromConfig(ctx context.Context, c *Config, client func(string) *http.Client) (*Auth, error) {\n\tauthenticator, err := jwtauth.AuthFromConfig(ctx, &c.Config, client)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Auth{\n\t\tHeaders: c.Headers,\n\t\tAuthenticator: authenticator,\n\t\tUnauthHandler: DefaultUnauthHandler,\n\t}, nil\n}", "func newAuthenticator(secret string) *authenticator {\n\treturn &authenticator{\n\t\tSecret: secret,\n\t}\n}", "func (a *AuthCommand) Initialize(app *kingpin.Application, config *servicecfg.Config) {\n\ta.config = config\n\t// operations with authorities\n\tauth := app.Command(\"auth\", \"Operations with user and host certificate authorities (CAs).\").Hidden()\n\ta.authExport = auth.Command(\"export\", \"Export public cluster (CA) keys to stdout.\")\n\ta.authExport.Flag(\"keys\", \"if set, will print private keys\").BoolVar(&a.exportPrivateKeys)\n\ta.authExport.Flag(\"fingerprint\", \"filter authority by fingerprint\").StringVar(&a.exportAuthorityFingerprint)\n\ta.authExport.Flag(\"compat\", \"export certificates compatible with specific version of Teleport\").StringVar(&a.compatVersion)\n\ta.authExport.Flag(\"type\",\n\t\tfmt.Sprintf(\"export certificate type (%v)\", strings.Join(allowedCertificateTypes, \", \"))).\n\t\tEnumVar(&a.authType, allowedCertificateTypes...)\n\n\ta.authGenerate = auth.Command(\"gen\", \"Generate a new SSH keypair.\").Hidden()\n\ta.authGenerate.Flag(\"pub-key\", \"path to the public key\").Required().StringVar(&a.genPubPath)\n\ta.authGenerate.Flag(\"priv-key\", \"path to the private key\").Required().StringVar(&a.genPrivPath)\n\n\ta.authSign = auth.Command(\"sign\", \"Create an identity file(s) for a given user.\")\n\ta.authSign.Flag(\"user\", \"Teleport user name\").StringVar(&a.genUser)\n\ta.authSign.Flag(\"host\", \"Teleport host name\").StringVar(&a.genHost)\n\ta.authSign.Flag(\"out\", \"Identity output\").Short('o').Required().StringVar(&a.output)\n\ta.authSign.Flag(\"format\",\n\t\tfmt.Sprintf(\"Identity format: %s. %s is the default.\",\n\t\t\tidentityfile.KnownFileFormats.String(), identityfile.DefaultFormat)).\n\t\tDefault(string(identityfile.DefaultFormat)).\n\t\tStringVar((*string)(&a.outputFormat))\n\ta.authSign.Flag(\"ttl\", \"TTL (time to live) for the generated certificate.\").\n\t\tDefault(fmt.Sprintf(\"%v\", apidefaults.CertDuration)).\n\t\tDurationVar(&a.genTTL)\n\ta.authSign.Flag(\"compat\", \"OpenSSH compatibility flag\").StringVar(&a.compatibility)\n\ta.authSign.Flag(\"proxy\", `Address of the Teleport proxy. When --format is set to \"kubernetes\", this address will be set as cluster address in the generated kubeconfig file`).StringVar(&a.proxyAddr)\n\ta.authSign.Flag(\"overwrite\", \"Whether to overwrite existing destination files. When not set, user will be prompted before overwriting any existing file.\").BoolVar(&a.signOverwrite)\n\ta.authSign.Flag(\"tar\", \"Create a tarball of the resulting certificates and stream to stdout.\").BoolVar(&a.streamTarfile)\n\t// --kube-cluster was an unfortunately chosen flag name, before teleport\n\t// supported kubernetes_service and registered kubernetes clusters that are\n\t// not trusted teleport clusters.\n\t// It's kept as an alias for --leaf-cluster for backwards-compatibility,\n\t// but hidden.\n\ta.authSign.Flag(\"kube-cluster\", `Leaf cluster to generate identity file for when --format is set to \"kubernetes\"`).Hidden().StringVar(&a.leafCluster)\n\ta.authSign.Flag(\"leaf-cluster\", `Leaf cluster to generate identity file for when --format is set to \"kubernetes\"`).StringVar(&a.leafCluster)\n\ta.authSign.Flag(\"kube-cluster-name\", `Kubernetes cluster to generate identity file for when --format is set to \"kubernetes\"`).StringVar(&a.kubeCluster)\n\ta.authSign.Flag(\"app-name\", `Application to generate identity file for. Mutually exclusive with \"--db-service\".`).StringVar(&a.appName)\n\ta.authSign.Flag(\"db-service\", `Database to generate identity file for. Mutually exclusive with \"--app-name\".`).StringVar(&a.dbService)\n\ta.authSign.Flag(\"db-user\", `Database user placed on the identity file. Only used when \"--db-service\" is set.`).StringVar(&a.dbUser)\n\ta.authSign.Flag(\"db-name\", `Database name placed on the identity file. Only used when \"--db-service\" is set.`).StringVar(&a.dbName)\n\ta.authSign.Flag(\"windows-user\", `Window user placed on the identity file. Only used when --format is set to \"windows\"`).StringVar(&a.windowsUser)\n\ta.authSign.Flag(\"windows-domain\", `Active Directory domain for which this cert is valid. Only used when --format is set to \"windows\"`).StringVar(&a.windowsDomain)\n\ta.authSign.Flag(\"windows-sid\", `Optional Security Identifier to embed in the certificate. Only used when --format is set to \"windows\"`).StringVar(&a.windowsSID)\n\n\ta.authRotate = auth.Command(\"rotate\", \"Rotate certificate authorities in the cluster.\")\n\ta.authRotate.Flag(\"grace-period\", \"Grace period keeps previous certificate authorities signatures valid, if set to 0 will force users to re-login and nodes to re-register.\").\n\t\tDefault(fmt.Sprintf(\"%v\", defaults.RotationGracePeriod)).\n\t\tDurationVar(&a.rotateGracePeriod)\n\ta.authRotate.Flag(\"manual\", \"Activate manual rotation , set rotation phases manually\").BoolVar(&a.rotateManualMode)\n\ta.authRotate.Flag(\"type\", fmt.Sprintf(\"Certificate authority to rotate, one of: %s\", strings.Join(getCertAuthTypes(), \", \"))).Required().EnumVar(&a.rotateType, getCertAuthTypes()...)\n\ta.authRotate.Flag(\"phase\", fmt.Sprintf(\"Target rotation phase to set, used in manual rotation, one of: %v\", strings.Join(types.RotatePhases, \", \"))).StringVar(&a.rotateTargetPhase)\n\n\ta.authLS = auth.Command(\"ls\", \"List connected auth servers.\")\n\ta.authLS.Flag(\"format\", \"Output format: 'yaml', 'json' or 'text'\").Default(teleport.YAML).StringVar(&a.format)\n\n\ta.authCRL = auth.Command(\"crl\", \"Export empty certificate revocation list (CRL) for certificate authorities.\")\n\ta.authCRL.Flag(\"type\", fmt.Sprintf(\"Certificate authority type, one of: %s\", strings.Join(allowedCRLCertificateTypes, \", \"))).Required().EnumVar(&a.caType, allowedCRLCertificateTypes...)\n}", "func requireBasicAuth(\n\tcheckFunc func(context.Context, *http.Request, string, string, string) error,\n\toptional bool,\n) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tauthenticated := false\n\n\t\t\tif checkFunc != nil {\n\t\t\t\taps := chi.URLParam(r, apsParamName)\n\t\t\t\tusername, password, _ := r.BasicAuth()\n\t\t\t\tif err := checkFunc(r.Context(), r, aps, username, password); err == nil {\n\t\t\t\t\tauthenticated = true\n\t\t\t\t}\n\t\t\t} else if optional {\n\t\t\t\tauthenticated = true\n\t\t\t}\n\n\t\t\tif !authenticated {\n\t\t\t\treqHost := r.Host\n\t\t\t\tif host, _, err := net.SplitHostPort(reqHost); err == nil {\n\t\t\t\t\treqHost = host\n\t\t\t\t}\n\n\t\t\t\tw.Header().Set(wwwAuthenticateHeader, fmt.Sprintf(`Basic realm=\"estserver@%s\"`,\n\t\t\t\t\turl.QueryEscape(reqHost)))\n\t\t\t\terrAuthRequired.Write(w)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func NewAuthInterceptor(jwtManager *skyutl.JwtManager) *AuthInterceptor {\n\treturn &AuthInterceptor{\n\t\tjwtManager: jwtManager,\n\t}\n}", "func New(config Config, ds *datastore.Datastore) *Authenticator {\n\te := gin.Default()\n\tconfig.AccessDuration = config.AccessDuration * time.Minute\n\treturn &Authenticator{\n\t\tconfig: config,\n\t\tds: ds,\n\t\tengine: e,\n\t}\n}", "func serverAuth(logger *zap.Logger, url string, user string, password string) (string, error) {\n\tlogger.Debug(\"serverAuth called\", zap.String(\"url\", url), zap.String(\"user\", user))\n\tbody := fmt.Sprintf(`{\"auth\":{\"identity\":{\"methods\":[\"password\"],\"password\":{\"user\":{\"domain\":{\"id\":\"default\"},\"name\":\"%s\",\"password\":\"%s\"}}}}}`, user, password)\n\trbody := strings.NewReader(body)\n\treq, err := http.NewRequest(\"POST\", url+\"v3/auth/tokens\", rbody)\n\tif err != nil {\n\t\tlogger.Debug(\"serverAuth error from NewRequest POST\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Error(err))\n\t\treturn \"\", err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tlogger.Debug(\"serverAuth error from DefaultClient.Do POST\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Error(err))\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 201 {\n\t\tlogger.Debug(\"serverAuth error from POST return status\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Int(\"status\", resp.StatusCode))\n\t\treturn \"\", fmt.Errorf(\"server auth token request gave status %d\", resp.StatusCode)\n\t}\n\trv := resp.Header.Get(\"X-Subject-Token\")\n\tif len(rv) == 0 {\n\t\tlogger.Debug(\"serverAuth succeeded, but ended up with zero-length token\")\n\t} else {\n\t\tlogger.Debug(\"serverAuth succeeded\")\n\t}\n\treturn rv, nil\n}" ]
[ "0.57816464", "0.56545347", "0.5547903", "0.5484067", "0.5483845", "0.5380546", "0.53680843", "0.5358327", "0.5313363", "0.528413", "0.5263241", "0.51906294", "0.51678735", "0.5165179", "0.5149926", "0.5122884", "0.5073951", "0.5071476", "0.5035328", "0.50271136", "0.502607", "0.500705", "0.49980494", "0.49961448", "0.49857497", "0.49857497", "0.498403", "0.49762425", "0.49643847", "0.49495068", "0.49415037", "0.49385473", "0.49382895", "0.49305367", "0.49270144", "0.48943305", "0.48894078", "0.48842877", "0.48689598", "0.48666012", "0.4851179", "0.4848042", "0.48467752", "0.48404285", "0.4839151", "0.48382208", "0.4836828", "0.48318678", "0.4831518", "0.4821355", "0.48174202", "0.48089775", "0.47887856", "0.47882402", "0.4782108", "0.47790065", "0.47612774", "0.4753492", "0.47523087", "0.47518402", "0.4748379", "0.47355413", "0.47344434", "0.4725433", "0.47253266", "0.472376", "0.47213635", "0.4720032", "0.47183356", "0.47135133", "0.4710221", "0.47035003", "0.47028214", "0.47005495", "0.46981645", "0.46977144", "0.4697671", "0.46975625", "0.46973783", "0.4690849", "0.46874547", "0.46874318", "0.4678184", "0.46773592", "0.46663868", "0.46649534", "0.4662035", "0.4660257", "0.46521914", "0.46393073", "0.46341625", "0.4631977", "0.4628598", "0.4627031", "0.46265063", "0.4624483", "0.46213597", "0.46172562", "0.46136868", "0.46069285" ]
0.4852183
40
Return a new interceptor function that authorizes RPCs using a password stored in the config.
func streamAuthInterceptor(auth Authenticate, access Access) grpc.StreamServerInterceptor { return func(srv interface{}, ss grpc.ServerStream, info *grpc.StreamServerInfo, handler grpc.StreamHandler) error { //fmt.Printf("Streaming query: %#v\n", info) md, _ := metadata.FromIncomingContext(ss.Context()) //fmt.Printf("Metadata: %#v\n", md) metaData := MetaData{} for i := range md { metaData[i] = md[i] } user, err := auth.Validate(metaData) if err != nil { return status.Error(codes.Unauthenticated, "PermissionDenied") } //current GripQL schema does not support bi-directional streaming //mainly because it can't be offered via HTTP based interface if info.IsServerStream { //ssWrapper := ServerStreamWrapper{ss} switch info.FullMethod { case "/gripql.Query/Traversal": w, err := NewStreamOutWrapper[gripql.GraphQuery](ss) if err != nil { return status.Error(codes.Unknown, "Request error") } err = access.Enforce(user, w.Request.Graph, Query) if err != nil { return status.Error(codes.PermissionDenied, "PermissionDenied") } return handler(srv, w) case "/gripql.Job/ListJobs": //TODO: filter list of jobs return handler(srv, ss) case "/gripql.Job/ResumeJob": //TODO: filter list of jobs return handler(srv, ss) case "/gripql.Job/ViewJob": //TODO: filter list of jobs return handler(srv, ss) case "/gripql.Job/SearchJobs": //TODO: filter list of jobs return handler(srv, ss) } log.Errorf("Unknown streaming output: %#v", info) return handler(srv, ss) } else if info.IsClientStream { if info.FullMethod == "/gripql.Edit/BulkAdd" { //This checks permission on a per entity basis //unfortunatly because of limitations in Protobuf input //stream URL formatting, each write request can //reference a different graph return handler(srv, &BulkWriteFilter{ss, user, access}) } else { log.Errorf("Unknown input streaming op %#v!!!", info) return handler(srv, ss) } } return status.Error(codes.Unknown, "Unknown method") } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *server) authInterceptor(c context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\tif s.configs.Authorization {\n\t\tmethods := strings.Split(info.FullMethod, \"/\")\n\t\tif md, ok := metadata.FromIncomingContext(c); !ok {\n\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t} else {\n\t\t\tif methods[len(methods)-1] == \"UserLogin\" {\n\t\t\t\tr := req.(*pb.AuthInfo)\n\t\t\t\tif result, err := s.gdb.userLogin(authInfo{\n\t\t\t\t\tUserName: r.GetUserName(),\n\t\t\t\t\tPassWord: r.GetPassWord(),\n\t\t\t\t}); err != nil {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t} else {\n\t\t\t\t\treturn &pb.UserToken{Token: result.Token}, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tvar au string\n\t\t\t\tif d, ok := md[\"authorization\"]; ok {\n\t\t\t\t\tau = d[0]\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t}\n\t\t\t\tif userName, token, ok := parseBasicAuth(au); !ok {\n\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t} else {\n\t\t\t\t\tif r, err := s.gdb.query(\"select token from user_cfg where userName='\" + userName + \"'\"); err != nil || len(r) == 0 {\n\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif token != r[0][\"token\"] {\n\t\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tsub, obj, act := userName, methods[len(methods)-1], \"POST\" // replace gRCP with POST\n\t\t\t\t\t\t\tif ok, _ := s.gdb.e.Enforce(sub, obj, act); !ok {\n\t\t\t\t\t\t\t\treturn nil, status.Errorf(codes.Unauthenticated, \"invalid token\")\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn handler(c, req)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t} else {\n\t\treturn handler(c, req)\n\t}\n}", "func New(c Config) lilty.ChainHandler {\n\treturn func(next lilty.HandlerFunc) lilty.HandlerFunc {\n\t\treturn func(ctxt *lilty.Context) {\n\t\t\tusername, password, ok := ctxt.Request.BasicAuth()\n\n\t\t\tmatch := c.Username == username && c.Password == password\n\n\t\t\tif !ok || !match {\n\t\t\t\tctxt.SetResponseHeader(lilty.WWWAuthenticate, `Basic realm=\"`+c.Realm+`\"`)\n\t\t\t\tctxt.SetStatusCode(401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext(ctxt)\n\t\t}\n\t}\n}", "func authWrapFunc(f func(http.ResponseWriter, *http.Request)) http.Handler {\n\treturn negroni.New(&AuthMiddleware{}, negroni.Wrap(http.HandlerFunc(f)))\n}", "func authWrapper(handler http.HandlerFunc, secrets auth.SecretProvider, host string) http.HandlerFunc {\n\tauthenticator := &auth.BasicAuth{Realm: host, Secrets: secrets}\n\treturn auth.JustCheck(authenticator, handler)\n}", "func authUnaryIntercept(\n\tctx context.Context,\n\treq interface{},\n\tinfo *grpc.UnaryServerInfo,\n\thandler grpc.UnaryHandler,\n) (resp interface{}, err error) {\n\t//bypass auth if method is /hahiye.AuthService/Login\n\tif info.FullMethod == \"/hahiye.AuthService/Login\" {\n\t\tfmt.Println(\"bypassing auth cz it's login action\")\n\t\treturn handler(ctx, req)\n\t}\n\tif err := auth(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(\"authorization OK\")\n\treturn handler(ctx, req)\n}", "func New(username, password string) *client {\n\tr := newRequestHandler()\n\tc := &client{\n\t\tusername: username,\n\t\tpassword: password,\n\t\toauth: AccessTokenStruct{},\n\t\thandler: r,\n\t\twsConn: nil,\n\t\treconnectToken: uuid.Nil,\n\t}\n\terr := c.authenticate()\n\tif err != nil {\n\t\tc.handler.override = responseOverride{override: true, data: *err}\n\t}\n\tgo c.authenticator() // Launch authenticator\n\treturn c\n}", "func interceptor(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\n\tif err := auth(ctx); err != nil {\n\t\tfmt.Println(\"111\")\n\t\treturn nil, err\n\t}\n\t//继续处理请求\n\treturn handler(ctx, req)\n\n}", "func authEndpoint(rw http.ResponseWriter, req *http.Request) {\n\n\t// request has to be POST\n\tif req.Method != \"POST\" {\n\t\thttp.Error(rw, \"bad method, only post allowed\", http.StatusBadRequest)\n\t}\n\n\t// has to be authenticated, in a real we would use soemthing more\n\t// secure like certificates etc.\n\tuser, _, ok := req.BasicAuth()\n\n\tif !ok {\n\t\thttp.Error(rw, \"authentication required\", http.StatusForbidden)\n\t}\n\n\tlog.Println(\"basic authentication successful for \", user)\n\n\t// now we issue token and return it\n\n\t// This context will be passed to all methods.\n\tctx := req.Context()\n\n\t// Create an empty session object which will be passed to the request handlers\n\tmySessionData := newSession(\"\")\n\n\t// This will create an access request object and iterate through the registered TokenEndpointHandlers to validate the request.\n\taccessRequest, err := fositeInstance.NewAccessRequest(ctx, req, mySessionData)\n\n\t// Catch any errors, e.g.:\n\t// * unknown client\n\t// * invalid redirect\n\t// * ...\n\tif err != nil {\n\t\tlog.Printf(\"Error occurred in NewAccessRequest: %+v\", err)\n\t\tfositeInstance.WriteAccessError(rw, accessRequest, err)\n\t\treturn\n\t}\n\n\t// If this is a client_credentials grant, grant all requested scopes\n\t// NewAccessRequest validated that all requested scopes the client is allowed to perform\n\t// based on configured scope matching strategy.\n\tif accessRequest.GetGrantTypes().ExactOne(\"client_credentials\") {\n\t\tfor _, scope := range accessRequest.GetRequestedScopes() {\n\t\t\taccessRequest.GrantScope(scope)\n\t\t}\n\t}\n\n\t// Next we create a response for the access request. Again, we iterate through the TokenEndpointHandlers\n\t// and aggregate the result in response.\n\tresponse, err := fositeInstance.NewAccessResponse(ctx, accessRequest)\n\tif err != nil {\n\t\tlog.Printf(\"Error occurred in NewAccessResponse: %+v\", err)\n\t\tfositeInstance.WriteAccessError(rw, accessRequest, err)\n\t\treturn\n\t}\n\n\t// All done, send the response.\n\tfositeInstance.WriteAccessResponse(rw, accessRequest, response)\n\n}", "func (fn AuthenticatorFunc) Authenticate(username, password, otp string) (*Session, error) {\n\treturn fn(username, password, otp)\n}", "func AuthInterceptor(authURL string, insecure bool) grpc.UnaryServerInterceptor {\n\tauth := newKeystoneAuth(authURL, insecure)\n\n\treturn func(ctx context.Context, req interface{},\n\t\tinfo *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\tmd, ok := metadata.FromIncomingContext(ctx)\n\t\tif !ok {\n\t\t\treturn nil, errutil.ErrorUnauthenticated\n\t\t}\n\t\ttoken := md[\"x-auth-token\"]\n\t\tif len(token) == 0 {\n\t\t\treturn nil, errutil.ErrorUnauthenticated\n\t\t}\n\t\tnewCtx, err := authenticate(ctx, auth, token[0])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn handler(newCtx, req)\n\t}\n}", "func New(authenticator auth.Authenticator) clevergo.MiddlewareFunc {\n\treturn func(next clevergo.Handle) clevergo.Handle {\n\t\treturn func(c *clevergo.Context) error {\n\t\t\tidentity, err := authenticator.Authenticate(c.Request, c.Response)\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err)\n\t\t\t\tauthenticator.Challenge(c.Request, c.Response)\n\t\t\t} else {\n\t\t\t\tc.WithValue(auth.IdentityKey, identity)\n\t\t\t}\n\t\t\treturn next(c)\n\t\t}\n\t}\n}", "func AuthWrapper(fn server.HandlerFunc) server.HandlerFunc {\n\treturn func(ctx context.Context, req server.Request, resp interface{}) error {\n\t\tmeta, ok := metadata.FromContext(ctx)\n\t\tif !ok {\n\t\t\treturn errors.New(\"no auth meta-data found in request\")\n\t\t}\n\n\t\t// Note this is now uppercase (not entirely sure why this is...)\n\t\ttoken := meta[\"token\"]\n\n\t\tif token == \"\" {\n\t\t\treturn errors.New(\"token is empty\")\n\t\t}\n\n\t\tlog.Println(\"Authenticating with token: \", token)\n\n\t\t// Auth here\n\t\tauthClient := userService.NewUserServiceClient(\"user\", k8s.NewService().Client())\n\t\t_, err := authClient.ValidateToken(context.Background(), &userService.Token{\n\t\t\tToken: token,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = fn(ctx, req, resp)\n\t\treturn err\n\t}\n}", "func streamAuthIntercept(\n\tserver interface{},\n\tstream grpc.ServerStream,\n\tinfo *grpc.StreamServerInfo,\n\thandler grpc.StreamHandler,\n) error {\n\t//bypass auth if method is /hahiye.AuthService/Login\n\tif info.FullMethod == \"/hahiye.AuthService/Login\" {\n\t\tfmt.Println(\"bypassing auth cz it's login action\")\n\t\treturn handler(server, stream)\n\t}\n\tif err := auth(stream.Context()); err != nil {\n\t\treturn err\n\t}\n\tlog.Println(\"authorization OK\")\n\treturn handler(server, stream)\n}", "func NewAuthInterceptor(\n\tjwtToken string,\n\tauthMethods map[string]bool,\n) (*AuthInterceptor, error) {\n\tinterceptor := &AuthInterceptor{\n\t\tauthMethods: authMethods,\n\t\taccessToken: jwtToken,\n\t}\n\tlog.Printf(\"Inside Client AuthIntercept\")\n\treturn interceptor, nil\n}", "func New(realm string, credentials map[string]string, protectedHTTPMethods []string, enabled bool) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tif !enabled {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tmethodIsProtected := false\n\t\t\tfor _, protectedHTTPMethod := range protectedHTTPMethods {\n\t\t\t\tif r.Method == protectedHTTPMethod {\n\t\t\t\t\tmethodIsProtected = true\n\t\t\t\t}\n\t\t\t}\n\t\t\tif !methodIsProtected {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tusername, password, ok := r.BasicAuth()\n\t\t\tif !ok {\n\t\t\t\tunauthorized(w, realm)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tvalidPassword, userFound := credentials[username]\n\t\t\tif userFound {\n\t\t\t\tif password == validPassword {\n\t\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tunauthorized(w, realm)\n\t\t})\n\t}\n}", "func AuthAndCallAPI(w http.ResponseWriter, r *http.Request, service string, method string, version string) {\n\te := Execution{name: \"AuthAndCallAPI \" + service}\n\te.Start()\n\n\tauthorization := r.Header.Get(\"authorization\")\n\n\ttoken := \"\"\n\ts := strings.Split(authorization, \" \")\n\tif len(s) >= 2 {\n\t\ttoken = s[1]\n\t}\n\n\tconfig := config.GetConfig()\n\tresp, _ := resty.R().\n\t\tSetFormData(map[string]string{\n\t\t\t\"token\": token,\n\t\t\t\"service\": service,\n\t\t}).\n\t\tSetResult(&Respon{}).\n\t\tPost(config.API.Auth + \"v100/auth/check_token\")\n\n\tvar respon Respon\n\t_ = json.Unmarshal(resp.Body(), &respon)\n\n\tif respon.Code != 200 {\n\t\trespond := Respon{\n\t\t\tStatus: respon.Status,\n\t\t\tCode: respon.Code,\n\t\t\tMessage: respon.Message,\n\t\t\tExeTime: respon.ExeTime,\n\t\t\tData: respon.Data,\n\t\t\tError: respon.Error,\n\t\t}\n\t\tRespondJson(w, resp.StatusCode(), respond)\n\t\treturn\n\t}\n\n\tCallAPI(w, r, service, method, version)\n}", "func (p *UserPool) AuthUser(name string, password string) {\n\n}", "func Authenticator(t *Token) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\tfn := func(w http.ResponseWriter, req *http.Request) {\n\t\t\ttokenString, err := t.GetToken(req)\n\t\t\tif err != nil {\n\t\t\t\thttp.Error(w, err.Error(), http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tdata, err := t.CheckToken(tokenString)\n\t\t\tif err != nil {\n\t\t\t\thttp.Error(w, err.Error(), http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// If we get here, everything worked and we can set the\n\t\t\t// user data in context\n\t\t\tnewReq := req.WithContext(context.WithValue(req.Context(), t.options.UserProperty, data))\n\t\t\t// update the current request with the new context information\n\t\t\t*req = *newReq\n\t\t\tnext.ServeHTTP(w, req)\n\t\t}\n\t\treturn http.HandlerFunc(fn)\n\t}\n}", "func (b *BasicAuthenticationBackend) Wrap(wrapped auth.AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := tokenFromRequest(r)\n\n\t\t// if not coming from cookie or X-Auth-Token\n\t\ts := strings.SplitN(token, \" \", 2)\n\t\tif len(s) == 2 || s[0] == \"Basic\" {\n\t\t\ttoken = s[1]\n\t\t}\n\n\t\t// add \"fake\" header to let the basic auth library do the authentication\n\t\tr.Header.Set(\"Authorization\", \"Basic \"+token)\n\n\t\tif username := b.CheckAuth(r); username == \"\" {\n\t\t\tUnauthorized(w, r, ErrWrongCredentials)\n\t\t} else {\n\t\t\tauthCallWrapped(w, r, username, wrapped)\n\t\t}\n\t}\n}", "func (*SpecialAuth) AuthFunc(_ context.Context, apiKey string, _ map[string]string) bool {\n\treturn apiKey == \"12345\"\n}", "func NewAuthInterceptor(\n\tauthn authn.AuthenticationServiceClient,\n\tauthz GRPCAuthorizationHandler,\n) AuthorizationInterceptor {\n\treturn &authInterceptor{authn: authn, authz: authz}\n}", "func CustomAuth(cfg *types.Config) gin.HandlerFunc {\n\tbasicAuthHandler := gin.BasicAuth(gin.Accounts{\n\t\t// Use the config's username and password for basic auth\n\t\tcfg.Username: cfg.Password,\n\t})\n\n\toidcHandler := getOIDCMiddleware(cfg.OIDCIssuer, cfg.OIDCSubject, cfg.OIDCGroups)\n\n\treturn func(c *gin.Context) {\n\t\tauthHeader := c.GetHeader(\"Authorization\")\n\t\tif strings.HasPrefix(authHeader, \"Bearer \") {\n\t\t\toidcHandler(c)\n\t\t} else {\n\t\t\tbasicAuthHandler(c)\n\t\t}\n\t}\n}", "func WrapAuthenticate(hfn http.Handler) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlVars := mux.Vars(r)\n\t\turlValues := r.URL.Query()\n\n\t\t// if the url parameter 'key' is empty or absent, end the request with an unauthorized response\n\t\tif urlValues.Get(\"key\") == \"\" {\n\t\t\terr := APIErrorUnauthorized()\n\t\t\trespondErr(w, err)\n\t\t\treturn\n\t\t}\n\n\t\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\t\tserviceToken := gorillaContext.Get(r, \"auth_service_token\").(string)\n\n\t\tprojectName := urlVars[\"project\"]\n\t\tprojectUUID := projects.GetUUIDByName(urlVars[\"project\"], refStr)\n\n\t\t// In all cases instead of project create\n\t\tif \"projects:create\" != mux.CurrentRoute(r).GetName() {\n\t\t\t// Check if given a project name the project wasn't found\n\t\t\tif projectName != \"\" && projectUUID == \"\" {\n\t\t\t\tapiErr := APIErrorNotFound(\"project\")\n\t\t\t\trespondErr(w, apiErr)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t// Check first if service token is used\n\t\tif serviceToken != \"\" && serviceToken == urlValues.Get(\"key\") {\n\t\t\tgorillaContext.Set(r, \"auth_roles\", []string{\"service_admin\"})\n\t\t\tgorillaContext.Set(r, \"auth_user\", \"\")\n\t\t\tgorillaContext.Set(r, \"auth_user_uuid\", \"\")\n\t\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\t\thfn.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\n\t\troles, user := auth.Authenticate(projectUUID, urlValues.Get(\"key\"), refStr)\n\n\t\tif len(roles) > 0 {\n\t\t\tuserUUID := auth.GetUUIDByName(user, refStr)\n\t\t\tgorillaContext.Set(r, \"auth_roles\", roles)\n\t\t\tgorillaContext.Set(r, \"auth_user\", user)\n\t\t\tgorillaContext.Set(r, \"auth_user_uuid\", userUUID)\n\t\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\t\thfn.ServeHTTP(w, r)\n\t\t} else {\n\t\t\terr := APIErrorUnauthorized()\n\t\t\trespondErr(w, err)\n\t\t}\n\n\t})\n}", "func AuthFunc(ctx context.Context) (context.Context, error) {\n\treturn ctx, nil\n}", "func New(addr string, c auth.Cache, opts ...auth.Option) auth.Strategy {\n\tfn := GetAuthenticateFunc(addr, opts...)\n\treturn token.New(fn, c, opts...)\n}", "func New(addr string, c auth.Cache, opts ...auth.Option) auth.Strategy {\n\tfn := GetAuthenticateFunc(addr, opts...)\n\treturn token.New(fn, c, opts...)\n}", "func (b *KeystoneAuthenticationBackend) Wrap(wrapped auth.AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := tokenFromRequest(r)\n\n\t\tif username, err := b.CheckUser(token); username == \"\" {\n\t\t\tUnauthorized(w, r, err)\n\t\t} else {\n\t\t\tauthCallWrapped(w, r, username, wrapped)\n\t\t}\n\t}\n}", "func authWrap(h http.Handler) http.Handler {\n\treturn negroni.New(&AuthMiddleware{}, negroni.Wrap(h))\n}", "func AuthenticationInterceptor(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (out interface{}, err error) {\n\tmethodParts := strings.Split(info.FullMethod, \"/\")\n\texceptions := []string{\"Auth\", \"Register\", \"Items\", \"Item\", \"AddOrder\"}\n\t_, found := utils.Find(exceptions, methodParts[len(methodParts)-1])\n\tif found {\n\t\treturn handler(ctx, req)\n\t}\n\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif !ok {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\ttokenString, ok := md[\"authorization\"]\n\tif !ok || len(tokenString) < 1 || len(tokenString[0]) < 8 {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\tclaims := &auth.Claims{}\n\ttoken, err := jwt.ParseWithClaims(tokenString[0][7:], claims, func(token *jwt.Token) (interface{}, error) {\n\t\treturn jwtKey, nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tuserID := claims.UserID\n\n\t// load userID into context\n\tauthContext := auth.AuthContext{\n\t\tUserID: userID,\n\t}\n\n\tif token.Valid {\n\t\tctx = context.WithValue(ctx, auth.UserAuthKey, authContext)\n\t} else {\n\t\treturn nil, errGrpcUnauthenticated\n\t}\n\n\treturn handler(ctx, req)\n}", "func login(service Service, logger log.Logger) routing.Handler {\n return func(c *routing.Context) error {\n var req struct {\n Username string `json:\"username\"`\n Password string `json:\"password\"`\n }\n\n if err := c.Read(&req); err != nil {\n logger.With(c.Request.Context()).Errorf(\"invalid request: %v\", err)\n return errors.BadRequest(\"\")\n }\n\n token, err := service.Login(c.Request.Context(), req.Username, req.Password)\n if err != nil {\n return err\n }\n\n return c.Write(struct {\n Token string `json:\"token\"`\n }{token})\n }\n}", "func (c *Commander) SignIn(w http.ResponseWriter, r *http.Request) {\n\tvar client model.Client\n\tvar err error\n\tvar error model.Error\n\tdb := database.DbConn()\n\tdefer db.Close()\n\t// create a new client\n\tif client, err = New(model.Config{\n\t\tBaseDN: \"DC=sls,DC=ads,DC=valuelabs,DC=net\",\n\t\t//BaseDN: \"cn=ltest,ou=SERVICE ACCOUNTS,ou=SLS,dc=SLS,dc=ads,dc=valuelabs,dc=net\",\n\t\tFilter: \"userPrincipalName\",\n\t\tROUser: model.User{Name: \"L test\", Password: \"Welcome@123\"},\n\t\tTitle: \"title\",\n\t\tHost: \"10.10.52.113:389\",\n\t}); err != nil {\n\t\tWriteLogFile(err)\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\tvar creds model.Credentials\n\t//\tvar pass string\n\tSetupResponse(&w, r)\n\tif (*r).Method == \"OPTIONS\" {\n\t\tw.Header().Set(\"Access-Control-Max-Age\", \"86400\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t// Get the JSON body and decode into credentials\n\terr = json.NewDecoder(r.Body).Decode(&creds)\n\tif err != nil {\n\t\t// If the structure of the body is wrong, return an HTTP error\n\t\tWriteLogFile(err)\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\t// var usr = creds.Username\n\t// var bytePassword = []byte(creds.Password)\n\tusername := creds.Username\n\tpassword := creds.Password\n\tRole = creds.Role\n\tsplitUser := strings.Split(username, \"@\")\n\tprint := splitUser[0]\n\tuser1 := fmt.Sprintf(\"%[email protected]\", print)\n\tuser2 := fmt.Sprintf(\"%[email protected]\", print)\n\terr = client.Auth(user2, password)\n\tif err == nil {\n\t\tfmt.Println(\"Success!\")\n\t\ttoken(w, user1)\n\t} else if err.Error() == \"not found\" {\n\t\tfmt.Println(\"H2\")\n\t\tif errr := client.Auth(user1, password); errr != nil {\n\t\t\tfmt.Println(\"H3\")\n\t\t\tWriteLogFile(errr)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\terror.Code = \"401\"\n\t\t\terror.Message = \"Invalid Username or Password\"\n\t\t\tjson.NewEncoder(w).Encode(error)\n\t\t\treturn\n\t\t} //else {\n\t\tfmt.Println(\"Success!\")\n\t\ttoken(w, user1)\n\t\t//}\n\t} else {\n\t\tfmt.Println(\"H4\")\n\t\tWriteLogFile(err)\n\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\terror.Code = \"401\"\n\t\terror.Message = \"Invalid Username or Password\"\n\t\tjson.NewEncoder(w).Encode(error)\n\t\treturn\n\t}\n}", "func Credentials(addr, username, password string) InfluxReporterFunc {\n\treturn func(c *InfluxReporter) {\n\t\tc.addr = addr\n\t\tc.username = username\n\t\tc.password = password\n\t}\n}", "func (b *basicAuth) set(r *http.Request) { r.SetBasicAuth(b.username, b.password) }", "func makeAuthFunc(schemes map[string]bool) func(ctx context.Context, input *AuthenticationInput) error {\n\treturn func(ctx context.Context, input *AuthenticationInput) error {\n\t\t// If the scheme is valid and present in the schemes\n\t\tvalid, present := schemes[input.SecuritySchemeName]\n\t\tif valid && present {\n\t\t\treturn nil\n\t\t}\n\n\t\t// If the scheme is present in che schemes\n\t\tif present {\n\t\t\t// Return an unmet scheme error\n\t\t\treturn fmt.Errorf(\"security scheme for %q wasn't met\", input.SecuritySchemeName)\n\t\t}\n\t\t// Return an unknown scheme error\n\t\treturn fmt.Errorf(\"security scheme for %q is unknown\", input.SecuritySchemeName)\n\t}\n}", "func (m *BasicAuthentication) SetPassword(value *string)() {\n m.password = value\n}", "func setBasicAuth(u *vcs.URL, extSvcType, username, password string) error {\n\tswitch extSvcType {\n\tcase extsvc.TypeGitHub, extsvc.TypeGitLab:\n\t\treturn errors.New(\"need token to push commits to \" + extSvcType)\n\n\tcase extsvc.TypeBitbucketServer, extsvc.TypeBitbucketCloud:\n\t\tu.User = url.UserPassword(username, password)\n\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"setBasicAuth: invalid external service type %q\", extSvcType))\n\t}\n\treturn nil\n}", "func (s *Server) AuthFuncOverride(ctx context.Context, fullMethodName string) (context.Context, error) {\n\treturn ctx, nil\n}", "func NewSetAuth() context.Handler {\n\treturn func(ctx context.Context) {\n\t\taddr := ctx.RemoteAddr()\n\t\tif _, ok := tempForbidRemote[addr]; ok {\n\t\t\tctx.StatusCode(404)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\n\t\tkey := ctx.Params().Get(\"name\")\n\t\ttoken := ctx.GetHeader(\"token\")\n\t\tif token == \"\" || getServerToken(key) != token {\n\t\t\ttempForbidRemote[addr] = 1\n\t\t\tctx.StatusCode(404)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\t\tctx.Next()\n\t}\n}", "func newTokenInjector(meta *metadata.Client, acc string) func(*http.Request) error {\n\tif acc == \"\" {\n\t\tacc = \"default\"\n\t}\n\tacc = url.PathEscape(acc)\n\treturn func(req *http.Request) error {\n\t\taud := fmt.Sprintf(\"%s://%s\", req.URL.Scheme, req.URL.Host)\n\t\taud = url.QueryEscape(aud)\n\t\t// TODO(smut): Cache the token and reuse if not yet expired.\n\t\t// Currently the only user of this package only makes one\n\t\t// request per boot so caching isn't too important yet.\n\t\ttok, err := meta.Get(fmt.Sprintf(tokMetadata, acc, aud))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treq.Header.Set(vmtoken.Header, tok)\n\t\treturn nil\n\t}\n}", "func ClientPassword(password string) func(*Client) {\n\treturn func(c *Client) { c.password = password }\n}", "func unaryAuthInterceptor(auth Authenticate, access Access) grpc.UnaryServerInterceptor {\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\t//fmt.Printf(\"AuthInt: %#v\\n\", ctx)\n\t\tmd, _ := metadata.FromIncomingContext(ctx)\n\t\t//fmt.Printf(\"Metadata: %#v\\n\", md)\n\t\t//omd, _ := metadata.FromOutgoingContext(ctx)\n\t\t//fmt.Printf(\"Raw: %#v\\n\", omd)\n\n\t\tmetaData := MetaData{}\n\t\tfor i := range md {\n\t\t\tmetaData[i] = md[i]\n\t\t}\n\n\t\tuser, err := auth.Validate(metaData)\n\t\tif err != nil {\n\t\t\treturn nil, status.Error(codes.Unauthenticated, \"PermissionDenied\")\n\t\t}\n\n\t\tif op, ok := MethodMap[info.FullMethod]; ok {\n\t\t\tgraph, err := getUnaryRequestGraph(req, info)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Error(codes.Unknown, fmt.Sprintf(\"Unable to get graph: %s\", err))\n\t\t\t}\n\t\t\terr = access.Enforce(user, graph, op)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, status.Error(codes.PermissionDenied, \"PermissionDenied\")\n\t\t\t}\n\t\t\treturn handler(ctx, req)\n\t\t}\n\t\treturn nil, status.Error(codes.Unknown, \"Unknown method\")\n\t}\n}", "func (ba BasicAuthenticator) Wrap(handler http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tusername, password, ok := r.BasicAuth()\n\n\t\tif !ok {\n\t\t\tvar strRequest string\n\t\t\tdumpRequest, err := httputil.DumpRequest(r, true)\n\t\t\tif err == nil {\n\t\t\t\tstrRequest = string(dumpRequest)\n\t\t\t}\n\t\t\terrStr := fmt.Sprint(\"credentials error\", \" error:\", ErrInvalidCredFormat, \" request:\", strRequest)\n\t\t\tw.Header().Set(\"WWW-Authenticate\", \"Basic realm=Authorization Required\")\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tdtlogger.Error(errStr)\n\t\t\t//http.Error(w, errStr, http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\tif username != ba.username || password != ba.password {\n\t\t\tvar strRequest string\n\t\t\tdumpRequest, err := httputil.DumpRequest(r, true)\n\t\t\tif err == nil {\n\t\t\t\tstrRequest = string(dumpRequest)\n\t\t\t}\n\t\t\terrStr := fmt.Sprint(\"credentials error\", \" error:\", ErrInvalidCreds, \" request:\", strRequest)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\tdtlogger.Error(errStr)\n\t\t\t//http.Error(w, errStr, http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\thandler.ServeHTTP(w, r)\n\t})\n}", "func Authenticate(endpoint func(http.ResponseWriter, *http.Request)) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\tendpoint(w, r)\n\t})\n}", "func CredentialCtx(s *Server) func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t// Get session cookie.\n\t\t\tsession := \"\"\n\t\t\tif cookie, err := r.Cookie(\"session\"); err == nil {\n\t\t\t\tsession = cookie.Value\n\t\t\t}\n\n\t\t\t// Get API key header.\n\t\t\tapiKey := \"\"\n\t\t\tif apiKeys, ok := r.Header[\"X-Api-Key\"]; ok {\n\t\t\t\tapiKey = apiKeys[0]\n\t\t\t}\n\t\t\tapiSecret := \"\"\n\t\t\tif apiSecrets, ok := r.Header[\"X-Api-Secret\"]; ok {\n\t\t\t\tapiSecret = apiSecrets[0]\n\t\t\t}\n\n\t\t\tif session == \"\" && apiKey == \"\" {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tauthInfo := chassis.AuthInfo{}\n\t\t\tif session != \"\" {\n\t\t\t\tif userID, _, isAdmin, err := s.db.LookupSession(session); err == nil {\n\t\t\t\t\tauthInfo.AuthMethod = chassis.SessionAuth\n\t\t\t\t\tauthInfo.UserID = userID\n\t\t\t\t\tauthInfo.UserIsAdmin = isAdmin\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t//TODO: CACHE MAY BE WORTH IT\n\t\t\t\tif user, err := s.userSvc.GetUserByApiKey(apiKey, apiSecret); err == nil {\n\t\t\t\t\tauthInfo.AuthMethod = chassis.SessionAuth\n\t\t\t\t\tauthInfo.UserID = user.ID\n\t\t\t\t\tauthInfo.UserIsAdmin = user.IsAdmin\n\t\t\t\t}\n\t\t\t}\n\t\t\tnext.ServeHTTP(w, r.WithContext(chassis.NewAuthContext(r.Context(), &authInfo)))\n\t\t})\n\t}\n}", "func (a *BasicAuth) Wrap(wrapped AuthenticatedHandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tif username := a.CheckAuth(r); username == \"\" {\n\t\t\ta.RequireAuth(w, r)\n\t\t} else {\n\t\t\tar := &AuthenticatedRequest{Request: *r, Username: username}\n\t\t\twrapped(w, ar)\n\t\t}\n\t}\n}", "func (c Clients) Auth(ctx context.Context, username, password string) error {\n\tvar req *request\n\tif username == \"\" {\n\t\treq = newRequest(\"*2\\r\\n$4\\r\\nAUTH\\r\\n$\")\n\t\treq.addString(password)\n\t} else {\n\t\treq = newRequest(\"*3\\r\\n$4\\r\\nAUTH\\r\\n$\")\n\t\treq.addString2(username, password)\n\t}\n\treturn c.c.cmdSimple(ctx, req)\n}", "func CustomAuth(handler http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\ttoken := r.Header.Get(\"Authorization\")\n\t\ttokenPieces := strings.Split(token, \" \")\n\t\tif len(tokenPieces) == 2 && tokenPieces[0] == \"Basic\" {\n\t\t\t// TODO: do basic auth\n\t\t\thandler.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\t\t// TODO: do session auth\n\n\t\t// if both validations fails\n\t\thttp.Error(w, \"error while trying to authenticate\", http.StatusUnauthorized)\n\n\t})\n}", "func New(config Config) App {\n\treturn App{\n\t\treq: request.New().Post(strings.TrimSpace(*config.url)).BasicAuth(strings.TrimSpace(*config.user), *config.pass),\n\t}\n}", "func newProxy(config *Config) (*oauthProxy, error) {\n\t// create the service logger\n\tlog, err := createLogger(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.Info(\"starting the service\", zap.String(\"prog\", prog), zap.String(\"author\", author), zap.String(\"version\", version))\n\tsvc := &oauthProxy{\n\t\tconfig: config,\n\t\tlog: log,\n\t\tmetricsHandler: prometheus.Handler(),\n\t}\n\n\t// parse the upstream endpoint\n\tif svc.endpoint, err = url.Parse(config.Upstream); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// initialize the store if any\n\tif config.StoreURL != \"\" {\n\t\tif svc.store, err = createStorage(config.StoreURL); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// initialize the openid client\n\tif !config.SkipTokenVerification {\n\t\tif svc.client, svc.idp, svc.idpClient, err = svc.newOpenIDClient(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tlog.Warn(\"TESTING ONLY CONFIG - the verification of the token have been disabled\")\n\t}\n\n\tif config.ClientID == \"\" && config.ClientSecret == \"\" {\n\t\tlog.Warn(\"client credentials are not set, depending on provider (confidential|public) you might be unable to auth\")\n\t}\n\n\t// are we running in forwarding mode?\n\tif config.EnableForwarding {\n\t\tif err := svc.createForwardingProxy(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tif err := svc.createReverseProxy(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn svc, nil\n}", "func Auth(fn func(ctx *system.Context)) func(ctx *system.Context) {\n\treturn func(ctx *system.Context) {\n\t\tgconfig, err := ctx.System.DB.CreateGuildIfNotExists(ctx.Msg.GuildID)\n\t\tif err != nil {\n\t\t\tctx.ReplyError(\"Error getting guild configuration: \", err)\n\t\t\treturn\n\t\t}\n\n\t\tisAdmin, err := ctx.IsAdmin()\n\t\tif err != nil {\n\t\t\tctx.ReplyError(\"Error checking administrator status: \", err)\n\t\t\treturn\n\t\t}\n\n\t\tif !isAdmin {\n\t\t\tctx.ReplyError(\"You need to be an administrator or own the guild to configure guild settings\")\n\t\t\treturn\n\t\t}\n\n\t\tctx.Set(\"gconfig\", gconfig)\n\t\tfn(ctx)\n\t}\n}", "func (r *RPC) Auth(c context.Context, arg *rpc.Auth, res *struct{}) (err error) {\n\treturn\n}", "func WrapHandlerInBasicAuth(h http.Handler, b BasicAuth) http.HandlerFunc {\n\tif b.Username == \"\" || b.Password == \"\" {\n\t\tlogrus.Warn(\"Metrics are exposed without protection. Make sure you set up protection at proxy level.\")\n\t}\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\t// Serve without authorization if either Username or Password is unset\n\t\tif b.Username == \"\" || b.Password == \"\" {\n\t\t\th.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\t\tuser, pass, ok := r.BasicAuth()\n\n\t\tif !ok || user != b.Username || pass != b.Password {\n\t\t\thttp.Error(w, http.StatusText(http.StatusForbidden), http.StatusForbidden)\n\t\t\treturn\n\t\t}\n\t\th.ServeHTTP(w, r)\n\t}\n}", "func UnaryAuth() grpc.UnaryServerInterceptor {\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (resp interface{}, err error) {\n\t\t//Get auth data that's passed from client in context.\n\t\tmd, ok := metadata.FromContext(ctx)\n\t\tif !ok {\n\t\t\tfmt.Println(\"Could not get metadata\", err)\n\t\t}\n\n\t\t//Check it to make sure it's good\n\t\tcred := md[authKey]\n\t\tfmt.Println(\"cred\", cred)\n\t\tif cred == nil {\n\t\t\t//Reject call if not\n\t\t\treturn nil, errors.New(\"Not authorized to make this call!\")\n\t\t}\n\n\t\t//Add user data to ctx.\n\t\tctx = context.WithValue(ctx, \"user\", cred)\n\n\t\t//Pass to next handler\n\t\treturn handler(ctx, req)\n\t}\n}", "func New() iris2.HandlerFunc {\n\tl := &authMiddleware{}\n\treturn l.Serve\n}", "func JWTAuthenticationInterceptor(publicKey string, excludingPath []string) grpc.UnaryServerInterceptor {\n\treturn func(ctx context.Context, req interface{}, info *grpc.UnaryServerInfo, handler grpc.UnaryHandler) (interface{}, error) {\n\t\treturn handler(ctx, req)\n\t}\n}", "func getToken(urlStr string, creds []byte)string{\n\n\tvar urlBuffer bytes.Buffer\n\n\tproxyStr := os.Getenv(\"HTTPS_PROXY\")\n\tproxyURL, err := url.Parse(proxyStr)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tfmt.Println(\"Authenticating with CloudBolt API....\")\n\turlBuffer.WriteString(urlStr)\n\tfmt.Println(urlStr)\n\turlBuffer.WriteString(\"/api/v2/api-token-auth/\")\n\treq, err := http.NewRequest(\"POST\", urlBuffer.String(), bytes.NewBuffer(creds))\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t\tProxy: http.ProxyURL(proxyURL),\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer resp.Body.Close()\n\tfmt.Println(resp.StatusCode)\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\ttoken := new(Token)\n\terr = json.Unmarshal(body, &token)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\treturn token.Token\n\n}", "func runAuthProxy(s *tsnet.Server, rt http.RoundTripper, logf logger.Logf) {\n\tln, err := s.Listen(\"tcp\", \":443\")\n\tif err != nil {\n\t\tlog.Fatalf(\"could not listen on :443: %v\", err)\n\t}\n\tu, err := url.Parse(fmt.Sprintf(\"https://%s:%s\", os.Getenv(\"KUBERNETES_SERVICE_HOST\"), os.Getenv(\"KUBERNETES_SERVICE_PORT_HTTPS\")))\n\tif err != nil {\n\t\tlog.Fatalf(\"runAuthProxy: failed to parse URL %v\", err)\n\t}\n\n\tlc, err := s.LocalClient()\n\tif err != nil {\n\t\tlog.Fatalf(\"could not get local client: %v\", err)\n\t}\n\tap := &authProxy{\n\t\tlogf: logf,\n\t\tlc: lc,\n\t\trp: &httputil.ReverseProxy{\n\t\t\tDirector: func(r *http.Request) {\n\t\t\t\t// We want to proxy to the Kubernetes API, but we want to use\n\t\t\t\t// the caller's identity to do so. We do this by impersonating\n\t\t\t\t// the caller using the Kubernetes User Impersonation feature:\n\t\t\t\t// https://kubernetes.io/docs/reference/access-authn-authz/authentication/#user-impersonation\n\n\t\t\t\t// Out of paranoia, remove all authentication headers that might\n\t\t\t\t// have been set by the client.\n\t\t\t\tr.Header.Del(\"Authorization\")\n\t\t\t\tr.Header.Del(\"Impersonate-Group\")\n\t\t\t\tr.Header.Del(\"Impersonate-User\")\n\t\t\t\tr.Header.Del(\"Impersonate-Uid\")\n\t\t\t\tfor k := range r.Header {\n\t\t\t\t\tif strings.HasPrefix(k, \"Impersonate-Extra-\") {\n\t\t\t\t\t\tr.Header.Del(k)\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// Now add the impersonation headers that we want.\n\t\t\t\twho := r.Context().Value(whoIsKey{}).(*apitype.WhoIsResponse)\n\t\t\t\tif who.Node.IsTagged() {\n\t\t\t\t\t// Use the nodes FQDN as the username, and the nodes tags as the groups.\n\t\t\t\t\t// \"Impersonate-Group\" requires \"Impersonate-User\" to be set.\n\t\t\t\t\tr.Header.Set(\"Impersonate-User\", strings.TrimSuffix(who.Node.Name, \".\"))\n\t\t\t\t\tfor _, tag := range who.Node.Tags {\n\t\t\t\t\t\tr.Header.Add(\"Impersonate-Group\", tag)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tr.Header.Set(\"Impersonate-User\", who.UserProfile.LoginName)\n\t\t\t\t}\n\n\t\t\t\t// Replace the URL with the Kubernetes APIServer.\n\t\t\t\tr.URL.Scheme = u.Scheme\n\t\t\t\tr.URL.Host = u.Host\n\t\t\t},\n\t\t\tTransport: rt,\n\t\t},\n\t}\n\ths := &http.Server{\n\t\t// Kubernetes uses SPDY for exec and port-forward, however SPDY is\n\t\t// incompatible with HTTP/2; so disable HTTP/2 in the proxy.\n\t\tTLSConfig: &tls.Config{\n\t\t\tGetCertificate: lc.GetCertificate,\n\t\t\tNextProtos: []string{\"http/1.1\"},\n\t\t},\n\t\tTLSNextProto: make(map[string]func(*http.Server, *tls.Conn, http.Handler)),\n\t\tHandler: ap,\n\t}\n\tif err := hs.ServeTLS(ln, \"\", \"\"); err != nil {\n\t\tlog.Fatalf(\"runAuthProxy: failed to serve %v\", err)\n\t}\n}", "func basicAuth(password string) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tw.Header().Set(\"WWW-Authenticate\", `Basic realm=\"Restricted\"`)\n\n\t\t\ts := strings.SplitN(r.Header.Get(\"Authorization\"), \" \", 2)\n\t\t\tif len(s) != 2 {\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tb, err := base64.StdEncoding.DecodeString(s[1])\n\t\t\tif err != nil {\n\t\t\t\tlog.Printf(\"base64.StdEncoding.DecodeString() Error: %s\\n\", err)\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tpair := strings.SplitN(string(b), \":\", 2)\n\t\t\tif len(pair) != 2 {\n\t\t\t\tlog.Printf(\"strings.SplitN() Error: %s\\n\", err)\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif pair[0] != \"admin\" || pair[1] != password {\n\t\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func Authenticator(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tctx := r.Context()\n\n\t\tjwtToken, err := TokenFromContext(ctx)\n\t\tif err != nil {\n\t\t\thttp.Error(w, http.StatusText(401), 401)\n\t\t\treturn\n\t\t}\n\n\t\t_ = jwtToken\n\n\t\t// Token is authenticated, pass it through\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func RequestCredentials(w http.ResponseWriter, r *http.Request) {\n\tvar (\n\t\tidentity AuthIdentity\n\t\tcredentials AuthCredentials\n\t)\n\n\teaaCtx := r.Context().Value(contextKey(\"appliance-ctx\")).(*Context)\n\n\tconst fName = \"/Auth RequestCredentials \"\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\n\terr := json.NewDecoder(r.Body).Decode(&identity)\n\tif err != nil {\n\t\tlog.Errf(fName+\"decode failed: %v\", err.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\thost, port, err := net.SplitHostPort(r.RemoteAddr)\n\tif err != nil {\n\t\tlog.Errf(fName+\"Cannot retrieve IP from RemoteAddr: %v [%v:%v] %v\",\n\t\t\tr.RemoteAddr, host, port, err)\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tcert, err := SignCSR(identity.Csr, eaaCtx)\n\tif err != nil {\n\t\tlog.Errf(fName+\"failed: %v\", err.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tsignedCertBlock := pem.EncodeToMemory(\n\t\t&pem.Block{Type: \"CERTIFICATE\", Bytes: cert.Raw})\n\tif signedCertBlock == nil {\n\t\tlog.Err(fName + \"/failed to enode signed cert\")\n\t\treturn\n\t}\n\trcaBlock := pem.EncodeToMemory(\n\t\t&pem.Block{Type: \"CERTIFICATE\",\n\t\t\tBytes: eaaCtx.certsEaaCa.rca.x509Cert.Raw})\n\tif rcaBlock == nil {\n\t\tlog.Err(fName + \"failed to enode rca cert\")\n\t\treturn\n\t}\n\n\tcredentials.ID = cert.Subject.CommonName\n\tcredentials.Certificate = string(signedCertBlock)\n\tcredentials.CaChain = []string{string(rcaBlock)}\n\tcredentials.CaPool = []string{string(rcaBlock)}\n\n\tencoder := json.NewEncoder(w)\n\terr = encoder.Encode(credentials)\n\tif err != nil {\n\t\tlog.Errf(fName+\"encoding output to JSON failed: %s\",\n\t\t\terr.Error())\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tlog.Info(fName + \" request from CN: \" + credentials.ID + \", from IP: \" +\n\t\thost + \" properly handled\")\n}", "func redirectPolicyFunc(req *http.Request, via []*http.Request) error {\n\treq.Header.Add(\"Authorization\", \"Basic \"+globalBasicAuthString)\n\treturn nil\n}", "func WrapMockAuthConfig(hfn http.HandlerFunc, cfg *config.APICfg, brk brokers.Broker, str stores.Store, mgr *oldPush.Manager, c push.Client, roles ...string) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlVars := mux.Vars(r)\n\n\t\tuserRoles := []string{\"publisher\", \"consumer\"}\n\t\tif len(roles) > 0 {\n\t\t\tuserRoles = roles\n\t\t}\n\n\t\tnStr := str.Clone()\n\t\tdefer nStr.Close()\n\n\t\tprojectUUID := projects.GetUUIDByName(urlVars[\"project\"], nStr)\n\t\tgorillaContext.Set(r, \"auth_project_uuid\", projectUUID)\n\t\tgorillaContext.Set(r, \"brk\", brk)\n\t\tgorillaContext.Set(r, \"str\", nStr)\n\t\tgorillaContext.Set(r, \"mgr\", mgr)\n\t\tgorillaContext.Set(r, \"apsc\", c)\n\t\tgorillaContext.Set(r, \"auth_resource\", cfg.ResAuth)\n\t\tgorillaContext.Set(r, \"auth_user\", \"UserA\")\n\t\tgorillaContext.Set(r, \"auth_user_uuid\", \"uuid1\")\n\t\tgorillaContext.Set(r, \"auth_roles\", userRoles)\n\t\tgorillaContext.Set(r, \"push_worker_token\", cfg.PushWorkerToken)\n\t\tgorillaContext.Set(r, \"push_enabled\", cfg.PushEnabled)\n\t\thfn.ServeHTTP(w, r)\n\n\t})\n}", "func Auth() func(http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tgivenAPIKey := r.Header.Get(\"API-KEY\")\n\n\t\t\t// Pass through unauthenticated requests.\n\t\t\tif givenAPIKey == \"\" {\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Inject admin user.\n\t\t\tif givenAPIKey == adminAPIKey {\n\t\t\t\tuser := &User{Name: \"Eric\", IsAdmin: true}\n\t\t\t\tctx := context.WithValue(r.Context(), userKey, user)\n\t\t\t\tr = r.WithContext(ctx)\n\t\t\t\tnext.ServeHTTP(w, r)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Inject non admin user.\n\t\t\tuser := &User{Name: \"Greg\", IsAdmin: false}\n\t\t\tctx := context.WithValue(r.Context(), userKey, user)\n\t\t\tr = r.WithContext(ctx)\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func (c *Client) Auth() (string, error) {\n\t// First do an empty get to get the auth challenge\n\treq, err := http.NewRequest(http.MethodGet, c.BaseURL+\"/v2/\", nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\trsp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed sending auth request: %w\", err)\n\t}\n\tdefer rsp.Body.Close()\n\tio.Copy(io.Discard, rsp.Body)\n\n\tif rsp.StatusCode == http.StatusOK {\n\t\t// no auth needed\n\t\treturn \"\", nil\n\t}\n\n\tif rsp.StatusCode != http.StatusUnauthorized {\n\t\treturn \"\", fmt.Errorf(\"unexpected status %s\", rsp.Status)\n\t}\n\n\t// The Www-Authenticate header tells us where to go to get a token\n\tvals, err := parseWWWAuthenticate(rsp.Header.Get(\"Www-Authenticate\"))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tu, err := url.Parse(vals[\"realm\"])\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"could not parse authentication realm: %w\", err)\n\t}\n\tq := u.Query()\n\tq.Set(\"service\", vals[\"service\"])\n\tq.Set(\"scope\", \"repository:\"+c.Name+\":pull,push\")\n\tu.RawQuery = q.Encode()\n\n\tfmt.Printf(\"get %s\\n\", u)\n\n\treq, err = http.NewRequest(http.MethodGet, u.String(), nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treq.SetBasicAuth(c.User, c.Password)\n\n\trsp, err = http.DefaultClient.Do(req)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed sending auth request: %w\", err)\n\t}\n\tdefer rsp.Body.Close()\n\tif rsp.StatusCode != http.StatusOK {\n\t\treturn \"\", fmt.Errorf(\"unexpected status %s\", rsp.Status)\n\t}\n\tbody, err := io.ReadAll(rsp.Body)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"could not read auth response body: %w\", err)\n\t}\n\n\ttype token struct {\n\t\tToken string `json:\"token\"`\n\t}\n\tvar tok token\n\tif err := json.Unmarshal(body, &tok); err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to unmarshal token: %w\", err)\n\t}\n\n\treturn tok.Token, nil\n}", "func Auth() gin.HandlerFunc {\r\n\tif gin.Mode() == \"debug\" {\r\n\t\treturn func(c *gin.Context) { c.Next() }\r\n\t}\r\n\treturn func(c *gin.Context) {\r\n\t\tAccessKey := c.GetHeader(\"AccessKey\")\r\n\t\tif c.GetHeader(\"AccessKey\") == \"\" {\r\n\t\t\tAccessKey = c.GetHeader(\"Token\")\r\n\t\t}\r\n\r\n\t\tsession := sessions.Default(c)\r\n\t\tLoginUserID := session.Get(\"UserID\")\r\n\t\tIsLeader := session.Get(\"IsLeader\")\r\n\r\n\t\tfmt.Println(\"AccessKey: \", AccessKey)\r\n\t\tswitch AccessKey {\r\n\t\tcase \"\":\r\n\t\t\tif LoginUserID != nil {\r\n\t\t\t\tc.Set(\"UserID\", LoginUserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\t\t\t} else {\r\n\t\t\t\tsession := sessions.Default(c)\r\n\t\t\t\tUserID := session.Get(\"UserID\")\r\n\t\t\t\tIsLeader := session.Get(\"IsLeader\")\r\n\r\n\t\t\t\tfmt.Println(\"UserID, IsLeader\", UserID, IsLeader)\r\n\t\t\t\tif UserID == nil {\r\n\t\t\t\t\tc.JSON(http.StatusUnauthorized, gin.H{\"message\": \"Empty AccessKey Please authorize before requesting\"})\r\n\t\t\t\t\tc.Abort()\r\n\t\t\t\t}\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\r\n\t\t\t}\r\n\r\n\t\tdefault:\r\n\t\t\tUserID, IsLeader, err := utils.LoadAccessKey(AccessKey)\r\n\r\n\t\t\tif LoginUserID != nil {\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\t\t\t\tc.Next()\r\n\t\t\t} else {\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\tc.JSON(http.StatusUnauthorized, gin.H{\"message\": \"Please authorize before requesting\"})\r\n\t\t\t\t\tc.Abort()\r\n\t\t\t\t}\r\n\r\n\t\t\t\tc.Set(\"UserID\", UserID)\r\n\t\t\t\tc.Set(\"IsLeader\", IsLeader)\r\n\r\n\t\t\t\tc.Next()\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n}", "func NewServerInterceptor(authenticator Authenticator) rpc.ServerInterceptor {\n\treturn rpc.ServerInterceptorFunc(\n\t\tfunc(\n\t\t\tctx context.Context,\n\t\t\trequest interface{},\n\t\t\tserverInfo *rpc.ServerInfo,\n\t\t\tserverHandler rpc.ServerHandler,\n\t\t) (interface{}, error) {\n\t\t\tuser, ok := authenticate(ctx, authenticator)\n\t\t\tif ok {\n\t\t\t\tctx = context.WithValue(ctx, authContextKey{}, user)\n\t\t\t}\n\t\t\treturn serverHandler.Handle(ctx, request)\n\t\t},\n\t)\n}", "func clientInterceptor(ctx context.Context, method string, req, reply interface{}, cc *grpc.ClientConn, invoker grpc.UnaryInvoker, opts ...grpc.CallOption) error {\n log.Printf(\"Method Called: %s\", method)\n return invoker(ctx, method, req, reply, cc, opts...)\n}", "func (b *Basic) InjectAuth(req *http.Request) error {\n\tif _, _, ok := req.BasicAuth(); ok {\n\t\treturn errors.New(\"Basic Auth header already exists\")\n\t}\n\n\treq.SetBasicAuth(b.Username, b.Password)\n\n\treturn nil\n}", "func Auth(route string) gin.HandlerFunc {\n\n\treturn func(c *gin.Context) {\n\n\t\tvar auth JwtPayload\n\n\t\t//authenticate jwt\n\t\tsecret := config.Config().GetString(\"hmac_key\")\n\n\t\ttoken := strings.Split(c.GetHeader(\"Authorization\"), \" \")\n\t\tif (len(token) != 2) || (token[0] != \"JWT\") {\n\t\t\tabort(c, 401, msg.Get(\"7\")) //JWT authorization header malformed\n\t\t\treturn\n\t\t}\n\n\t\tvar e *msg.Message\n\t\tauth, e = JwtAuth(token[1], secret)\n\t\tif e != nil {\n\t\t\tabort(c, 401, *e)\n\t\t\treturn\n\t\t}\n\n\t\ta := ACL()\n\t\tg := Grant{RoleID: auth.RoleID, Route: route, Method: c.Request.Method}\n\t\tr, ok := a[g]\n\t\tif !ok {\n\t\t\tabort(c, 401, msg.Get(\"8\")) //Not enough permissions\n\t\t\treturn\n\t\t}\n\n\t\tnow := time.Now()\n\t\tif now.Before(r.From) || now.After(r.To) {\n\t\t\tabort(c, 401, msg.Get(\"9\")) //Role access expired or not yet valid\n\t\t\treturn\n\t\t}\n\n\t\tif tps.IsEnabled() && tps.Transaction(auth.UserID, auth.TPS) != nil {\n\t\t\tabort(c, 401, msg.Get(\"10\")) //TPS limit exceeded\n\t\t\treturn\n\t\t}\n\n\t\tif _, ok := DeletedUsersMap[auth.UserID]; ok {\n\t\t\tabort(c, 401, msg.Get(\"11\")) //Unauthorized\n\t\t\treturn\n\t\t}\n\n\t\tc.Set(\"Auth\", auth)\n\t\tc.Next()\n\t}\n}", "func New(config *Config, next http.RoundTripper) http.RoundTripper {\n\t// Need to delay fetching auth settings until env vars have had a chance to propagate\n\tif authSettings == nil {\n\t\tauthSettings = awsds.ReadAuthSettingsFromEnvironmentVariables()\n\t}\n\n\treturn RoundTripperFunc(func(r *http.Request) (*http.Response, error) {\n\t\tif next == nil {\n\t\t\tnext = http.DefaultTransport\n\t\t}\n\t\treturn (&middleware{\n\t\t\tconfig: config,\n\t\t\tnext: next,\n\t\t}).exec(r)\n\t})\n}", "func AuthorizedAPIClientFor(username string, password string) *baloo.Client {\n\tjwt, err := Authenticate(username, password)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\treturn baloo.New(constants.API_URL).\n\t\tUseRequest(AuthMiddleware(jwt))\n}", "func GetLoginFunc(db *sqlx.DB) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusername := \"\"\n\t\tpassword := \"\"\n\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error reading body: \", err.Error())\n\t\t\thttp.Error(w, \"Error reading body: \"+err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tvar lj loginJson\n\t\tlog.Println(body)\n\t\terr = json.Unmarshal(body, &lj)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error unmarshalling JSON: \", err.Error())\n\t\t\thttp.Error(w, \"Invalid JSON: \"+err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tusername = lj.U\n\t\tpassword = lj.P\n\t\tuserInterface, err := api.GetUser(username, db)\n\t\tif err != nil {\n\t\t\thttp.Error(w, \"Invalid user: \"+err.Error(), http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\t\tu, ok := userInterface.(api.Users)\n\t\tif !ok {\n\t\t\thttp.Error(w, \"Error GetUser returned a non-user.\", http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tencBytes := sha1.Sum([]byte(password))\n\t\tencString := hex.EncodeToString(encBytes[:])\n\t\tif err != nil {\n\t\t\tctx.Set(r, \"user\", nil)\n\t\t\tlog.Println(\"Invalid password\")\n\t\t\thttp.Error(w, \"Invalid password: \"+err.Error(), http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\t\tif u.LocalPassword.String != encString {\n\t\t\tctx.Set(r, \"user\", nil)\n\t\t\tlog.Println(\"Invalid password\")\n\t\t\thttp.Error(w, \"Invalid password\", http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\t// Create the token\n\t\ttoken := jwt.New(jwt.SigningMethodHS256)\n\t\t// Set some claims\n\t\ttoken.Claims[\"userid\"] = u.Username\n\t\ttoken.Claims[\"role\"] = u.Links.RolesLink.ID\n\t\ttoken.Claims[\"exp\"] = time.Now().Add(time.Hour * 72).Unix()\n\t\t// Sign and get the complete encoded token as a string\n\t\ttokenString, err := token.SignedString([]byte(\"mySigningKey\")) // TODO JvD\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tjs, err := json.Marshal(TokenResponse{Token: tokenString})\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.Write(js)\n\t}\n}", "func AuthFunc() types.AuthFunc {\n\treturn globalBusiness.AuthFunc\n}", "func Handler(handler http.HandlerFunc, security *Config) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusernameEntered, passwordEntered, ok := r.BasicAuth()\n\t\tif !ok || usernameEntered != security.Basic.Username || Sha512(passwordEntered) != strings.ToLower(security.Basic.PasswordSha512Hash) {\n\t\t\tw.Header().Set(\"WWW-Authenticate\", \"Basic\")\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\t_, _ = w.Write([]byte(\"Unauthorized\"))\n\t\t\treturn\n\t\t}\n\t\thandler(w, r)\n\t}\n}", "func makeAuthInfoWriterCustom(cmd *cobra.Command) (runtime.ClientAuthInfoWriter, error) {\n\tauths := []runtime.ClientAuthInfoWriter{}\n\t/*x-auth-token User authorization token*/\n\tif viper.IsSet(\"uisp.x-auth-token\") {\n\t\tXAuthTokenKey := viper.GetString(\"uisp.x-auth-token\")\n\t\tauths = append(auths, httptransport.APIKeyAuth(\"x-auth-token\", \"header\", XAuthTokenKey))\n\t}\n\tif len(auths) == 0 {\n\t\tlogDebugf(\"Warning: No auth params detected.\")\n\t\treturn nil, nil\n\t}\n\t// compose all auths together\n\treturn httptransport.Compose(auths...), nil\n}", "func basicAuthHandler(handler func(string, string) http.HandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tusername, password, ok := r.BasicAuth()\n\n\t\tif !ok {\n\t\t\tw.Header().Add(\"WWW-Authenticate\", `Basic realm=\"cwmp-proxy\"`)\n\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\n\t\thandler(username, password).ServeHTTP(w, r)\n\t}\n}", "func (proto *Protocol) AUTHENTICATE(command *Command) {\n\targs := strings.Split(command.args, \" \")\n\n\tif len(args) < 1 {\n\t\t// FIXME what error?\n\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\treturn\n\t}\n\n\tswitch strings.ToUpper(args[0]) {\n\tcase \"PLAIN\":\n\t\tif len(args) > 1 {\n\t\t\t// Do auth now\n\t\t\treturn\n\t\t}\n\t\tproto.readIntercept = func(l string) {\n\t\t\tproto.readIntercept = nil\n\t\t\tuser, pass, err := auth.DecodePLAIN(l)\n\t\t\tif err != nil {\n\t\t\t\t// FIXME what error?\n\t\t\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif ok := proto.ValidateAuthenticationHandler(\"PLAIN\", user, pass); !ok {\n\t\t\t\t// FIXME what error?\n\t\t\t\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n\t\t\t\treturn\n\t\t\t}\n\t\t\t// OK\n\t\t\tproto.State = AUTH\n\t\t\tproto.Responses <- &Response{command.tag, Status(ResponseOK), nil, \"\", nil}\n\t\t\treturn\n\t\t}\n\t\tproto.Responses <- &Response{\"+\", nil, nil, \"\", nil}\n\t\treturn\n\t}\n\n\t// FIXME what error?\n\tproto.Responses <- ResponseUnrecognisedCommand(command.tag)\n}", "func NewProvider(username string, password string) *Provider {\n transport := httptransport.New(\"api.autodns.com\", \"/v1\", []string{\"https\"})\n transport.DefaultAuthentication = httptransport.BasicAuth(username, password)\n\n formats := strfmt.Default\n\n return &Provider{\n username: username,\n password: password,\n client: zone_tasks.New(transport, formats),\n }\n}", "func WrapAuthorize(hfn http.Handler, routeName string) http.HandlerFunc {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\turlValues := r.URL.Query()\n\n\t\trefStr := gorillaContext.Get(r, \"str\").(stores.Store)\n\t\trefRoles := gorillaContext.Get(r, \"auth_roles\").([]string)\n\t\tserviceToken := gorillaContext.Get(r, \"auth_service_token\").(string)\n\n\t\t// Check first if service token is used\n\t\tif serviceToken != \"\" && serviceToken == urlValues.Get(\"key\") {\n\t\t\thfn.ServeHTTP(w, r)\n\t\t\treturn\n\t\t}\n\n\t\tif auth.Authorize(routeName, refRoles, refStr) {\n\t\t\thfn.ServeHTTP(w, r)\n\t\t} else {\n\t\t\terr := APIErrorForbidden()\n\t\t\trespondErr(w, err)\n\t\t}\n\t})\n}", "func StreamAuth() grpc.StreamServerInterceptor {\n\n\treturn func(srv interface{}, ss grpc.ServerStream, info *grpc.StreamServerInfo, handler grpc.StreamHandler) error {\n\t\tcred := ss.Context().Value(authKey)\n\t\tif cred == nil {\n\t\t\treturn errors.New(\"Not authorized to make this call!\")\n\t\t}\n\n\t\treturn handler(srv, ss)\n\t}\n}", "func (a *App) Auth() negroni.HandlerFunc {\n\treturn negroni.HandlerFunc(func(w http.ResponseWriter, req *http.Request, next http.HandlerFunc) {\n\t\tdb, ok := context.Get(req, \"db\").(*mgo.Database)\n\t\tif !ok {\n\t\t\ta.R.JSON(w, http.StatusInternalServerError, &Response{Status: \"Error\", Message: \"Internal server error\"})\n\t\t\treturn\n\t\t}\n\t\tauthHeader := req.Header.Get(\"Authorization\")\n\t\tif authHeader == \"\" {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tdata, err := base64.StdEncoding.DecodeString(strings.Replace(authHeader, \"Basic \", \"\", 1))\n\t\tif err != nil {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tuser := &User{}\n\t\tparts := strings.Split(string(data), \":\")\n\t\tif len(parts) < 2 {\n\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\treturn\n\t\t}\n\t\tif parts[0] == parts[1] {\n\t\t\tshaHash := sha256.New()\n\t\t\tif _, err := shaHash.Write([]byte(parts[0])); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t\ttoken := base64.StdEncoding.EncodeToString(shaHash.Sum(nil))\n\t\t\tif err := db.C(\"users\").Find(bson.M{\n\t\t\t\t\"services.resume.loginTokens\": bson.M{\"$elemMatch\": bson.M{\"hashedToken\": token}},\n\t\t\t}).One(&user); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t} else {\n\t\t\tin := []bson.M{bson.M{\"address\": parts[0], \"verified\": false}}\n\t\t\tif err := db.C(\"users\").Find(bson.M{\"emails\": bson.M{\"$in\": in}}).One(&user); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t\tshaHash := sha256.New()\n\t\t\tif _, err := shaHash.Write([]byte(parts[1])); err != nil {\n\t\t\t\thttp.Error(w, \"Not Authorized\", http.StatusUnauthorized)\n\t\t\t\treturn\n\t\t\t}\n\t\t\th := hex.EncodeToString(shaHash.Sum(nil))\n\t\t\tif err := bcrypt.CompareHashAndPassword([]byte(user.Services.Password.Bcrypt), []byte(h)); err != nil {\n\t\t\t\ta.R.JSON(w, http.StatusUnauthorized, &Response{Status: \"Error\", Message: \"Not Authorized\"})\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tcontext.Set(req, \"user\", user)\n\t\tnext(w, req)\n\t})\n}", "func (srv *targetServiceHandler) auth(h http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, req *http.Request) {\n\t\tctx := httpbakery.ContextWithRequest(context.TODO(), req)\n\t\tops, err := opsForRequest(req)\n\t\tif err != nil {\n\t\t\tfail(w, http.StatusInternalServerError, \"%v\", err)\n\t\t\treturn\n\t\t}\n\t\tauthChecker := srv.checker.Auth(httpbakery.RequestMacaroons(req)...)\n\t\tif _, err = authChecker.Allow(ctx, ops...); err != nil {\n\t\t\thttpbakery.WriteError(ctx, w, srv.oven.Error(ctx, req, err))\n\t\t\treturn\n\t\t}\n\t\th.ServeHTTP(w, req)\n\t})\n}", "func auth(conn net.Conn, username string, password string) error {\n\tbufConn := bufio.NewReader(conn)\n\n\t// Read the version byte\n\tversion := []byte{0}\n\tif _, err := bufConn.Read(version); err != nil {\n\t\terr := fmt.Errorf(\"[AUTH] socks: Failed to get version byte: %v\", err)\n\t\treturn err\n\t}\n\n\t// Ensure we are compatible\n\tif version[0] != socks5Version {\n\t\terr := fmt.Errorf(\"[AUTH] Unsupported SOCKS version: %v\", version)\n\t\treturn err\n\t}\n\n\t// Get the methods\n\tmethods, err := readMethods(bufConn)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"[AUTH] Failed to get auth methods: %v\", err)\n\t}\n\n\t// Select a usable method (only auth for us here)\n\tfor _, method := range methods {\n\t\tif method == UserPassAuth {\n\t\t\t// Tell the client to use user/pass auth\n\t\t\tif _, err := conn.Write([]byte{socks5Version, UserPassAuth}); err != nil {\n\t\t\t\treturn fmt.Errorf(\"[AUTH] Can't write method reply: %v\", err)\n\t\t\t}\n\n\t\t\t//read username/password\n\t\t\tu, p, err := getauthdata(bufConn, conn)\n\t\t\tif (err != nil) || (u != username || p != password) {\n\t\t\t\tconn.Write([]byte{userAuthVersion, authFailure})\n\t\t\t\treturn fmt.Errorf(\"[AUTH] Username/password auth failed: %v\", err)\n\t\t\t}\n\n\t\t\t//correct auth\n\t\t\tconn.Write([]byte{userAuthVersion, authSuccess})\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// No usable method found\n\tconn.Write([]byte{socks5Version, noAcceptable})\n\treturn fmt.Errorf(\"[AUTH] No acceptable auth method\")\n}", "func (a *AuthCommand) Initialize(app *kingpin.Application, config *service.Config) {\n\ta.config = config\n\n\t// operations with authorities\n\tauth := app.Command(\"auth\", \"Operations with user and host certificate authorities (CAs)\").Hidden()\n\ta.authExport = auth.Command(\"export\", \"Export public cluster (CA) keys to stdout\")\n\ta.authExport.Flag(\"keys\", \"if set, will print private keys\").BoolVar(&a.exportPrivateKeys)\n\ta.authExport.Flag(\"fingerprint\", \"filter authority by fingerprint\").StringVar(&a.exportAuthorityFingerprint)\n\ta.authExport.Flag(\"compat\", \"export cerfiticates compatible with specific version of Teleport\").StringVar(&a.compatVersion)\n\ta.authExport.Flag(\"type\", \"certificate type: 'user', 'host' or 'tls'\").StringVar(&a.authType)\n\n\ta.authGenerate = auth.Command(\"gen\", \"Generate a new SSH keypair\").Hidden()\n\ta.authGenerate.Flag(\"pub-key\", \"path to the public key\").Required().StringVar(&a.genPubPath)\n\ta.authGenerate.Flag(\"priv-key\", \"path to the private key\").Required().StringVar(&a.genPrivPath)\n\n\ta.authSign = auth.Command(\"sign\", \"Create an identity file(s) for a given user\")\n\ta.authSign.Flag(\"user\", \"Teleport user name\").StringVar(&a.genUser)\n\ta.authSign.Flag(\"host\", \"Teleport host name\").StringVar(&a.genHost)\n\ta.authSign.Flag(\"out\", \"identity output\").Short('o').StringVar(&a.output)\n\ta.authSign.Flag(\"format\", \"identity format: 'file' (default) or 'dir'\").Default(string(client.DefaultIdentityFormat)).StringVar((*string)(&a.outputFormat))\n\ta.authSign.Flag(\"ttl\", \"TTL (time to live) for the generated certificate\").Default(fmt.Sprintf(\"%v\", defaults.CertDuration)).DurationVar(&a.genTTL)\n\ta.authSign.Flag(\"compat\", \"OpenSSH compatibility flag\").StringVar(&a.compatibility)\n}", "func AuthLoginWrapper(ctx *fasthttp.RequestCtx, mgoClient *mgo.Session, redisClient *redis.Client, cfg datastructures.Configuration) {\n\tlog.Info(\"AuthLoginWrapper | Starting authentication | Parsing authentication credentials\")\n\tctx.Response.Header.SetContentType(\"application/json; charset=utf-8\")\n\tusername, password := ParseAuthenticationCoreHTTP(ctx) // Retrieve the username and password encoded in the request from BasicAuth headers, GET & POST\n\tif authutils.ValidateCredentials(username, password) { // Verify if the input parameter respect the rules ...\n\t\tlog.Debug(\"AuthLoginWrapper | Input validated | User: \", username, \" | Pass: \", password, \" | Calling core functionalities ...\")\n\t\tcheck := authutils.LoginUserHTTPCore(username, password, mgoClient, cfg.Mongo.Users.DB, cfg.Mongo.Users.Collection) // Login phase\n\t\tif strings.Compare(check, \"OK\") == 0 { // Login Succeed\n\t\t\tlog.Debug(\"AuthLoginWrapper | Login succesfully! Generating token!\")\n\t\t\ttoken := basiccrypt.GenerateToken(username, password) // Generate a simple md5 hashed token\n\t\t\tlog.Info(\"AuthLoginWrapper | Inserting token into Redis \", token)\n\t\t\tbasicredis.InsertIntoClient(redisClient, username, token, cfg.Redis.Token.Expire) // insert the token into the DB\n\t\t\tlog.Info(\"AuthLoginWrapper | Token inserted! All operation finished correctly! | Setting token into response\")\n\t\t\tauthcookie := authutils.CreateCookie(\"GoLog-Token\", token, cfg.Redis.Token.Expire)\n\t\t\tctx.Response.Header.SetCookie(authcookie) // Set the token into the cookie headers\n\t\t\tctx.Response.Header.Set(\"GoLog-Token\", token) // Set the token into a custom headers for future security improvments\n\t\t\tlog.Warn(\"AuthLoginWrapper | Client logged in succesfully!! | \", username, \":\", password, \" | Token: \", token)\n\t\t\terr := json.NewEncoder(ctx).Encode(datastructures.Response{Status: true, Description: \"User logged in!\", ErrorCode: username + \":\" + password, Data: token})\n\t\t\tcommonutils.Check(err, \"AuthLoginWrapper\")\n\t\t} else {\n\t\t\tcommonutils.AuthLoginWrapperErrorHelper(ctx, check, username, password)\n\t\t}\n\t} else { // error parsing credential\n\t\tlog.Info(\"AuthLoginWrapper | Error parsing credential!! |\", username+\":\"+password)\n\t\tctx.Response.Header.DelCookie(\"GoLog-Token\")\n\t\tctx.Error(fasthttp.StatusMessage(fasthttp.StatusUnauthorized), fasthttp.StatusUnauthorized)\n\t\tctx.Response.Header.Set(\"WWW-Authenticate\", \"Basic realm=Restricted\")\n\t\t//err := json.NewEncoder(ctx).Encode(datastructures.Response{Status: false, Description: \"Error parsing credential\", ErrorCode: \"Missing or manipulated input\", Data: nil})\n\t\t//commonutils.Check(err, \"AuthLoginWrapper\")\n\t}\n}", "func Auth() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\trawCallerID := mlauth.GetCaller(c.Request)\n\t\tisAdmin := mlauth.IsCallerAdmin(c.Request)\n\n\t\tcallerID, err := strconv.ParseUint(rawCallerID, 10, 64)\n\n\t\t// If request is not from an admin, and we failed parsing caller ID, fail\n\t\tif !isAdmin && err != nil {\n\t\t\terrors.ReturnError(c, &errors.Error{\n\t\t\t\tCode: errors.BadRequestApiError,\n\t\t\t\tCause: \"parsing header value\",\n\t\t\t\tMessage: \"invalid caller.id\",\n\t\t\t\tValues: map[string]string{\n\t\t\t\t\t\"caller.id\": rawCallerID,\n\t\t\t\t},\n\t\t\t})\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tc.Set(\"callerID\", callerID)\n\t\tc.Set(\"isAdmin\", isAdmin)\n\t\tc.Next()\n\t}\n}", "func (a *AbstractNetworkConnectionHandler) OnAuthPassword(_ string, _ []byte) (response AuthResponse, reason error) {\n\treturn AuthResponseUnavailable, nil\n}", "func (s *HTTPServer) wrap(handler endpoint, methods []string) http.HandlerFunc {\n\treturn func(resp http.ResponseWriter, req *http.Request) {\n\t\tsetHeaders(resp, s.agent.config.HTTPResponseHeaders)\n\t\tsetTranslateAddr(resp, s.agent.config.TranslateWANAddrs)\n\n\t\t// Obfuscate any tokens from appearing in the logs\n\t\tformVals, err := url.ParseQuery(req.URL.RawQuery)\n\t\tif err != nil {\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Failed to decode query: %s from=%s\", err, req.RemoteAddr)\n\t\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tlogURL := req.URL.String()\n\t\tif tokens, ok := formVals[\"token\"]; ok {\n\t\t\tfor _, token := range tokens {\n\t\t\t\tif token == \"\" {\n\t\t\t\t\tlogURL += \"<hidden>\"\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tlogURL = strings.Replace(logURL, token, \"<hidden>\", -1)\n\t\t\t}\n\t\t}\n\t\tlogURL = aclEndpointRE.ReplaceAllString(logURL, \"$1<hidden>$4\")\n\n\t\tif s.blacklist.Block(req.URL.Path) {\n\t\t\terrMsg := \"Endpoint is blocked by agent configuration\"\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Request %s %v, error: %v from=%s\", req.Method, logURL, err, req.RemoteAddr)\n\t\t\tresp.WriteHeader(http.StatusForbidden)\n\t\t\tfmt.Fprint(resp, errMsg)\n\t\t\treturn\n\t\t}\n\n\t\tisForbidden := func(err error) bool {\n\t\t\tif acl.IsErrPermissionDenied(err) || acl.IsErrNotFound(err) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\t_, ok := err.(ForbiddenError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisMethodNotAllowed := func(err error) bool {\n\t\t\t_, ok := err.(MethodNotAllowedError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisBadRequest := func(err error) bool {\n\t\t\t_, ok := err.(BadRequestError)\n\t\t\treturn ok\n\t\t}\n\n\t\tisTooManyRequests := func(err error) bool {\n\t\t\t// Sadness net/rpc can't do nice typed errors so this is all we got\n\t\t\treturn err.Error() == consul.ErrRateLimited.Error()\n\t\t}\n\n\t\taddAllowHeader := func(methods []string) {\n\t\t\tresp.Header().Add(\"Allow\", strings.Join(methods, \",\"))\n\t\t}\n\n\t\thandleErr := func(err error) {\n\t\t\ts.agent.logger.Printf(\"[ERR] http: Request %s %v, error: %v from=%s\", req.Method, logURL, err, req.RemoteAddr)\n\t\t\tswitch {\n\t\t\tcase isForbidden(err):\n\t\t\t\tresp.WriteHeader(http.StatusForbidden)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase structs.IsErrRPCRateExceeded(err):\n\t\t\t\tresp.WriteHeader(http.StatusTooManyRequests)\n\t\t\tcase isMethodNotAllowed(err):\n\t\t\t\t// RFC2616 states that for 405 Method Not Allowed the response\n\t\t\t\t// MUST include an Allow header containing the list of valid\n\t\t\t\t// methods for the requested resource.\n\t\t\t\t// https://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html\n\t\t\t\taddAllowHeader(err.(MethodNotAllowedError).Allow)\n\t\t\t\tresp.WriteHeader(http.StatusMethodNotAllowed) // 405\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase isBadRequest(err):\n\t\t\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tcase isTooManyRequests(err):\n\t\t\t\tresp.WriteHeader(http.StatusTooManyRequests)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\tdefault:\n\t\t\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\t\t\tfmt.Fprint(resp, err.Error())\n\t\t\t}\n\t\t}\n\n\t\tstart := time.Now()\n\t\tdefer func() {\n\t\t\ts.agent.logger.Printf(\"[DEBUG] http: Request %s %v (%v) from=%s\", req.Method, logURL, time.Since(start), req.RemoteAddr)\n\t\t}()\n\n\t\tvar obj interface{}\n\n\t\t// if this endpoint has declared methods, respond appropriately to OPTIONS requests. Otherwise let the endpoint handle that.\n\t\tif req.Method == \"OPTIONS\" && len(methods) > 0 {\n\t\t\taddAllowHeader(append([]string{\"OPTIONS\"}, methods...))\n\t\t\treturn\n\t\t}\n\n\t\t// if this endpoint has declared methods, check the request method. Otherwise let the endpoint handle that.\n\t\tmethodFound := len(methods) == 0\n\t\tfor _, method := range methods {\n\t\t\tif method == req.Method {\n\t\t\t\tmethodFound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif !methodFound {\n\t\t\terr = MethodNotAllowedError{req.Method, append([]string{\"OPTIONS\"}, methods...)}\n\t\t} else {\n\t\t\terr = s.checkWriteAccess(req)\n\n\t\t\tif err == nil {\n\t\t\t\t// Invoke the handler\n\t\t\t\tobj, err = handler(resp, req)\n\t\t\t}\n\t\t}\n\t\tcontentType := \"application/json\"\n\t\thttpCode := http.StatusOK\n\t\tif err != nil {\n\t\t\tif errPayload, ok := err.(CodeWithPayloadError); ok {\n\t\t\t\thttpCode = errPayload.StatusCode\n\t\t\t\tif errPayload.ContentType != \"\" {\n\t\t\t\t\tcontentType = errPayload.ContentType\n\t\t\t\t}\n\t\t\t\tif errPayload.Reason != \"\" {\n\t\t\t\t\tresp.Header().Add(\"X-Consul-Reason\", errPayload.Reason)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\thandleErr(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif obj == nil {\n\t\t\treturn\n\t\t}\n\t\tvar buf []byte\n\t\tif contentType == \"application/json\" {\n\t\t\tbuf, err = s.marshalJSON(req, obj)\n\t\t\tif err != nil {\n\t\t\t\thandleErr(err)\n\t\t\t\treturn\n\t\t\t}\n\t\t} else {\n\t\t\tif strings.HasPrefix(contentType, \"text/\") {\n\t\t\t\tif val, ok := obj.(string); ok {\n\t\t\t\t\tbuf = []byte(val)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tresp.Header().Set(\"Content-Type\", contentType)\n\t\tresp.WriteHeader(httpCode)\n\t\tresp.Write(buf)\n\t}\n}", "func BasicAuth(username, password string) Credentials {\n\treturn &basicAuth{username, password}\n}", "func AuthMiddlewareHandlerFunc(authenticator *Authenticator, handleFunc http.HandlerFunc) func(w http.ResponseWriter, r *http.Request) {\n\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\tuser := authenticator.AuthenticateHttpRequest(r)\n\t\tif user != nil {\n\t\t\thandleFunc(w, r.WithContext(context.WithValue(r.Context(), \"User\", user)))\n\t\t} else {\n\t\t\tw.Header().Add(\"WWW-Authenticate\", \"Basic realm=restricted\")\n\t\t\thttp.Error(w, http.StatusText(http.StatusUnauthorized),\n\t\t\t\thttp.StatusUnauthorized)\n\t\t}\n\n\t}\n\n}", "func (t *OpetCode) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n fmt.Printf(\"Invoke function\\n\")\n // Retrieve the requested Smart Contract function and arguments\n function, args := APIstub.GetFunctionAndParameters()\n // Route to the appropriate handler function to interact with the ledger appropriately\n if function == \"initLedger\" {\n return t.initLedger(APIstub)\n } else if function == \"createUser\" {\n return t.createUser(APIstub, args)\n } else if function == \"retrieveUser\" {\n return t.retrieveUser(APIstub, args)\n } else if function == \"createDocument\" {\n return t.createDocument(APIstub, args)\n } else if function == \"retrieveDocument\" {\n return t.retrieveDocument(APIstub, args)\n }\n return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func Auth() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tc.Writer.Write([]byte(\"auth required for endpoint\"))\n\t}\n}", "func NewGetAuth() context.Handler {\n\treturn func(ctx context.Context) {\n\t\tname := ctx.Params().Get(\"name\")\n\t\ttoken := ctx.GetHeader(\"token\")\n\t\tif token == \"\" || getClientToken(name) != token {\n\t\t\tctx.StatusCode(403)\n\t\t\tctx.StopExecution()\n\t\t\treturn\n\t\t}\n\t\tctx.Next()\n\t}\n}", "func NewMiddleware(creds *secrets.BasicAuthCredentials) es.CommandHandlerMiddleware {\n\treturn func(handler es.CommandHandler) es.CommandHandler {\n\t\treturn es.CommandHandlerFunc(func(ctx context.Context, cmd es.Command) error {\n\t\t\tcur, err := secrets.AuthFromContext(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif !creds.Equals(cur) {\n\t\t\t\treturn ErrInvalidCredentials\n\t\t\t}\n\n\t\t\treturn handler.HandleCommand(ctx, cmd)\n\t\t})\n\t}\n}", "func AuthFromConfig(ctx context.Context, c *Config, client func(string) *http.Client) (*Auth, error) {\n\tauthenticator, err := jwtauth.AuthFromConfig(ctx, &c.Config, client)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Auth{\n\t\tHeaders: c.Headers,\n\t\tAuthenticator: authenticator,\n\t\tUnauthHandler: DefaultUnauthHandler,\n\t}, nil\n}", "func newAuthenticator(secret string) *authenticator {\n\treturn &authenticator{\n\t\tSecret: secret,\n\t}\n}", "func (a *AuthCommand) Initialize(app *kingpin.Application, config *servicecfg.Config) {\n\ta.config = config\n\t// operations with authorities\n\tauth := app.Command(\"auth\", \"Operations with user and host certificate authorities (CAs).\").Hidden()\n\ta.authExport = auth.Command(\"export\", \"Export public cluster (CA) keys to stdout.\")\n\ta.authExport.Flag(\"keys\", \"if set, will print private keys\").BoolVar(&a.exportPrivateKeys)\n\ta.authExport.Flag(\"fingerprint\", \"filter authority by fingerprint\").StringVar(&a.exportAuthorityFingerprint)\n\ta.authExport.Flag(\"compat\", \"export certificates compatible with specific version of Teleport\").StringVar(&a.compatVersion)\n\ta.authExport.Flag(\"type\",\n\t\tfmt.Sprintf(\"export certificate type (%v)\", strings.Join(allowedCertificateTypes, \", \"))).\n\t\tEnumVar(&a.authType, allowedCertificateTypes...)\n\n\ta.authGenerate = auth.Command(\"gen\", \"Generate a new SSH keypair.\").Hidden()\n\ta.authGenerate.Flag(\"pub-key\", \"path to the public key\").Required().StringVar(&a.genPubPath)\n\ta.authGenerate.Flag(\"priv-key\", \"path to the private key\").Required().StringVar(&a.genPrivPath)\n\n\ta.authSign = auth.Command(\"sign\", \"Create an identity file(s) for a given user.\")\n\ta.authSign.Flag(\"user\", \"Teleport user name\").StringVar(&a.genUser)\n\ta.authSign.Flag(\"host\", \"Teleport host name\").StringVar(&a.genHost)\n\ta.authSign.Flag(\"out\", \"Identity output\").Short('o').Required().StringVar(&a.output)\n\ta.authSign.Flag(\"format\",\n\t\tfmt.Sprintf(\"Identity format: %s. %s is the default.\",\n\t\t\tidentityfile.KnownFileFormats.String(), identityfile.DefaultFormat)).\n\t\tDefault(string(identityfile.DefaultFormat)).\n\t\tStringVar((*string)(&a.outputFormat))\n\ta.authSign.Flag(\"ttl\", \"TTL (time to live) for the generated certificate.\").\n\t\tDefault(fmt.Sprintf(\"%v\", apidefaults.CertDuration)).\n\t\tDurationVar(&a.genTTL)\n\ta.authSign.Flag(\"compat\", \"OpenSSH compatibility flag\").StringVar(&a.compatibility)\n\ta.authSign.Flag(\"proxy\", `Address of the Teleport proxy. When --format is set to \"kubernetes\", this address will be set as cluster address in the generated kubeconfig file`).StringVar(&a.proxyAddr)\n\ta.authSign.Flag(\"overwrite\", \"Whether to overwrite existing destination files. When not set, user will be prompted before overwriting any existing file.\").BoolVar(&a.signOverwrite)\n\ta.authSign.Flag(\"tar\", \"Create a tarball of the resulting certificates and stream to stdout.\").BoolVar(&a.streamTarfile)\n\t// --kube-cluster was an unfortunately chosen flag name, before teleport\n\t// supported kubernetes_service and registered kubernetes clusters that are\n\t// not trusted teleport clusters.\n\t// It's kept as an alias for --leaf-cluster for backwards-compatibility,\n\t// but hidden.\n\ta.authSign.Flag(\"kube-cluster\", `Leaf cluster to generate identity file for when --format is set to \"kubernetes\"`).Hidden().StringVar(&a.leafCluster)\n\ta.authSign.Flag(\"leaf-cluster\", `Leaf cluster to generate identity file for when --format is set to \"kubernetes\"`).StringVar(&a.leafCluster)\n\ta.authSign.Flag(\"kube-cluster-name\", `Kubernetes cluster to generate identity file for when --format is set to \"kubernetes\"`).StringVar(&a.kubeCluster)\n\ta.authSign.Flag(\"app-name\", `Application to generate identity file for. Mutually exclusive with \"--db-service\".`).StringVar(&a.appName)\n\ta.authSign.Flag(\"db-service\", `Database to generate identity file for. Mutually exclusive with \"--app-name\".`).StringVar(&a.dbService)\n\ta.authSign.Flag(\"db-user\", `Database user placed on the identity file. Only used when \"--db-service\" is set.`).StringVar(&a.dbUser)\n\ta.authSign.Flag(\"db-name\", `Database name placed on the identity file. Only used when \"--db-service\" is set.`).StringVar(&a.dbName)\n\ta.authSign.Flag(\"windows-user\", `Window user placed on the identity file. Only used when --format is set to \"windows\"`).StringVar(&a.windowsUser)\n\ta.authSign.Flag(\"windows-domain\", `Active Directory domain for which this cert is valid. Only used when --format is set to \"windows\"`).StringVar(&a.windowsDomain)\n\ta.authSign.Flag(\"windows-sid\", `Optional Security Identifier to embed in the certificate. Only used when --format is set to \"windows\"`).StringVar(&a.windowsSID)\n\n\ta.authRotate = auth.Command(\"rotate\", \"Rotate certificate authorities in the cluster.\")\n\ta.authRotate.Flag(\"grace-period\", \"Grace period keeps previous certificate authorities signatures valid, if set to 0 will force users to re-login and nodes to re-register.\").\n\t\tDefault(fmt.Sprintf(\"%v\", defaults.RotationGracePeriod)).\n\t\tDurationVar(&a.rotateGracePeriod)\n\ta.authRotate.Flag(\"manual\", \"Activate manual rotation , set rotation phases manually\").BoolVar(&a.rotateManualMode)\n\ta.authRotate.Flag(\"type\", fmt.Sprintf(\"Certificate authority to rotate, one of: %s\", strings.Join(getCertAuthTypes(), \", \"))).Required().EnumVar(&a.rotateType, getCertAuthTypes()...)\n\ta.authRotate.Flag(\"phase\", fmt.Sprintf(\"Target rotation phase to set, used in manual rotation, one of: %v\", strings.Join(types.RotatePhases, \", \"))).StringVar(&a.rotateTargetPhase)\n\n\ta.authLS = auth.Command(\"ls\", \"List connected auth servers.\")\n\ta.authLS.Flag(\"format\", \"Output format: 'yaml', 'json' or 'text'\").Default(teleport.YAML).StringVar(&a.format)\n\n\ta.authCRL = auth.Command(\"crl\", \"Export empty certificate revocation list (CRL) for certificate authorities.\")\n\ta.authCRL.Flag(\"type\", fmt.Sprintf(\"Certificate authority type, one of: %s\", strings.Join(allowedCRLCertificateTypes, \", \"))).Required().EnumVar(&a.caType, allowedCRLCertificateTypes...)\n}", "func requireBasicAuth(\n\tcheckFunc func(context.Context, *http.Request, string, string, string) error,\n\toptional bool,\n) func(next http.Handler) http.Handler {\n\treturn func(next http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tauthenticated := false\n\n\t\t\tif checkFunc != nil {\n\t\t\t\taps := chi.URLParam(r, apsParamName)\n\t\t\t\tusername, password, _ := r.BasicAuth()\n\t\t\t\tif err := checkFunc(r.Context(), r, aps, username, password); err == nil {\n\t\t\t\t\tauthenticated = true\n\t\t\t\t}\n\t\t\t} else if optional {\n\t\t\t\tauthenticated = true\n\t\t\t}\n\n\t\t\tif !authenticated {\n\t\t\t\treqHost := r.Host\n\t\t\t\tif host, _, err := net.SplitHostPort(reqHost); err == nil {\n\t\t\t\t\treqHost = host\n\t\t\t\t}\n\n\t\t\t\tw.Header().Set(wwwAuthenticateHeader, fmt.Sprintf(`Basic realm=\"estserver@%s\"`,\n\t\t\t\t\turl.QueryEscape(reqHost)))\n\t\t\t\terrAuthRequired.Write(w)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext.ServeHTTP(w, r)\n\t\t})\n\t}\n}", "func NewAuthInterceptor(jwtManager *skyutl.JwtManager) *AuthInterceptor {\n\treturn &AuthInterceptor{\n\t\tjwtManager: jwtManager,\n\t}\n}", "func New(config Config, ds *datastore.Datastore) *Authenticator {\n\te := gin.Default()\n\tconfig.AccessDuration = config.AccessDuration * time.Minute\n\treturn &Authenticator{\n\t\tconfig: config,\n\t\tds: ds,\n\t\tengine: e,\n\t}\n}", "func serverAuth(logger *zap.Logger, url string, user string, password string) (string, error) {\n\tlogger.Debug(\"serverAuth called\", zap.String(\"url\", url), zap.String(\"user\", user))\n\tbody := fmt.Sprintf(`{\"auth\":{\"identity\":{\"methods\":[\"password\"],\"password\":{\"user\":{\"domain\":{\"id\":\"default\"},\"name\":\"%s\",\"password\":\"%s\"}}}}}`, user, password)\n\trbody := strings.NewReader(body)\n\treq, err := http.NewRequest(\"POST\", url+\"v3/auth/tokens\", rbody)\n\tif err != nil {\n\t\tlogger.Debug(\"serverAuth error from NewRequest POST\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Error(err))\n\t\treturn \"\", err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := http.DefaultClient.Do(req)\n\tif err != nil {\n\t\tlogger.Debug(\"serverAuth error from DefaultClient.Do POST\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Error(err))\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 201 {\n\t\tlogger.Debug(\"serverAuth error from POST return status\", zap.String(\"url\", url), zap.String(\"user\", user), zap.Int(\"status\", resp.StatusCode))\n\t\treturn \"\", fmt.Errorf(\"server auth token request gave status %d\", resp.StatusCode)\n\t}\n\trv := resp.Header.Get(\"X-Subject-Token\")\n\tif len(rv) == 0 {\n\t\tlogger.Debug(\"serverAuth succeeded, but ended up with zero-length token\")\n\t} else {\n\t\tlogger.Debug(\"serverAuth succeeded\")\n\t}\n\treturn rv, nil\n}" ]
[ "0.57816464", "0.56545347", "0.5547903", "0.5484067", "0.5483845", "0.5380546", "0.53680843", "0.5358327", "0.5313363", "0.528413", "0.5263241", "0.51906294", "0.51678735", "0.5165179", "0.5149926", "0.5122884", "0.5073951", "0.5071476", "0.5035328", "0.50271136", "0.502607", "0.500705", "0.49980494", "0.49961448", "0.49857497", "0.49857497", "0.498403", "0.49762425", "0.49643847", "0.49495068", "0.49415037", "0.49385473", "0.49382895", "0.49305367", "0.49270144", "0.48943305", "0.48894078", "0.48842877", "0.48689598", "0.48666012", "0.4852183", "0.4851179", "0.4848042", "0.48467752", "0.48404285", "0.4839151", "0.48382208", "0.4836828", "0.48318678", "0.4831518", "0.4821355", "0.48174202", "0.48089775", "0.47887856", "0.47882402", "0.4782108", "0.47790065", "0.47612774", "0.4753492", "0.47523087", "0.47518402", "0.4748379", "0.47355413", "0.47344434", "0.4725433", "0.47253266", "0.472376", "0.47213635", "0.4720032", "0.47183356", "0.47135133", "0.4710221", "0.47035003", "0.47028214", "0.47005495", "0.46981645", "0.46977144", "0.4697671", "0.46975625", "0.46973783", "0.4690849", "0.46874547", "0.46874318", "0.4678184", "0.46773592", "0.46663868", "0.46649534", "0.4662035", "0.4660257", "0.46521914", "0.46393073", "0.46341625", "0.4631977", "0.4628598", "0.4627031", "0.46265063", "0.4624483", "0.46213597", "0.46172562", "0.46136868", "0.46069285" ]
0.0
-1
FilterFileWatcher filters for file watcher and returns only watcher which are type of filesystem
func FilterFileWatcher(watchers config.Watchers) config.Watchers { filewatchers := config.Watchers{} for _, watcher := range watchers { for _, tag := range watcher.Tags { if tag["key"] == "Type" && tag["value"] == "filesystem" { filewatchers = append(filewatchers, watcher) } } } return filewatchers }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func FileFilter() Filter {\n\treturn fileFilter{}\n}", "func fileFilterSingleton() *FileFilter {\n\treturn fileFilter\n}", "func (f *FsWatcher) Watch() {\r\n\tif !f.started {\r\n\t\tf.started = true\r\n\t\tgo func() {\r\n\t\t\t// Reference time used to dectect file changes\r\n\t\t\tstart := time.Now()\r\n\t\t\tfor {\r\n\t\t\t\tfilepath.Walk(f.root, func(path string, info os.FileInfo, err error) error {\r\n\t\t\t\t\tif f.stopPending {\r\n\t\t\t\t\t\treturn errors.New(\"Stopping\")\r\n\t\t\t\t\t}\r\n\t\t\t\t\tb := filepath.Base(path)\r\n\t\t\t\t\tdir, ok := f.ignored[b]\r\n\t\t\t\t\tif ok && b == dir {\r\n\t\t\t\t\t\treturn filepath.SkipDir\r\n\t\t\t\t\t}\r\n\r\n\t\t\t\t\t// TODO : This is not optimized.\r\n\t\t\t\t\t// Using the * handler, a directory will only be skipped if no files\r\n\t\t\t\t\tif info.ModTime().After(start) {\r\n\t\t\t\t\t\thandler, ok := f.handlers[filepath.Ext(path)]\r\n\t\t\t\t\t\tif ok == true {\r\n\t\t\t\t\t\t\tif handler(path, info) == true {\r\n\t\t\t\t\t\t\t\treturn filepath.SkipDir\r\n\t\t\t\t\t\t\t}\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t}\r\n\t\t\t\t\treturn nil\r\n\t\t\t\t})\r\n\r\n\t\t\t\tif f.stopPending {\r\n\t\t\t\t\tf.started = false\r\n\t\t\t\t\treturn\r\n\t\t\t\t}\r\n\t\t\t\tstart = time.Now()\r\n\t\t\t\ttime.Sleep(time.Duration(f.interval) * time.Millisecond)\r\n\t\t\t}\r\n\t\t}()\r\n\t}\r\n}", "func FileFilter(path string, info os.FileInfo) bool {\n\treturn !info.IsDir()\n}", "func watch(dispatchChan chan string) {\n\n\t// Start a file system watcher\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tlog.Fatal(\"Error creating watcher\", err)\n\t}\n\n\tsignalSourceChanged := debounce(800*time.Millisecond, func() {\n\t\tdispatchChan <- SOURCECHANGED\n\t})\n\n\t// Start waiting for file system messages to receive...\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\t// On any event\n\t\t\tcase ev := <-watcher.Event:\n\t\t\t\tif isGoFile(ev.Name) && (ev.IsCreate() || ev.IsDelete()) {\n\t\t\t\t\tsignalSourceChanged()\n\t\t\t\t}\n\n\t\t\t// Stop when encountering errors...\n\t\t\tcase err := <-watcher.Error:\n\t\t\t\tlog.Fatal(\"Error start watching\", err)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Create a list of all directories to watch...\n\tdirectories := append(findDirectoriesIn(DIR_TO_WATCH), DIR_TO_WATCH)\n\tfor _, directory := range directories {\n\t\t// Configure to watcher to watch the files we want to...\n\t\tif err := watcher.Watch(directory); err != nil {\n\t\t\tlog.Fatal(\"Error start watching\", err)\n\t\t}\n\t}\n}", "func FilteredFileServer(pattern *regexp.Regexp, fs http.FileSystem) http.Handler {\n\n\tif pattern == nil {\n\t\tpanic(fmt.Errorf(\"pattern is nil\"))\n\t}\n\n\tif fs == nil {\n\t\tpanic(fmt.Errorf(\"fs is nil\"))\n\t}\n\n\tfserver := http.FileServer(fs)\n\n\tret := http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\n\t\tp := path.Clean(\"/\" + r.URL.Path)\n\n\t\tif !strings.HasPrefix(p, \"/\") { // should never happen after Clean above, but just being extra cautious\n\t\t\thttp.NotFound(w, r)\n\t\t\treturn\n\t\t}\n\n\t\tif !pattern.MatchString(p) {\n\t\t\thttp.NotFound(w, r)\n\t\t\treturn\n\t\t}\n\n\t\t// delegate to the regular file-serving behavior\n\t\tfserver.ServeHTTP(w, r)\n\n\t})\n\n\treturn ret\n}", "func Filter(c *C, pattern string) Stage {\n\treturn FilterFunc(c, func(f File) bool {\n\t\tm, err := glob.Match(pattern, f.Stat.Name())\n\t\tif err != nil {\n\t\t\tc.Println(err)\n\t\t}\n\t\treturn m\n\t})\n}", "func FileWalkerGenericFilter(filepath string, rule Rule) bool {\n\n\tfileMeta, errs := os.Stat(filepath)\n\t// not exists or permission not allow\n\tif os.IsNotExist(errs) || os.IsPermission(errs) {\n\t\treturn false\n\t}\n\n\t// file content is empty\n\tif rule.AllowEmpty && fileMeta.Size()-0 == 0 {\n\t\treturn true\n\t}\n\n\t// file size is out of limito\n\tif fileMeta.Size()-rule.FileSizeLimit > 0 {\n\t\treturn false\n\t}\n\n\t// shadow file\n\tif strings.HasPrefix(fileMeta.Name(), \".\") {\n\t\treturn false\n\t}\n\n\t// can't be read now\n\tif fileMeta.ModTime().Unix()+int64(rule.CollectWaitTime) > time.Now().Unix() {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func DirFilter() Filter {\n\treturn dirFilter{}\n}", "func TestWatcher(t *testing.T) {\n\tw := NewWatcherTest(t, \"testdata/vfs.txt\")\n\tdefer w.Close()\n\n\tcases := [...]WCase{\n\t\tcreate(w, \"src/github.com/ppknap/link/include/coost/.link.hpp.swp\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/fs_test.go\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/binfs/\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/binfs.go\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/binfs_test.go\"),\n\t\tremove(w, \"src/github.com/rjeczalik/fs/binfs/\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/binfs/\"),\n\t\tcreate(w, \"src/github.com/rjeczalik/fs/virfs\"),\n\t\tremove(w, \"src/github.com/rjeczalik/fs/virfs\"),\n\t\tcreate(w, \"file\"),\n\t\tcreate(w, \"dir/\"),\n\t}\n\n\tw.ExpectAny(cases[:])\n}", "func (c *watchImpl) Watch(handler WatcherHandler) (chan struct{}, error) {\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := watcher.Add(path.Dir(c.filename)); err != nil {\n\t\treturn nil, err\n\t}\n\tstopCh := make(chan struct{}, 0)\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tgo handler.Error(err)\n\t\t\tcase ev := <-watcher.Events:\n\t\t\t\tif ev.Op&fsnotify.Write == fsnotify.Write || ev.Op&fsnotify.Create == fsnotify.Create {\n\t\t\t\t\tif ev.Name == c.filename {\n\t\t\t\t\t\tgo handler.Updated()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\tcase <-stopCh:\n\t\t\t\twatcher.Close()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn stopCh, nil\n}", "func FilterUnchanged(c cookoo.Context, p *cookoo.Params) (interface{}, cookoo.Interrupt) {\n\tfiles := p.Get(\"files\", []string{}).([]string)\n\tsince := p.Get(\"since\", time.Now().Add(time.Second * -10)).(time.Time)\n\n\tmodified := []string{}\n\tfor _, file := range files {\n\t\tstat, err := os.Stat(file)\n\t\tif err == nil && stat.ModTime().After(since) {\n\t\t\tmodified = append(modified, file)\n\t\t}\n\t}\n\n\treturn modified, nil\n}", "func NewWatcher(file string, deadTime time.Duration) (*Watcher, error) {\n\tfileToWatch, err := filepath.Abs(file)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Couldn't get absolute path of file. \" + err.Error())\n\t}\n\tdirectoryToWatch := filepath.Dir(fileToWatch)\n\n\tw := Watcher{\n\t\tC: make(chan struct{}),\n\t\tstopChan: make(chan struct{}),\n\t}\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, errors.New(\"Couldn't establish watcher. \" + err.Error())\n\t}\n\n\tgo func() {\n\t\tt := time.NewTimer(deadTime)\n\t\ttimerRunning := true\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\t// Received an event. Check it's for our file.\n\t\t\t\teventFile, evErr := filepath.Abs(event.Name)\n\t\t\t\tif evErr != nil || eventFile != fileToWatch {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t// It's for our file so stop and restart the timer.\n\t\t\t\tif timerRunning {\n\t\t\t\t\tif !t.Stop() {\n\t\t\t\t\t\t// empty the timer chan if we failed to stop it\n\t\t\t\t\t\t<-t.C\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tt.Reset(deadTime)\n\t\t\t\ttimerRunning = true\n\t\t\tcase watcherErr := <-watcher.Errors:\n\t\t\t\tlog.Println(\"Throttled Watcher error:\", watcherErr)\n\t\t\tcase <-t.C:\n\t\t\t\ttimerRunning = false\n\t\t\t\tw.C <- struct{}{}\n\t\t\tcase <-w.stopChan:\n\t\t\t\tif timerRunning {\n\t\t\t\t\tt.Stop()\n\t\t\t\t}\n\t\t\t\twatcher.Close()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\terr = watcher.Add(directoryToWatch)\n\tif err != nil {\n\t\tw.stopChan <- struct{}{}\n\t\treturn nil, errors.New(\"Couldn't watch directory. \" + err.Error())\n\t}\n\n\treturn &w, nil\n\n}", "func dirWatcher(watcher *fsnotify.Watcher) {\n\tfor {\n\t\tselect {\n\t\tcase event, ok := <-watcher.Events:\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlog.Debug(\"Event: \", event)\n\t\t\tif arrContains(event.Op.String()) && !creatingDirectory(event.Op.String(), event.Name) {\n\t\t\t\terr := messaging.ClientSend(event.Op.String(), event.Name)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Error(\"Error sending event to server: \", err)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\tcase err, ok := <-watcher.Errors:\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlog.Error(\"Error: \", err)\n\t\t}\n\t}\n}", "func (fw *FileWatcher) Watch() error {\n\tvar err error\n\tfw.watcher, err = fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := fw.subscribeToDirRecursively(\".\"); err != nil {\n\t\treturn err\n\t}\n\n\tfw.Handler.OnStartWatching()\n\tdefer fw.Handler.OnStopWatching()\n\n\tlog.Println(\"Scanning for files\")\n\tfor {\n\t\tselect {\n\t\tcase event := <-fw.watcher.Events:\n\n\t\t\t// TODO: If it's a new directory, subscribe\n\n\t\t\tif event.Op&fsnotify.Write == fsnotify.Write {\n\t\t\t\tif filepath.Ext(event.Name) == \".go\" {\n\t\t\t\t\tlog.Println(\"modified file:\", event.Name, \" Rebuilding\")\n\t\t\t\t\tfw.Handler.OnFileModified(event.Name)\n\t\t\t\t}\n\t\t\t}\n\t\tcase err := <-fw.watcher.Errors:\n\t\t\tlog.Println(\"error:\", err)\n\t\t}\n\t}\n}", "func (r *Runner) Watch(shutdown <-chan struct{}, ignore []string) (chan bool, error) {\n\trestartChan := make(chan bool)\n\n\twatcher, err := NewRecursiveWatcher(r.WatchPath, ignore)\n\tif err != nil {\n\t\tlog.Println(\"ERRO:\", err)\n\t\treturn nil, err\n\t}\n\n\tgo func(restart chan<- bool) {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-shutdown:\n\t\t\t\twatcher.Close()\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\t//log.Println(\"DEBU: event:\", event)\n\t\t\t\tswitch {\n\t\t\t\tcase event.Op&fsnotify.Create == fsnotify.Create:\n\t\t\t\t\t// create a file or directory\n\t\t\t\t\tfi, err := os.Stat(event.Name)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t// eg. stat .subl513.tmp : no such file or directory\n\t\t\t\t\t\tlog.Println(\"ERRO:\", err)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tif watcher.ShouldIgnoreFile(event.Name) {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif fi.IsDir() {\n\t\t\t\t\t\t//log.Printf(\"DEBU: detected new directory %s\\n\", event.Name)\n\t\t\t\t\t\twatcher.AddFolder(event.Name)\n\t\t\t\t\t\trestart <- true\n\t\t\t\t\t\tlog.Printf(\"INFO: added new folder: %s\\n\", event.Name)\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// created a file\n\t\t\t\t\t\trestart <- true\n\t\t\t\t\t\tlog.Printf(\"INFO: added new file: %s\\n\", event.Name)\n\t\t\t\t\t}\n\t\t\t\tcase event.Op&fsnotify.Write == fsnotify.Write:\n\t\t\t\t\tif !watcher.ShouldIgnoreFile(event.Name) {\n\t\t\t\t\t\trestart <- true\n\t\t\t\t\t\tlog.Printf(\"INFO: modified file: %s\\n\", event.Name)\n\t\t\t\t\t}\n\t\t\t\tcase event.Op&fsnotify.Remove == fsnotify.Remove:\n\t\t\t\t\twatcher.Remove(event.Name)\n\t\t\t\t\trestart <- true\n\t\t\t\t\tlog.Printf(\"INFO: removed file: %s\\n\", event.Name)\n\t\t\t\tcase event.Op&fsnotify.Chmod == fsnotify.Chmod:\n\t\t\t\t\tlog.Printf(\"chmod file: %s\", event.Name)\n\t\t\t\tcase event.Op&fsnotify.Rename == fsnotify.Rename:\n\t\t\t\t\t// renaming a file triggers a create event\n\t\t\t\t\t//log.Printf(\"DEBU: renamed file: %s\", event.Name)\n\t\t\t\t}\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tlog.Println(\"ERRO:\", err)\n\t\t\t}\n\t\t}\n\t}(restartChan)\n\n\treturn restartChan, nil\n}", "func watch(ctx watchContext) error {\n\tw := watcher.New()\n\tw.FilterOps(watcher.Write, watcher.Create, watcher.Move, watcher.Remove, watcher.Rename)\n\n\tgo func() {\n\twatcherLoop:\n\t\tfor {\n\n\t\t\tselect {\n\t\t\tcase event, ok := <-w.Event:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif filepath.Ext(event.Path) == \"\" {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tfor _, ignorePath := range ctx.IgnorePaths {\n\t\t\t\t\tp, err := filepath.Abs(ignorePath)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tout.Err(style.Exclamation, \"failed to read ignored filepath: %s\", err.Error())\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif strings.HasPrefix(event.Path, p) {\n\t\t\t\t\t\tcontinue watcherLoop\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tctx.ChangedCh <- event.Path\n\n\t\t\tcase err, ok := <-w.Error:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tout.Err(style.Exclamation, \"failed to watch project files: %s\", err.Error())\n\n\t\t\tcase _, ok := <-ctx.StopCh:\n\t\t\t\tif !ok {\n\t\t\t\t\tw.Close()\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\n\tif err := w.AddRecursive(ctx.Path); err != nil {\n\t\treturn err\n\t}\n\n\tif err := w.Ignore(ctx.IgnorePaths...); err != nil {\n\t\treturn err\n\t}\n\n\tvar err error\n\n\tgo func() {\n\t\terr = w.Start(time.Millisecond * 100)\n\t}()\n\n\treturn err\n}", "func (w *filePoller) watch(f *os.File, lastFi os.FileInfo, chClose chan struct{}) {\n\tdefer f.Close()\n\n\ttimer := time.NewTimer(watchWaitTime)\n\tif !timer.Stop() {\n\t\t<-timer.C\n\t}\n\tdefer timer.Stop()\n\n\tfor {\n\t\ttimer.Reset(watchWaitTime)\n\n\t\tselect {\n\t\tcase <-timer.C:\n\t\tcase <-chClose:\n\t\t\tlogrus.Debugf(\"watch for %s closed\", f.Name())\n\t\t\treturn\n\t\t}\n\n\t\tfi, err := os.Stat(f.Name())\n\t\tif err != nil {\n\t\t\t// if we got an error here and lastFi is not set, we can presume that nothing has changed\n\t\t\t// This should be safe since before `watch()` is called, a stat is performed, there is any error `watch` is not called\n\t\t\tif lastFi == nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// If it doesn't exist at this point, it must have been removed\n\t\t\t// no need to send the error here since this is a valid operation\n\t\t\tif os.IsNotExist(err) {\n\t\t\t\tif err := w.sendEvent(fsnotify.Event{Op: fsnotify.Remove, Name: f.Name()}, chClose); err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tlastFi = nil\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// at this point, send the error\n\t\t\tif err := w.sendErr(err, chClose); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tif lastFi == nil {\n\t\t\tif err := w.sendEvent(fsnotify.Event{Op: fsnotify.Create, Name: fi.Name()}, chClose); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlastFi = fi\n\t\t\tcontinue\n\t\t}\n\n\t\tif fi.Mode() != lastFi.Mode() {\n\t\t\tif err := w.sendEvent(fsnotify.Event{Op: fsnotify.Chmod, Name: fi.Name()}, chClose); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlastFi = fi\n\t\t\tcontinue\n\t\t}\n\n\t\tif fi.ModTime() != lastFi.ModTime() || fi.Size() != lastFi.Size() {\n\t\t\tif err := w.sendEvent(fsnotify.Event{Op: fsnotify.Write, Name: fi.Name()}, chClose); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tlastFi = fi\n\t\t\tcontinue\n\t\t}\n\t}\n}", "func New() (*FileWatcher, error) {\n\t// Create a new file watcher\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tf := &FileWatcher{\n\t\twatcher: watcher,\n\t\ttriggerClose: make(chan struct{}),\n\t}\n\n\t// Start the goroutine to watch the files\n\tgo func() {\n\t\tdefer f.watcher.Close()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\t// Skip if the event function is not defined\n\t\t\t\tif f.onEvent == nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// Create a new event object\n\t\t\t\te := &Event{\n\t\t\t\t\tPath: event.Name,\n\t\t\t\t\tevent: &event,\n\t\t\t\t}\n\n\t\t\t\t// Call the callback in a safe way\n\t\t\t\tfunc() {\n\t\t\t\t\t// Recover panics and log the error\n\t\t\t\t\tdefer func() {\n\t\t\t\t\t\tif e := recover(); e != nil {\n\t\t\t\t\t\t\tlog.L.Error(\"filewatcher callback panic: %v\", e)\n\t\t\t\t\t\t}\n\t\t\t\t\t}()\n\n\t\t\t\t\tf.onEvent(e)\n\t\t\t\t}()\n\n\t\t\t\t// If a new folder was created, then add it to monitor recursive actions\n\t\t\t\tif e.IsCreate() {\n\t\t\t\t\tgo func() {\n\t\t\t\t\t\tif stat, err := os.Stat(e.Path); err == nil && stat.IsDir() {\n\t\t\t\t\t\t\terr = f.Add(e.Path)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tlog.L.Error(\"filewatcher: failed to add recursive directory: %v\", err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}()\n\t\t\t\t}\n\n\t\t\t\t// We don't have to remove deleted directories from the watcher,\n\t\t\t\t// because they are removed automatically...\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tlog.L.Error(\"a file watcher error occurred: %s\", err.Error())\n\t\t\tcase <-f.triggerClose:\n\t\t\t\t// Just exit the loop\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn f, nil\n}", "func (a *Application) FileFilterOperator() pipeline.Operator {\n\treturn func(in, out chan interface{}) {\n\t\tfrom, to := a.dateFrom.Truncate(24*time.Hour), a.dateTo.Truncate(24*time.Hour).AddDate(0, 0, 1)\n\t\tfor i := range in {\n\t\t\tif item, ok := i.(walker.WalkItem); ok {\n\t\t\t\tif fd, err := time.ParseInLocation(\"u_ex060102.log\", item.Name(), time.UTC); err == nil {\n\t\t\t\t\t// Check if the file date fits with searched date range\n\t\t\t\t\tif from.Before(fd) && fd.Before(to) {\n\t\t\t\t\t\tout <- item\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\titem.Close()\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tpanic(\"Expecting walker.WalkItem in pipeline.Operator filter\")\n\t\t\t}\n\t\t}\n\t}\n}", "func (d *Discovery) watchFiles() {\n\tif d.watcher == nil {\n\t\tpanic(\"no watcher configured\")\n\t}\n\tfor _, p := range d.paths {\n\t\tif dir, _ := filepath.Split(p); dir != \"\" {\n\t\t\tp = dir\n\t\t} else {\n\t\t\tp = \"./\"\n\t\t}\n\t\tif err := d.watcher.Add(p); err != nil {\n\t\t\tlevel.Error(d.logger).Log(\"msg\", \"Error adding file watch\", \"path\", p, \"err\", err)\n\t\t}\n\t}\n}", "func GatherFiletype(root, filetype string) (*FileList, error) {\n\tlist := &FileList{}\n\tgather := func(path string, file os.FileInfo, err error) error {\n\t\tif filepath.Ext(path) == filetype {\n\t\t\tlist.Add(path)\n\t\t}\n\t\treturn nil\n\t}\n\n\terr := filepath.Walk(root, gather)\n\tif err != nil {\n\t\treturn list, err\n\t}\n\treturn list, nil\n}", "func FilteredFiles(ctx context.Context, c *client.Client, t time.Time, org string, project string, num int) ([]*github.CommitFile, error) {\n\tlogrus.Infof(\"Fetching file list for #%d\", num)\n\n\tchanged, err := ghcache.PullRequestsListFiles(ctx, c.Cache, c.GitHubClient, t, org, project, num)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogrus.Infof(\"%s/%s #%d had %d changed files\", org, project, num, len(changed))\n\n\tfiles := []*github.CommitFile{}\n\tfor _, cf := range changed {\n\t\tif ignorePathRe.MatchString(cf.GetFilename()) {\n\t\t\tlogrus.Infof(\"ignoring %s\", cf.GetFilename())\n\t\t\tcontinue\n\t\t}\n\t\tlogrus.Errorf(\"#%d changed: %s\", num, cf.GetFilename())\n\n\t\tfiles = append(files, cf)\n\t}\n\n\treturn files, err\n}", "func StandardFilterMatcher(pattern, path string) bool {\n\tmatched, err := filepath.Match(pattern, path)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tif matched {\n\t\treturn true\n\t}\n\n\tisDir, err := isDir(pattern)\n\tif err != nil || !isDir {\n\t\treturn false\n\t}\n\n\tif len(path) < len(pattern) {\n\t\treturn false\n\t}\n\n\tif path[:len(pattern)] == pattern {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (m *RdmaDevPlugin) Watch() {\n\tlog.Println(\"Starting FS watcher.\")\n\twatcher, err := newFSWatcher(deprecatedSockDir)\n\tif err != nil {\n\t\tlog.Println(\"Failed to created FS watcher.\")\n\t\tos.Exit(1)\n\t}\n\tdefer watcher.Close()\n\n\tselect {\n\tcase event := <-watcher.Events:\n\t\tif event.Name == m.socketPath && event.Op&fsnotify.Create == fsnotify.Create {\n\t\t\tlog.Printf(\"inotify: %s created, restarting.\", m.socketPath)\n\t\t\tif err = m.Restart(); err != nil {\n\t\t\t\tlog.Fatalf(\"unable to restart server %v\", err)\n\t\t\t}\n\t\t}\n\n\tcase err := <-watcher.Errors:\n\t\tlog.Printf(\"inotify: %s\", err)\n\n\tcase stop := <-m.stopWatcher:\n\t\tif stop {\n\t\t\tlog.Println(\"kubelet watcher stopped\")\n\t\t\twatcher.Close()\n\t\t\treturn\n\t\t}\n\t}\n}", "func (li TriceIDLookUpLI) FileWatcher(w io.Writer, fSys *afero.Afero) {\n\n\t// creates a new file watcher\n\twatcher, err := fsnotify.NewWatcher()\n\tmsg.FatalOnErr(err)\n\tdefer func() { msg.OnErr(watcher.Close()) }()\n\n\tdone := make(chan bool)\n\tgo func() {\n\t\tvar now, last time.Time\n\t\tfor {\n\t\t\tselect {\n\t\t\t// watch for events\n\t\t\tcase event, ok := <-watcher.Events:\n\t\t\t\tfmt.Fprintln(w, \"EVENT:\", event, ok, time.Now().UTC())\n\n\t\t\t\tnow = time.Now()\n\t\t\t\tdiff := now.Sub(last)\n\t\t\t\tif diff > 5000*time.Millisecond {\n\t\t\t\t\tfmt.Fprintln(w, \"refreshing li list\")\n\t\t\t\t\tmsg.FatalOnErr(li.fromFile(fSys, LIFnJSON))\n\t\t\t\t\tlast = time.Now()\n\t\t\t\t}\n\n\t\t\t// watch for errors\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tfmt.Fprintln(w, \"ERROR1\", err, time.Now().UTC())\n\t\t\t}\n\t\t}\n\t}()\n\n\t// out of the box fsnotify can watch a single file, or a single directory\n\tmsg.InfoOnErr(watcher.Add(LIFnJSON), \"ERROR2\")\n\tif Verbose {\n\t\tfmt.Fprintln(w, LIFnJSON, \"watched now for changes\")\n\t}\n\t<-done\n}", "func Watch(ctx context.Context, cliEngine *engine.Engine, task string, t ox.Task) {\n\ttaskCtx, cancel := context.WithCancel(ctx)\n\n\tfiles, err := getWatcherFiles(t.Sources, t.Dir)\n\tif err != nil {\n\t\tutils.PrintError(err)\n\t\treturn\n\t}\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tutils.PrintError(err)\n\t\treturn\n\t}\n\tdefer watcher.Close()\n\n\tfor _, file := range files {\n\t\terr = watcher.Add(file)\n\t\tif err != nil {\n\t\t\tutils.PrintError(err)\n\t\t\treturn\n\t\t}\n\t}\n\n\trunOnWatch := func() {\n\t\terr := cliEngine.Run(taskCtx, task)\n\t\tif err != nil {\n\t\t\tutils.PrintError(err)\n\t\t}\n\t}\n\n\tgo runOnWatch()\n\n\tfor {\n\t\tselect {\n\t\tcase event := <-watcher.Events:\n\t\t\tswitch {\n\t\t\tcase event.Op&fsnotify.Write == fsnotify.Write:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Create == fsnotify.Create:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Remove == fsnotify.Remove:\n\t\t\t\tfallthrough\n\t\t\tcase event.Op&fsnotify.Rename == fsnotify.Rename:\n\t\t\t\tcancel()\n\t\t\t\ttaskCtx, cancel = context.WithCancel(ctx)\n\t\t\t\tgo runOnWatch()\n\t\t\t}\n\t\tcase <-ctx.Done():\n\t\t\tcancel()\n\t\t\treturn\n\t\tcase err := <-watcher.Errors:\n\t\t\tutils.PrintError(err)\n\t\t\tcancel()\n\t\t\treturn\n\t\t}\n\t}\n}", "func (c *Client) getLockableFilter() *filepathfilter.Filter {\n\tc.ensureLockablesLoaded()\n\treturn c.lockableFilter\n}", "func filter(fi os.FileInfo) bool {\n\treturn !strings.Contains(fi.Name(), \"test\")\n}", "func (f *FormatterFilter) Filter(inFile *hclwrite.File) (*hclwrite.File, error) {\n\tformatter := NewDefaultFormatter()\n\ttmp, err := formatter.Format(inFile)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// The hclwrite package doesn't provide a token-base interface, so we need to\n\t// parse it again. It's obviously inefficient, but the only way to match the\n\t// type signature.\n\toutFile, err := safeParseConfig(tmp, \"generated_by_FormatterFilter\", hcl.Pos{Line: 1, Column: 1})\n\tif err != nil {\n\t\t// should never happen.\n\t\treturn nil, fmt.Errorf(\"failed to parse formatted bytes: %s\", err)\n\t}\n\n\treturn outFile, nil\n}", "func FS(config *httpConfig.Static) http.FileSystem {\n\treturn FileSystem{NewExtensionFilter(config.Allow, config.Forbid), http.Dir(config.Dir)}\n}", "func (w *Watcher) Watch() {\n\tset := make(map[fs.FileInfo]bool)\n\tfor {\n\t\tselect {\n\t\t//If watcher is closed it should stop watching directory\n\t\tcase <-w.stopCh:\n\t\t\treturn\n\t\tdefault:\n\t\t\tfis, err := ioutil.ReadDir(w.watchDirectory)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t//TODO add more advanced watching of changes\n\t\t\t//If there is no file , add it\n\t\t\tfor _, fi := range fis {\n\t\t\t\tif _, ok := set[fi]; !ok {\n\t\t\t\t\tset[fi] = true\n\t\t\t\t\tw.notify(Event{\n\t\t\t\t\t\tType: UpdateEventType,\n\t\t\t\t\t\tCaller: fi.Name(),\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func TestWatcher(t *testing.T) {\n\tdir := tempMkdir(t, \"crun-watcher\")\n\n\tif err := os.Chdir(dir); err != nil {\n\t\tt.Fatal(\"Failed to change to temp dir\")\n\t}\n\n\tc := make(chan string, 2)\n\twatch(c)\n\n\tnotified := 0\n\n\tgo func() {\n\t\tfor msg := range c {\n\t\t\tif msg == SOURCECHANGED {\n\t\t\t\tnotified++\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Should only do one, since debounced\n\tioutil.WriteFile(\"testfile-1.go\", []byte(\"something\"), 0644)\n\tioutil.WriteFile(\"testfile-2.go\", []byte(\"something\"), 0644)\n\n\ttime.Sleep(1 * time.Second)\n\tif notified != 1 {\n\t\tt.Fatal(\"Watcher should have notified 1 time\", notified)\n\t}\n\n\tioutil.WriteFile(\"testfile-3.go\", []byte(\"something\"), 0644)\n\ttime.Sleep(1 * time.Second)\n\tif notified != 2 {\n\t\tt.Fatal(\"Watcher should have notified 2 times\", notified)\n\t}\n\n\tioutil.WriteFile(\"testfile-1.md\", []byte(\"something\"), 0644)\n\ttime.Sleep(1 * time.Second)\n\tif notified != 2 {\n\t\tt.Fatal(\"Watcher should only notify on md files\", notified)\n\t}\n\n\tioutil.WriteFile(\"testfile-4.go\", []byte(\"something\"), 0644)\n\tioutil.WriteFile(\"testfile-5.go\", []byte(\"something\"), 0644)\n\n\ttime.Sleep(1 * time.Second)\n\tif notified != 3 {\n\t\tt.Fatal(\"Watcher should have notified 3 times\", notified)\n\t}\n\tclose(c)\n}", "func (lu TriceIDLookUp) FileWatcher(w io.Writer, fSys *afero.Afero, m *sync.RWMutex) {\n\n\t// creates a new file watcher\n\twatcher, err := fsnotify.NewWatcher()\n\tmsg.FatalOnErr(err)\n\tdefer func() { msg.OnErr(watcher.Close()) }()\n\n\tdone := make(chan bool)\n\tgo func() {\n\t\tvar now, last time.Time\n\t\tfor {\n\t\t\tselect {\n\t\t\t// watch for events\n\t\t\tcase event, ok := <-watcher.Events:\n\t\t\t\tfmt.Fprintln(w, \"EVENT:\", event, ok, time.Now().UTC())\n\n\t\t\t\tnow = time.Now()\n\t\t\t\tdiff := now.Sub(last)\n\t\t\t\tif diff > 5000*time.Millisecond {\n\t\t\t\t\tfmt.Fprintln(w, \"refreshing id.List\")\n\t\t\t\t\tm.Lock()\n\t\t\t\t\tmsg.FatalOnErr(lu.fromFile(fSys, FnJSON))\n\t\t\t\t\tlu.AddFmtCount(w)\n\t\t\t\t\tm.Unlock()\n\t\t\t\t\tlast = time.Now()\n\t\t\t\t}\n\n\t\t\t// watch for errors\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tfmt.Fprintln(w, \"ERROR1\", err, time.Now().UTC())\n\t\t\t}\n\t\t}\n\t}()\n\n\t// out of the box fsnotify can watch a single file, or a single directory\n\tmsg.InfoOnErr(watcher.Add(FnJSON), \"ERROR2\")\n\tif Verbose {\n\t\tfmt.Fprintln(w, FnJSON, \"watched now for changes\")\n\t}\n\t<-done\n}", "func monitorLocalChanges(rootdir string, cafile string, server string, listFileInProcess *ListFileInProcess) {\n\tfmt.Println(\"*** Recursively monitoring folder\", rootdir)\n\twatcher, err := watch.NewWatcher(rootdir, hasher.PROCESSING_DIR)\n\t//watcher, err := watch.NewRecursiveWatcher(rootdir, hasher.PROCESSING_DIR)\n\tif err != nil {\n\t\tlog.Println(\"Watcher create error : \", err)\n\t}\n\tdefer watcher.Close()\n\t_done := make(chan bool)\n\n\tgo func() {\n\t\twatcher.start()\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\tswitch {\n\t\t\t\tcase event.Op&fsnotify.Create == fsnotify.Create:\n\t\t\t\t\tfi, err := os.Stat(event.Name)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t// eg. stat .subl513.tmp : no such file or directory\n\t\t\t\t\t\tfmt.Println(err)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t} else if fi.IsDir() {\n\t\t\t\t\t\tfmt.Println(\"Detected new directory\", event.Name)\n\t\t\t\t\t\tif !watch.ShouldIgnoreFile(filepath.Base(event.Name), hasher.PROCESSING_DIR) {\n\t\t\t\t\t\t\tfmt.Println(\"Monitoring new folder...\")\n\t\t\t\t\t\t\twatcher.AddFolder(event.Name)\n\t\t\t\t\t\t\tconnsender := connectToServer(cafile, server)\n\t\t\t\t\t\t\tgo sendClientFolderChanges(connsender, event.Name, listFileInProcess)\n\t\t\t\t\t\t\t//watcher.Folders <- event.Name\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfmt.Println(\"Detected new file, for now do nothing\", event.Name)\n\t\t\t\t\t\t// watcher.Files <- event.Name // created a file\n\t\t\t\t\t\t// TODO\n\t\t\t\t\t}\n\n\t\t\t\tcase event.Op&fsnotify.Write == fsnotify.Write:\n\t\t\t\t\t// modified a file, assuming that you don't modify folders\n\t\t\t\t\tfmt.Println(\"Detected file modification\", event.Name)\n\t\t\t\t\t// Don't handle folder change, since they receive notification\n\t\t\t\t\t// when a file they contain is changed\n\t\t\t\t\tfi, err := os.Stat(event.Name)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Println(err)\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tif fi.Mode().IsRegular() {\n\t\t\t\t\t\t// watcher.Files <- event.Name\n\t\t\t\t\t\tlog.Println(\"Modified file: \", event.Name)\n\t\t\t\t\t\t// connsender := connectToServer(cafile, server)\n\t\t\t\t\t\t// go sendClientChanges(connsender, event.Name, listFileInProcess)\n\t\t\t\t\t}\n\t\t\t\tcase event.Op&fsnotify.Remove == fsnotify.Remove:\n\t\t\t\t\tlog.Println(\"Removed file: \", event.Name)\n\t\t\t\t\t// connsender := connectToServer(cafile, server)\n\t\t\t\t\t// go sendClientDelete(connsender, event.Name, listFileInProcess)\n\t\t\t\tcase event.Op&fsnotify.Rename == fsnotify.Rename:\n\t\t\t\t\tlog.Println(\"Renamed file: \", event.Name)\n\t\t\t\t\t// The following is to handle an issue in fsnotify\n\t\t\t\t\t// On rename, fsnotify sends three events on linux: RENAME(old), CREATE(new), RENAME(new)\n\t\t\t\t\t// fsnotify sends two events on windows: RENAME(old), CREATE(new)\n\t\t\t\t\t// The way we handle this is:\n\t\t\t\t\t// 1. If there is a second rename, skip it\n\t\t\t\t\t// 2. When the first rename happens, remove old file/folder\n\t\t\t\t\t// 3. We'll re-add it when the new create comes in\n\t\t\t\t\t// Step 2 and 3 might be optimized later by remembering which was old/new and performing simple move\n\t\t\t\t\t_, err := os.Stat(event.Name)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t// Rename talks about a file/folder now gone, send a remove request to server\n\t\t\t\t\t\tlog.Println(\"Rename leading to delete\", event.Name)\n\t\t\t\t\t\tconnsender := connectToServer(cafile, server)\n\t\t\t\t\t\tgo sendClientDelete(connsender, event.Name, listFileInProcess)\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// Rename talks about a file/folder already existing, skip it (do nothing)\n\t\t\t\t\t}\n\t\t\t\tcase event.Op&fsnotify.Chmod == fsnotify.Chmod:\n\t\t\t\t\tlog.Println(\"File changed permission: \", event.Name)\n\t\t\t\t}\n\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tlog.Println(\"Watcher watching error : \", err)\n\t\t\t\t_done <- true\n\t\t\t\tdone <- true\n\t\t\t}\n\t\t}\n\n\t}()\n\n\t<-_done\n}", "func exampleNewWatcher(t *testing.T) {\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tdefer watcher.Close()\n\n\terr = watcher.Add(\".\")\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\tsigs := make(chan os.Signal)\n\tsignal.Notify(sigs, syscall.SIGINT, syscall.SIGTERM)\nMAIN:\n\tfor {\n\t\tselect {\n\t\tcase event, ok := <-watcher.Events:\n\t\t\tif !ok { // when close channel ex) close(watcher.Events)\n\t\t\t\tbreak MAIN\n\t\t\t}\n\t\t\tlog.Println(\"event:\", event)\n\t\t\tif event.Op&fsnotify.Write == fsnotify.Write {\n\t\t\t\tlog.Println(\"modified file:\", event.Name)\n\t\t\t}\n\n\t\tcase err, ok := <-watcher.Errors:\n\t\t\tif !ok { // when close channel ex) close(watcher.Errors)\n\t\t\t\tbreak MAIN\n\t\t\t}\n\t\t\tlog.Println(\"error:\", err)\n\n\t\tcase signal := <-sigs:\n\t\t\tif signal == os.Interrupt {\n\t\t\t\tbreak MAIN\n\t\t\t\t// close(watcher.Events) or close(watcher.Errors)\n\t\t\t}\n\t\t}\n\t}\n}", "func (r *Reflex) filterMatching(out chan<- string, in <-chan string) {\n\tfor name := range in {\n\t\tif !r.matcher.Match(name) {\n\t\t\tcontinue\n\t\t}\n\n\t\tif r.onlyFiles || r.onlyDirs {\n\t\t\tstat, err := os.Stat(name)\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif (r.onlyFiles && stat.IsDir()) || (r.onlyDirs && !stat.IsDir()) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tout <- name\n\t}\n}", "func watchFilesystem(pathnames []string, outFilename string, buildCallback func()) {\n\tvar (\n\t\tbuildRate = time.Millisecond * 500\n\t\tpollRate = buildRate * 2\n\t)\n\n\t// Create a new watcher instance.\n\tw := watcher.New()\n\n\t// Only notify on certain events.\n\tw.FilterOps(\n\t\twatcher.Create,\n\t\twatcher.Write,\n\t\twatcher.Remove,\n\t\twatcher.Rename,\n\t\twatcher.Move,\n\t)\n\n\t// Ignore the output file.\n\tw.Ignore(outFilename)\n\n\t// Start a goroutine to handle the event loop.\n\tgo func() {\n\t\tbuild := false\n\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-time.After(buildRate):\n\t\t\t\tif build {\n\t\t\t\t\tbuildCallback()\n\t\t\t\t\tbuild = false\n\t\t\t\t}\n\t\t\tcase event := <-w.Event:\n\t\t\t\tif event.FileInfo != nil {\n\t\t\t\t\tisDir := event.IsDir()\n\n\t\t\t\t\tif event.Op == watcher.Write && isDir {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\n\t\t\t\t\tvar pathname string\n\t\t\t\t\tswitch event.Op {\n\t\t\t\t\tcase watcher.Move, watcher.Rename:\n\t\t\t\t\t\tpathname = fmt.Sprintf(\"%s -> %s\", relPath(event.OldPath), relPath(event.Path))\n\t\t\t\t\t\tif !build && !isDir {\n\t\t\t\t\t\t\tbuild = knownFileType(event.OldPath) || knownFileType(event.Path)\n\t\t\t\t\t\t}\n\t\t\t\t\tdefault:\n\t\t\t\t\t\tpathname = relPath(event.Path)\n\t\t\t\t\t\tif !build && !isDir {\n\t\t\t\t\t\t\tbuild = knownFileType(event.Path)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tlog.Printf(\"%s: %s\", event.Op, pathname)\n\t\t\t\t}\n\t\t\tcase err := <-w.Error:\n\t\t\t\tlog.Fatalln(err)\n\t\t\tcase <-w.Closed:\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Recursively watch the specified paths for changes.\n\tfor _, pathname := range pathnames {\n\t\tif err := w.AddRecursive(pathname); err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t}\n\t}\n\n\t// Print a message telling the user how to cancel watching\n\t// and list all paths being watched.\n\tlog.Print()\n\tlog.Print(\"Watch mode started. Press CTRL+C to stop.\")\n\tlog.Print()\n\tlog.Printf(\"Recursively watched paths: %d\", len(pathnames))\n\tfor _, pathname := range pathnames {\n\t\tlog.Printf(\" %s\", relPath(pathname))\n\t}\n\tlog.Print()\n\n\t// Build the ouput once before the watcher starts.\n\tbuildCallback()\n\n\t// Start watching.\n\tif err := w.Start(pollRate); err != nil {\n\t\tlog.Fatalln(err)\n\t}\n}", "func (t *Manager) withFilter(input func(in io.WriteCloser)) (string, error) {\n\tshell := os.Getenv(\"SHELL\")\n\tif shell == \"\" {\n\t\tshell = \"sh\"\n\t}\n\tcmd := exec.Command(shell, \"-c\", fzfPath)\n\tcmd.Stderr = os.Stderr\n\tin, _ := cmd.StdinPipe()\n\tgo func() {\n\t\tinput(in)\n\t\tin.Close()\n\t}()\n\tresult, err := cmd.Output()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tres := strings.Split(string(result), \"\\n\")\n\tif res[len(res)-1] == \"\" {\n\t\tres = res[0 : len(res)-1]\n\t}\n\treturn res[0], nil\n}", "func NewWatcher() (*Watcher, error) {\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Watcher{*watcher}, nil\n}", "func (w *FileWatch) stopWatcher() {\n\tlog.Trace(\">>>>> stopWatcher\")\n\tdefer log.Trace(\"<<<<< stopWatcher\")\n\tclose(w.watchStop)\n}", "func testsFilter(info os.FileInfo) bool {\n\treturn !strings.HasSuffix(info.Name(), \"_test.go\")\n}", "func DirFilter(path string, info os.FileInfo) bool {\n\treturn info.IsDir()\n}", "func FilterFunc(c *C, filter func(File) bool) Stage {\n\treturn func(files <-chan File, out chan<- File) {\n\t\tfor f := range files {\n\t\t\tif filter(f) {\n\t\t\t\tf.Close()\n\t\t\t} else {\n\t\t\t\tout <- f\n\t\t\t}\n\t\t}\n\t}\n}", "func NewWatcher() *Watcher {\n\tw := &Watcher{\n\t\tEvents: make(chan Event),\n\t\tErrors: make(chan error),\n\t\tclosed: make(chan struct{}),\n\t\tnames: make(map[string]struct{}),\n\t\tfiles: make(map[string]os.FileInfo),\n\t}\n\treturn w\n}", "func (fc *FilterConfig) initFsnotifyEventHandler() {\n\tconst pauseDelay = 5 * time.Second // used to let all changes be done before reloading the file\n\tgo func() {\n\t\ttimer := time.NewTimer(0)\n\t\tdefer timer.Stop()\n\t\tfirstTime := true\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-timer.C:\n\t\t\t\tif firstTime {\n\t\t\t\t\tfirstTime = false\n\t\t\t\t} else {\n\t\t\t\t\tfc.reloadFile()\n\t\t\t\t}\n\t\t\tcase <-fc.watcher.Events:\n\t\t\t\ttimer.Reset(pauseDelay)\n\t\t\tcase err := <-fc.watcher.Errors:\n\t\t\t\tgoglog.Logger.Errorf(\"ip2location: %s\", err.Error())\n\t\t\tcase <-fc.ctx.Done():\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n}", "func (task *Task) isWatchedFile(path string) bool {\n\tfilename, err := filepath.Rel(wd, path)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tfilename = filepath.ToSlash(filename)\n\t//util.Debug(\"task\", \"checking for match %s\\n\", filename)\n\n\tmatched := false\n\tfor _, info := range task.EffectiveWatchRegexps {\n\t\tif info.Negate {\n\t\t\tif matched {\n\t\t\t\tmatched = !info.MatchString(filename)\n\t\t\t\t//util.Debug(\"task\", \"negated match? %s %s\\n\", filename, matched)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t} else if info.MatchString(filename) {\n\t\t\tmatched = true\n\t\t\t//util.Debug(\"task\", \"matched %s %s\\n\", filename, matched)\n\t\t\tcontinue\n\t\t}\n\t}\n\treturn matched\n}", "func (event FileEvent) GetType() string {\n\treturn \"file\"\n}", "func (plugin *PluginConfig) watch(inChan utils.InputChannel) (err error) {\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tutils.Logger.Errorf(\"File input plugin watch error %s\", err)\n\t\t}\n\t}()\n\n\tvar (\n\t\tallfiles = make([]string, 0)\n\t\tfi os.FileInfo\n\t)\n\n\tif err = plugin.loadSinceDB(); err != nil {\n\t\tutils.Logger.Errorf(\"loadSinceDB return error %s\", err)\n\t\treturn\n\t}\n\n\tif len(plugin.DirsPath) < 1 {\n\t\tutils.Logger.Errorf(\"No director need to watch.\")\n\t\treturn\n\t}\n\n\t// find all log file path.\n\tfor _, dir := range plugin.DirsPath {\n\t\tfl, err := utils.FileList(dir, plugin.FileType)\n\t\tif err != nil {\n\t\t\tutils.Logger.Errorln(err)\n\t\t}\n\t\tallfiles = append(allfiles, fl...)\n\t}\n\n\t// loop save sincdb\n\tgo func() {\n\t\tplugin.wgExit.Add(1)\n\t\tdefer plugin.wgExit.Done()\n\n\t\tfor plugin.running {\n\t\t\ttime.Sleep(time.Duration(plugin.Intervals) * time.Second)\n\t\t\tif err = plugin.checkAndSaveSinceDB(); err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor _, fp := range allfiles {\n\t\t// get all sysmlinks.\n\t\tif fp, err = filepath.EvalSymlinks(fp); err != nil {\n\t\t\tutils.Logger.Warnf(\"Get symlinks failed: %s error %s\", fp, err)\n\t\t\tcontinue\n\t\t}\n\t\t// check file status.\n\t\tif fi, err = os.Stat(fp); err != nil {\n\t\t\tutils.Logger.Warnf(\"Get file status %s error %s\", fp, err)\n\t\t\tcontinue\n\t\t}\n\t\t// skip directory\n\t\tif fi.IsDir() {\n\t\t\tutils.Logger.Warnf(\"Skipping directory %s\", fi.Name())\n\t\t\tcontinue\n\t\t}\n\t\t// monitor file.\n\t\tutils.Logger.Info(\"Watching \", fp)\n\t\treadEventChan := make(chan fsnotify.Event, 10)\n\t\tgo plugin.loopRead(readEventChan, fp, inChan)\n\t\tgo plugin.loopWatch(readEventChan, fp, fsnotify.Create|fsnotify.Write)\n\t}\n\n\treturn\n}", "func (w *watcher) Watch() {\n\tfor {\n\t\tfor watchPath := range w.watchItems {\n\t\t\tfileChanged, err := w.scanChange(watchPath)\n\t\t\tif err != nil {\n\t\t\t\tw.errors <- err\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif fileChanged != \"\" {\n\t\t\t\tw.events <- fileChanged\n\t\t\t\tstartTime = time.Now()\n\t\t\t}\n\t\t}\n\n\t\ttime.Sleep(time.Duration(w.pollInterval) * time.Millisecond)\n\t}\n}", "func FileExtFilterForLang(lang string) (regex, glob string) {\n\tswitch strings.ToLower(lang) {\n\tcase \"go\", \"golang\":\n\t\treturn \".*\\\\.go\", \"*.go\"\n\t}\n\treturn \".*\", \"*\"\n}", "func (f *FileCache) watchproc() {\n\tvar evt fsnotify.Event\n\tfor {\n\t\tevt = <-f.wh.Events\n\t\tif evt.Op&fsnotify.Create == fsnotify.Create {\n\t\t\tif !strings.HasSuffix(evt.Name, f.c.Suffix) {\n\t\t\t\tlog.Warn(\"create invalid file: %s\", evt.Name)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfi, err := os.Stat(evt.Name)\n\t\t\tif err != nil {\n\t\t\t\tlog.Error(\"os.Stat(%s) error(%v)\", evt.Name, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tf.eLock.Lock()\n\t\t\tf.logs[evt.Name] = fi\n\t\t\tf.eLock.Unlock()\n\t\t\tlog.Info(\"create file: %s\", evt.Name)\n\t\t}\n\t\tif evt.Op&fsnotify.Remove == fsnotify.Remove {\n\t\t\tf.eLock.Lock()\n\t\t\tdelete(f.logs, evt.Name)\n\t\t\tf.eLock.Unlock()\n\t\t\tlog.Info(\"remove file: %s\", evt.Name)\n\t\t}\n\t}\n}", "func (_Univ2 *Univ2Filterer) WatchSync(opts *bind.WatchOpts, sink chan<- *Univ2Sync) (event.Subscription, error) {\n\n\tlogs, sub, err := _Univ2.contract.WatchLogs(opts, \"Sync\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(Univ2Sync)\n\t\t\t\tif err := _Univ2.contract.UnpackLog(event, \"Sync\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (mod Mod) Filter(root string, includes []string, excludes []string) (*Mod, error) {\n\tchanged, err := filter.Files(mod.Changed, includes, excludes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdeleted, err := filter.Files(mod.Deleted, includes, excludes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tadded, err := filter.Files(mod.Added, includes, excludes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Mod{Changed: changed, Deleted: deleted, Added: added}, nil\n}", "func filterFiles(bctx build.Context, inputs []string) ([]string, error) {\n\tvar outputs []string\n\tfor _, input := range inputs {\n\t\tif match, _, _, err := matchFile(bctx, input, false); err != nil {\n\t\t\treturn nil, err\n\t\t} else if match {\n\t\t\toutputs = append(outputs, input)\n\t\t}\n\t}\n\treturn outputs, nil\n}", "func evaluateFileChanges(events []fsnotify.Event, path string, fileIgnores []string, watcher *fsnotify.Watcher) ([]string, []string) {\n\tvar changedFiles []string\n\tvar deletedPaths []string\n\n\tignoreMatcher := gitignore.CompileIgnoreLines(fileIgnores...)\n\n\tfor _, event := range events {\n\t\tklog.V(4).Infof(\"filesystem watch event: %s\", event)\n\t\tisIgnoreEvent := shouldIgnoreEvent(event)\n\n\t\t// add file name to changedFiles only once\n\t\talreadyInChangedFiles := false\n\t\tfor _, cfile := range changedFiles {\n\t\t\tif cfile == event.Name {\n\t\t\t\talreadyInChangedFiles = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\t// Filter out anything in ignores list from the list of changed files\n\t\t// This is important in spite of not watching the\n\t\t// ignores paths because, when a directory that is ignored, is deleted,\n\t\t// because its parent is watched, the fsnotify automatically raises an event\n\t\t// for it.\n\t\tvar watchError error\n\t\trel, err := filepath.Rel(path, event.Name)\n\t\tif err != nil {\n\t\t\twatchError = fmt.Errorf(\"unable to get relative path of %q on %q\", event.Name, path)\n\t\t}\n\t\tmatched := ignoreMatcher.MatchesPath(rel)\n\t\tif !alreadyInChangedFiles && !matched && !isIgnoreEvent {\n\t\t\t// Append the new file change event to changedFiles if and only if the event is not a file remove event\n\t\t\tif event.Op&fsnotify.Remove != fsnotify.Remove {\n\t\t\t\tchangedFiles = append(changedFiles, event.Name)\n\t\t\t}\n\t\t}\n\n\t\t// Rename operation triggers RENAME event on old path + CREATE event for renamed path so delete old path in case of rename\n\t\t// Also weirdly, fsnotify raises a RENAME event for deletion of files/folders with space in their name so even that should be handled here\n\t\tif event.Op&fsnotify.Remove == fsnotify.Remove || event.Op&fsnotify.Rename == fsnotify.Rename {\n\t\t\t// On remove/rename, stop watching the resource\n\t\t\tif e := watcher.Remove(event.Name); e != nil {\n\t\t\t\tklog.V(4).Infof(\"error removing watch for %s: %v\", event.Name, e)\n\t\t\t}\n\t\t\t// Append the file to list of deleted files\n\t\t\t// When a file/folder is deleted, it raises 2 events:\n\t\t\t//\ta. RENAME with event.Name empty\n\t\t\t//\tb. REMOVE with event.Name as file name\n\t\t\tif !alreadyInChangedFiles && !matched && event.Name != \"\" {\n\t\t\t\tdeletedPaths = append(deletedPaths, event.Name)\n\t\t\t}\n\t\t} else {\n\t\t\t// On other ops, recursively watch the resource (if applicable)\n\t\t\tif e := addRecursiveWatch(watcher, path, event.Name, fileIgnores); e != nil && watchError == nil {\n\t\t\t\tklog.V(4).Infof(\"Error occurred in addRecursiveWatch, setting watchError to %v\", e)\n\t\t\t\twatchError = e\n\t\t\t}\n\t\t}\n\t}\n\tdeletedPaths = removeDuplicates(deletedPaths)\n\n\treturn changedFiles, deletedPaths\n}", "func NewWorktreeFilter(Filter Filter, Dirty, Clean bool) Filter {\n\treturn predicateFilter{\n\t\tFilter: Filter,\n\t\tPredicate: func(env Env, clonePath string) bool {\n\t\t\tdirty, err := env.Git.IsDirty(clonePath)\n\t\t\treturn err == nil && dirty\n\t\t},\n\n\t\tIncludeTrue: Dirty,\n\t\tIncludeFalse: Clean,\n\t}\n}", "func (o *WatchClient) eventWatcher(\n\tctx context.Context,\n\tparameters WatchParameters,\n\tevaluateChangesHandler evaluateChangesFunc,\n\tprocessEventsHandler processEventsFunc,\n\tcomponentStatus ComponentStatus,\n) error {\n\n\tvar (\n\t\tdevfilePath = odocontext.GetDevfilePath(ctx)\n\t\tpath = filepath.Dir(devfilePath)\n\t\tcomponentName = odocontext.GetComponentName(ctx)\n\t\tappName = odocontext.GetApplication(ctx)\n\t\tout = parameters.StartOptions.Out\n\t)\n\n\tvar events []fsnotify.Event\n\n\t// sourcesTimer helps collect multiple events that happen in a quick succession. We start with 1ms as we don't care much\n\t// at this point. In the select block, however, every time we receive an event, we reset the sourcesTimer to watch for\n\t// 100ms since receiving that event. This is done because a single filesystem event by the user triggers multiple\n\t// events for fsnotify. It's a known-issue, but not really bug. For more info look at below issues:\n\t// - https://github.com/fsnotify/fsnotify/issues/122\n\t// - https://github.com/fsnotify/fsnotify/issues/344\n\tsourcesTimer := time.NewTimer(time.Millisecond)\n\t<-sourcesTimer.C\n\n\t// devfileTimer has the same usage as sourcesTimer, for file events coming from devfileWatcher\n\tdevfileTimer := time.NewTimer(time.Millisecond)\n\t<-devfileTimer.C\n\n\t// deployTimer has the same usage as sourcesTimer, for events coming from watching Deployments, from deploymentWatcher\n\tdeployTimer := time.NewTimer(time.Millisecond)\n\t<-deployTimer.C\n\n\tpodsPhases := NewPodPhases()\n\n\tfor {\n\t\tselect {\n\t\tcase event := <-o.sourcesWatcher.Events:\n\t\t\tevents = append(events, event)\n\t\t\t// We are waiting for more events in this interval\n\t\t\tsourcesTimer.Reset(100 * time.Millisecond)\n\n\t\tcase <-sourcesTimer.C:\n\t\t\t// timer has fired\n\t\t\tif !componentCanSyncFile(componentStatus.GetState()) {\n\t\t\t\tklog.V(4).Infof(\"State of component is %q, don't sync sources\", componentStatus.GetState())\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tvar changedFiles, deletedPaths []string\n\t\t\tif !o.forceSync {\n\t\t\t\t// first find the files that have changed (also includes the ones newly created) or deleted\n\t\t\t\tchangedFiles, deletedPaths = evaluateChangesHandler(events, path, parameters.StartOptions.IgnorePaths, o.sourcesWatcher)\n\t\t\t\t// process the changes and sync files with remote pod\n\t\t\t\tif len(changedFiles) == 0 && len(deletedPaths) == 0 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tcomponentStatus.SetState(StateSyncOutdated)\n\t\t\tfmt.Fprintf(out, \"Pushing files...\\n\\n\")\n\t\t\terr := processEventsHandler(ctx, parameters, changedFiles, deletedPaths, &componentStatus)\n\t\t\to.forceSync = false\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\t// empty the events to receive new events\n\t\t\tif componentStatus.GetState() == StateReady {\n\t\t\t\tevents = []fsnotify.Event{} // empty the events slice to capture new events\n\t\t\t}\n\n\t\tcase watchErr := <-o.sourcesWatcher.Errors:\n\t\t\treturn watchErr\n\n\t\tcase key := <-o.keyWatcher:\n\t\t\tif key == 'p' {\n\t\t\t\to.forceSync = true\n\t\t\t\tsourcesTimer.Reset(100 * time.Millisecond)\n\t\t\t}\n\n\t\tcase <-parameters.StartOptions.PushWatcher:\n\t\t\to.forceSync = true\n\t\t\tsourcesTimer.Reset(100 * time.Millisecond)\n\n\t\tcase ev := <-o.deploymentWatcher.ResultChan():\n\t\t\tswitch obj := ev.Object.(type) {\n\t\t\tcase *appsv1.Deployment:\n\t\t\t\tklog.V(4).Infof(\"deployment watcher Event: Type: %s, name: %s, rv: %s, generation: %d, pods: %d\\n\",\n\t\t\t\t\tev.Type, obj.GetName(), obj.GetResourceVersion(), obj.GetGeneration(), obj.Status.ReadyReplicas)\n\t\t\t\tif obj.GetGeneration() > o.deploymentGeneration || obj.Status.ReadyReplicas != o.readyReplicas {\n\t\t\t\t\to.deploymentGeneration = obj.GetGeneration()\n\t\t\t\t\to.readyReplicas = obj.Status.ReadyReplicas\n\t\t\t\t\tdeployTimer.Reset(300 * time.Millisecond)\n\t\t\t\t}\n\n\t\t\tcase *metav1.Status:\n\t\t\t\tklog.V(4).Infof(\"Status: %+v\\n\", obj)\n\t\t\t}\n\n\t\tcase <-deployTimer.C:\n\t\t\terr := processEventsHandler(ctx, parameters, nil, nil, &componentStatus)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\tcase <-o.devfileWatcher.Events:\n\t\t\tdevfileTimer.Reset(100 * time.Millisecond)\n\n\t\tcase <-devfileTimer.C:\n\t\t\tfmt.Fprintf(out, \"Updating Component...\\n\\n\")\n\t\t\terr := processEventsHandler(ctx, parameters, nil, nil, &componentStatus)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\tcase ev := <-o.podWatcher.ResultChan():\n\t\t\tswitch ev.Type {\n\t\t\tcase watch.Deleted:\n\t\t\t\tpod, ok := ev.Object.(*corev1.Pod)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn errors.New(\"unable to decode watch event\")\n\t\t\t\t}\n\t\t\t\tpodsPhases.Delete(out, pod)\n\t\t\tcase watch.Added, watch.Modified:\n\t\t\t\tpod, ok := ev.Object.(*corev1.Pod)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn errors.New(\"unable to decode watch event\")\n\t\t\t\t}\n\t\t\t\tpodsPhases.Add(out, pod.GetCreationTimestamp(), pod)\n\t\t\t}\n\n\t\tcase ev := <-o.warningsWatcher.ResultChan():\n\t\t\tswitch kevent := ev.Object.(type) {\n\t\t\tcase *corev1.Event:\n\t\t\t\tpodName := kevent.InvolvedObject.Name\n\t\t\t\tselector := labels.GetSelector(componentName, appName, labels.ComponentDevMode, true)\n\t\t\t\tmatching, err := o.kubeClient.IsPodNameMatchingSelector(ctx, podName, selector)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif matching {\n\t\t\t\t\tlog.Fwarning(out, kevent.Message)\n\t\t\t\t}\n\t\t\t}\n\n\t\tcase watchErr := <-o.devfileWatcher.Errors:\n\t\t\treturn watchErr\n\n\t\tcase <-ctx.Done():\n\t\t\tklog.V(2).Info(\"Dev mode interrupted by user\")\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func TestWatcher() {\n\tw, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tlog.Fatalf(\"Error creating Watcher: %s\", err.Error())\n\t}\n\tdefer w.Close()\n\tw.Add(\"/media/peza\")\n\n\tvar wg sync.WaitGroup\n\twg.Add(1)\n\tgo func() {\n\t\tdefer wg.Done()\n\t\tlog.Println(\"Starting watch...\")\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase ev := <-w.Events:\n\t\t\t\tlog.Printf(\"Event: %s, %d\", ev.Name, ev.Op)\n\t\t\t\t// case err := <- w.Errors:\n\t\t\t}\n\t\t}\n\n\t}()\n\twg.Wait()\n}", "func watcher(configModel model.Config) {\n\t// Set the client variable\n\tconfig.Client = configModel.Client.Name\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer watcher.Close()\n\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event := <-watcher.Events:\n\t\t\t\tif event.Op&fsnotify.Write == fsnotify.Write {\n\t\t\t\t\tlogs.INFO.Println(\"Modified file -> \", event.Name)\n\t\t\t\t\t// When the file name has not been defined, it is time to\n\t\t\t\t\t// use the SetFile() method to add a new file to read.\n\t\t\t\t\tif filename == \"\" {\n\t\t\t\t\t\tstore.SetFile(event.Name)\n\t\t\t\t\t\tfilename = event.Name\n\t\t\t\t\t}\n\t\t\t\t\tif filename != \"\" && filename != event.Name {\n\t\t\t\t\t\tlogs.INFO.Println(\"Reset seek\")\n\t\t\t\t\t\tseek = 0\n\t\t\t\t\t}\n\t\t\t\t\treadLines(event.Name)\n\t\t\t\t}\n\t\t\tcase err := <-watcher.Errors:\n\t\t\t\tlogs.CRITICAL.Println(\"Error on watcher: \", err)\n\t\t\t}\n\t\t}\n\t}()\n\terr = watcher.Add(configModel.Pathlog.Name)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\t<-done\n}", "func FileViewDirOnlyFilter(fv *FileView, fi *FileInfo) bool {\n\treturn fi.IsDir()\n}", "func FileViewExtOnlyFilter(fv *FileView, fi *FileInfo) bool {\n\tif fi.IsDir() {\n\t\treturn true\n\t}\n\text := strings.ToLower(filepath.Ext(fi.Name))\n\t_, has := fv.ExtMap[ext]\n\treturn has\n}", "func (g *Gonf) WatchFile() bool {\n\treturn g.watchFile\n}", "func (ms Mounts) Filter(f FilterFunc) Mounts {\n\tfiltered := make([]*Mount, 0, len(ms))\n\n\tfor _, m := range ms {\n\t\tif f(m) {\n\t\t\tfiltered = append(filtered, m)\n\t\t}\n\t}\n\n\treturn Mounts(filtered)\n}", "func (c Configuration) getListOfFiles(fs afero.Fs, pathList []string) []FileInfo {\n\tlogger := c.logger()\n\tvar filesToMonitor []FileInfo\n\tcompleteListOfPaths := c.getCompleteListOfPaths(pathList)\n\n\tfor _, fullPath := range completeListOfPaths {\n\t\tfullPath := fullPath\n\t\tpkgFile := pkg.NewFile(func(file *pkg.File) {\n\t\t\tfile.Fs, file.Path, file.Logger = fs, fullPath, logger\n\t\t})\n\n\t\tPathFull := \"\"\n\t\tif baseFile, ok := pkgFile.Fs.(*afero.BasePathFs); ok {\n\t\t\tPathFull, _ = baseFile.RealPath(fullPath)\n\t\t}\n\t\tif PathFull == \"\" {\n\t\t\tPathFull = fullPath\n\t\t}\n\t\tlogger.Debug().Msgf(\"file to watch: %v\", PathFull)\n\t\tPathFull, fi := c.resolvePath(PathFull)\n\t\tif PathFull == \"\" {\n\t\t\tcontinue // could not resolve the file. skip for now.\n\t\t}\n\n\t\tswitch mode := fi.Mode(); {\n\t\tcase mode.IsDir():\n\t\t\tlogger.Debug().Msg(\"Path is a dir\")\n\t\t\terr := filepath.Walk(PathFull, func(path string, info os.FileInfo, err error) error {\n\t\t\t\twalkPath, resolvedInfo := c.resolvePath(path)\n\t\t\t\tif walkPath == \"\" {\n\t\t\t\t\treturn nil // path could not be resolved skip for now\n\t\t\t\t}\n\t\t\t\tisDir := resolvedInfo.IsDir()\n\t\t\t\tlogger.Debug().Msgf(\"Path: %v\", path)\n\t\t\t\tfilesToMonitor = append(filesToMonitor, FileInfo{File: path, IsDir: isDir})\n\t\t\t\treturn nil\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\tlogger.Error().Err(err).Msgf(\"error walking dir: %v\", PathFull)\n\t\t\t}\n\t\tcase mode.IsRegular():\n\t\t\tlogger.Debug().Msg(\"Path is a file\")\n\t\t\tlogger.Debug().Msgf(\"Path: %v\", PathFull)\n\t\t\tfilesToMonitor = append(filesToMonitor, FileInfo{File: PathFull, IsDir: false})\n\t\tdefault:\n\t\t\tlogger.Debug().Msg(\"Path is a dir\")\n\t\t}\n\t}\n\treturn filesToMonitor\n}", "func NewWatcher(bufsize, sysBufSize int, sleepTime time.Duration, fn func([]*WatchEvent),\n) (w *Watcher, err error) {\n\tfd, err := syscall.InotifyInit()\n\tif err != nil {\n\t\treturn\n\t}\n\tif fd == -1 {\n\t\terr = os.NewSyscallError(\"inotify_init\", err)\n\t\treturn\n\t}\n\tif useNonBlock {\n\t\tsyscall.SetNonblock(fd, true)\n\t}\n\tw = &Watcher{\n\t\tfd: fd,\n\t\tfn: fn,\n\t\tev: make(chan []*WatchEvent, bufsize),\n\t\twds: make(map[int32]string),\n\t\tflags: make(map[string]uint32),\n\t\tsl: sleepTime,\n\t\tsysbufsize: sysBufSize,\n\t}\n\tgo w.readEvents()\n\tgo w.handleEvents()\n\treturn\n}", "func Watch(s Selector, a Action) (*Watcher, error) {\n\tfsw, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tw := &Watcher{\n\t\tfsw: fsw,\n\t\tcache: make(map[string]string),\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event, ok := <-w.fsw.Events:\n\t\t\t\tif !ok {\n\t\t\t\t\tlog.Println(\"There was an error in an event consumer [events].\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\te := Event{event}\n\t\t\t\tcached := w.getCache(e.Name)\n\t\t\t\tw.setCache(e.Name, a(e, cached))\n\t\t\tcase err, ok := <-w.fsw.Errors:\n\t\t\t\tif !ok {\n\t\t\t\t\tlog.Println(\"There was an error in an event consumer [errs].\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tlog.Println(\"error:\", err)\n\t\t\t}\n\t\t}\n\t}()\n\n\tfor _, name := range s() {\n\t\terr = w.fsw.Add(name)\n\t}\n\n\treturn w, err\n}", "func waitWatchEvent(realPath string, op fsnotify.Op, lock *sync.RWMutex) (event fsnotify.Event, err error) {\n\tvar (\n\t\tdir string\n\t\twatcher *fsnotify.Watcher\n\t\tok bool\n\t)\n\n\tdir = filepath.Dir(realPath)\n\t// one dir on watcher, use the existied watcher.\n\tfunc() {\n\t\tlock.Lock()\n\t\tdefer lock.Unlock()\n\t\tif watcher, ok = mapWatcher[dir]; !ok {\n\t\t\tif watcher, err = fsnotify.NewWatcher(); err != nil {\n\t\t\t\terr = errors.New(\"fsnotify create new watcher failed: \" + dir)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif err = watcher.Add(dir); err != nil {\n\t\t\t\terr = errors.New(\"add new watch path failed: \" + dir)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tmapWatcher[dir] = watcher\n\t\t}\n\t}()\n\n\tfor {\n\t\tselect {\n\t\tcase event = <-watcher.Events:\n\t\t\tfmt.Println(event)\n\t\t\tif event.Name == realPath {\n\t\t\t\tif op > 0 {\n\t\t\t\t\t// if this is create or write event\n\t\t\t\t\tif event.Op&op > 0 {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// any type of event\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\tcase err = <-watcher.Errors:\n\t\t\terr = errors.New(\"watcher error \" + realPath)\n\t\t\treturn\n\t\t}\n\t}\n}", "func getFilteredFiles(folder string) ([]model.File, error) {\n\tfiles := make([]model.File, 0)\n\tif filesInFolder, err := helper.GetFiles(folder, isRecursive); err == nil {\n\t\tfor _, file := range filesInFolder {\n\t\t\tif prefix != \"\" && !strings.HasPrefix(file.FileName, prefix) {\n\t\t\t\tcontinue\n\t\t\t} else if suffix != \"\" && !strings.HasSuffix(file.FileName, suffix) {\n\t\t\t\tcontinue\n\t\t\t} else if regexPattern != \"\" && !regexp.MustCompile(regexPattern).MatchString(file.FileName) {\n\t\t\t\tcontinue\n\t\t\t} else if extension != \"\" && strings.TrimSuffix(file.Ext, extension) != \".\" {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tfiles = append(files, file)\n\t\t}\n\t} else {\n\t\treturn files, err\n\t}\n\n\treturn files, nil\n}", "func (o FolderSinkOutput) Filter() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FolderSink) pulumi.StringPtrOutput { return v.Filter }).(pulumi.StringPtrOutput)\n}", "func (w *Watcher) pollEvents(currFileList map[string]os.FileInfo) {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tcreates := make(map[string]os.FileInfo)\n\tremoves := make(map[string]os.FileInfo)\n\n\t// check for remove\n\tfor latestFp, latestFi := range w.files {\n\t\tif _, ok := currFileList[latestFp]; !ok {\n\t\t\tremoves[latestFp] = latestFi\n\t\t}\n\t}\n\n\t// check for create / modify / chmod\n\tfor fp, currFi := range currFileList {\n\t\tlatestFi, ok := w.files[fp]\n\t\tif !ok {\n\t\t\t// create\n\t\t\tcreates[fp] = currFi\n\t\t\tcontinue\n\t\t}\n\n\t\t// ModTime may return timestamp in second level on some file system\n\t\t// So use ModTime + Size to judge modify event will be more precisely\n\t\tif !latestFi.ModTime().Equal(currFi.ModTime()) ||\n\t\t\tlatestFi.Size() != currFi.Size() {\n\t\t\t// modify\n\t\t\tselect {\n\t\t\tcase <-w.closed:\n\t\t\t\treturn\n\t\t\tcase w.Events <- Event{Path: fp, Op: Modify, FileInfo: currFi}:\n\t\t\t}\n\t\t}\n\n\t\tif latestFi.Mode() != currFi.Mode() {\n\t\t\t// chmod\n\t\t\tselect {\n\t\t\tcase <-w.closed:\n\t\t\t\treturn\n\t\t\tcase w.Events <- Event{Path: fp, Op: Chmod, FileInfo: currFi}:\n\t\t\t}\n\t\t}\n\t}\n\n\t// check for rename / move\n\tfor removeFp, removeFi := range removes {\n\t\tfor createFp, createFi := range creates {\n\t\t\tif os.SameFile(removeFi, createFi) {\n\t\t\t\tev := Event{\n\t\t\t\t\tPath: removeFp, // for Move, use from-path\n\t\t\t\t\tOp: Move,\n\t\t\t\t\tFileInfo: removeFi,\n\t\t\t\t}\n\t\t\t\tif filepath.Dir(removeFp) == filepath.Dir(createFp) {\n\t\t\t\t\tev.Op = Rename\n\t\t\t\t}\n\n\t\t\t\tdelete(removes, removeFp)\n\t\t\t\tdelete(creates, createFp)\n\n\t\t\t\tselect {\n\t\t\t\tcase <-w.closed:\n\t\t\t\t\treturn\n\t\t\t\tcase w.Events <- ev:\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// send events for create\n\tfor fp, fi := range creates {\n\t\tselect {\n\t\tcase <-w.closed:\n\t\t\treturn\n\t\tcase w.Events <- Event{Path: fp, Op: Create, FileInfo: fi}:\n\t\t}\n\t}\n\n\t// send events for remove\n\tfor fp, fi := range removes {\n\t\tselect {\n\t\tcase <-w.closed:\n\t\t\treturn\n\t\tcase w.Events <- Event{Path: fp, Op: Remove, FileInfo: fi}:\n\t\t}\n\t}\n}", "func watchForNewFiles(watchPath string, fileName chan string) {\n\n\tlogrus.Debugf(\"Watching for changes at %s\", watchPath)\n\n\t// if the folder doesnt exist, create it\n\tif _, err := os.Stat(watchPath); os.IsNotExist(err) {\n\t\tos.MkdirAll(watchPath, 0777)\n\t}\n\n\t// Creates a new file watcher\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\tlogrus.Errorf(\"Error %s\", err)\n\t}\n\tdefer watcher.Close()\n\n\t// adds the path as parameter to be watched\n\tif err := watcher.Add(watchPath); err != nil {\n\t\tlogrus.Errorf(\"Error adding folder to watch. Folder: %s. Error details: %s\", watchPath, err)\n\t\treturn\n\t}\n\n\t// Watch folder loop\n\tfor {\n\t\tselect {\n\t\t// watch for events fired on the folder watch loop\n\t\tcase event := <-watcher.Events:\n\t\t\t// send the file name only if the detected change was a\n\t\t\t// file creation\n\t\t\tif event.Op == fsnotify.Create || event.Op == fsnotify.Rename || event.Op == fsnotify.Write {\n\t\t\t\tfileName <- event.Name\n\t\t\t}\n\n\t\t// watch for errors\n\t\tcase err := <-watcher.Errors:\n\t\t\tlogrus.Errorf(\"Error on watching folder/path. Path: %s. Error: %s\", watchPath, err)\n\t\t}\n\t}\n\n}", "func (watcher *Watcher) Start() error {\n\t// fsnotify is our secret sauce here.\n\t// NewWatcher() starts a monitor for\n\t// FS activities. It creates two channels:\n\t// `w.Events` and `w.Errors` for normal\n\t// events and errenous events.\n\tw, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// really cool go feature. Now, we never keep to\n\t// worry about forgetting to close the websocket\n\t// connections.\n\tdefer w.Close()\n\n\t// this channel just monitors the exit condition\n\t// of the goroutine that we are about to start.\n\tdone := make(chan bool)\n\n\t// here we go. We are starting the real-time\n\t// content authoring magic here.\n\tgo func() {\n\t\t// a blocking loop reading the fs events\n\t\t// from the fsnotify channels.\n\t\t// I manually convert the type of events\n\t\t// to a string `method`.\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase ev := <-w.Events:\n\t\t\t\tvar method string\n\t\t\t\tswitch {\n\t\t\t\tcase ev.Op&fsnotify.Create > 0:\n\t\t\t\t\t// a file is created\n\t\t\t\t\tmethod = \"Create\"\n\t\t\t\tcase ev.Op&fsnotify.Write > 0:\n\t\t\t\t\t// a file is updated\n\t\t\t\t\tmethod = \"Write\"\n\t\t\t\tcase ev.Op&fsnotify.Remove > 0:\n\t\t\t\t\t// a file is removed\n\t\t\t\t\tmethod = \"Remove\"\n\t\t\t\tcase ev.Op&fsnotify.Rename > 0:\n\t\t\t\t\t// a file has been renamed\n\t\t\t\t\t// to a new name\n\t\t\t\t\tmethod = \"Rename\"\n\t\t\t\tcase ev.Op&fsnotify.Chmod > 0:\n\t\t\t\t\t// the modtime is changed\n\t\t\t\t\tmethod = \"Chmod\"\n\t\t\t\tdefault:\n\t\t\t\t\tmethod = \"Unknown\"\n\t\t\t\t}\n\t\t\t\t// notify the client the filename and\n\t\t\t\t// the type of event through the\n\t\t\t\t// websocket\n\t\t\t\tif notify(watcher.Conn, ev.Name, method) != nil {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\tcase err = <-w.Errors:\n\t\t\t\t// when we see a fs error, we just break\n\t\t\t\t// out of the loop.\n\t\t\t\t// I am not sure if there is what we\n\t\t\t\t// should do.\n\t\t\t\tlog.Printf(\"Modified: %s\", err.Error())\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\t// before exiting the goroutine, we need to\n\t\t// inform the parent that we are done.\n\t\tdone <- true\n\t}()\n\n\t// cool magic here: tell fsnotify to start\n\t// monitoring the directory that is the workspace\n\terr = w.Add(watcher.Dir)\n\tif err != nil {\n\t\tlog.Printf(\"[Watcher] %s\", err.Error())\n\t\treturn err\n\t}\n\n\t// block until the goroutine is done\n\t<-done\n\n\t// return. The deferred will close the websocket\n\t// for us.\n\treturn nil\n}", "func Watch(paths ...string) (*Watcher, error) {\n\tevent := make(chan EventItem)\n\terr := watch(paths, event)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Watcher{\n\t\tEvent: event,\n\t}, nil\n}", "func (w *Watcher) Watch() {\n\tfor {\n\t\tselect {\n\t\tcase ev := <-w.watcher.Event:\n\t\t\tfor _, handler := range w.modifiedHandlers {\n\t\t\t\tif strings.HasPrefix(ev.Name, handler.path) {\n\t\t\t\t\tfmt.Println(handler)\n\t\t\t\t\thandler.callback(ev.Name)\n\t\t\t\t}\n\t\t\t}\n\t\t\tlog.Println(\"event:\", ev)\n\t\t\tlog.Println(\"handlers:\", w.modifiedHandlers)\n\t\t\t//case addreq :=\n\t\tcase err := <-w.watcher.Error:\n\t\t\tlog.Println(\"error:\", err)\n\t\t}\n\t}\n}", "func FilterPaths(fileInfo os.FileInfo) bool {\n\tname := fileInfo.Name()\n\tfor _, pattern := range IgnoredPaths {\n\t\tmatch, _ := filepath.Match(pattern, name)\n\t\tif match {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (w *FileWatch) StartWatcher() {\n\tlog.Trace(\">>>>> StartWatcher\")\n\tdefer log.Trace(\"<<<<< StartWatcher\")\n\tpid := os.Getpid()\n\tlog.Tracef(\"Watcher [%d PID] is successfully started\", pid)\n\n\t// Control the ticker interval, dont want to frequently wakeup\n\t// watcher as it is only needed when there is event notification. So if there is\n\t// event notification, ticker is set to wake up every one minute otherwise sleep\n\t// for 1 hour.\n\tvar delayControlFlag time.Duration = tickerDefaultDelay\n\n\t// This is used to control the flow of events, we dont want to process frequent update\n\t// If there are multiple update within 1 min, only process one event and ignore the rest of the events\n\tisSpuriousUpdate:=false\n\t// forever\n\tfor {\n\t\tselect {\n\t\tcase <-w.watchStop:\n\t\t\tlog.Infof(\"Stopping [%d PID ] csi watcher\", pid)\n\t\t\tw.wg.Done()\n\t\t\tw.watchList.Close()\n\t\t\treturn\n\t\tcase <-w.watchList.Events:\n\t\t\t// There might be spurious update, ignore the event if it occurs within 1 min.\n\t\t\tif !isSpuriousUpdate {\n\t\t\t\tlog.Infof(\"Watcher [%d PID], received notification\", pid)\n\t\t\t\tw.watchRun()\n\t\t\t\tlog.Infof(\"Watcher [%d PID], notification served\", pid)\n\t\t\t\tisSpuriousUpdate = true\n\t\t\t\tdelayControlFlag = 1\n\t\t\t} else {\n\t\t\t\tlog.Warnf(\"Watcher [%d PID], received spurious notification, ignore\", pid)\n\t\t\t}\n\t\tcase <-time.NewTicker(time.Minute * delayControlFlag).C:\n\t\t\tisSpuriousUpdate = false\n\t\t\tdelayControlFlag = tickerDefaultDelay\n\t\t}\n\t}\n}", "func NewWatcher(cfg WatcherConfig) (Watcher, error) {\n\tif cfg.PollInterval == 0 {\n\t\tcfg.PollInterval = DefaultPoolInterval\n\t}\n\n\tif len(cfg.Extensions) == 0 {\n\t\tcfg.Extensions = DefaultExtensions\n\t}\n\n\tallowedExts := make(map[string]bool)\n\tfor _, ext := range cfg.Extensions {\n\t\tallowedExts[\".\"+ext] = true\n\t}\n\n\twatchPaths, err := resolvePaths(cfg.WatchItems, allowedExts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tignorePaths, err := resolvePaths(cfg.IgnoreItems, allowedExts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlogger.Debugf(\"Resolved watch paths: %v\", watchPaths)\n\tlogger.Debugf(\"Resolved ignore paths: %v\", ignorePaths)\n\treturn &watcher{\n\t\tevents: make(chan string),\n\t\terrors: make(chan error),\n\t\tdefaultIgnore: cfg.DefaultIgnore,\n\t\tpollInterval: cfg.PollInterval,\n\t\twatchItems: watchPaths,\n\t\tignoreItems: ignorePaths,\n\t\tallowedExtensions: allowedExts,\n\t}, nil\n}", "func (_BaseAccessWallet *BaseAccessWalletFilterer) WatchContentTypeRemoved(opts *bind.WatchOpts, sink chan<- *BaseAccessWalletContentTypeRemoved) (event.Subscription, error) {\n\n\tlogs, sub, err := _BaseAccessWallet.contract.WatchLogs(opts, \"ContentTypeRemoved\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(BaseAccessWalletContentTypeRemoved)\n\t\t\t\tif err := _BaseAccessWallet.contract.UnpackLog(event, \"ContentTypeRemoved\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (e *EventLog) getFilesWithMatchDatePattern() []string {\n\tvar validFileNames []string\n\tif allFiles, err := e.fileSystem.ReadDir(e.eventLogPath); err == nil {\n\t\tfor _, fileInfo := range allFiles {\n\t\t\tfileName := fileInfo.Name()\n\t\t\tif !fileInfo.Mode().IsDir() && e.isValidFileName(fileName) {\n\t\t\t\tvalidFileNames = append(validFileNames, fileName)\n\t\t\t}\n\t\t}\n\t}\n\treturn validFileNames\n}", "func (node *Node) checkFile(name, path string, filters []string) {\n\n\tinfo, err := os.Lstat(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t// check if the file is a directory\n\tif info.Mode() & os.ModeSymlink != 0 {\n\t\tlinked, err := os.Readlink(path)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\ts := strings.Split(path, \"/\")\n\t\tpath = strings.Join(\n\t\t\tappend(\n\t\t\t\ts[:len(s)-1],\n\t\t\t\tlinked,\n\t\t\t),\n\t\t\t\"/\",\n\t\t)\n\t\tinfo, err = os.Lstat(path)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\t// check if path is a directory\n\tif info.IsDir() {\n\t\tnode.Add(name).StaticFolder(path)\n\t\treturn\n\t}\n\n\tfor _, filter := range filters {\n\t\tif !strings.Contains(name, filter) { return }\n\t}\n\n\tnode.Add(name).File(path)\n\n}", "func watchFile(ctx context.Context, cancel func()) error {\n\tn, err := os.Executable()\n\tif err != nil {\n\t\treturn err\n\t}\n\tfi, err := os.Stat(n)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmod0 := fi.ModTime()\n\tw, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err = w.Add(n); err != nil {\n\t\treturn err\n\t}\n\tgo func() {\n\t\tdefer w.Close()\n\t\tdone := ctx.Done()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-done:\n\t\t\t\treturn\n\t\t\tcase err := <-w.Errors:\n\t\t\t\tlog.Printf(\"watching %s failed: %v\", n, err)\n\t\t\t\treturn\n\t\t\tcase <-w.Events:\n\t\t\t\tif fi, err = os.Stat(n); err != nil || !fi.ModTime().Equal(mod0) {\n\t\t\t\t\tlog.Printf(\"%s was modified, exiting\", n)\n\t\t\t\t\tcancel()\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\treturn nil\n}", "func (_BaseLibrary *BaseLibraryFilterer) WatchContentTypeRemoved(opts *bind.WatchOpts, sink chan<- *BaseLibraryContentTypeRemoved) (event.Subscription, error) {\n\n\tlogs, sub, err := _BaseLibrary.contract.WatchLogs(opts, \"ContentTypeRemoved\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn event.NewSubscription(func(quit <-chan struct{}) error {\n\t\tdefer sub.Unsubscribe()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase log := <-logs:\n\t\t\t\t// New log arrived, parse the event and forward to the user\n\t\t\t\tevent := new(BaseLibraryContentTypeRemoved)\n\t\t\t\tif err := _BaseLibrary.contract.UnpackLog(event, \"ContentTypeRemoved\", log); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tevent.Raw = log\n\n\t\t\t\tselect {\n\t\t\t\tcase sink <- event:\n\t\t\t\tcase err := <-sub.Err():\n\t\t\t\t\treturn err\n\t\t\t\tcase <-quit:\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\tcase err := <-sub.Err():\n\t\t\t\treturn err\n\t\t\tcase <-quit:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\t}), nil\n}", "func (c *Catalog) SimpleWatcher() eirinix.Watcher {\n\treturn &SimpleWatch{}\n}", "func NewFileFilter(t mockConstructorTestingTNewFileFilter) *FileFilter {\n\tmock := &FileFilter{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func startfileWatcherDemo(ch chan bool) {\r\n\tch <- true\r\n\tfor {\r\n\t\td, _ := os.Open(watchedPath)\r\n\t\tfiles, _ := d.Readdir(-1)\r\n\t\tfor _, fi := range files {\r\n\t\t\tfilePath := watchedPath + \"/\" + fi.Name()\r\n\t\t\tf, _ := os.Open(filePath)\r\n\t\t\tdata, _ := ioutil.ReadAll(f)\r\n\t\t\tf.Close()\r\n\t\t\tif data != nil {\r\n\t\t\t\tos.Remove(filePath)\r\n\t\t\t}\r\n\r\n\t\t\tgo func(data string) {\r\n\t\t\t\tfmt.Println(\"Record proccessed - \", data)\r\n\t\t\t}(string(data))\r\n\t\t}\r\n\t}\r\n}", "func DirWatcher() {\n\terr := helpers.CheckDir(HomeDir)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tgoto initwatcher\n\t} else {\n\t\tfmt.Println(\"Creating Directory :\" + HomeDir)\n\t\tos.Mkdir(HomeDir, 0700)\n\t\tos.Mkdir(HomeDir+\"/tmp\", 0700)\n\t\tos.Mkdir(HomeDir+\"/meta\", 0700)\n\t\tos.Mkdir(HomeDir+\"/blocks\", 0700)\n\t\tgoto initwatcher\n\t}\n\ninitwatcher:\n\tlog.Println(\"Adding watcher to directory\")\n\twatcher, err := fsnotify.NewWatcher()\n\tdefer watcher.Close()\n\tdone := make(chan bool)\n\n\tif err != nil {\n\t\tlog.Println(\"Error Occured in Watcher\")\n\t}\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase event, ok := <-watcher.Events:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\thandleAction(event)\n\n\t\t\tcase err, ok := <-watcher.Errors:\n\t\t\t\tif !ok {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tlog.Println(\"error:\", err)\n\t\t\t}\n\t\t}\n\t}()\n\n\terr = watcher.Add(HomeDir)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\t<-done\n\n}", "func NewFileSystemWatch(files []string, ev func(fsnotify.Event), errs func(error)) *FileSystemWatch {\n\treturn &FileSystemWatch{\n\t\tfiles: files,\n\t\tevents: ev,\n\t\terrors: errs,\n\t}\n}", "func (s *VolumeStore) filter(vols []volume.Volume, f filterFunc) []volume.Volume {\n\tvar ls []volume.Volume\n\tfor _, v := range vols {\n\t\tif f(v) {\n\t\t\tls = append(ls, v)\n\t\t}\n\t}\n\treturn ls\n}", "func (e EventTypeFilter) Filter(msg bus.Message) error {\n\tfor _, eventType := range e.AllowedEventTypes {\n\t\tif eventType == msg.Event().Type {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\treturn dispatcher.ErrNack\n}", "func SafeFilter(hdr *tar.Header) bool {\n\tif hdr.Typeflag == tar.TypeDir {\n\t\thdr.Mode = 0o770\n\t\treturn true\n\t}\n\tif hdr.Typeflag == tar.TypeReg {\n\t\thdr.Mode = 0o660\n\t\treturn true\n\t}\n\treturn false\n}", "func (uc *Userclient) iMonitorLocalChanges() {\r\n\tfor {\r\n\t\tselect {\r\n\t\tcase ev := <-uc.watcher.Event:\r\n\t\t\tif ev.IsCreate() {\r\n\t\t\t\t//Making the key at which to store the Syncfile.\r\n\t\t\t\t//First, take the path (given by watcher.Event) and parse out the /'s to replace with ?'s\r\n\t\t\t\tpatharray := strings.Split(ev.Name, \"/\")\r\n\t\t\t\tfoundWhite := false\r\n\t\t\t\t//if for some resaons Whiteboard is in the path, take it out and all prefix\r\n\t\t\t\tvar i int\r\n\t\t\t\tfor i = 0; foundWhite == false && i < len(patharray); i++ {\r\n\t\t\t\t\tif patharray[i] == \"whiteboard\" {\r\n\t\t\t\t\t\tfoundWhite = true\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\t\t\t\t//If it wasn't...reset i\r\n\t\t\t\tif foundWhite == false {\r\n\t\t\t\t\ti = 0\r\n\t\t\t\t}\r\n\t\t\t\t// The key is user:class?path to file (delimited by ?'s')\r\n\t\t\t\tintermed := strings.Join(patharray[i:], \"?\")\r\n\t\t\t\tclass := patharray[i]\r\n\t\t\t\tif class == \".permkey\" {\r\n\t\t\t\t\tbreak\r\n\t\t\t\t}\r\n\t\t\t\tfmt.Println(\"Class \", class)\r\n\t\t\t\tkey := fmt.Sprintf(\"%v:%v\", uc.user.Username, intermed)\r\n\r\n\t\t\t\t//Changing directory madness starts here\r\n\t\t\t\t//cos apparently you have to be ni that directory (parameter is filename, not filepath)\r\n\t\t\t\t<-uc.wdChangeMutex\r\n\t\t\t\tpwd, _ := os.Getwd()\r\n\t\t\t\tfmt.Printf(\"WD %v for %v\\n\", pwd, ev.Name)\r\n\t\t\t\tfor j := i; j < len(patharray)-1; j++ {\r\n\t\t\t\t\tcdErr := os.Chdir(patharray[j])\r\n\t\t\t\t\tif cdErr != nil {\r\n\t\t\t\t\t\tfmt.Println(\"Couldn't cd into \", patharray[j])\r\n\t\t\t\t\t\tbreak\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\r\n\t\t\t\tfile, fileErr := os.Open(patharray[len(patharray)-1])\r\n\t\t\t\tif fileErr != nil {\r\n\t\t\t\t\tfmt.Println(\"Fail \", patharray[len(patharray)-1])\r\n\t\t\t\t\tbreak\r\n\t\t\t\t}\r\n\r\n\t\t\t\texistSyncFile := uc.iGet(key)\r\n\t\t\t\tvar permissions map[string]int\r\n\r\n\t\t\t\t//if it already exists, we're just copying it over from server (don't overwrite server)\r\n\t\t\t\tif existSyncFile == nil { //if it doesn't exist\r\n\t\t\t\t\t//Make the sync file to store\r\n\t\t\t\t\tfi, statErr := file.Stat()\r\n\t\t\t\t\tif statErr != nil {\r\n\t\t\t\t\t\tbreak\r\n\t\t\t\t\t}\r\n\r\n\t\t\t\t\tpermissions = make(map[string]int)\r\n\t\t\t\t\tpermissions[uc.user.Username] = storageproto.WRITE\r\n\t\t\t\t\tvar files []string\r\n\t\t\t\t\tif fi.IsDir() {\r\n\t\t\t\t\t\tfiles = []string{}\r\n\t\t\t\t\t} else {\r\n\t\t\t\t\t\tfiles = nil\r\n\t\t\t\t\t}\r\n\r\n\t\t\t\t\t//Finding content of file\r\n\t\t\t\t\tbuffer := bytes.NewBuffer(make([]byte, 0))\r\n\t\t\t\t\t<-uc.permkeyFileMutex\r\n\t\t\t\t\treader := bufio.NewReader(file)\r\n\t\t\t\t\tfor {\r\n\t\t\t\t\t\tline, _, readErr := reader.ReadLine()\r\n\t\t\t\t\t\tif readErr != nil {\r\n\t\t\t\t\t\t\tbreak\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\tbuffer.Write(line)\r\n\r\n\t\t\t\t\t}\r\n\t\t\t\t\tuc.permkeyFileMutex <- 1\r\n\r\n\t\t\t\t\tsyncFile := &storageproto.SyncFile{Owner: uc.user, Class: class, UpdateTime: time.Now().Nanosecond(), Contents: buffer.Bytes(), Files: files, Permissions: permissions, Synced: true}\r\n\r\n\t\t\t\t\t//Give to midclient\r\n\t\t\t\t\tuc.iPush(key, syncFile)\r\n\t\t\t\t\t//Get parent to add to Files list\r\n\t\t\t\t\tparentfilepath := uc.homedir + \"/\" + strings.Join(patharray[:(len(patharray)-1)], \"/\")\r\n\t\t\t\t\t<-uc.fileKeyMutex\r\n\t\t\t\t\t//parent has to be in map...otherwise this would make literally 0 sense\r\n\t\t\t\t\tparentkeyperm, _ := uc.fileKeyMap[parentfilepath]\r\n\t\t\t\t\tfmt.Println(\"My parent: \", parentfilepath)\r\n\t\t\t\t\tuc.fileKeyMutex <- 1\r\n\t\t\t\t\tparentSync := uc.iGet(parentkeyperm.Key)\r\n\t\t\t\t\tif parentSync != nil {\r\n\t\t\t\t\t\tparentSync.Files = append(parentSync.Files, key)\r\n\t\t\t\t\t\tuc.iPush(parentkeyperm.Key, parentSync)\r\n\t\t\t\t\t}\r\n\t\t\t\t} else {\r\n\t\t\t\t\tpermissions = existSyncFile.Permissions\r\n\t\t\t\t}\r\n\t\t\t\t//cd back to whiteboard/\r\n\t\t\t\tfor j := i; j < len(patharray)-1; j++ {\r\n\t\t\t\t\tos.Chdir(\"..\")\r\n\t\t\t\t}\r\n\t\t\t\tfile.Close()\r\n\r\n\t\t\t\t//also add to file\r\n\t\t\t\t<-uc.permkeyFileMutex\r\n\t\t\t\ttoFile := fmt.Sprintf(\"%v %v %v:%v\\n\", ev.Name, key, uc.user.Username, storageproto.WRITE)\r\n\t\t\t\t_, writeErr := uc.permkeyFile.WriteString(toFile)\r\n\t\t\t\tif writeErr != nil {\r\n\t\t\t\t\tfmt.Println(\"Write error!\")\r\n\t\t\t\t}\r\n\t\t\t\tuc.permkeyFileMutex <- 1\r\n\r\n\t\t\t\t//Hash for easy access later\r\n\t\t\t\tuc.fileKeyMap[ev.Name] = KeyPermissions{key, permissions}\r\n\r\n\t\t\t\tuc.wdChangeMutex <- 1\r\n\t\t\t\tfmt.Printf(\"Create Event: %v\\n\", ev)\r\n\t\t\t} else if ev.IsModify() {\r\n\r\n\t\t\t} else if ev.IsDelete() {\r\n\t\t\t\tfmt.Printf(\"Delete Event: %v\\n\", ev)\r\n\t\t\t} else if ev.IsRename() {\r\n\t\t\t\tfmt.Printf(\"Rename Event: %v\\n\", ev)\r\n\t\t\t}\r\n\t\tcase err := <-uc.watcher.Error:\r\n\t\t\tfmt.Println(\"File error: \", err)\r\n\t\t}\r\n\t}\r\n}", "func scan_pass(dirname string, redisClient *redis.Client) ([]*WatchRecord, error){\n files, err := ioutil.ReadDir(dirname)\n\n if(err != nil) {\n return nil, err\n }\n\n out_records := make([]*WatchRecord, len(files))\n\n for i, f := range(files) {\n //see https://golang.org/pkg/os/#FileInfo for description of \"f\"'s fields\n record, watchErr := get_watch_record_with_retry(dirname, f.Name(), redisClient)\n if(watchErr != nil){\n return nil, watchErr\n }\n new_record, updateErr := update_watch_record(record, &f, redisClient)\n out_records[i] = new_record\n if(updateErr != nil){\n return nil, updateErr\n }\n }\n return out_records, nil\n}", "func FileSystemWatchFromGlob(filesGlob []string, ev func(fsnotify.Event), errs func(error)) (*FileSystemWatch, error) {\n\tvar watches []string\n\n\tfor _, file := range filesGlob {\n\t\tfiles, err := filepath.Glob(file)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\twatches = append(watches, files...)\n\t}\n\n\treturn NewFileSystemWatch(watches, ev, errs), nil\n}", "func (di *directoryInfo) watch() {\n\tdi.root.Watch()\n\n\tdi.removeTicker = time.NewTicker(2 * time.Second)\n\n\tdefer func() {\n\t\tdi.root.Close()\n\t}()\n\n\t// event listener\n\tfor {\n\t\tselect {\n\t\tcase e := <-di.root.Ch:\n\t\t\t// file event\n\t\t\tdi.updatePsiMap(e)\n\t\tcase <-di.removeTicker.C:\n\t\t}\n\t}\n}", "func (loop *Loop) Watch() error {\n\twatch, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error creating watcher: %w\", err)\n\t}\n\tdefer watch.Close()\n\terr = loop.watch(watch, \".\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error watching: %w\", err)\n\t}\n\tfor {\n\t\tselect {\n\t\tcase event := <-watch.Events:\n\t\t\tname := filepath.Base(event.Name)\n\t\t\tmatch, err := loop.match(name)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"error matching name %s: %w\", name, err)\n\t\t\t}\n\t\t\tif match {\n\t\t\t\treturn nil\n\t\t\t}\n\t\tcase err := <-watch.Errors:\n\t\t\treturn fmt.Errorf(\"error from watch: %w\", err)\n\t\t}\n\t}\n}", "func NewWatcher() (*Watcher, error) {\n\tfd, errno := syscall.InotifyInit()\n\tif fd == -1 {\n\t\treturn nil, os.NewSyscallError(\"inotify_init\", errno)\n\t}\n\tw := &Watcher{\n\t\tfd: fd,\n\t\twatches: make(map[string]*watch),\n\t\tfsnFlags: make(map[string]uint32),\n\t\tpaths: make(map[int]string),\n\t\tinternalEvent: make(chan *FileEvent),\n\t\tEvent: make(chan *FileEvent),\n\t\tError: make(chan error),\n\t\tdone: make(chan bool, 1),\n\t}\n\n\tgo w.readEvents()\n\tgo w.purgeEvents()\n\treturn w, nil\n}", "func filterDirFunc(dirEntry os.DirEntry) bool {\r\n\treturn !dirEntry.IsDir()\r\n}", "func Watch(c cookoo.Context, p *cookoo.Params) (interface{}, cookoo.Interrupt) {\n\tdir := cookoo.GetString(\"dir\", \".\", p)\n\troute := cookoo.GetString(\"update\", \"@update\", p)\n\n\tr, ok := c.Has(\"router\")\n\tif !ok {\n\t\treturn time.Now(), fmt.Errorf(\"Could not find 'router' in context.\")\n\t}\n\n\trouter := r.(*cookoo.Router)\n\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer watcher.Close()\n\twatcher.Add(dir)\n\n\tfmt.Printf(\"[INFO] Watching %s for changes to .codl files.\\n\", dir)\n\n\t// Watch for updates to files.\n\tfor {\n\t\tselect {\n\t\tcase good := <-watcher.Events:\n\n\t\t\t// Look for create, write, and rename events.\n\t\t\tswitch good.Op {\n\t\t\t//case fsnotify.Create, fsnotify.Write, fsnotify.Rename:\n\t\t\tcase fsnotify.Write, fsnotify.Create:\n\t\t\t\tif path.Ext(good.Name) != \".codl\" {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tfmt.Printf(\"[INFO] %s has changed. Updating. (%s)\\n\", good.Name, good.String())\n\t\t\t\tc.Put(\"files\", []string{good.Name})\n\t\t\t\terr := router.HandleRequest(route, c, false)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Printf(\"[ERROR] %s\\n\", err)\n\t\t\t\t\t//return time.Now(), err\n\t\t\t\t}\n\t\t\t\tc.Put(\"lastUpdated\", time.Now())\n\n\t\t\t// Log but otherwise ignore Remove.\n\t\t\tcase fsnotify.Remove:\n\t\t\t\tfmt.Printf(\"[INFO] %s has been removed.\\n\", good.Name)\n\t\t\t}\n\t\tcase bad := <-watcher.Errors:\n\t\t\tc.Logf(\"warn\", \"Error watching: %s\", bad.Error())\n\t\t}\n\t}\n}", "func (_Univ2 *Univ2Filterer) FilterSync(opts *bind.FilterOpts) (*Univ2SyncIterator, error) {\n\n\tlogs, sub, err := _Univ2.contract.FilterLogs(opts, \"Sync\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Univ2SyncIterator{contract: _Univ2.contract, event: \"Sync\", logs: logs, sub: sub}, nil\n}" ]
[ "0.6076233", "0.5818836", "0.57764274", "0.572449", "0.5580477", "0.55674976", "0.5498557", "0.54695123", "0.54118043", "0.5334793", "0.5292012", "0.5247912", "0.5230286", "0.52064234", "0.5187119", "0.5153706", "0.51411194", "0.51167303", "0.5104982", "0.510158", "0.5098717", "0.50670946", "0.5054977", "0.50500995", "0.50496954", "0.49785596", "0.49746686", "0.49687406", "0.4965476", "0.4962313", "0.4947566", "0.4941133", "0.4936991", "0.49354526", "0.4910474", "0.49090523", "0.48943895", "0.48655483", "0.48523247", "0.48356405", "0.48341265", "0.48277667", "0.48267913", "0.48201227", "0.48193842", "0.48162332", "0.4813655", "0.48098952", "0.48064405", "0.47987896", "0.47907916", "0.47905222", "0.4780633", "0.47627774", "0.47575322", "0.47521305", "0.4747326", "0.47382185", "0.4738125", "0.472833", "0.47245467", "0.47053787", "0.46954718", "0.46948573", "0.4687983", "0.46877405", "0.46766797", "0.46750945", "0.46680045", "0.46623883", "0.4653436", "0.4651607", "0.4638463", "0.46256566", "0.46233734", "0.46211028", "0.46188232", "0.46131426", "0.46096465", "0.45901692", "0.4586097", "0.4582213", "0.4579257", "0.45725715", "0.45667395", "0.4565964", "0.45545894", "0.455437", "0.4546147", "0.4535318", "0.4528002", "0.45247477", "0.45169607", "0.4510956", "0.45092526", "0.45066443", "0.45005304", "0.44965172", "0.44910765", "0.448542" ]
0.81828123
0
GetSeekInfo gets the current file size
func GetSeekInfo(name string) int64 { info, err := os.Stat(name) if err != nil { logger.Error(err) return 0 } return info.Size() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (fi *fileInfo) Size() int64 { return fi.size }", "func (i *FileInfo) Size() int64 { return i.size }", "func FileSeek(f *os.File, offset int64, whence int) (int64, error)", "func (m *wasiSnapshotPreview1Impl) fdSeek(pfd wasiFd, poffset wasiFiledelta, pwhence wasiWhence) (rv wasiFilesize, err wasiErrno) {\n\tf, err := m.files.getFile(pfd, wasiRightsFdSeek)\n\tif err != wasiErrnoSuccess {\n\t\treturn 0, err\n\t}\n\n\tpos, ferr := f.Seek(int64(poffset), int(pwhence))\n\tif ferr != nil {\n\t\treturn 0, fileErrno(ferr)\n\t}\n\treturn wasiFilesize(pos), wasiErrnoSuccess\n}", "func (f FixedFileinfo) Size() int { return binary.Size(f) }", "func getSeek(dirPath string, name string) int64 {\n\tfullname := path.Join(dirPath, seekPrefix+name)\n\ts, err := ioutil.ReadFile(fullname)\n\tif err != nil {\n\t\treturn int64(0)\n\t}\n\ts2 := string(s)\n\tsPos, _ := strconv.ParseInt(s2, 10, 64)\n\treturn sPos\n\n}", "func (e FATFileInfo) Size() int64 {\n\treturn int64(e.currentFile.Size)\n}", "func (info *FileInfo) Size() int64 {\n\treturn info.FileSize\n}", "func (info *unixFileInfo) Size() int64 {\n\tif info == nil {\n\t\treturn 0\n\t}\n\treturn info.sys.Size\n}", "func (s *Stream) GetSize() uint { return s.cur_size }", "func (f *file) Size() int64 { return f.sz }", "func (o SiaFileInfo) Size() int64 {\n\treturn o.FileSize\n}", "func (fi *fileInfo) Size() int64 {\n\treturn fi.size\n}", "func (f *file) SizeNow() int64 { return f.sz }", "func seekerSize(seeker io.ReadSeeker) (int64, error) {\n\tcurrent, err := seeker.Seek(0, os.SEEK_CUR)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tend, err := seeker.Seek(0, os.SEEK_END)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tresumed, err := seeker.Seek(current, os.SEEK_SET)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tif resumed != current {\n\t\treturn 0, fmt.Errorf(\"error returning seeker to original state, could not seek back to original location\")\n\t}\n\n\treturn end, nil\n}", "func (h *Hook) CurrentFileSize(fileLoc string) int64 {\n\tfl, err := osStat(fileLoc)\n\tif err != nil {\n\t\th.WLogger.Println(err)\n\t\treturn -1\n\t}\n\treturn fl.Size()\n}", "func (z *ZipReadSeeker) Seek(offset int64, whence int) (absOffset int64, err error) {\n\t// Calculate the desired absolute offset\n\tswitch whence {\n\tcase io.SeekStart:\n\t\tabsOffset = offset\n\tcase io.SeekCurrent:\n\t\tabsOffset = int64(z.offset) + offset\n\tcase io.SeekEnd:\n\t\tabsOffset = int64(z.f.FileHeader.UncompressedSize64) + offset\n\tdefault:\n\t\treturn 0, errors.New(\"devportal.ZipReadSeeker.Seek: invalid whence\")\n\t}\n\n\tlogging.Logger().Tracef(\"ZIP SEEK: current: %d, offset: %d, whence: %d, new: %d\", z.offset, offset, whence, absOffset)\n\n\t// cannot seek before BOF\n\tif absOffset < 0 {\n\t\treturn 0, errors.New(\"devportal.ZipReadSeeker.Seek: negative position\")\n\t}\n\n\t// Don't do anything if the position wouldn't change\n\tif uint64(absOffset) == z.offset {\n\t\tlogging.Logger().Tracef(\"ZIP SEEK: noop\")\n\t\treturn int64(z.offset), nil\n\t}\n\n\t// Re-open the file\n\tz.Close()\n\tz.ReadCloser, err = z.f.Open()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tz.offset = 0\n\n\t// Read a bunch of bytes, but only up to the end of the file\n\tnToRead := absOffset\n\tif nToRead > int64(z.f.FileHeader.UncompressedSize64) {\n\t\tnToRead = int64(z.f.FileHeader.UncompressedSize64)\n\t}\n\n\t_, err = io.CopyN(ioutil.Discard, z, nToRead)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tlogging.Logger().Tracef(\"NEW ZIP OFFSET: %d, %+v\", z.offset, err)\n\treturn absOffset, nil\n}", "func (f *File) seek(offset int64, whence int) (ret int64, err error) {\n\tif f.dirinfo != nil {\n\t\t// Free cached dirinfo, so we allocate a new one if we\n\t\t// access this file as a directory again. See #35767 and #37161.\n\t\tf.dirinfo.close()\n\t\tf.dirinfo = nil\n\t}\n\tret, err = f.pfd.Seek(offset, whence)\n\truntime.KeepAlive(f)\n\treturn ret, err\n}", "func (file *Remote) Seek(offset int64, whence int) (int64, error) {\n\tfile.m.Lock()\n\tdefer file.m.Unlock()\n\n\tswitch whence {\n\tcase io.SeekStart:\n\t\tif offset < 0 {\n\t\t\treturn int64(file.pos), errors.New(\"negative offset\")\n\t\t}\n\n\t\tfile.pos = uint64(offset)\n\t\treturn offset, nil\n\n\tcase io.SeekCurrent:\n\t\tnpos := int64(file.pos) + offset\n\t\tif npos < 0 {\n\t\t\treturn int64(file.pos), errors.New(\"negative offset\")\n\t\t}\n\n\t\tfile.pos = uint64(npos)\n\t\treturn npos, nil\n\n\tcase io.SeekEnd:\n\t\tstat, err := file.Stat(\"\")\n\t\tif err != nil {\n\t\t\treturn int64(file.pos), err\n\t\t}\n\n\t\tnpos := int64(stat.Length) + offset\n\t\tif npos < 0 {\n\t\t\treturn int64(file.pos), errors.New(\"negative offset\")\n\t\t}\n\n\t\tfile.pos = uint64(npos)\n\t\treturn npos, nil\n\t}\n\n\tpanic(util.Errorf(\"Invalid whence: %v\", whence))\n}", "func(this*Window)Seek(offset int64,whence int)(ret int64,err error){\nf,err:=this.File(\"body\")\nif err!=nil{\nreturn 0,err\n}\nreturn f.Seek(offset,whence)\n}", "func (f *IndexFile) Size() int64 { return int64(len(f.data)) }", "func (f *MemFile) Seek(off int64, whence int) (int64, error) {\n\tf.mtx.Lock()\n\tdefer f.mtx.Unlock()\n\n\tvar abs int64\n\tswitch whence {\n\tcase os.SEEK_SET: // Relative to the origin of the file\n\t\tabs = off\n\tcase os.SEEK_CUR: // Relative to the current offset\n\t\tabs = int64(f.offset) + off\n\tcase os.SEEK_END: // Relative to the end\n\t\tabs = f.Size() + off\n\tdefault:\n\t\treturn 0, ErrInvalidSeekWhence\n\t}\n\tif abs < 0 {\n\t\treturn 0, ErrNegativeSeekLocation\n\t}\n\tf.offset = abs\n\treturn abs, nil\n}", "func (f *File) Seek(offset int64, whence int) (int64, error) {\n\tif f.IsClosed {\n\t\treturn 0, os.ErrClosed\n\t}\n\n\tswitch whence {\n\tcase io.SeekCurrent:\n\t\tf.Position += offset\n\tcase io.SeekStart:\n\t\tf.Position = offset\n\tcase io.SeekEnd:\n\t\tf.Position = int64(len(f.Content)) + offset\n\t}\n\n\treturn f.Position, nil\n}", "func _file_size(call otto.FunctionCall) otto.Value {\n\tfilepath, _ := call.Argument(0).ToString()\n\n\tfile, err := os.Open(filepath)\n\tif err != nil {\n\t\tjsThrow(call, err)\n\t}\n\tdefer file.Close()\n\n\tinfo, err := file.Stat()\n\tif err != nil {\n\t\tjsThrow(call, err)\n\t}\n\tv, _ := otto.ToValue(info.Size())\n\treturn v\n}", "func (f FileEntry) Seek(offset int64, whence int) (int64, error) {\n\treturn f.Reader.Seek(offset+f.File.Offset(), whence)\n}", "func (o *LargeObject) Seek(offset int64, whence int) (n int64, err error) {\n\terr = o.tx.QueryRow(o.ctx, \"select lo_lseek64($1, $2, $3)\", o.fd, offset, whence).Scan(&n)\n\treturn n, err\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}", "func (fi bindataFileInfo) Size() int64 {\n\treturn fi.size\n}" ]
[ "0.67846054", "0.6705082", "0.6643975", "0.6616", "0.6550556", "0.6441409", "0.642182", "0.6405816", "0.63920975", "0.6344479", "0.63371134", "0.6336542", "0.63301504", "0.6324967", "0.6209124", "0.6172135", "0.61436087", "0.6119679", "0.6110798", "0.61034346", "0.60987127", "0.60824454", "0.60545444", "0.6050547", "0.60422516", "0.6025172", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891", "0.5984891" ]
0.84307337
0
ToPropertyValueModel transforms the ProtocolProperties DTO to the ProtocolProperties model
func ToProtocolPropertiesModel(p ProtocolProperties) models.ProtocolProperties { protocolProperties := make(models.ProtocolProperties) for k, v := range p { protocolProperties[k] = v } return protocolProperties }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ToProtocolModels(protocolDTOs map[string]ProtocolProperties) map[string]models.ProtocolProperties {\n\tprotocolModels := make(map[string]models.ProtocolProperties)\n\tfor k, protocolProperties := range protocolDTOs {\n\t\tprotocolModels[k] = ToProtocolPropertiesModel(protocolProperties)\n\t}\n\treturn protocolModels\n}", "func ToPropertyValueModel(p PropertyValue) models.PropertyValue {\n\treturn models.PropertyValue{\n\t\tDataType: p.DataType,\n\t\tType: p.Type,\n\t\tReadWrite: p.ReadWrite,\n\t\tUnits: p.Units,\n\t\tMinimum: p.Minimum,\n\t\tMaximum: p.Maximum,\n\t\tDefaultValue: p.DefaultValue,\n\t\tMask: p.Mask,\n\t\tShift: p.Shift,\n\t\tScale: p.Scale,\n\t\tOffset: p.Offset,\n\t\tBase: p.Base,\n\t\tAssertion: p.Assertion,\n\t\tMediaType: p.MediaType,\n\t}\n}", "func FromPropertyValueModelToDTO(p models.PropertyValue) PropertyValue {\n\treturn PropertyValue{\n\t\tDataType: p.DataType,\n\t\tType: p.Type,\n\t\tReadWrite: p.ReadWrite,\n\t\tUnits: p.Units,\n\t\tMinimum: p.Minimum,\n\t\tMaximum: p.Maximum,\n\t\tDefaultValue: p.DefaultValue,\n\t\tMask: p.Mask,\n\t\tShift: p.Shift,\n\t\tScale: p.Scale,\n\t\tOffset: p.Offset,\n\t\tBase: p.Base,\n\t\tAssertion: p.Assertion,\n\t\tMediaType: p.MediaType,\n\t}\n}", "func ProtoToModel(p *vertexaipb.VertexaiModel) *vertexai.Model {\n\tobj := &vertexai.Model{\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tVersionId: dcl.StringOrNil(p.GetVersionId()),\n\t\tVersionCreateTime: dcl.StringOrNil(p.GetVersionCreateTime()),\n\t\tVersionUpdateTime: dcl.StringOrNil(p.GetVersionUpdateTime()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tVersionDescription: dcl.StringOrNil(p.GetVersionDescription()),\n\t\tTrainingPipeline: dcl.StringOrNil(p.GetTrainingPipeline()),\n\t\tOriginalModelInfo: ProtoToVertexaiModelOriginalModelInfo(p.GetOriginalModelInfo()),\n\t\tContainerSpec: ProtoToVertexaiModelContainerSpec(p.GetContainerSpec()),\n\t\tArtifactUri: dcl.StringOrNil(p.GetArtifactUri()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tUpdateTime: dcl.StringOrNil(p.GetUpdateTime()),\n\t\tEtag: dcl.StringOrNil(p.GetEtag()),\n\t\tEncryptionSpec: ProtoToVertexaiModelEncryptionSpec(p.GetEncryptionSpec()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t}\n\tfor _, r := range p.GetSupportedExportFormats() {\n\t\tobj.SupportedExportFormats = append(obj.SupportedExportFormats, *ProtoToVertexaiModelSupportedExportFormats(r))\n\t}\n\tfor _, r := range p.GetSupportedDeploymentResourcesTypes() {\n\t\tobj.SupportedDeploymentResourcesTypes = append(obj.SupportedDeploymentResourcesTypes, *ProtoToVertexaiModelSupportedDeploymentResourcesTypesEnum(r))\n\t}\n\tfor _, r := range p.GetSupportedInputStorageFormats() {\n\t\tobj.SupportedInputStorageFormats = append(obj.SupportedInputStorageFormats, r)\n\t}\n\tfor _, r := range p.GetSupportedOutputStorageFormats() {\n\t\tobj.SupportedOutputStorageFormats = append(obj.SupportedOutputStorageFormats, r)\n\t}\n\tfor _, r := range p.GetDeployedModels() {\n\t\tobj.DeployedModels = append(obj.DeployedModels, *ProtoToVertexaiModelDeployedModels(r))\n\t}\n\treturn obj\n}", "func VertexaiModelEncryptionSpecToProto(o *vertexai.ModelEncryptionSpec) *vertexaipb.VertexaiModelEncryptionSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelEncryptionSpec{}\n\tp.SetKmsKeyName(dcl.ValueOrEmptyString(o.KmsKeyName))\n\treturn p\n}", "func (fieldType SimpleType) ConvertToModel(value interface{}) (interface{}, error) {\n\tif value == nil {\n\t\treturn nil, nil\n\t}\n\tvalueType := reflect.TypeOf(value)\n\tswitch fieldType.GetKind() {\n\tcase KindString, KindUser:\n\t\tif valueType.Kind() != reflect.String {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"string\", valueType.Name())\n\t\t}\n\t\treturn value, nil\n\tcase KindURL:\n\t\tif valueType.Kind() == reflect.String && govalidator.IsURL(value.(string)) {\n\t\t\treturn value, nil\n\t\t}\n\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"URL\", valueType.Name())\n\tcase KindFloat:\n\t\tif valueType.Kind() != reflect.Float64 {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"float64\", valueType.Name())\n\t\t}\n\t\treturn value, nil\n\tcase KindInteger, KindDuration:\n\t\tif valueType.Kind() != reflect.Int {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"int\", valueType.Name())\n\t\t}\n\t\treturn value, nil\n\tcase KindInstant:\n\t\t// instant == milliseconds\n\t\tif !valueType.Implements(timeType) {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"time.Time\", valueType.Name())\n\t\t}\n\t\treturn value.(time.Time).UnixNano(), nil\n\tcase KindWorkitemReference:\n\t\tif valueType.Kind() != reflect.String {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"string\", valueType.Name())\n\t\t}\n\t\tidValue, err := strconv.Atoi(value.(string))\n\t\treturn idValue, err\n\tcase KindList:\n\t\tif (valueType.Kind() != reflect.Array) && (valueType.Kind() != reflect.Slice) {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s,\", value, \"array/slice\", valueType.Kind())\n\t\t}\n\t\treturn value, nil\n\tcase KindEnum:\n\t\t// to be done yet | not sure what to write here as of now.\n\t\treturn value, nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unexpected type constant: %d\", fieldType.GetKind())\n\t}\n}", "func VertexaiModelOriginalModelInfoToProto(o *vertexai.ModelOriginalModelInfo) *vertexaipb.VertexaiModelOriginalModelInfo {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelOriginalModelInfo{}\n\tp.SetModel(dcl.ValueOrEmptyString(o.Model))\n\treturn p\n}", "func (w *NotificationPolicy) ConvertToDBModel() error {\n\tif len(w.Targets) != 0 {\n\t\ttargets, err := json.Marshal(w.Targets)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tw.TargetsDB = string(targets)\n\t}\n\tif len(w.EventTypes) != 0 {\n\t\teventTypes, err := json.Marshal(w.EventTypes)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tw.EventTypesDB = string(eventTypes)\n\t}\n\n\treturn nil\n}", "func (shipType ShipParameter) ToModel(battle model.Battle) model.ShipType {\n\treturn model.ShipType{\n\t\tName: shipType.Name,\n\t\tLength: shipType.Length,\n\t}\n}", "func ProtoToVertexaiModelEncryptionSpec(p *vertexaipb.VertexaiModelEncryptionSpec) *vertexai.ModelEncryptionSpec {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelEncryptionSpec{\n\t\tKmsKeyName: dcl.StringOrNil(p.GetKmsKeyName()),\n\t}\n\treturn obj\n}", "func ToKindProps[T KindProperties](v cue.Value) (T, error) {\n\tprops := new(T)\n\tif !v.Exists() {\n\t\treturn *props, ErrValueNotExist\n\t}\n\n\tfw := CUEFramework(v.Context())\n\tvar kdef cue.Value\n\n\tanyprops := any(*props).(SomeKindProperties)\n\tswitch anyprops.(type) {\n\tcase CoreProperties:\n\t\tkdef = fw.LookupPath(cue.MakePath(cue.Str(\"Core\")))\n\tcase CustomProperties:\n\t\tkdef = fw.LookupPath(cue.MakePath(cue.Str(\"Custom\")))\n\tcase ComposableProperties:\n\t\tkdef = fw.LookupPath(cue.MakePath(cue.Str(\"Composable\")))\n\tdefault:\n\t\t// unreachable so long as all the possibilities in KindProperties have switch branches\n\t\tpanic(\"unreachable\")\n\t}\n\n\titem := v.Unify(kdef)\n\tif item.Err() != nil {\n\t\treturn *props, errors.Wrap(errors.Promote(ErrValueNotAKind, \"\"), item.Err())\n\t}\n\n\tif err := item.Decode(props); err != nil {\n\t\t// Should only be reachable if CUE and Go framework types have diverged\n\t\tpanic(errors.Details(err, nil))\n\t}\n\n\treturn *props, nil\n}", "func (params Parameters) ToModel(battle model.Battle) model.Parameters {\n\tshipTypes := make([]model.ShipType, 0, len(params.Ships))\n\n\tfor _, shipType := range params.Ships {\n\t\tfor i := 0; i < shipType.Count; i++ {\n\t\t\tshipTypes = append(shipTypes, shipType.ToModel(battle))\n\t\t}\n\t}\n\n\treturn model.Parameters{\n\t\tWidth: params.Width,\n\t\tHeight: params.Height,\n\t\tShips: shipTypes,\n\t}\n}", "func (s SecretWatchableSet) ConvertToModel(in interface{}) WatchableResource {\n\tk8Secret := in.(*api.Secret)\n\tsecret := &Secret{\n\t\tNamespace: k8Secret.Namespace,\n\t\tFields: k8Secret.Data,\n\t}\n\n\tif routingKey, ok := k8Secret.Data[s.Config.APIKeySecretDataField]; ok {\n\t\tsecret.RoutingKey = &routingKey\n\t}\n\n\t// Pre calculdate hash\n\tsecret.hash = calculateSecretHash(secret)\n\n\treturn secret\n}", "func (mepm MachineExtensionPropertiesModel) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif mepm.ForceUpdateTag != nil {\n\t\tobjectMap[\"forceUpdateTag\"] = mepm.ForceUpdateTag\n\t}\n\tif mepm.Publisher != nil {\n\t\tobjectMap[\"publisher\"] = mepm.Publisher\n\t}\n\tif mepm.Type != nil {\n\t\tobjectMap[\"type\"] = mepm.Type\n\t}\n\tif mepm.TypeHandlerVersion != nil {\n\t\tobjectMap[\"typeHandlerVersion\"] = mepm.TypeHandlerVersion\n\t}\n\tif mepm.AutoUpgradeMinorVersion != nil {\n\t\tobjectMap[\"autoUpgradeMinorVersion\"] = mepm.AutoUpgradeMinorVersion\n\t}\n\tif mepm.Settings != nil {\n\t\tobjectMap[\"settings\"] = mepm.Settings\n\t}\n\tif mepm.ProtectedSettings != nil {\n\t\tobjectMap[\"protectedSettings\"] = mepm.ProtectedSettings\n\t}\n\tif mepm.InstanceView != nil {\n\t\tobjectMap[\"instanceView\"] = mepm.InstanceView\n\t}\n\treturn json.Marshal(objectMap)\n}", "func ProtoToVertexaiModelContainerSpecPorts(p *vertexaipb.VertexaiModelContainerSpecPorts) *vertexai.ModelContainerSpecPorts {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelContainerSpecPorts{\n\t\tContainerPort: dcl.Int64OrNil(p.GetContainerPort()),\n\t}\n\treturn obj\n}", "func ProtoToVertexaiModelOriginalModelInfo(p *vertexaipb.VertexaiModelOriginalModelInfo) *vertexai.ModelOriginalModelInfo {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelOriginalModelInfo{\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t}\n\treturn obj\n}", "func VertexaiModelContainerSpecPortsToProto(o *vertexai.ModelContainerSpecPorts) *vertexaipb.VertexaiModelContainerSpecPorts {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelContainerSpecPorts{}\n\tp.SetContainerPort(dcl.ValueOrEmptyInt64(o.ContainerPort))\n\treturn p\n}", "func AppengineDomainMappingSslSettingsToProto(o *appengine.DomainMappingSslSettings) *appenginepb.AppengineDomainMappingSslSettings {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &appenginepb.AppengineDomainMappingSslSettings{\n\t\tCertificateId: dcl.ValueOrEmptyString(o.CertificateId),\n\t\tSslManagementType: AppengineDomainMappingSslSettingsSslManagementTypeEnumToProto(o.SslManagementType),\n\t\tPendingManagedCertificateId: dcl.ValueOrEmptyString(o.PendingManagedCertificateId),\n\t}\n\treturn p\n}", "func ProtoToModelDeployment(p *betapb.VertexaiBetaModelDeployment) *beta.ModelDeployment {\n\tobj := &beta.ModelDeployment{\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t\tDeployedModelId: dcl.StringOrNil(p.GetDeployedModelId()),\n\t\tDedicatedResources: ProtoToVertexaiBetaModelDeploymentDedicatedResources(p.GetDedicatedResources()),\n\t\tEndpoint: dcl.StringOrNil(p.GetEndpoint()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t}\n\treturn obj\n}", "func Convert2ConnectionProperty(wwns []string, lunID string) connector.ConnectionProperty {\n\tvar property connector.ConnectionProperty\n\tproperty.TargetWwns = wwns\n\tproperty.TargetLun, _ = strconv.Atoi(lunID)\n\tproperty.StorageProtocol = connector.FcProtocol\n\n\treturn property\n}", "func ModelProtocolToListenerProtocol(p protocol.Instance,\n\ttrafficDirection core.TrafficDirection) ListenerProtocol {\n\t// If protocol sniffing is not enabled, the default value is TCP\n\tif p == protocol.Unsupported {\n\t\tswitch trafficDirection {\n\t\tcase core.TrafficDirection_INBOUND:\n\t\t\tif !features.EnableProtocolSniffingForInbound {\n\t\t\t\tp = protocol.TCP\n\t\t\t}\n\t\tcase core.TrafficDirection_OUTBOUND:\n\t\t\tif !features.EnableProtocolSniffingForOutbound {\n\t\t\t\tp = protocol.TCP\n\t\t\t}\n\t\tdefault:\n\t\t\t// Should not reach here.\n\t\t}\n\t}\n\n\tswitch p {\n\tcase protocol.HTTP, protocol.HTTP2, protocol.GRPC, protocol.GRPCWeb:\n\t\treturn ListenerProtocolHTTP\n\tcase protocol.TCP, protocol.HTTPS, protocol.TLS,\n\t\tprotocol.Mongo, protocol.Redis, protocol.MySQL:\n\t\treturn ListenerProtocolTCP\n\tcase protocol.Thrift:\n\t\tif features.EnableThriftFilter {\n\t\t\treturn ListenerProtocolThrift\n\t\t}\n\t\treturn ListenerProtocolTCP\n\tcase protocol.UDP:\n\t\treturn ListenerProtocolUnknown\n\tcase protocol.Unsupported:\n\t\treturn ListenerProtocolAuto\n\tdefault:\n\t\t// Should not reach here.\n\t\treturn ListenerProtocolAuto\n\t}\n}", "func (t *Translator) fromModel(req *TranslationRequest) (*JSONMessage, error) {\n\tcmpTable := t.Tables.Component\n\n\t// Get all of the components associated with that message.\n\tcomps := make([]*Component, 0)\n\terr := cmpTable.Get().Where(\"message\", req.Model.Id).All(t.Store, &comps)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Translate the components into JSON.\n\tparsedComponents := make(map[string]*JSONComponent)\n\tfor _, c := range comps {\n\t\tparsedComponents[c.Name] = &JSONComponent{\n\t\t\tBinary: c.Data,\n\t\t\tString: string(c.Data),\n\t\t}\n\t}\n\n\ttoAddrs := strings.Split(req.Model.To, \",\")\n\n\t// Download the profiles of all recipients of the message.\n\tprofiles := make([]*JSONProfile, len(toAddrs))\n\tfor i, toAddr := range toAddrs {\n\t\tprofiles[i] = req.Profiles[toAddr]\n\t}\n\n\tmyProfile := &JSONProfile{\n\t\tName: req.Me.Profile.Name,\n\t\tAvatar: req.Me.Profile.Image,\n\t\tAlias: req.Me.Alias.String(),\n\t}\n\n\t// Export the finished product.\n\treturn &JSONMessage{\n\t\tName: req.Model.Name,\n\t\tDate: time.Unix(req.Model.Date, 0),\n\n\t\t// To and From Info\n\t\tFrom: myProfile,\n\t\tTo: profiles,\n\n\t\t// Components\n\t\tComponents: parsedComponents,\n\n\t\t// Meta\n\t\tSelf: true,\n\t\tPublic: !req.Model.Alert,\n\t}, nil\n}", "func (fieldType SimpleType) ConvertFromModel(value interface{}) (interface{}, error) {\n\tvalueType := reflect.TypeOf(value)\n\tswitch fieldType.GetKind() {\n\tcase KindString, KindURL, KindUser, KindInteger, KindFloat, KindDuration:\n\t\treturn value, nil\n\tcase KindInstant:\n\t\treturn time.Unix(0, value.(int64)), nil\n\tcase KindWorkitemReference:\n\t\tif valueType.Kind() != reflect.String {\n\t\t\treturn nil, fmt.Errorf(\"value %v should be %s, but is %s\", value, \"string\", valueType.Name())\n\t\t}\n\t\treturn strconv.FormatUint(value.(uint64), 10), nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unexpected type constant: %d\", fieldType.GetKind())\n\t}\n}", "func (r Role) MapToModel(m map[string]interface{}) Role {\n\tr.Id = m[\"id\"].(int)\n\tr.Name, _ = m[\"name\"].(string)\n\tr.Slug, _ = m[\"slug\"].(string)\n\tr.CreatedAt, _ = m[\"created_at\"].(string)\n\tr.UpdatedAt, _ = m[\"updated_at\"].(string)\n\treturn r\n}", "func ModelDeploymentToProto(resource *beta.ModelDeployment) *betapb.VertexaiBetaModelDeployment {\n\tp := &betapb.VertexaiBetaModelDeployment{}\n\tp.SetModel(dcl.ValueOrEmptyString(resource.Model))\n\tp.SetDeployedModelId(dcl.ValueOrEmptyString(resource.DeployedModelId))\n\tp.SetDedicatedResources(VertexaiBetaModelDeploymentDedicatedResourcesToProto(resource.DedicatedResources))\n\tp.SetEndpoint(dcl.ValueOrEmptyString(resource.Endpoint))\n\tp.SetLocation(dcl.ValueOrEmptyString(resource.Location))\n\tp.SetProject(dcl.ValueOrEmptyString(resource.Project))\n\n\treturn p\n}", "func ModelToProto(resource *vertexai.Model) *vertexaipb.VertexaiModel {\n\tp := &vertexaipb.VertexaiModel{}\n\tp.SetName(dcl.ValueOrEmptyString(resource.Name))\n\tp.SetVersionId(dcl.ValueOrEmptyString(resource.VersionId))\n\tp.SetVersionCreateTime(dcl.ValueOrEmptyString(resource.VersionCreateTime))\n\tp.SetVersionUpdateTime(dcl.ValueOrEmptyString(resource.VersionUpdateTime))\n\tp.SetDisplayName(dcl.ValueOrEmptyString(resource.DisplayName))\n\tp.SetDescription(dcl.ValueOrEmptyString(resource.Description))\n\tp.SetVersionDescription(dcl.ValueOrEmptyString(resource.VersionDescription))\n\tp.SetTrainingPipeline(dcl.ValueOrEmptyString(resource.TrainingPipeline))\n\tp.SetOriginalModelInfo(VertexaiModelOriginalModelInfoToProto(resource.OriginalModelInfo))\n\tp.SetContainerSpec(VertexaiModelContainerSpecToProto(resource.ContainerSpec))\n\tp.SetArtifactUri(dcl.ValueOrEmptyString(resource.ArtifactUri))\n\tp.SetCreateTime(dcl.ValueOrEmptyString(resource.CreateTime))\n\tp.SetUpdateTime(dcl.ValueOrEmptyString(resource.UpdateTime))\n\tp.SetEtag(dcl.ValueOrEmptyString(resource.Etag))\n\tp.SetEncryptionSpec(VertexaiModelEncryptionSpecToProto(resource.EncryptionSpec))\n\tp.SetProject(dcl.ValueOrEmptyString(resource.Project))\n\tp.SetLocation(dcl.ValueOrEmptyString(resource.Location))\n\tsSupportedExportFormats := make([]*vertexaipb.VertexaiModelSupportedExportFormats, len(resource.SupportedExportFormats))\n\tfor i, r := range resource.SupportedExportFormats {\n\t\tsSupportedExportFormats[i] = VertexaiModelSupportedExportFormatsToProto(&r)\n\t}\n\tp.SetSupportedExportFormats(sSupportedExportFormats)\n\tsSupportedDeploymentResourcesTypes := make([]vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum, len(resource.SupportedDeploymentResourcesTypes))\n\tfor i, r := range resource.SupportedDeploymentResourcesTypes {\n\t\tsSupportedDeploymentResourcesTypes[i] = vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum(vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum_value[string(r)])\n\t}\n\tp.SetSupportedDeploymentResourcesTypes(sSupportedDeploymentResourcesTypes)\n\tsSupportedInputStorageFormats := make([]string, len(resource.SupportedInputStorageFormats))\n\tfor i, r := range resource.SupportedInputStorageFormats {\n\t\tsSupportedInputStorageFormats[i] = r\n\t}\n\tp.SetSupportedInputStorageFormats(sSupportedInputStorageFormats)\n\tsSupportedOutputStorageFormats := make([]string, len(resource.SupportedOutputStorageFormats))\n\tfor i, r := range resource.SupportedOutputStorageFormats {\n\t\tsSupportedOutputStorageFormats[i] = r\n\t}\n\tp.SetSupportedOutputStorageFormats(sSupportedOutputStorageFormats)\n\tsDeployedModels := make([]*vertexaipb.VertexaiModelDeployedModels, len(resource.DeployedModels))\n\tfor i, r := range resource.DeployedModels {\n\t\tsDeployedModels[i] = VertexaiModelDeployedModelsToProto(&r)\n\t}\n\tp.SetDeployedModels(sDeployedModels)\n\tmLabels := make(map[string]string, len(resource.Labels))\n\tfor k, r := range resource.Labels {\n\t\tmLabels[k] = r\n\t}\n\tp.SetLabels(mLabels)\n\n\treturn p\n}", "func ProtoToVertexaiModelContainerSpec(p *vertexaipb.VertexaiModelContainerSpec) *vertexai.ModelContainerSpec {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelContainerSpec{\n\t\tImageUri: dcl.StringOrNil(p.GetImageUri()),\n\t\tPredictRoute: dcl.StringOrNil(p.GetPredictRoute()),\n\t\tHealthRoute: dcl.StringOrNil(p.GetHealthRoute()),\n\t}\n\tfor _, r := range p.GetCommand() {\n\t\tobj.Command = append(obj.Command, r)\n\t}\n\tfor _, r := range p.GetArgs() {\n\t\tobj.Args = append(obj.Args, r)\n\t}\n\tfor _, r := range p.GetEnv() {\n\t\tobj.Env = append(obj.Env, *ProtoToVertexaiModelContainerSpecEnv(r))\n\t}\n\tfor _, r := range p.GetPorts() {\n\t\tobj.Ports = append(obj.Ports, *ProtoToVertexaiModelContainerSpecPorts(r))\n\t}\n\treturn obj\n}", "func (Files) ToModel(data interface{}, model *Files) error {\n\tbsonBytes, err := bson.Marshal(data.(bson.M))\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = bson.Unmarshal(bsonBytes, &model)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (policy *Servers_ConnectionPolicy_Spec) ConvertSpecTo(destination genruntime.ConvertibleSpec) error {\n\tdst, ok := destination.(*v20211101s.Servers_ConnectionPolicy_Spec)\n\tif ok {\n\t\t// Populate destination from our instance\n\t\treturn policy.AssignProperties_To_Servers_ConnectionPolicy_Spec(dst)\n\t}\n\n\t// Convert to an intermediate form\n\tdst = &v20211101s.Servers_ConnectionPolicy_Spec{}\n\terr := policy.AssignProperties_To_Servers_ConnectionPolicy_Spec(dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"initial step of conversion in ConvertSpecTo()\")\n\t}\n\n\t// Update dst from our instance\n\terr = dst.ConvertSpecTo(destination)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"final step of conversion in ConvertSpecTo()\")\n\t}\n\n\treturn nil\n}", "func (h *Handler) nodeConfigToProto(nodeConfig *v1.NodeConfig) *model.NodeConfig {\n\tnodeConfigProto := &model.NodeConfig{}\n\tnodeConfigProto.NodeName = nodeConfig.Name\n\tif nodeConfig.Spec.MainVPPInterface.InterfaceName != \"\" {\n\t\tnodeConfigProto.MainVppInterface = h.interfaceConfigToProto(nodeConfig.Spec.MainVPPInterface)\n\t}\n\tnodeConfigProto.Gateway = nodeConfig.Spec.Gateway\n\tnodeConfigProto.StealInterface = nodeConfig.Spec.StealInterface\n\tnodeConfigProto.NatExternalTraffic = nodeConfig.Spec.NatExternalTraffic\n\tfor _, otherNode := range nodeConfig.Spec.OtherVPPInterfaces {\n\t\tnodeConfigProto.OtherVppInterfaces = append(nodeConfigProto.OtherVppInterfaces,\n\t\t\th.interfaceConfigToProto(otherNode))\n\t}\n\n\treturn nodeConfigProto\n}", "func (c CheckpointStorageConfig) ToModel() (*model.CheckpointStorageConfig, error) {\n\tvar m model.CheckpointStorageConfig\n\n\tif len(c) == 0 {\n\t\treturn &m, nil\n\t}\n\n\tdec := json.NewDecoder(bytes.NewReader(c))\n\tdec.DisallowUnknownFields()\n\n\tif err := dec.Decode(&m); err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\n\treturn &m, nil\n}", "func ProtoToVertexaiModelDeployedModels(p *vertexaipb.VertexaiModelDeployedModels) *vertexai.ModelDeployedModels {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelDeployedModels{\n\t\tEndpoint: dcl.StringOrNil(p.GetEndpoint()),\n\t\tDeployedModelId: dcl.StringOrNil(p.GetDeployedModelId()),\n\t}\n\treturn obj\n}", "func (m *PropertyMap) ToProto() map[string]*pb.Property {\n\t// This may seem wrong, but m is a pointer to a map, which is also a\n\t// nullable reference type.\n\tif m == nil || *m == nil {\n\t\treturn nil\n\t}\n\tret := make(map[string]*pb.Property)\n\tfor k, v := range *m {\n\t\tret[k] = v.ToProto()\n\t}\n\treturn ret\n}", "func (mpm MachinePropertiesModel) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif mpm.LocationData != nil {\n\t\tobjectMap[\"locationData\"] = mpm.LocationData\n\t}\n\tif mpm.OsProfile != nil {\n\t\tobjectMap[\"osProfile\"] = mpm.OsProfile\n\t}\n\tif mpm.VMID != nil {\n\t\tobjectMap[\"vmId\"] = mpm.VMID\n\t}\n\tif mpm.ClientPublicKey != nil {\n\t\tobjectMap[\"clientPublicKey\"] = mpm.ClientPublicKey\n\t}\n\treturn json.Marshal(objectMap)\n}", "func propValue(v *pb.Value) (interface{}, error) {\n\tswitch vt := v.ValueType.(type) {\n\tcase *pb.Value_IntegerValue:\n\t\treturn vt.IntegerValue, nil\n\tcase *pb.Value_TimestampValue:\n\t\treturn ptypes.Timestamp(vt.TimestampValue)\n\tcase *pb.Value_BooleanValue:\n\t\treturn vt.BooleanValue, nil\n\tcase *pb.Value_BytesValue:\n\t\treturn vt.BytesValue, nil\n\tcase *pb.Value_MapValue:\n\t\treturn protoToEntity(&pb.Document{Fields: vt.MapValue.Fields})\n\tcase *pb.Value_StringValue:\n\t\treturn vt.StringValue, nil\n\tcase *pb.Value_ArrayValue:\n\t\treturn nil, fmt.Errorf(\"datastore: unhandled ArrayValue\")\n\tcase *pb.Value_DoubleValue:\n\t\treturn vt.DoubleValue, nil\n\tcase *pb.Value_ReferenceValue:\n\t\tkey, err := referenceValueToKey(vt.ReferenceValue)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn key, nil\n\tcase *pb.Value_GeoPointValue:\n\t\treturn appengine.GeoPoint{Lat: vt.GeoPointValue.Latitude, Lng: vt.GeoPointValue.Longitude}, nil\n\t}\n\treturn nil, nil\n}", "func ProtoToAppengineDomainMappingSslSettings(p *appenginepb.AppengineDomainMappingSslSettings) *appengine.DomainMappingSslSettings {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &appengine.DomainMappingSslSettings{\n\t\tCertificateId: dcl.StringOrNil(p.CertificateId),\n\t\tSslManagementType: ProtoToAppengineDomainMappingSslSettingsSslManagementTypeEnum(p.GetSslManagementType()),\n\t\tPendingManagedCertificateId: dcl.StringOrNil(p.PendingManagedCertificateId),\n\t}\n\treturn obj\n}", "func ProtoMavenRuntimeLibraryToDependencyConfig(proto *registryv1alpha1.MavenConfig_RuntimeLibrary) bufpluginconfig.MavenDependencyConfig {\n\treturn bufpluginconfig.MavenDependencyConfig{\n\t\tGroupID: proto.GetGroupId(),\n\t\tArtifactID: proto.GetArtifactId(),\n\t\tVersion: proto.GetVersion(),\n\t\tClassifier: proto.GetClassifier(),\n\t\tExtension: proto.GetExtension(),\n\t}\n}", "func propertiesToParams(list smd.PropertyList) []Parameter {\n\tvar parameters []Parameter\n\tfor _, prop := range list {\n\t\tp := Parameter{\n\t\t\tName: prop.Name,\n\t\t\tOptional: prop.Optional,\n\t\t\tDescription: prop.Description,\n\t\t}\n\t\tpType := phpType(prop.Type)\n\t\tif prop.Type == smd.Object && prop.Ref != \"\" {\n\t\t\tpType = objectType(prop.Ref)\n\t\t}\n\t\tif prop.Type == smd.Array {\n\t\t\tpType = arrayType(prop.Items)\n\t\t}\n\n\t\tp.Type = pType\n\t\tparameters = append(parameters, p)\n\n\t}\n\treturn parameters\n}", "func VertexaiModelContainerSpecToProto(o *vertexai.ModelContainerSpec) *vertexaipb.VertexaiModelContainerSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelContainerSpec{}\n\tp.SetImageUri(dcl.ValueOrEmptyString(o.ImageUri))\n\tp.SetPredictRoute(dcl.ValueOrEmptyString(o.PredictRoute))\n\tp.SetHealthRoute(dcl.ValueOrEmptyString(o.HealthRoute))\n\tsCommand := make([]string, len(o.Command))\n\tfor i, r := range o.Command {\n\t\tsCommand[i] = r\n\t}\n\tp.SetCommand(sCommand)\n\tsArgs := make([]string, len(o.Args))\n\tfor i, r := range o.Args {\n\t\tsArgs[i] = r\n\t}\n\tp.SetArgs(sArgs)\n\tsEnv := make([]*vertexaipb.VertexaiModelContainerSpecEnv, len(o.Env))\n\tfor i, r := range o.Env {\n\t\tsEnv[i] = VertexaiModelContainerSpecEnvToProto(&r)\n\t}\n\tp.SetEnv(sEnv)\n\tsPorts := make([]*vertexaipb.VertexaiModelContainerSpecPorts, len(o.Ports))\n\tfor i, r := range o.Ports {\n\t\tsPorts[i] = VertexaiModelContainerSpecPortsToProto(&r)\n\t}\n\tp.SetPorts(sPorts)\n\treturn p\n}", "func (p *Properties) Decode(x interface{}) error {\n\tt, v := reflect.TypeOf(x), reflect.ValueOf(x)\n\tif t.Kind() != reflect.Ptr || v.Elem().Type().Kind() != reflect.Struct {\n\t\treturn fmt.Errorf(\"not a pointer to struct: %s\", t)\n\t}\n\tif err := dec(p, \"\", nil, nil, v); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c DestinationProtocol) ToPtr() *DestinationProtocol {\n\treturn &c\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue{}\n\tp.SetIntegerValue(dcl.ValueOrEmptyInt64(o.IntegerValue))\n\tp.SetFloatValue(dcl.ValueOrEmptyDouble(o.FloatValue))\n\tp.SetStringValue(dcl.ValueOrEmptyString(o.StringValue))\n\tp.SetBooleanValue(dcl.ValueOrEmptyBool(o.BooleanValue))\n\tp.SetTimestampValue(dcl.ValueOrEmptyString(o.TimestampValue))\n\tp.SetTimeValue(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValueToProto(o.TimeValue))\n\tp.SetDateValue(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueDateValueToProto(o.DateValue))\n\tp.SetDayOfWeekValue(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueDayOfWeekValueEnumToProto(o.DayOfWeekValue))\n\treturn p\n}", "func (c *Converter) buildProperties(properties map[string]spec.Schema) string {\n\n\tbody := make(map[string]interface{})\n\n\tkeys := []string{}\n\tfor key := range properties {\n\t\tkeys = append(keys, key)\n\t}\n\n\tsort.Strings(keys)\n\n\tfor _, key := range keys {\n\t\tprop := properties[key]\n\n\t\t//Property as an example value : we take it as value\n\t\tif prop.Example != nil {\n\t\t\tbody[key] = prop.Example\n\t\t\tcontinue\n\t\t}\n\n\t\t//Property as a Enum : we take the first possible value\n\t\t//Note: we only support string enum for now;\n\t\t//TODO : add support for other type enum.\n\t\tif prop.Type.Contains(\"string\") && len(prop.Enum) > 0 {\n\t\t\tbody[key] = prop.Enum[0]\n\t\t\tcontinue\n\t\t}\n\n\t\tif prop.Type.Contains(\"object\") {\n\t\t\tbody[key] = json.RawMessage(c.buildProperties(prop.Properties))\n\t\t\tcontinue\n\t\t}\n\n\t\tif prop.Type.Contains(\"array\") {\n\t\t\tif prop.Items.ContainsType(\"object\") {\n\t\t\t\tarray := []json.RawMessage{json.RawMessage(c.buildProperties(prop.Items.Schema.Properties))}\n\t\t\t\trawArray, _ := json.MarshalIndent(array, \"\", \"\\t\")\n\t\t\t\tbody[key] = json.RawMessage(rawArray)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tvar array []interface{}\n\t\t\tarray = append(array, buildPropertyDefaultValue(prop.Items.Schema.Type, prop.Items.Schema.Format))\n\t\t\trawArray, _ := json.MarshalIndent(array, \"\", \"\\t\")\n\t\t\tbody[key] = json.RawMessage(rawArray)\n\t\t\tcontinue\n\t\t}\n\n\t\tbody[key] = buildPropertyDefaultValue(prop.Type, prop.Format)\n\n\t}\n\n\tb, err := json.MarshalIndent(body, \"\", \"\\t\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn string(b)\n}", "func toModel(orgID uint, repository helm.Repository) repositoryModel {\n\treturn repositoryModel{\n\t\tOrganizationID: orgID,\n\t\tName: repository.Name,\n\t\tURL: repository.URL,\n\t\tPasswordSecretID: repository.PasswordSecretID,\n\t\tTlsSecretID: repository.TlsSecretID,\n\t}\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValue) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValue {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValue{}\n\tp.SetIntegerValue(dcl.ValueOrEmptyInt64(o.IntegerValue))\n\tp.SetFloatValue(dcl.ValueOrEmptyDouble(o.FloatValue))\n\tp.SetStringValue(dcl.ValueOrEmptyString(o.StringValue))\n\tp.SetBooleanValue(dcl.ValueOrEmptyBool(o.BooleanValue))\n\tp.SetTimestampValue(dcl.ValueOrEmptyString(o.TimestampValue))\n\tp.SetTimeValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValueToProto(o.TimeValue))\n\tp.SetDateValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueDateValueToProto(o.DateValue))\n\tp.SetDayOfWeekValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueDayOfWeekValueEnumToProto(o.DayOfWeekValue))\n\treturn p\n}", "func getModelValue(o interface{}) (reflect.Value, error) {\n\tvalue, ok := o.(reflect.Value)\n\tif !ok {\n\t\tvalue = reflect.ValueOf(o)\n\t}\n\tswitch value.Kind() {\n\tcase reflect.Struct:\n\t\tif _, ok := reflect.New(value.Type()).Interface().(IModel); ok {\n\t\t\treturn value, nil\n\t\t}\n\t\treturn value, errors.New(\"given object does not meet Model interface\")\n\tcase reflect.Ptr:\n\t\treturn getModelValue(value.Elem())\n\tcase reflect.Slice:\n\t\tif value.Len() == 0 {\n\t\t\telemType := value.Type().Elem()\n\t\t\tif elemType.Kind() == reflect.Ptr {\n\t\t\t\treturn getModelValue(reflect.New(elemType.Elem()).Elem())\n\t\t\t}\n\t\t\treturn value, errors.Errorf(\"slice should contain pointers to model\")\n\t\t}\n\t\treturn getModelValue(value.Index(0))\n\tdefault:\n\t\treturn value, errors.Errorf(\"expected pointer to model, got %T (kind: %v)\", o, value.Kind())\n\t}\n}", "func bsonProperties(f reflect.StructField, origName bool) *proto.Properties {\n\tvar prop proto.Properties\n\tprop.Init(f.Type, f.Name, f.Tag.Get(\"protobuf\"), &f)\n\tif origName || prop.JSONName == \"\" {\n\t\tprop.JSONName = prop.OrigName\n\t}\n\treturn &prop\n}", "func (policy *ServersConnectionPolicy) ConvertTo(hub conversion.Hub) error {\n\tdestination, ok := hub.(*v20211101s.ServersConnectionPolicy)\n\tif !ok {\n\t\treturn fmt.Errorf(\"expected sql/v1api20211101storage/ServersConnectionPolicy but received %T instead\", hub)\n\t}\n\n\treturn policy.AssignProperties_To_ServersConnectionPolicy(destination)\n}", "func mapToStruct(mi *modelInfo, structPtr interface{}, fMap FieldMap) {\n\tfMap = nestMap(fMap)\n\tval := reflect.ValueOf(structPtr)\n\tind := reflect.Indirect(val)\n\tif val.Kind() != reflect.Ptr || ind.Kind() != reflect.Struct {\n\t\ttools.LogAndPanic(log, \"structPtr must be a pointer to a struct\", \"structPtr\", structPtr)\n\t}\n\tfor i := 0; i < ind.NumField(); i++ {\n\t\tfVal := ind.Field(i)\n\t\tsf := ind.Type().Field(i)\n\t\tfi, ok := mi.fields.get(sf.Name)\n\t\tif !ok {\n\t\t\ttools.LogAndPanic(log, \"Unregistered field in model\", \"field\", sf.Name, \"model\", mi.name)\n\t\t}\n\t\tmValue, mValExists := fMap[fi.json]\n\t\tif sf.Type.Kind() == reflect.Ptr {\n\t\t\tif mValExists {\n\t\t\t\tfm, ok := mValue.(FieldMap)\n\t\t\t\tif !ok {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif !fVal.Elem().IsValid() {\n\t\t\t\t\t// Create the related struct if it does not exist\n\t\t\t\t\tfVal.Set(reflect.New(sf.Type.Elem()))\n\t\t\t\t}\n\t\t\t\tmapToStruct(fi.relatedModel, fVal.Interface(), fm)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif mValExists && mValue != nil {\n\t\t\tconvertedValue := reflect.ValueOf(mValue).Convert(fVal.Type())\n\t\t\tfVal.Set(convertedValue)\n\t\t}\n\t}\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValue{}\n\tp.SetIntegerValue(dcl.ValueOrEmptyInt64(o.IntegerValue))\n\tp.SetFloatValue(dcl.ValueOrEmptyDouble(o.FloatValue))\n\tp.SetStringValue(dcl.ValueOrEmptyString(o.StringValue))\n\tp.SetBooleanValue(dcl.ValueOrEmptyBool(o.BooleanValue))\n\tp.SetTimestampValue(dcl.ValueOrEmptyString(o.TimestampValue))\n\tp.SetTimeValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValueToProto(o.TimeValue))\n\tp.SetDateValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueDateValueToProto(o.DateValue))\n\tp.SetDayOfWeekValue(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueDayOfWeekValueEnumToProto(o.DayOfWeekValue))\n\treturn p\n}", "func (v VPNConnectionProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"connectionBandwidth\", v.ConnectionBandwidth)\n\tpopulate(objectMap, \"connectionStatus\", v.ConnectionStatus)\n\tpopulate(objectMap, \"dpdTimeoutSeconds\", v.DpdTimeoutSeconds)\n\tpopulate(objectMap, \"egressBytesTransferred\", v.EgressBytesTransferred)\n\tpopulate(objectMap, \"enableBgp\", v.EnableBgp)\n\tpopulate(objectMap, \"enableInternetSecurity\", v.EnableInternetSecurity)\n\tpopulate(objectMap, \"enableRateLimiting\", v.EnableRateLimiting)\n\tpopulate(objectMap, \"ipsecPolicies\", v.IPSecPolicies)\n\tpopulate(objectMap, \"ingressBytesTransferred\", v.IngressBytesTransferred)\n\tpopulate(objectMap, \"provisioningState\", v.ProvisioningState)\n\tpopulate(objectMap, \"remoteVpnSite\", v.RemoteVPNSite)\n\tpopulate(objectMap, \"routingConfiguration\", v.RoutingConfiguration)\n\tpopulate(objectMap, \"routingWeight\", v.RoutingWeight)\n\tpopulate(objectMap, \"sharedKey\", v.SharedKey)\n\tpopulate(objectMap, \"trafficSelectorPolicies\", v.TrafficSelectorPolicies)\n\tpopulate(objectMap, \"useLocalAzureIpAddress\", v.UseLocalAzureIPAddress)\n\tpopulate(objectMap, \"usePolicyBasedTrafficSelectors\", v.UsePolicyBasedTrafficSelectors)\n\tpopulate(objectMap, \"vpnConnectionProtocolType\", v.VPNConnectionProtocolType)\n\tpopulate(objectMap, \"vpnLinkConnections\", v.VPNLinkConnections)\n\treturn json.Marshal(objectMap)\n}", "func (Mapper) ModelToEntity(m, e interface{}) error {\r\n\tobj, err := json.Marshal(m)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\treturn json.Unmarshal(obj, e)\r\n}", "func (c ProtocolType) ToPtr() *ProtocolType {\n\treturn &c\n}", "func (object Object) To(value interface{}) Object {\n\treturn object.Property(as.PropertyTo, value)\n}", "func VertexaiModelDeployedModelsToProto(o *vertexai.ModelDeployedModels) *vertexaipb.VertexaiModelDeployedModels {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelDeployedModels{}\n\tp.SetEndpoint(dcl.ValueOrEmptyString(o.Endpoint))\n\tp.SetDeployedModelId(dcl.ValueOrEmptyString(o.DeployedModelId))\n\treturn p\n}", "func (me *Model) ToRPCMsg(escape bool) *rpc.Message {\n\tret := &rpc.Message{\n\t\tId: me.ID.Hex(),\n\t\tTopicId: me.TopicID.Hex(),\n\t\tSenderName: me.SenderName,\n\t\tPostTime: &timestamp.Timestamp{\n\t\t\tSeconds: me.PostTime.Unix(),\n\t\t\tNanos: int32(me.PostTime.Nanosecond()),\n\t\t},\n\t\tMessage: me.Message,\n\t\tBump: me.Bump,\n\t}\n\tif escape {\n\t\tret.SenderName = html.EscapeString(me.SenderName)\n\t\tret.Message = html.EscapeString(me.Message)\n\t}\n\treturn ret\n}", "func VertexaiEndpointEncryptionSpecToProto(o *vertexai.EndpointEncryptionSpec) *vertexaipb.VertexaiEndpointEncryptionSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiEndpointEncryptionSpec{}\n\tp.SetKmsKeyName(dcl.ValueOrEmptyString(o.KmsKeyName))\n\treturn p\n}", "func (pool *Workspaces_BigDataPool_Spec) ConvertSpecTo(destination genruntime.ConvertibleSpec) error {\n\tif destination == pool {\n\t\treturn errors.New(\"attempted conversion between unrelated implementations of github.com/Azure/azure-service-operator/v2/pkg/genruntime/ConvertibleSpec\")\n\t}\n\n\treturn destination.ConvertSpecFrom(pool)\n}", "func (source *PingSource) ConvertTo(ctx context.Context, obj apis.Convertible) error {\n\tswitch sink := obj.(type) {\n\tcase *v1beta1.PingSource:\n\t\tsink.ObjectMeta = source.ObjectMeta\n\t\tsink.Spec = v1beta1.PingSourceSpec{\n\t\t\tJsonData: source.Spec.JsonData,\n\t\t\tSourceSpec: source.Spec.SourceSpec,\n\t\t}\n\t\tsink.Status = v1beta1.PingSourceStatus{\n\t\t\tSourceStatus: source.Status.SourceStatus,\n\t\t}\n\n\t\t// in v1beta1, timezone has its own field\n\t\tschedule := source.Spec.Schedule\n\t\tif strings.HasPrefix(schedule, \"TZ=\") || strings.HasPrefix(schedule, \"CRON_TZ=\") {\n\t\t\ti := strings.Index(schedule, \" \")\n\t\t\teq := strings.Index(schedule, \"=\")\n\t\t\tsink.Spec.Timezone = schedule[eq+1 : i]\n\t\t\tsink.Spec.Schedule = strings.TrimSpace(schedule[i:])\n\t\t} else {\n\t\t\tsink.Spec.Schedule = schedule\n\t\t}\n\n\t\treturn nil\n\tdefault:\n\t\treturn apis.ConvertToViaProxy(ctx, source, &v1beta1.PingSource{}, sink)\n\t}\n}", "func ProtoToComputeBetaInstanceTemplateProperties(p *betapb.ComputeBetaInstanceTemplateProperties) *beta.InstanceTemplateProperties {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &beta.InstanceTemplateProperties{\n\t\tCanIPForward: dcl.Bool(p.CanIpForward),\n\t\tDescription: dcl.StringOrNil(p.Description),\n\t\tMachineType: dcl.StringOrNil(p.MachineType),\n\t\tMinCpuPlatform: dcl.StringOrNil(p.MinCpuPlatform),\n\t\tReservationAffinity: ProtoToComputeBetaInstanceTemplatePropertiesReservationAffinity(p.GetReservationAffinity()),\n\t\tShieldedInstanceConfig: ProtoToComputeBetaInstanceTemplatePropertiesShieldedInstanceConfig(p.GetShieldedInstanceConfig()),\n\t\tScheduling: ProtoToComputeBetaInstanceTemplatePropertiesScheduling(p.GetScheduling()),\n\t}\n\tfor _, r := range p.GetDisks() {\n\t\tobj.Disks = append(obj.Disks, *ProtoToComputeBetaInstanceTemplatePropertiesDisks(r))\n\t}\n\tfor _, r := range p.GetGuestAccelerators() {\n\t\tobj.GuestAccelerators = append(obj.GuestAccelerators, *ProtoToComputeBetaInstanceTemplatePropertiesGuestAccelerators(r))\n\t}\n\tfor _, r := range p.GetNetworkInterfaces() {\n\t\tobj.NetworkInterfaces = append(obj.NetworkInterfaces, *ProtoToComputeBetaInstanceTemplatePropertiesNetworkInterfaces(r))\n\t}\n\tfor _, r := range p.GetServiceAccounts() {\n\t\tobj.ServiceAccounts = append(obj.ServiceAccounts, *ProtoToComputeBetaInstanceTemplatePropertiesServiceAccounts(r))\n\t}\n\tfor _, r := range p.GetTags() {\n\t\tobj.Tags = append(obj.Tags, r)\n\t}\n\treturn obj\n}", "func (k *SdkShapeConverter) SdkPropertiesToRequestBody(props map[string]CloudAPIProperty,\n\tinputs map[string]interface{}, state map[string]interface{}) map[string]interface{} {\n\n\tbody := map[string]interface{}{}\n\tfor name, prop := range props {\n\t\tp := prop // https://github.com/golang/go/wiki/CommonMistakes#using-reference-to-loop-iterator-variable\n\t\tparent := body\n\t\tsdkName := name\n\t\tif prop.SdkName != \"\" {\n\t\t\tsdkName = prop.SdkName\n\t\t}\n\n\t\tvalues := inputs\n\t\tif prop.CopyFromOutputs && state != nil {\n\t\t\tvalues = state\n\t\t}\n\n\t\tif value, has := values[sdkName]; has {\n\t\t\tif prop.Container != \"\" {\n\t\t\t\tif v, has := body[prop.Container].(map[string]interface{}); has {\n\t\t\t\t\tparent = v\n\t\t\t\t} else {\n\t\t\t\t\tparent = map[string]interface{}{}\n\t\t\t\t\tbody[prop.Container] = parent\n\t\t\t\t}\n\t\t\t}\n\t\t\tparent[name] = k.convertPropValue(&p, value, state[sdkName], k.SdkPropertiesToRequestBody)\n\t\t}\n\t}\n\treturn body\n}", "func ComputeFirewallAllowedToProto(o *compute.FirewallAllowed) *computepb.ComputeFirewallAllowed {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &computepb.ComputeFirewallAllowed{\n\t\tIpProtocol: dcl.ValueOrEmptyString(o.IPProtocol),\n\t}\n\tfor _, r := range o.Ports {\n\t\tp.Ports = append(p.Ports, r)\n\t}\n\tfor _, r := range o.IPProtocolAlt {\n\t\tp.IpProtocolAlt = append(p.IpProtocolAlt, r)\n\t}\n\treturn p\n}", "func (v *VPNSiteLinkConnectionProperties) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"connectionBandwidth\":\n\t\t\terr = unpopulate(val, \"ConnectionBandwidth\", &v.ConnectionBandwidth)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"connectionStatus\":\n\t\t\terr = unpopulate(val, \"ConnectionStatus\", &v.ConnectionStatus)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"egressBytesTransferred\":\n\t\t\terr = unpopulate(val, \"EgressBytesTransferred\", &v.EgressBytesTransferred)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"egressNatRules\":\n\t\t\terr = unpopulate(val, \"EgressNatRules\", &v.EgressNatRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableBgp\":\n\t\t\terr = unpopulate(val, \"EnableBgp\", &v.EnableBgp)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"enableRateLimiting\":\n\t\t\terr = unpopulate(val, \"EnableRateLimiting\", &v.EnableRateLimiting)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ipsecPolicies\":\n\t\t\terr = unpopulate(val, \"IPSecPolicies\", &v.IPSecPolicies)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ingressBytesTransferred\":\n\t\t\terr = unpopulate(val, \"IngressBytesTransferred\", &v.IngressBytesTransferred)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"ingressNatRules\":\n\t\t\terr = unpopulate(val, \"IngressNatRules\", &v.IngressNatRules)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &v.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"routingWeight\":\n\t\t\terr = unpopulate(val, \"RoutingWeight\", &v.RoutingWeight)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedKey\":\n\t\t\terr = unpopulate(val, \"SharedKey\", &v.SharedKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"useLocalAzureIpAddress\":\n\t\t\terr = unpopulate(val, \"UseLocalAzureIPAddress\", &v.UseLocalAzureIPAddress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"usePolicyBasedTrafficSelectors\":\n\t\t\terr = unpopulate(val, \"UsePolicyBasedTrafficSelectors\", &v.UsePolicyBasedTrafficSelectors)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vpnConnectionProtocolType\":\n\t\t\terr = unpopulate(val, \"VPNConnectionProtocolType\", &v.VPNConnectionProtocolType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vpnGatewayCustomBgpAddresses\":\n\t\t\terr = unpopulate(val, \"VPNGatewayCustomBgpAddresses\", &v.VPNGatewayCustomBgpAddresses)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vpnLinkConnectionMode\":\n\t\t\terr = unpopulate(val, \"VPNLinkConnectionMode\", &v.VPNLinkConnectionMode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vpnSiteLink\":\n\t\t\terr = unpopulate(val, \"VPNSiteLink\", &v.VPNSiteLink)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v VPNSiteLinkConnectionProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"connectionBandwidth\", v.ConnectionBandwidth)\n\tpopulate(objectMap, \"connectionStatus\", v.ConnectionStatus)\n\tpopulate(objectMap, \"egressBytesTransferred\", v.EgressBytesTransferred)\n\tpopulate(objectMap, \"egressNatRules\", v.EgressNatRules)\n\tpopulate(objectMap, \"enableBgp\", v.EnableBgp)\n\tpopulate(objectMap, \"enableRateLimiting\", v.EnableRateLimiting)\n\tpopulate(objectMap, \"ipsecPolicies\", v.IPSecPolicies)\n\tpopulate(objectMap, \"ingressBytesTransferred\", v.IngressBytesTransferred)\n\tpopulate(objectMap, \"ingressNatRules\", v.IngressNatRules)\n\tpopulate(objectMap, \"provisioningState\", v.ProvisioningState)\n\tpopulate(objectMap, \"routingWeight\", v.RoutingWeight)\n\tpopulate(objectMap, \"sharedKey\", v.SharedKey)\n\tpopulate(objectMap, \"useLocalAzureIpAddress\", v.UseLocalAzureIPAddress)\n\tpopulate(objectMap, \"usePolicyBasedTrafficSelectors\", v.UsePolicyBasedTrafficSelectors)\n\tpopulate(objectMap, \"vpnConnectionProtocolType\", v.VPNConnectionProtocolType)\n\tpopulate(objectMap, \"vpnGatewayCustomBgpAddresses\", v.VPNGatewayCustomBgpAddresses)\n\tpopulate(objectMap, \"vpnLinkConnectionMode\", v.VPNLinkConnectionMode)\n\tpopulate(objectMap, \"vpnSiteLink\", v.VPNSiteLink)\n\treturn json.Marshal(objectMap)\n}", "func (c ForwardingProtocol) ToPtr() *ForwardingProtocol {\n\treturn &c\n}", "func (setting *Servers_AdvancedThreatProtectionSetting_Spec) ConvertSpecTo(destination genruntime.ConvertibleSpec) error {\n\tif destination == setting {\n\t\treturn errors.New(\"attempted conversion between unrelated implementations of github.com/Azure/azure-service-operator/v2/pkg/genruntime/ConvertibleSpec\")\n\t}\n\n\treturn destination.ConvertSpecFrom(setting)\n}", "func (w *NotificationPolicy) ConvertFromDBModel() error {\n\ttargets := []EventTarget{}\n\tif len(w.TargetsDB) != 0 {\n\t\terr := json.Unmarshal([]byte(w.TargetsDB), &targets)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tw.Targets = targets\n\n\ttypes := []string{}\n\tif len(w.EventTypesDB) != 0 {\n\t\terr := json.Unmarshal([]byte(w.EventTypesDB), &types)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tw.EventTypes = types\n\n\treturn nil\n}", "func (op *Operation) ToProto(wrappedID string) *api.UploadOperation {\n\tvar ref *api.ObjectRef\n\tif op.Status == api.UploadStatus_PUBLISHED {\n\t\tref = &api.ObjectRef{\n\t\t\tHashAlgo: op.HashAlgo,\n\t\t\tHexDigest: op.HexDigest,\n\t\t}\n\t}\n\treturn &api.UploadOperation{\n\t\tOperationId: wrappedID,\n\t\tUploadUrl: op.UploadURL,\n\t\tStatus: op.Status,\n\t\tObject: ref,\n\t\tErrorMessage: op.Error,\n\t}\n}", "func (u *UnknownPathAttribute) ToProto() *api.UnknownPathAttribute {\n\ta := &api.UnknownPathAttribute{\n\t\tOptional: u.Optional,\n\t\tTransitive: u.Transitive,\n\t\tPartial: u.Partial,\n\t\tTypeCode: uint32(u.TypeCode),\n\t\tValue: make([]byte, len(u.Value)),\n\t}\n\n\tcopy(a.Value, u.Value)\n\treturn a\n}", "func ProtoToComputeFirewallAllowed(p *computepb.ComputeFirewallAllowed) *compute.FirewallAllowed {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &compute.FirewallAllowed{\n\t\tIPProtocol: dcl.StringOrNil(p.IpProtocol),\n\t}\n\tfor _, r := range p.GetPorts() {\n\t\tobj.Ports = append(obj.Ports, r)\n\t}\n\tfor _, r := range p.GetIpProtocolAlt() {\n\t\tobj.IPProtocolAlt = append(obj.IPProtocolAlt, r)\n\t}\n\treturn obj\n}", "func (m *Metric) ToLineProtocol() string {\n\tvar fields strings.Builder\n\tisFirst := true\n\tfor _, field := range m.Fields {\n\t\t// Fields section doesn't have a leading comma per syntax above so need to skip the comma on the first field\n\t\tif isFirst {\n\t\t\tisFirst = false\n\t\t} else {\n\t\t\tfields.WriteString(\",\")\n\t\t}\n\t\tfields.WriteString(field.Name + \"=\" + formatLineProtocolValue(field.Value))\n\t}\n\n\t// Tags section does have a leading comma per syntax above\n\tvar tags strings.Builder\n\tfor _, tag := range m.Tags {\n\t\ttags.WriteString(\",\" + tag.Name + \"=\" + tag.Value)\n\t}\n\n\tresult := fmt.Sprintf(\"%s%s %s %d\", m.Name, tags.String(), fields.String(), m.Timestamp)\n\n\treturn result\n}", "func (in *V1SystemDeploymentParameters) DeepCopyToModelInto(out *models.V1SystemDeploymentParameters) {\n\t*out = models.V1SystemDeploymentParameters{}\n\tif in.DenyOnOpaFail != nil {\n\t\tin, out := &in.DenyOnOpaFail, &out.DenyOnOpaFail\n\t\t*out = new(bool)\n\t\t**out = **in\n\t}\n\tout.HTTPProxy = in.HTTPProxy\n\tout.HTTPSProxy = in.HTTPSProxy\n\tout.KubernetesVersion = in.KubernetesVersion\n\tout.Namespace = in.Namespace\n\tout.NoProxy = in.NoProxy\n\tout.TimeoutSeconds = in.TimeoutSeconds\n\tif in.TrustedCaCerts != nil {\n\t\tin, out := &in.TrustedCaCerts, &out.TrustedCaCerts\n\t\t*out = make([]string, len(*in))\n\t\tcopy(*out, *in)\n\t}\n\tout.TrustedContainerRegistry = in.TrustedContainerRegistry\n}", "func (sg *SubGraph) ToProtocolBuffer(l *Latency) (*graph.Node, error) {\n\tvar seedNode *protoOutputNode\n\tif sg.SrcUIDs == nil {\n\t\treturn seedNode.New(sg.Attr).(*protoOutputNode).Node, nil\n\t}\n\n\tx.AssertTrue(len(sg.uidMatrix) == 1)\n\tn := seedNode.New(sg.Attr)\n\tul := sg.uidMatrix[0]\n\tif sg.Params.GetUID || sg.Params.isDebug {\n\t\tn.SetUID(ul.Uids[0])\n\t}\n\n\tif rerr := sg.preTraverse(ul.Uids[0], n); rerr != nil {\n\t\treturn n.(*protoOutputNode).Node, rerr\n\t}\n\n\tl.ProtocolBuffer = time.Since(l.Start) - l.Parsing - l.Processing\n\treturn n.(*protoOutputNode).Node, nil\n}", "func VertexaiEndpointDeployedModelsToProto(o *vertexai.EndpointDeployedModels) *vertexaipb.VertexaiEndpointDeployedModels {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiEndpointDeployedModels{}\n\tp.SetDedicatedResources(VertexaiEndpointDeployedModelsDedicatedResourcesToProto(o.DedicatedResources))\n\tp.SetAutomaticResources(VertexaiEndpointDeployedModelsAutomaticResourcesToProto(o.AutomaticResources))\n\tp.SetId(dcl.ValueOrEmptyString(o.Id))\n\tp.SetModel(dcl.ValueOrEmptyString(o.Model))\n\tp.SetModelVersionId(dcl.ValueOrEmptyString(o.ModelVersionId))\n\tp.SetDisplayName(dcl.ValueOrEmptyString(o.DisplayName))\n\tp.SetCreateTime(dcl.ValueOrEmptyString(o.CreateTime))\n\tp.SetServiceAccount(dcl.ValueOrEmptyString(o.ServiceAccount))\n\tp.SetEnableAccessLogging(dcl.ValueOrEmptyBool(o.EnableAccessLogging))\n\tp.SetPrivateEndpoints(VertexaiEndpointDeployedModelsPrivateEndpointsToProto(o.PrivateEndpoints))\n\treturn p\n}", "func (p Properties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"allowedFqdnList\", p.AllowedFqdnList)\n\tpopulate(objectMap, \"authorizationPolicies\", p.AuthorizationPolicies)\n\tpopulate(objectMap, \"cloudToDevice\", p.CloudToDevice)\n\tpopulate(objectMap, \"comments\", p.Comments)\n\tpopulate(objectMap, \"disableDeviceSAS\", p.DisableDeviceSAS)\n\tpopulate(objectMap, \"disableLocalAuth\", p.DisableLocalAuth)\n\tpopulate(objectMap, \"disableModuleSAS\", p.DisableModuleSAS)\n\tpopulate(objectMap, \"enableDataResidency\", p.EnableDataResidency)\n\tpopulate(objectMap, \"enableFileUploadNotifications\", p.EnableFileUploadNotifications)\n\tpopulate(objectMap, \"eventHubEndpoints\", p.EventHubEndpoints)\n\tpopulate(objectMap, \"features\", p.Features)\n\tpopulate(objectMap, \"hostName\", p.HostName)\n\tpopulate(objectMap, \"ipFilterRules\", p.IPFilterRules)\n\tpopulate(objectMap, \"locations\", p.Locations)\n\tpopulate(objectMap, \"messagingEndpoints\", p.MessagingEndpoints)\n\tpopulate(objectMap, \"minTlsVersion\", p.MinTLSVersion)\n\tpopulate(objectMap, \"networkRuleSets\", p.NetworkRuleSets)\n\tpopulate(objectMap, \"privateEndpointConnections\", p.PrivateEndpointConnections)\n\tpopulate(objectMap, \"provisioningState\", p.ProvisioningState)\n\tpopulate(objectMap, \"publicNetworkAccess\", p.PublicNetworkAccess)\n\tpopulate(objectMap, \"restrictOutboundNetworkAccess\", p.RestrictOutboundNetworkAccess)\n\tpopulate(objectMap, \"routing\", p.Routing)\n\tpopulate(objectMap, \"state\", p.State)\n\tpopulate(objectMap, \"storageEndpoints\", p.StorageEndpoints)\n\treturn json.Marshal(objectMap)\n}", "func (peering *VirtualNetworks_VirtualNetworkPeering_Spec) ConvertSpecTo(destination genruntime.ConvertibleSpec) error {\n\tdst, ok := destination.(*v20201101s.VirtualNetworks_VirtualNetworkPeering_Spec)\n\tif ok {\n\t\t// Populate destination from our instance\n\t\treturn peering.AssignProperties_To_VirtualNetworks_VirtualNetworkPeering_Spec(dst)\n\t}\n\n\t// Convert to an intermediate form\n\tdst = &v20201101s.VirtualNetworks_VirtualNetworkPeering_Spec{}\n\terr := peering.AssignProperties_To_VirtualNetworks_VirtualNetworkPeering_Spec(dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"initial step of conversion in ConvertSpecTo()\")\n\t}\n\n\t// Update dst from our instance\n\terr = dst.ConvertSpecTo(destination)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"final step of conversion in ConvertSpecTo()\")\n\t}\n\n\treturn nil\n}", "func toFlowSpecParameters(obj Protection, protectionID int64) []db_models.FlowSpecParameter {\n\tresult := make([]db_models.FlowSpecParameter, 0)\n t, _ := obj.(*FlowSpec)\n\tfor _, target := range t.FlowSpecTargets() {\n\t\tresult = append(result, db_models.FlowSpecParameter{\n\t\t\tProtectionId: protectionID,\n\t\t\tFlowType: target.flowType,\n\t\t\tFlowSpec: target.flowSpec,\n\t\t})\n\t}\n\n\treturn result\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValueToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue{}\n\tp.SetHours(dcl.ValueOrEmptyInt64(o.Hours))\n\tp.SetMinutes(dcl.ValueOrEmptyInt64(o.Minutes))\n\tp.SetSeconds(dcl.ValueOrEmptyInt64(o.Seconds))\n\tp.SetNanos(dcl.ValueOrEmptyInt64(o.Nanos))\n\treturn p\n}", "func (m *modeKinds) ToProto() *api.ModeKinds {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tvar proto api.ModeKinds\n\n\tm.RLock()\n\tdefer m.RUnlock()\n\n\tif m.userPrefixes != nil {\n\t\tproto.UserPrefixes = make([]*api.ModeKinds_UserPrefix, len(m.userPrefixes))\n\t\tfor i, prefix := range m.userPrefixes {\n\t\t\tproto.UserPrefixes[i] = &api.ModeKinds_UserPrefix{\n\t\t\t\tSymbol: string(prefix[0]),\n\t\t\t\tChar: string(prefix[1]),\n\t\t\t}\n\t\t}\n\t}\n\tif m.channelModes != nil {\n\t\tproto.ChannelModes = make(map[string]int32, len(m.channelModes))\n\t\tfor k, v := range m.channelModes {\n\t\t\tproto.ChannelModes[string(k)] = int32(v)\n\t\t}\n\t}\n\n\treturn &proto\n}", "func ComputeBetaInstanceTemplatePropertiesToProto(o *beta.InstanceTemplateProperties) *betapb.ComputeBetaInstanceTemplateProperties {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.ComputeBetaInstanceTemplateProperties{\n\t\tCanIpForward: dcl.ValueOrEmptyBool(o.CanIPForward),\n\t\tDescription: dcl.ValueOrEmptyString(o.Description),\n\t\tMachineType: dcl.ValueOrEmptyString(o.MachineType),\n\t\tMinCpuPlatform: dcl.ValueOrEmptyString(o.MinCpuPlatform),\n\t\tReservationAffinity: ComputeBetaInstanceTemplatePropertiesReservationAffinityToProto(o.ReservationAffinity),\n\t\tShieldedInstanceConfig: ComputeBetaInstanceTemplatePropertiesShieldedInstanceConfigToProto(o.ShieldedInstanceConfig),\n\t\tScheduling: ComputeBetaInstanceTemplatePropertiesSchedulingToProto(o.Scheduling),\n\t}\n\tfor _, r := range o.Disks {\n\t\tp.Disks = append(p.Disks, ComputeBetaInstanceTemplatePropertiesDisksToProto(&r))\n\t}\n\tp.Labels = make(map[string]string)\n\tfor k, r := range o.Labels {\n\t\tp.Labels[k] = r\n\t}\n\tp.Metadata = make(map[string]string)\n\tfor k, r := range o.Metadata {\n\t\tp.Metadata[k] = r\n\t}\n\tfor _, r := range o.GuestAccelerators {\n\t\tp.GuestAccelerators = append(p.GuestAccelerators, ComputeBetaInstanceTemplatePropertiesGuestAcceleratorsToProto(&r))\n\t}\n\tfor _, r := range o.NetworkInterfaces {\n\t\tp.NetworkInterfaces = append(p.NetworkInterfaces, ComputeBetaInstanceTemplatePropertiesNetworkInterfacesToProto(&r))\n\t}\n\tfor _, r := range o.ServiceAccounts {\n\t\tp.ServiceAccounts = append(p.ServiceAccounts, ComputeBetaInstanceTemplatePropertiesServiceAccountsToProto(&r))\n\t}\n\tfor _, r := range o.Tags {\n\t\tp.Tags = append(p.Tags, r)\n\t}\n\treturn p\n}", "func (o BuildStrategySpecBuildStepsPortsOutput) Protocol() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BuildStrategySpecBuildStepsPorts) *string { return v.Protocol }).(pulumi.StringPtrOutput)\n}", "func (k *SdkShapeConverter) SdkPropertiesToRequestBody(props map[string]CloudAPIProperty,\n\tinputs map[string]interface{}, state map[string]interface{}) map[string]interface{} {\n\n\tbody := map[string]interface{}{}\n\tfor name, prop := range props {\n\t\tp := prop // https://github.com/golang/go/wiki/CommonMistakes#using-reference-to-loop-iterator-variable\n\t\tparent := body\n\t\tsdkName := name\n\t\tif prop.SdkName != \"\" {\n\t\t\tsdkName = prop.SdkName\n\t\t}\n\n\t\tvalues := inputs\n\t\tif prop.CopyFromOutputs && state != nil {\n\t\t\tvalues = state\n\t\t}\n\n\t\tif value, has := values[sdkName]; has {\n\t\t\tif prop.Flatten {\n\t\t\t\tchild := k.convertPropValue(&p, value, state[sdkName], k.SdkPropertiesToRequestBody)\n\t\t\t\tasMap := child.(map[string]interface{})\n\t\t\t\tfor k, v := range asMap {\n\t\t\t\t\tparent[k] = v\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif prop.Container != \"\" {\n\t\t\t\tif v, has := body[prop.Container].(map[string]interface{}); has {\n\t\t\t\t\tparent = v\n\t\t\t\t} else {\n\t\t\t\t\tparent = map[string]interface{}{}\n\t\t\t\t\tbody[prop.Container] = parent\n\t\t\t\t}\n\t\t\t}\n\t\t\tparent[name] = k.convertPropValue(&p, value, state[sdkName], k.SdkPropertiesToRequestBody)\n\t\t}\n\t}\n\treturn body\n}", "func (instance *Network) ToProtocol() (_ *protocol.Network, xerr fail.Error) {\n\tdefer fail.OnPanic(&xerr)\n\n\tif instance == nil || instance.IsNull() {\n\t\treturn nil, fail.InvalidInstanceError()\n\t}\n\n\tinstance.lock.RLock()\n\tdefer instance.lock.RUnlock()\n\n\tvar pn *protocol.Network\n\txerr = instance.Review(func(clonable data.Clonable, props *serialize.JSONProperties) fail.Error {\n\t\tan, ok := clonable.(*abstract.Network)\n\t\tif !ok {\n\t\t\treturn fail.InconsistentError(\"'*abstract.Networking' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\n\t\t}\n\n\t\tpn = &protocol.Network{\n\t\t\tId: an.ID,\n\t\t\tName: an.Name,\n\t\t\tCidr: an.CIDR,\n\t\t}\n\n\t\treturn props.Inspect(networkproperty.SubnetsV1, func(clonable data.Clonable) fail.Error {\n\t\t\tnsV1, ok := clonable.(*propertiesv1.NetworkSubnets)\n\t\t\tif !ok {\n\t\t\t\treturn fail.InconsistentError(\"'*propertiesv1.NetworkSubnets' expected, '%s' provided\", reflect.TypeOf(clonable).String())\n\t\t\t}\n\t\t\tfor k := range nsV1.ByName {\n\t\t\t\tpn.Subnets = append(pn.Subnets, k)\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t})\n\txerr = debug.InjectPlannedFail(xerr)\n\tif xerr != nil {\n\t\treturn nil, xerr\n\t}\n\n\treturn pn, nil\n}", "func PrivatecaBetaCertificateTemplatePredefinedValuesAdditionalExtensionsObjectIdToProto(o *beta.CertificateTemplatePredefinedValuesAdditionalExtensionsObjectId) *betapb.PrivatecaBetaCertificateTemplatePredefinedValuesAdditionalExtensionsObjectId {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.PrivatecaBetaCertificateTemplatePredefinedValuesAdditionalExtensionsObjectId{}\n\tsObjectIdPath := make([]int64, len(o.ObjectIdPath))\n\tfor i, r := range o.ObjectIdPath {\n\t\tsObjectIdPath[i] = r\n\t}\n\tp.SetObjectIdPath(sObjectIdPath)\n\treturn p\n}", "func (in MQTTDevicePropertyValue) ToUnstructured() interface{} {\n\tif len(in.Raw) > 0 {\n\t\treturn in.Raw\n\t}\n\treturn nil\n}", "func MavenProtoRuntimeConfigToRuntimeConfig(proto *registryv1alpha1.MavenConfig_RuntimeConfig) (bufpluginconfig.MavenRuntimeConfig, error) {\n\tlibraries := proto.GetRuntimeLibraries()\n\tvar dependencies []bufpluginconfig.MavenDependencyConfig\n\tfor _, library := range libraries {\n\t\tdependencies = append(dependencies, ProtoMavenRuntimeLibraryToDependencyConfig(library))\n\t}\n\treturn bufpluginconfig.MavenRuntimeConfig{\n\t\tName: proto.GetName(),\n\t\tDeps: dependencies,\n\t\tOptions: proto.GetOptions(),\n\t}, nil\n}", "func (e *Server) ToModel() base.ModelInterface {\n\tm := model.Server{}\n\tbase.EntityToModel(&e.Entity, &m.Model)\n\tcreateProductInfoModel(&e.ProductInfo, &m.ProductInfo)\n\tm.OriginURIs.Chassis = e.OriginURIsChassis\n\tm.OriginURIs.System = e.OriginURIsSystem\n\tm.PhysicalUUID = e.PhysicalUUID\n\tm.Name = e.Name\n\tm.Description = e.Description\n\tm.Hostname = e.Hostname\n\tm.Type = e.Type\n\tm.Protocol = e.Protocol\n\tm.Vender = e.Vender\n\tm.PowerState = e.PowerState\n\tm.IndicatorLED = e.IndicatorLED\n\tm.Credential = e.Credential\n\tm.State = e.State\n\tm.Health = e.Health\n\t// ComputerSystem.Processors\n\tprocessors := []model.Processor{}\n\tfor i := range e.Processors {\n\t\tprocessors = append(processors, *e.Processors[i].ToModel())\n\t}\n\tm.ComputerSystem.Processors = processors\n\t// ComputerSystem.Memory\n\tmemory := []model.Memory{}\n\tfor i := range e.Memory {\n\t\tmemory = append(memory, *e.Memory[i].ToModel())\n\t}\n\tm.ComputerSystem.Memory = memory\n\n\t// ComputerSystem.EthernetInterfaces\n\tethernetInterfaces := []model.EthernetInterface{}\n\tfor i := range e.EthernetInterfaces {\n\t\tethernetInterfaces = append(ethernetInterfaces, *e.EthernetInterfaces[i].ToModel())\n\t}\n\tm.ComputerSystem.EthernetInterfaces = ethernetInterfaces\n\t// ComputerSystem.NetworkInterfaces\n\tnetworkInterfaces := []model.NetworkInterface{}\n\tfor i := range e.NetworkInterfaces {\n\t\tnetworkInterfaces = append(networkInterfaces, *e.NetworkInterfaces[i].ToModel())\n\t}\n\tm.ComputerSystem.NetworkInterfaces = networkInterfaces\n\t// ComputerSystem.Storages\n\tstorages := []model.Storage{}\n\tfor i := range e.Storages {\n\t\tstorages = append(storages, *e.Storages[i].ToModel())\n\t}\n\tm.ComputerSystem.Storages = storages\n\t// Chassis.Power\n\tcreateResourceModel(&e.Power.EmbeddedResource, &m.Chassis.Power.Resource)\n\tpowerControl := []model.PowerControl{}\n\tfor i := range e.Power.PowerControl {\n\t\tpowerControl = append(powerControl, *e.Power.PowerControl[i].ToModel())\n\t}\n\tm.Chassis.Power.PowerControl = powerControl\n\n\tpowerSupplies := []model.PowerSupply{}\n\tfor i := range e.Power.PowerSupplies {\n\t\tpowerSupplies = append(powerSupplies, *e.Power.PowerSupplies[i].ToModel())\n\t}\n\tm.Chassis.Power.PowerSupplies = powerSupplies\n\n\tredundancy := []model.Redundancy{}\n\tfor i := range e.Power.Redundancy {\n\t\tredundancy = append(redundancy, *e.Power.Redundancy[i].ToModel())\n\t}\n\tm.Chassis.Power.Redundancy = redundancy\n\t// Chassis.Thermal\n\tcreateResourceModel(&e.Thermal.EmbeddedResource, &m.Chassis.Thermal.Resource)\n\tfans := []model.Fan{}\n\tfor i := range e.Thermal.Fans {\n\t\tfans = append(fans, *e.Thermal.Fans[i].ToModel())\n\t}\n\tm.Chassis.Thermal.Fans = fans\n\t// Chassis.Boards\n\tboards := []model.Board{}\n\tfor i := range e.Boards {\n\t\tboards = append(boards, *e.Boards[i].ToModel())\n\t}\n\tm.Chassis.Boards = boards\n\t// Chassis.NetworkAdapters\n\tnetworkAdapters := []model.NetworkAdapter{}\n\tfor i := range e.NetworkAdapters {\n\t\tnetworkAdapters = append(networkAdapters, *e.NetworkAdapters[i].ToModel())\n\t}\n\tm.Chassis.NetworkAdapters = networkAdapters\n\t// Chassis.Drives\n\tdrives := []model.Drive{}\n\tfor i := range e.Drives {\n\t\tdrives = append(drives, *e.Drives[i].ToModel())\n\t}\n\tm.Chassis.Drives = drives\n\t// Chassis.PCIeDevices\n\tpcieDevices := []model.PCIeDevice{}\n\tfor i := range e.PCIeDevices {\n\t\tpcieDevices = append(pcieDevices, *e.PCIeDevices[i].ToModel())\n\t}\n\tm.Chassis.PCIeDevices = pcieDevices\n\treturn &m\n}", "func (in *V1AllowedMapping) DeepCopyToModelInto(out *models.V1AllowedMapping) {\n\t*out = models.V1AllowedMapping{}\n\tif in.Negated != nil {\n\t\tin, out := &in.Negated, &out.Negated\n\t\t*out = new(bool)\n\t\t**out = **in\n\t}\n\tif in.Path != nil {\n\t\tin, out := &in.Path, &out.Path\n\t\t*out = new(string)\n\t\t**out = **in\n\t}\n}", "func ProtoToVertexaiEndpointDeployedModels(p *vertexaipb.VertexaiEndpointDeployedModels) *vertexai.EndpointDeployedModels {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.EndpointDeployedModels{\n\t\tDedicatedResources: ProtoToVertexaiEndpointDeployedModelsDedicatedResources(p.GetDedicatedResources()),\n\t\tAutomaticResources: ProtoToVertexaiEndpointDeployedModelsAutomaticResources(p.GetAutomaticResources()),\n\t\tId: dcl.StringOrNil(p.GetId()),\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t\tModelVersionId: dcl.StringOrNil(p.GetModelVersionId()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tServiceAccount: dcl.StringOrNil(p.GetServiceAccount()),\n\t\tEnableAccessLogging: dcl.Bool(p.GetEnableAccessLogging()),\n\t\tPrivateEndpoints: ProtoToVertexaiEndpointDeployedModelsPrivateEndpoints(p.GetPrivateEndpoints()),\n\t}\n\treturn obj\n}", "func (jUser JSONUser) ConvertToModelUser() *schema.User {\n\tuserIDConverted, _ := strconv.ParseInt(jUser.ID, 10, 0)\n\n\tuser := schema.User{\n\t\tID: int(userIDConverted),\n\t\tName: jUser.Name,\n\t\tEmail: jUser.Email,\n\t\tPassword: jUser.Password,\n\t}\n\n\treturn &user\n}", "func (profile *Profile) ConvertTo(hub conversion.Hub) error {\n\tdestination, ok := hub.(*v20210601s.Profile)\n\tif !ok {\n\t\treturn fmt.Errorf(\"expected cdn/v1api20210601storage/Profile but received %T instead\", hub)\n\t}\n\n\treturn profile.AssignProperties_To_Profile(destination)\n}", "func ProtoToTargetVpnGateway(p *betapb.ComputeBetaTargetVpnGateway) *beta.TargetVpnGateway {\n\tobj := &beta.TargetVpnGateway{\n\t\tId: dcl.Int64OrNil(p.Id),\n\t\tName: dcl.StringOrNil(p.Name),\n\t\tDescription: dcl.StringOrNil(p.Description),\n\t\tRegion: dcl.StringOrNil(p.Region),\n\t\tNetwork: dcl.StringOrNil(p.Network),\n\t\tStatus: ProtoToComputeBetaTargetVpnGatewayStatusEnum(p.GetStatus()),\n\t\tSelfLink: dcl.StringOrNil(p.SelfLink),\n\t\tProject: dcl.StringOrNil(p.Project),\n\t}\n\tfor _, r := range p.GetTunnel() {\n\t\tobj.Tunnel = append(obj.Tunnel, r)\n\t}\n\tfor _, r := range p.GetForwardingRule() {\n\t\tobj.ForwardingRule = append(obj.ForwardingRule, r)\n\t}\n\treturn obj\n}", "func ProtoToEndpoint(p *vertexaipb.VertexaiEndpoint) *vertexai.Endpoint {\n\tobj := &vertexai.Endpoint{\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tEtag: dcl.StringOrNil(p.GetEtag()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tUpdateTime: dcl.StringOrNil(p.GetUpdateTime()),\n\t\tEncryptionSpec: ProtoToVertexaiEndpointEncryptionSpec(p.GetEncryptionSpec()),\n\t\tNetwork: dcl.StringOrNil(p.GetNetwork()),\n\t\tModelDeploymentMonitoringJob: dcl.StringOrNil(p.GetModelDeploymentMonitoringJob()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t}\n\tfor _, r := range p.GetDeployedModels() {\n\t\tobj.DeployedModels = append(obj.DeployedModels, *ProtoToVertexaiEndpointDeployedModels(r))\n\t}\n\treturn obj\n}", "func (peering *VirtualNetworksVirtualNetworkPeering) ConvertTo(hub conversion.Hub) error {\n\tdestination, ok := hub.(*v20201101s.VirtualNetworksVirtualNetworkPeering)\n\tif !ok {\n\t\treturn fmt.Errorf(\"expected network/v1api20201101storage/VirtualNetworksVirtualNetworkPeering but received %T instead\", hub)\n\t}\n\n\treturn peering.AssignProperties_To_VirtualNetworksVirtualNetworkPeering(destination)\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValueToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationReplaceConfigNewValueTimeValue{}\n\tp.SetHours(dcl.ValueOrEmptyInt64(o.Hours))\n\tp.SetMinutes(dcl.ValueOrEmptyInt64(o.Minutes))\n\tp.SetSeconds(dcl.ValueOrEmptyInt64(o.Seconds))\n\tp.SetNanos(dcl.ValueOrEmptyInt64(o.Nanos))\n\treturn p\n}", "func (pcj ParamChangeJSON) ToParamChange() types.ParamChange {\n\treturn types.ParamChange{Subspace: pcj.Subspace, Key: pcj.Key, Value: string(pcj.Value)}\n}", "func (m *ccMetric) ToLineProtocol(metaAsTags map[string]bool) string {\n\n\treturn write.PointToLineProtocol(\n\t\tm.ToPoint(metaAsTags),\n\t\ttime.Nanosecond,\n\t)\n}", "func (l LabVirtualMachineProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"allowClaim\", l.AllowClaim)\n\tpopulate(objectMap, \"applicableSchedule\", l.ApplicableSchedule)\n\tpopulate(objectMap, \"artifactDeploymentStatus\", l.ArtifactDeploymentStatus)\n\tpopulate(objectMap, \"artifacts\", l.Artifacts)\n\tpopulate(objectMap, \"computeId\", l.ComputeID)\n\tpopulate(objectMap, \"computeVm\", l.ComputeVM)\n\tpopulate(objectMap, \"createdByUser\", l.CreatedByUser)\n\tpopulate(objectMap, \"createdByUserId\", l.CreatedByUserID)\n\tpopulateTimeRFC3339(objectMap, \"createdDate\", l.CreatedDate)\n\tpopulate(objectMap, \"customImageId\", l.CustomImageID)\n\tpopulate(objectMap, \"dataDiskParameters\", l.DataDiskParameters)\n\tpopulate(objectMap, \"disallowPublicIpAddress\", l.DisallowPublicIPAddress)\n\tpopulate(objectMap, \"environmentId\", l.EnvironmentID)\n\tpopulateTimeRFC3339(objectMap, \"expirationDate\", l.ExpirationDate)\n\tpopulate(objectMap, \"fqdn\", l.Fqdn)\n\tpopulate(objectMap, \"galleryImageReference\", l.GalleryImageReference)\n\tpopulate(objectMap, \"isAuthenticationWithSshKey\", l.IsAuthenticationWithSSHKey)\n\tpopulate(objectMap, \"labSubnetName\", l.LabSubnetName)\n\tpopulate(objectMap, \"labVirtualNetworkId\", l.LabVirtualNetworkID)\n\tpopulate(objectMap, \"lastKnownPowerState\", l.LastKnownPowerState)\n\tpopulate(objectMap, \"networkInterface\", l.NetworkInterface)\n\tpopulate(objectMap, \"notes\", l.Notes)\n\tpopulate(objectMap, \"osType\", l.OSType)\n\tpopulate(objectMap, \"ownerObjectId\", l.OwnerObjectID)\n\tpopulate(objectMap, \"ownerUserPrincipalName\", l.OwnerUserPrincipalName)\n\tpopulate(objectMap, \"password\", l.Password)\n\tpopulate(objectMap, \"planId\", l.PlanID)\n\tpopulate(objectMap, \"provisioningState\", l.ProvisioningState)\n\tpopulate(objectMap, \"sshKey\", l.SSHKey)\n\tpopulate(objectMap, \"scheduleParameters\", l.ScheduleParameters)\n\tpopulate(objectMap, \"size\", l.Size)\n\tpopulate(objectMap, \"storageType\", l.StorageType)\n\tpopulate(objectMap, \"uniqueIdentifier\", l.UniqueIdentifier)\n\tpopulate(objectMap, \"userName\", l.UserName)\n\tpopulate(objectMap, \"virtualMachineCreationSource\", l.VirtualMachineCreationSource)\n\treturn json.Marshal(objectMap)\n}", "func (pr *PolicyReflector) labelSelectorToProto(selector *clientapi_metav1.LabelSelector) *proto.Policy_LabelSelector {\n\tselectorProto := &proto.Policy_LabelSelector{}\n\t// MatchLabels\n\tif selector.MatchLabels != nil {\n\t\tfor key, val := range selector.MatchLabels {\n\t\t\tselectorProto.MatchLabel = append(selectorProto.MatchLabel, &proto.Policy_Label{Key: key, Value: val})\n\t\t}\n\t}\n\t// MatchExpressions\n\tif selector.MatchExpressions != nil {\n\t\tfor _, expression := range selector.MatchExpressions {\n\t\t\texpressionProto := &proto.Policy_LabelSelector_LabelExpression{}\n\t\t\t// Key\n\t\t\texpressionProto.Key = expression.Key\n\t\t\t// Operator\n\t\t\tswitch expression.Operator {\n\t\t\tcase clientapi_metav1.LabelSelectorOpIn:\n\t\t\t\texpressionProto.Operator = proto.Policy_LabelSelector_LabelExpression_IN\n\t\t\tcase clientapi_metav1.LabelSelectorOpNotIn:\n\t\t\t\texpressionProto.Operator = proto.Policy_LabelSelector_LabelExpression_NOT_IN\n\t\t\tcase clientapi_metav1.LabelSelectorOpExists:\n\t\t\t\texpressionProto.Operator = proto.Policy_LabelSelector_LabelExpression_EXISTS\n\t\t\tcase clientapi_metav1.LabelSelectorOpDoesNotExist:\n\t\t\t\texpressionProto.Operator = proto.Policy_LabelSelector_LabelExpression_DOES_NOT_EXIST\n\n\t\t\t}\n\t\t\t// Values\n\t\t\tif expression.Values != nil {\n\t\t\t\tfor _, val := range expression.Values {\n\t\t\t\t\texpressionProto.Value = append(expressionProto.Value, val)\n\t\t\t\t}\n\t\t\t}\n\t\t\t// append expression\n\t\t\tselectorProto.MatchExpression = append(selectorProto.MatchExpression, expressionProto)\n\t\t}\n\t}\n\treturn selectorProto\n}", "func (o *ReservationModel) GetPropertyOk() (*EmbeddedPropertyModel, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Property, true\n}" ]
[ "0.6432634", "0.6401385", "0.6242521", "0.5165841", "0.51026475", "0.51025736", "0.5087682", "0.50459445", "0.4919206", "0.49014685", "0.4898491", "0.48583665", "0.4844294", "0.47884327", "0.4761342", "0.47591895", "0.47278386", "0.46968642", "0.46868664", "0.46734816", "0.46618605", "0.46424326", "0.46074793", "0.46073508", "0.4604718", "0.45994678", "0.45841476", "0.45827797", "0.4580468", "0.45609123", "0.45464587", "0.45266685", "0.45206386", "0.45129362", "0.45083344", "0.4489998", "0.44824252", "0.44758216", "0.4468918", "0.44654846", "0.4455204", "0.4445053", "0.44313136", "0.44302166", "0.4428461", "0.44076017", "0.44070777", "0.4398312", "0.43930566", "0.4392248", "0.43918702", "0.4390326", "0.43892908", "0.43879744", "0.4377167", "0.43763334", "0.43564418", "0.43560228", "0.43531597", "0.43530965", "0.43510422", "0.43406618", "0.43371707", "0.43362013", "0.433532", "0.4334532", "0.43274254", "0.43269408", "0.43242034", "0.43149823", "0.43136603", "0.4311794", "0.43064982", "0.4305984", "0.42972547", "0.42948836", "0.4292646", "0.42923364", "0.42907745", "0.42877", "0.42852786", "0.42831403", "0.4278173", "0.4272382", "0.4270221", "0.42691302", "0.4263922", "0.42612985", "0.4260703", "0.42565995", "0.42545673", "0.42534307", "0.4250296", "0.42461437", "0.4245868", "0.42374048", "0.42363742", "0.42312452", "0.42284372", "0.42272562" ]
0.76288784
0
ToProtocolModels transforms the Protocol DTO map to the Protocol model map
func ToProtocolModels(protocolDTOs map[string]ProtocolProperties) map[string]models.ProtocolProperties { protocolModels := make(map[string]models.ProtocolProperties) for k, protocolProperties := range protocolDTOs { protocolModels[k] = ToProtocolPropertiesModel(protocolProperties) } return protocolModels }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ToProtocolPropertiesModel(p ProtocolProperties) models.ProtocolProperties {\n\tprotocolProperties := make(models.ProtocolProperties)\n\tfor k, v := range p {\n\t\tprotocolProperties[k] = v\n\t}\n\treturn protocolProperties\n}", "func Map2Protocol(m map[string]string) Protocol {\n\tt, err := time.Parse(dateFormat, m[\"Time\"])\n\tif err != nil {\n\t\tfmt.Printf(\"err by parse date: %v\", err)\n\t}\n\tid, _ := strconv.Atoi(m[\"HeroID\"])\n\treturn Protocol{\n\t\tAction: m[\"Action\"],\n\t\tHeroID: int64(id),\n\t\tNote: m[\"Note\"],\n\t\tTime: t,\n\t}\n}", "func MapToFields(ctx context.Context, cmd SofaRpcCmd) (map[string]string, error) {\n\tprotocol := cmd.ProtocolCode()\n\n\tif conv, ok := sofaConvFactory[protocol]; ok {\n\t\treturn conv.MapToFields(ctx, cmd)\n\t}\n\treturn nil, rpc.ErrUnrecognizedCode\n}", "func Protocol2Map(p Protocol) map[string]string {\n\treturn map[string]string{\n\t\t\"Action\": p.Action,\n\t\t\"HeroID\": strconv.Itoa(int(p.HeroID)),\n\t\t\"Note\": p.Note,\n\t\t\"Time\": p.GetTimeString(),\n\t}\n}", "func ProtoToModel(p *vertexaipb.VertexaiModel) *vertexai.Model {\n\tobj := &vertexai.Model{\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tVersionId: dcl.StringOrNil(p.GetVersionId()),\n\t\tVersionCreateTime: dcl.StringOrNil(p.GetVersionCreateTime()),\n\t\tVersionUpdateTime: dcl.StringOrNil(p.GetVersionUpdateTime()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tVersionDescription: dcl.StringOrNil(p.GetVersionDescription()),\n\t\tTrainingPipeline: dcl.StringOrNil(p.GetTrainingPipeline()),\n\t\tOriginalModelInfo: ProtoToVertexaiModelOriginalModelInfo(p.GetOriginalModelInfo()),\n\t\tContainerSpec: ProtoToVertexaiModelContainerSpec(p.GetContainerSpec()),\n\t\tArtifactUri: dcl.StringOrNil(p.GetArtifactUri()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tUpdateTime: dcl.StringOrNil(p.GetUpdateTime()),\n\t\tEtag: dcl.StringOrNil(p.GetEtag()),\n\t\tEncryptionSpec: ProtoToVertexaiModelEncryptionSpec(p.GetEncryptionSpec()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t}\n\tfor _, r := range p.GetSupportedExportFormats() {\n\t\tobj.SupportedExportFormats = append(obj.SupportedExportFormats, *ProtoToVertexaiModelSupportedExportFormats(r))\n\t}\n\tfor _, r := range p.GetSupportedDeploymentResourcesTypes() {\n\t\tobj.SupportedDeploymentResourcesTypes = append(obj.SupportedDeploymentResourcesTypes, *ProtoToVertexaiModelSupportedDeploymentResourcesTypesEnum(r))\n\t}\n\tfor _, r := range p.GetSupportedInputStorageFormats() {\n\t\tobj.SupportedInputStorageFormats = append(obj.SupportedInputStorageFormats, r)\n\t}\n\tfor _, r := range p.GetSupportedOutputStorageFormats() {\n\t\tobj.SupportedOutputStorageFormats = append(obj.SupportedOutputStorageFormats, r)\n\t}\n\tfor _, r := range p.GetDeployedModels() {\n\t\tobj.DeployedModels = append(obj.DeployedModels, *ProtoToVertexaiModelDeployedModels(r))\n\t}\n\treturn obj\n}", "func ProtoToVertexaiModelDeployedModels(p *vertexaipb.VertexaiModelDeployedModels) *vertexai.ModelDeployedModels {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelDeployedModels{\n\t\tEndpoint: dcl.StringOrNil(p.GetEndpoint()),\n\t\tDeployedModelId: dcl.StringOrNil(p.GetDeployedModelId()),\n\t}\n\treturn obj\n}", "func ProtoToVertexaiEndpointDeployedModels(p *vertexaipb.VertexaiEndpointDeployedModels) *vertexai.EndpointDeployedModels {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.EndpointDeployedModels{\n\t\tDedicatedResources: ProtoToVertexaiEndpointDeployedModelsDedicatedResources(p.GetDedicatedResources()),\n\t\tAutomaticResources: ProtoToVertexaiEndpointDeployedModelsAutomaticResources(p.GetAutomaticResources()),\n\t\tId: dcl.StringOrNil(p.GetId()),\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t\tModelVersionId: dcl.StringOrNil(p.GetModelVersionId()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tServiceAccount: dcl.StringOrNil(p.GetServiceAccount()),\n\t\tEnableAccessLogging: dcl.Bool(p.GetEnableAccessLogging()),\n\t\tPrivateEndpoints: ProtoToVertexaiEndpointDeployedModelsPrivateEndpoints(p.GetPrivateEndpoints()),\n\t}\n\treturn obj\n}", "func mapToStruct(mi *modelInfo, structPtr interface{}, fMap FieldMap) {\n\tfMap = nestMap(fMap)\n\tval := reflect.ValueOf(structPtr)\n\tind := reflect.Indirect(val)\n\tif val.Kind() != reflect.Ptr || ind.Kind() != reflect.Struct {\n\t\ttools.LogAndPanic(log, \"structPtr must be a pointer to a struct\", \"structPtr\", structPtr)\n\t}\n\tfor i := 0; i < ind.NumField(); i++ {\n\t\tfVal := ind.Field(i)\n\t\tsf := ind.Type().Field(i)\n\t\tfi, ok := mi.fields.get(sf.Name)\n\t\tif !ok {\n\t\t\ttools.LogAndPanic(log, \"Unregistered field in model\", \"field\", sf.Name, \"model\", mi.name)\n\t\t}\n\t\tmValue, mValExists := fMap[fi.json]\n\t\tif sf.Type.Kind() == reflect.Ptr {\n\t\t\tif mValExists {\n\t\t\t\tfm, ok := mValue.(FieldMap)\n\t\t\t\tif !ok {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif !fVal.Elem().IsValid() {\n\t\t\t\t\t// Create the related struct if it does not exist\n\t\t\t\t\tfVal.Set(reflect.New(sf.Type.Elem()))\n\t\t\t\t}\n\t\t\t\tmapToStruct(fi.relatedModel, fVal.Interface(), fm)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif mValExists && mValue != nil {\n\t\t\tconvertedValue := reflect.ValueOf(mValue).Convert(fVal.Type())\n\t\t\tfVal.Set(convertedValue)\n\t\t}\n\t}\n}", "func MapToCmd(ctx context.Context, headerMap map[string]string) (SofaRpcCmd, error) {\n\n\t// TODO: temporary use bolt.HeaderProtocolCode, need to use common definition\n\tif proto, exist := headerMap[SofaPropertyHeader(HeaderProtocolCode)]; exist {\n\t\tprotoValue := ConvertPropertyValueUint8(proto)\n\t\tprotocolCode := protoValue\n\n\t\tif conv, ok := sofaConvFactory[protocolCode]; ok {\n\t\t\t//TODO: delete this copy\n\t\t\t// proxy downstream maybe retry, use the headerMap many times\n\t\t\t// if proxy downstream can keep a encoded header(iobuf)\n\t\t\t// map to cmd will be called only once, so we can modify the map without a copy\n\t\t\theaderCopy := make(map[string]string, len(headerMap))\n\t\t\tfor k, v := range headerMap {\n\t\t\t\theaderCopy[k] = v\n\t\t\t}\n\t\t\treturn conv.MapToCmd(ctx, headerCopy)\n\t\t}\n\t\treturn nil, rpc.ErrUnrecognizedCode\n\t}\n\treturn nil, rpc.ErrNoProtocolCode\n}", "func ProtoToDomainMapping(p *appenginepb.AppengineDomainMapping) *appengine.DomainMapping {\n\tobj := &appengine.DomainMapping{\n\t\tSelfLink: dcl.StringOrNil(p.SelfLink),\n\t\tName: dcl.StringOrNil(p.Name),\n\t\tSslSettings: ProtoToAppengineDomainMappingSslSettings(p.GetSslSettings()),\n\t\tApp: dcl.StringOrNil(p.App),\n\t}\n\tfor _, r := range p.GetResourceRecords() {\n\t\tobj.ResourceRecords = append(obj.ResourceRecords, *ProtoToAppengineDomainMappingResourceRecords(r))\n\t}\n\treturn obj\n}", "func VertexaiModelDeployedModelsToProto(o *vertexai.ModelDeployedModels) *vertexaipb.VertexaiModelDeployedModels {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelDeployedModels{}\n\tp.SetEndpoint(dcl.ValueOrEmptyString(o.Endpoint))\n\tp.SetDeployedModelId(dcl.ValueOrEmptyString(o.DeployedModelId))\n\treturn p\n}", "func VertexaiModelOriginalModelInfoToProto(o *vertexai.ModelOriginalModelInfo) *vertexaipb.VertexaiModelOriginalModelInfo {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelOriginalModelInfo{}\n\tp.SetModel(dcl.ValueOrEmptyString(o.Model))\n\treturn p\n}", "func VertexaiEndpointDeployedModelsToProto(o *vertexai.EndpointDeployedModels) *vertexaipb.VertexaiEndpointDeployedModels {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiEndpointDeployedModels{}\n\tp.SetDedicatedResources(VertexaiEndpointDeployedModelsDedicatedResourcesToProto(o.DedicatedResources))\n\tp.SetAutomaticResources(VertexaiEndpointDeployedModelsAutomaticResourcesToProto(o.AutomaticResources))\n\tp.SetId(dcl.ValueOrEmptyString(o.Id))\n\tp.SetModel(dcl.ValueOrEmptyString(o.Model))\n\tp.SetModelVersionId(dcl.ValueOrEmptyString(o.ModelVersionId))\n\tp.SetDisplayName(dcl.ValueOrEmptyString(o.DisplayName))\n\tp.SetCreateTime(dcl.ValueOrEmptyString(o.CreateTime))\n\tp.SetServiceAccount(dcl.ValueOrEmptyString(o.ServiceAccount))\n\tp.SetEnableAccessLogging(dcl.ValueOrEmptyBool(o.EnableAccessLogging))\n\tp.SetPrivateEndpoints(VertexaiEndpointDeployedModelsPrivateEndpointsToProto(o.PrivateEndpoints))\n\treturn p\n}", "func ProtoToModelDeployment(p *betapb.VertexaiBetaModelDeployment) *beta.ModelDeployment {\n\tobj := &beta.ModelDeployment{\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t\tDeployedModelId: dcl.StringOrNil(p.GetDeployedModelId()),\n\t\tDedicatedResources: ProtoToVertexaiBetaModelDeploymentDedicatedResources(p.GetDedicatedResources()),\n\t\tEndpoint: dcl.StringOrNil(p.GetEndpoint()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t}\n\treturn obj\n}", "func VertexaiModelContainerSpecPortsToProto(o *vertexai.ModelContainerSpecPorts) *vertexaipb.VertexaiModelContainerSpecPorts {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelContainerSpecPorts{}\n\tp.SetContainerPort(dcl.ValueOrEmptyInt64(o.ContainerPort))\n\treturn p\n}", "func ProtoToVertexaiModelContainerSpecPorts(p *vertexaipb.VertexaiModelContainerSpecPorts) *vertexai.ModelContainerSpecPorts {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelContainerSpecPorts{\n\t\tContainerPort: dcl.Int64OrNil(p.GetContainerPort()),\n\t}\n\treturn obj\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformations{}\n\tsTransformations := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformations, len(o.Transformations))\n\tfor i, r := range o.Transformations {\n\t\tsTransformations[i] = DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsToProto(&r)\n\t}\n\tp.SetTransformations(sTransformations)\n\treturn p\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformations{}\n\tp.SetPrimitiveTransformation(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationToProto(o.PrimitiveTransformation))\n\tsInfoTypes := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsInfoTypes, len(o.InfoTypes))\n\tfor i, r := range o.InfoTypes {\n\t\tsInfoTypes[i] = DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsInfoTypesToProto(&r)\n\t}\n\tp.SetInfoTypes(sInfoTypes)\n\treturn p\n}", "func ProtoToVertexaiModelContainerSpec(p *vertexaipb.VertexaiModelContainerSpec) *vertexai.ModelContainerSpec {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelContainerSpec{\n\t\tImageUri: dcl.StringOrNil(p.GetImageUri()),\n\t\tPredictRoute: dcl.StringOrNil(p.GetPredictRoute()),\n\t\tHealthRoute: dcl.StringOrNil(p.GetHealthRoute()),\n\t}\n\tfor _, r := range p.GetCommand() {\n\t\tobj.Command = append(obj.Command, r)\n\t}\n\tfor _, r := range p.GetArgs() {\n\t\tobj.Args = append(obj.Args, r)\n\t}\n\tfor _, r := range p.GetEnv() {\n\t\tobj.Env = append(obj.Env, *ProtoToVertexaiModelContainerSpecEnv(r))\n\t}\n\tfor _, r := range p.GetPorts() {\n\t\tobj.Ports = append(obj.Ports, *ProtoToVertexaiModelContainerSpecPorts(r))\n\t}\n\treturn obj\n}", "func DomainMappingToProto(resource *appengine.DomainMapping) *appenginepb.AppengineDomainMapping {\n\tp := &appenginepb.AppengineDomainMapping{\n\t\tSelfLink: dcl.ValueOrEmptyString(resource.SelfLink),\n\t\tName: dcl.ValueOrEmptyString(resource.Name),\n\t\tSslSettings: AppengineDomainMappingSslSettingsToProto(resource.SslSettings),\n\t\tApp: dcl.ValueOrEmptyString(resource.App),\n\t}\n\tfor _, r := range resource.ResourceRecords {\n\t\tp.ResourceRecords = append(p.ResourceRecords, AppengineDomainMappingResourceRecordsToProto(&r))\n\t}\n\n\treturn p\n}", "func (c Client) GetPortToProtocolMappingForService(svc service.MeshService) (map[uint32]string, error) {\n\tportToProtocolMap := make(map[uint32]string)\n\n\t// TODO\n\n\treturn portToProtocolMap, nil\n}", "func ModelProtocolToListenerProtocol(p protocol.Instance,\n\ttrafficDirection core.TrafficDirection) ListenerProtocol {\n\t// If protocol sniffing is not enabled, the default value is TCP\n\tif p == protocol.Unsupported {\n\t\tswitch trafficDirection {\n\t\tcase core.TrafficDirection_INBOUND:\n\t\t\tif !features.EnableProtocolSniffingForInbound {\n\t\t\t\tp = protocol.TCP\n\t\t\t}\n\t\tcase core.TrafficDirection_OUTBOUND:\n\t\t\tif !features.EnableProtocolSniffingForOutbound {\n\t\t\t\tp = protocol.TCP\n\t\t\t}\n\t\tdefault:\n\t\t\t// Should not reach here.\n\t\t}\n\t}\n\n\tswitch p {\n\tcase protocol.HTTP, protocol.HTTP2, protocol.GRPC, protocol.GRPCWeb:\n\t\treturn ListenerProtocolHTTP\n\tcase protocol.TCP, protocol.HTTPS, protocol.TLS,\n\t\tprotocol.Mongo, protocol.Redis, protocol.MySQL:\n\t\treturn ListenerProtocolTCP\n\tcase protocol.Thrift:\n\t\tif features.EnableThriftFilter {\n\t\t\treturn ListenerProtocolThrift\n\t\t}\n\t\treturn ListenerProtocolTCP\n\tcase protocol.UDP:\n\t\treturn ListenerProtocolUnknown\n\tcase protocol.Unsupported:\n\t\treturn ListenerProtocolAuto\n\tdefault:\n\t\t// Should not reach here.\n\t\treturn ListenerProtocolAuto\n\t}\n}", "func MapToStruct(request map[string]interface{}, result interface{}) error {\n\tcfg := &mapstructure.DecoderConfig{\n\t\tMetadata: nil,\n\t\tResult: result,\n\t\tTagName: \"json\",\n\t}\n\n\tdecoder, err := mapstructure.NewDecoder(cfg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = decoder.Decode(request)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func MapToStruct(params map[string]interface{}, contr interface{}) {\n\tmapstructure.Decode(params, &contr)\n}", "func ModelDeploymentToProto(resource *beta.ModelDeployment) *betapb.VertexaiBetaModelDeployment {\n\tp := &betapb.VertexaiBetaModelDeployment{}\n\tp.SetModel(dcl.ValueOrEmptyString(resource.Model))\n\tp.SetDeployedModelId(dcl.ValueOrEmptyString(resource.DeployedModelId))\n\tp.SetDedicatedResources(VertexaiBetaModelDeploymentDedicatedResourcesToProto(resource.DedicatedResources))\n\tp.SetEndpoint(dcl.ValueOrEmptyString(resource.Endpoint))\n\tp.SetLocation(dcl.ValueOrEmptyString(resource.Location))\n\tp.SetProject(dcl.ValueOrEmptyString(resource.Project))\n\n\treturn p\n}", "func (m *modeKinds) ToProto() *api.ModeKinds {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tvar proto api.ModeKinds\n\n\tm.RLock()\n\tdefer m.RUnlock()\n\n\tif m.userPrefixes != nil {\n\t\tproto.UserPrefixes = make([]*api.ModeKinds_UserPrefix, len(m.userPrefixes))\n\t\tfor i, prefix := range m.userPrefixes {\n\t\t\tproto.UserPrefixes[i] = &api.ModeKinds_UserPrefix{\n\t\t\t\tSymbol: string(prefix[0]),\n\t\t\t\tChar: string(prefix[1]),\n\t\t\t}\n\t\t}\n\t}\n\tif m.channelModes != nil {\n\t\tproto.ChannelModes = make(map[string]int32, len(m.channelModes))\n\t\tfor k, v := range m.channelModes {\n\t\t\tproto.ChannelModes[string(k)] = int32(v)\n\t\t}\n\t}\n\n\treturn &proto\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformations{}\n\tsTransformations := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformations, len(o.Transformations))\n\tfor i, r := range o.Transformations {\n\t\tsTransformations[i] = DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsToProto(&r)\n\t}\n\tp.SetTransformations(sTransformations)\n\treturn p\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformations{}\n\tp.SetPrimitiveTransformation(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsPrimitiveTransformationToProto(o.PrimitiveTransformation))\n\tsInfoTypes := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsInfoTypes, len(o.InfoTypes))\n\tfor i, r := range o.InfoTypes {\n\t\tsInfoTypes[i] = DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsTransformationsInfoTypesToProto(&r)\n\t}\n\tp.SetInfoTypes(sInfoTypes)\n\treturn p\n}", "func (r Role) MapToModel(m map[string]interface{}) Role {\n\tr.Id = m[\"id\"].(int)\n\tr.Name, _ = m[\"name\"].(string)\n\tr.Slug, _ = m[\"slug\"].(string)\n\tr.CreatedAt, _ = m[\"created_at\"].(string)\n\tr.UpdatedAt, _ = m[\"updated_at\"].(string)\n\treturn r\n}", "func ModelToProto(resource *vertexai.Model) *vertexaipb.VertexaiModel {\n\tp := &vertexaipb.VertexaiModel{}\n\tp.SetName(dcl.ValueOrEmptyString(resource.Name))\n\tp.SetVersionId(dcl.ValueOrEmptyString(resource.VersionId))\n\tp.SetVersionCreateTime(dcl.ValueOrEmptyString(resource.VersionCreateTime))\n\tp.SetVersionUpdateTime(dcl.ValueOrEmptyString(resource.VersionUpdateTime))\n\tp.SetDisplayName(dcl.ValueOrEmptyString(resource.DisplayName))\n\tp.SetDescription(dcl.ValueOrEmptyString(resource.Description))\n\tp.SetVersionDescription(dcl.ValueOrEmptyString(resource.VersionDescription))\n\tp.SetTrainingPipeline(dcl.ValueOrEmptyString(resource.TrainingPipeline))\n\tp.SetOriginalModelInfo(VertexaiModelOriginalModelInfoToProto(resource.OriginalModelInfo))\n\tp.SetContainerSpec(VertexaiModelContainerSpecToProto(resource.ContainerSpec))\n\tp.SetArtifactUri(dcl.ValueOrEmptyString(resource.ArtifactUri))\n\tp.SetCreateTime(dcl.ValueOrEmptyString(resource.CreateTime))\n\tp.SetUpdateTime(dcl.ValueOrEmptyString(resource.UpdateTime))\n\tp.SetEtag(dcl.ValueOrEmptyString(resource.Etag))\n\tp.SetEncryptionSpec(VertexaiModelEncryptionSpecToProto(resource.EncryptionSpec))\n\tp.SetProject(dcl.ValueOrEmptyString(resource.Project))\n\tp.SetLocation(dcl.ValueOrEmptyString(resource.Location))\n\tsSupportedExportFormats := make([]*vertexaipb.VertexaiModelSupportedExportFormats, len(resource.SupportedExportFormats))\n\tfor i, r := range resource.SupportedExportFormats {\n\t\tsSupportedExportFormats[i] = VertexaiModelSupportedExportFormatsToProto(&r)\n\t}\n\tp.SetSupportedExportFormats(sSupportedExportFormats)\n\tsSupportedDeploymentResourcesTypes := make([]vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum, len(resource.SupportedDeploymentResourcesTypes))\n\tfor i, r := range resource.SupportedDeploymentResourcesTypes {\n\t\tsSupportedDeploymentResourcesTypes[i] = vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum(vertexaipb.VertexaiModelSupportedDeploymentResourcesTypesEnum_value[string(r)])\n\t}\n\tp.SetSupportedDeploymentResourcesTypes(sSupportedDeploymentResourcesTypes)\n\tsSupportedInputStorageFormats := make([]string, len(resource.SupportedInputStorageFormats))\n\tfor i, r := range resource.SupportedInputStorageFormats {\n\t\tsSupportedInputStorageFormats[i] = r\n\t}\n\tp.SetSupportedInputStorageFormats(sSupportedInputStorageFormats)\n\tsSupportedOutputStorageFormats := make([]string, len(resource.SupportedOutputStorageFormats))\n\tfor i, r := range resource.SupportedOutputStorageFormats {\n\t\tsSupportedOutputStorageFormats[i] = r\n\t}\n\tp.SetSupportedOutputStorageFormats(sSupportedOutputStorageFormats)\n\tsDeployedModels := make([]*vertexaipb.VertexaiModelDeployedModels, len(resource.DeployedModels))\n\tfor i, r := range resource.DeployedModels {\n\t\tsDeployedModels[i] = VertexaiModelDeployedModelsToProto(&r)\n\t}\n\tp.SetDeployedModels(sDeployedModels)\n\tmLabels := make(map[string]string, len(resource.Labels))\n\tfor k, r := range resource.Labels {\n\t\tmLabels[k] = r\n\t}\n\tp.SetLabels(mLabels)\n\n\treturn p\n}", "func Msg2Map(msg proto.Message) map[string]interface{} {\n\tt := reflect.TypeOf(msg)\n\tv := reflect.ValueOf(msg)\n\n\treturn buildStruct(t, v)\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformations{}\n\tp.SetCondition(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsConditionToProto(o.Condition))\n\tp.SetPrimitiveTransformation(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsPrimitiveTransformationToProto(o.PrimitiveTransformation))\n\tp.SetInfoTypeTransformations(DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsInfoTypeTransformationsToProto(o.InfoTypeTransformations))\n\tsFields := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsFields, len(o.Fields))\n\tfor i, r := range o.Fields {\n\t\tsFields[i] = DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsFieldsToProto(&r)\n\t}\n\tp.SetFields(sFields)\n\treturn p\n}", "func (Files) ToModels(data interface{}, model *[]Files) error {\n\tbsonBytes, err := bson.Marshal(data.(bson.M))\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = bson.Unmarshal(bsonBytes, &model)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *Input) ToMap() map[string]interface{} {\n\treturn map[string]interface{}{\n\t\t\"message\": o.Message,\n\t}\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsToProto(o *beta.DeidentifyTemplateDeidentifyConfigRecordTransformations) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformations {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformations{}\n\tsFieldTransformations := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformations, len(o.FieldTransformations))\n\tfor i, r := range o.FieldTransformations {\n\t\tsFieldTransformations[i] = DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsFieldTransformationsToProto(&r)\n\t}\n\tp.SetFieldTransformations(sFieldTransformations)\n\tsRecordSuppressions := make([]*betapb.DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsRecordSuppressions, len(o.RecordSuppressions))\n\tfor i, r := range o.RecordSuppressions {\n\t\tsRecordSuppressions[i] = DlpBetaDeidentifyTemplateDeidentifyConfigRecordTransformationsRecordSuppressionsToProto(&r)\n\t}\n\tp.SetRecordSuppressions(sRecordSuppressions)\n\treturn p\n}", "func (p Pipeline) ToProto(pb *pipelinepb.AppliedPipeline) error {\n\tnumOps := len(p.Operations)\n\tif cap(pb.Ops) >= numOps {\n\t\tpb.Ops = pb.Ops[:numOps]\n\t} else {\n\t\tpb.Ops = make([]pipelinepb.AppliedPipelineOp, numOps)\n\t}\n\tfor i := 0; i < numOps; i++ {\n\t\tif err := p.Operations[i].ToProto(&pb.Ops[i]); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func VertexaiModelContainerSpecToProto(o *vertexai.ModelContainerSpec) *vertexaipb.VertexaiModelContainerSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelContainerSpec{}\n\tp.SetImageUri(dcl.ValueOrEmptyString(o.ImageUri))\n\tp.SetPredictRoute(dcl.ValueOrEmptyString(o.PredictRoute))\n\tp.SetHealthRoute(dcl.ValueOrEmptyString(o.HealthRoute))\n\tsCommand := make([]string, len(o.Command))\n\tfor i, r := range o.Command {\n\t\tsCommand[i] = r\n\t}\n\tp.SetCommand(sCommand)\n\tsArgs := make([]string, len(o.Args))\n\tfor i, r := range o.Args {\n\t\tsArgs[i] = r\n\t}\n\tp.SetArgs(sArgs)\n\tsEnv := make([]*vertexaipb.VertexaiModelContainerSpecEnv, len(o.Env))\n\tfor i, r := range o.Env {\n\t\tsEnv[i] = VertexaiModelContainerSpecEnvToProto(&r)\n\t}\n\tp.SetEnv(sEnv)\n\tsPorts := make([]*vertexaipb.VertexaiModelContainerSpecPorts, len(o.Ports))\n\tfor i, r := range o.Ports {\n\t\tsPorts[i] = VertexaiModelContainerSpecPortsToProto(&r)\n\t}\n\tp.SetPorts(sPorts)\n\treturn p\n}", "func (sg *SubGraph) ToProtocolBuffer(l *Latency) (*graph.Node, error) {\n\tvar seedNode *protoOutputNode\n\tif sg.SrcUIDs == nil {\n\t\treturn seedNode.New(sg.Attr).(*protoOutputNode).Node, nil\n\t}\n\n\tx.AssertTrue(len(sg.uidMatrix) == 1)\n\tn := seedNode.New(sg.Attr)\n\tul := sg.uidMatrix[0]\n\tif sg.Params.GetUID || sg.Params.isDebug {\n\t\tn.SetUID(ul.Uids[0])\n\t}\n\n\tif rerr := sg.preTraverse(ul.Uids[0], n); rerr != nil {\n\t\treturn n.(*protoOutputNode).Node, rerr\n\t}\n\n\tl.ProtocolBuffer = time.Since(l.Start) - l.Parsing - l.Processing\n\treturn n.(*protoOutputNode).Node, nil\n}", "func (c *_Comments) ToModels(input ...*Comment) []mapping.Model {\n\toutput := make([]mapping.Model, len(input))\n\tfor i := range input {\n\t\toutput[i] = input[i]\n\t}\n\treturn output\n}", "func AppengineDomainMappingSslSettingsToProto(o *appengine.DomainMappingSslSettings) *appenginepb.AppengineDomainMappingSslSettings {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &appenginepb.AppengineDomainMappingSslSettings{\n\t\tCertificateId: dcl.ValueOrEmptyString(o.CertificateId),\n\t\tSslManagementType: AppengineDomainMappingSslSettingsSslManagementTypeEnumToProto(o.SslManagementType),\n\t\tPendingManagedCertificateId: dcl.ValueOrEmptyString(o.PendingManagedCertificateId),\n\t}\n\treturn p\n}", "func (Files) ToModel(data interface{}, model *Files) error {\n\tbsonBytes, err := bson.Marshal(data.(bson.M))\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = bson.Unmarshal(bsonBytes, &model)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *ormPlugin) generateMapFunctions(message *generator.Descriptor) {\n\tccTypeNamePb := generator.CamelCaseSlice(message.TypeName())\n\tccTypeNameBase := lintName(ccTypeNamePb)\n\tccTypeNameOrm := fmt.Sprintf(\"%sORM\", ccTypeNameBase)\n\t///// To Orm\n\tp.P(`// Convert`, ccTypeNameBase, `ToORM takes a pb object and returns an orm object`)\n\tp.P(`func Convert`, ccTypeNameBase, `ToORM (from `,\n\t\tccTypeNamePb, `) `, ccTypeNameOrm, ` {`)\n\tp.P(`to := `, ccTypeNameOrm, `{}`)\n\tfor _, field := range message.Field {\n\t\t// Checking if field is skipped\n\t\tif field.Options != nil {\n\t\t\tv, err := proto.GetExtension(field.Options, gorm.E_Field)\n\t\t\tif err == nil && v.(*gorm.GormFieldOptions) != nil {\n\t\t\t\tif v.(*gorm.GormFieldOptions).Drop != nil && *v.(*gorm.GormFieldOptions).Drop {\n\t\t\t\t\tp.P(`// Skipping field: `, p.GetOneOfFieldName(message, field))\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tp.generateFieldMap(message, field, true)\n\t}\n\tp.P(`return to`)\n\tp.P(`}`)\n\n\tp.P()\n\t///// To Pb\n\tp.P(`// Convert`, ccTypeNameBase, `FromORM takes an orm object and returns a pb object`)\n\tp.P(`func Convert`, ccTypeNameBase, `FromORM (from `, ccTypeNameOrm, `) `,\n\t\tccTypeNamePb, ` {`)\n\tp.P(`to := `, ccTypeNamePb, `{}`)\n\tfor _, field := range message.Field {\n\t\t// Checking if field is skipped\n\t\tif field.Options != nil {\n\t\t\tv, err := proto.GetExtension(field.Options, gorm.E_Field)\n\t\t\tif err == nil && v.(*gorm.GormFieldOptions) != nil {\n\t\t\t\tif v.(*gorm.GormFieldOptions).Drop != nil && *v.(*gorm.GormFieldOptions).Drop {\n\t\t\t\t\tp.P(`// Skipping field: `, p.GetOneOfFieldName(message, field))\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tp.generateFieldMap(message, field, false)\n\t}\n\tp.P(`return to`)\n\tp.P(`}`)\n}", "func VertexaiEndpointDeployedModelsDedicatedResourcesMachineSpecToProto(o *vertexai.EndpointDeployedModelsDedicatedResourcesMachineSpec) *vertexaipb.VertexaiEndpointDeployedModelsDedicatedResourcesMachineSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiEndpointDeployedModelsDedicatedResourcesMachineSpec{}\n\tp.SetMachineType(dcl.ValueOrEmptyString(o.MachineType))\n\tp.SetAcceleratorType(VertexaiEndpointDeployedModelsDedicatedResourcesMachineSpecAcceleratorTypeEnumToProto(o.AcceleratorType))\n\tp.SetAcceleratorCount(dcl.ValueOrEmptyInt64(o.AcceleratorCount))\n\treturn p\n}", "func (p *_Posts) ToModels(input ...*Post) []mapping.Model {\n\toutput := make([]mapping.Model, len(input))\n\tfor i := range input {\n\t\toutput[i] = input[i]\n\t}\n\treturn output\n}", "func MapPodsToAPIModel(pods []model.Pod) (result []*pods.Pod) {\n\tfor _, pod := range pods {\n\t\tresult = append(result, MapPodToAPIModel(pod))\n\t}\n\treturn result\n}", "func (c *TypeConverter) GenStructConverter(\n\tfromFields []*compile.FieldSpec,\n\ttoFields []*compile.FieldSpec,\n\tfieldMap map[string]FieldMapperEntry,\n) error {\n\t// Add compiled FieldSpecs to the FieldMapperEntry\n\tfieldMap = addSpecToMap(fieldMap, fromFields, \"\")\n\t// Check for vlaues not populated recursively by addSpecToMap\n\tfor k, v := range fieldMap {\n\t\tif fieldMap[k].Field == nil {\n\t\t\treturn errors.Errorf(\n\t\t\t\t\"Failed to find field ( %s ) for transform.\",\n\t\t\t\tv.QualifiedName,\n\t\t\t)\n\t\t}\n\t}\n\n\tc.useRecurGen = c.isRecursiveStruct(toFields) || c.isRecursiveStruct(fromFields)\n\n\tif c.useRecurGen && len(fieldMap) != 0 {\n\t\tc.append(\"inOriginal := in; _ = inOriginal\")\n\t\tc.append(\"outOriginal := out; _ = outOriginal\")\n\t}\n\n\terr := c.genStructConverter(\"\", \"\", \"\", fromFields, toFields, fieldMap, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func BuildProtoTypeMap() (map[string]*types.Package, map[string]*ProtoType, error) {\n\tpackages := make(map[string]*types.Package)\n\t// imp := importer.Default()\n\tfset := token.NewFileSet()\n\timp := importer.ForCompiler(fset, \"source\", nil)\n\n\timportPak := func(pak string) error {\n\t\tipak, err := imp.Import(pak)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tpackages[pak] = ipak\n\t\treturn nil\n\t}\n\n\t// import all packages\n\tallPackages := []string{\n\t\t\"github.com/paralin/go-dota2/protocol\",\n\t}\n\tfor _, pak := range allPackages {\n\t\tif err := importPak(pak); err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t}\n\n\tif err := importPak(\"github.com/paralin/go-steam/steamid\"); err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// build a map of proto types\n\tprotoMap := make(map[string]*ProtoType)\n\tfor pakStr, pak := range packages {\n\t\tscope := pak.Scope()\n\t\tfor _, nam := range scope.Names() {\n\t\t\tobj := scope.Lookup(nam)\n\t\t\tobjStr := obj.String()\n\t\t\tif !obj.Exported() {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif !TypeRegex.MatchString(objStr) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tprotoTyp := &ProtoType{\n\t\t\t\tPakStr: pakStr,\n\t\t\t\tPak: pak,\n\t\t\t\tTypeName: obj.Name(),\n\t\t\t\tObj: obj,\n\t\t\t\tTypeStr: obj.String(),\n\t\t\t}\n\n\t\t\tprotoMap[obj.Name()] = protoTyp\n\t\t}\n\t}\n\n\treturn packages, protoMap, nil\n}", "func (pro Protocol) ToK8s() v1.Protocol {\n\tif r := pros[pro]; r != \"\" {\n\t\treturn r\n\t}\n\treturn v1.ProtocolTCP\n}", "func ProtoToVertexaiModelOriginalModelInfo(p *vertexaipb.VertexaiModelOriginalModelInfo) *vertexai.ModelOriginalModelInfo {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelOriginalModelInfo{\n\t\tModel: dcl.StringOrNil(p.GetModel()),\n\t}\n\treturn obj\n}", "func mapProtoTaskToTask(obj *proto.Task) *Task {\n\tulid, _ := ulid.Parse(obj.Id)\n\treturn &Task{ulid, obj.Title, obj.Description, int32(obj.Completed)}\n}", "func (l *Link) ToMap() util.Map {\n\tm := util.Map{\n\t\t\"msgtype\": l.MsgType.String(),\n\t\t\"touser\": l.ToUserName.Value,\n\t\tl.MsgType.String(): util.Map{\n\t\t\t\"title\": l.Title,\n\t\t\t\"description\": l.Description,\n\t\t\t\"url\": l.URL,\n\t\t\t\"thumb_url\": l.ThumbURL,\n\t\t},\n\t}\n\treturn m\n}", "func InputTypeToStructMap(t string) IInputSchema {\n\tswitch t {\n\tcase TextInputType:\n\t\treturn TextInputSchema{}\n\tcase TableInputType:\n\t\treturn TableInputSchema{}\n\tcase ImageInputType:\n\t\treturn ImageInputSchema{}\n\t}\n\treturn nil\n}", "func (b *_Blogs) ToModels(input ...*Blog) []mapping.Model {\n\toutput := make([]mapping.Model, len(input))\n\tfor i := range input {\n\t\toutput[i] = input[i]\n\t}\n\treturn output\n}", "func AppengineDomainMappingResourceRecordsToProto(o *appengine.DomainMappingResourceRecords) *appenginepb.AppengineDomainMappingResourceRecords {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &appenginepb.AppengineDomainMappingResourceRecords{\n\t\tName: dcl.ValueOrEmptyString(o.Name),\n\t\tRrdata: dcl.ValueOrEmptyString(o.Rrdata),\n\t\tType: AppengineDomainMappingResourceRecordsTypeEnumToProto(o.Type),\n\t}\n\treturn p\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation{}\n\tp.SetReplaceConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfigToProto(o.ReplaceConfig))\n\tp.SetRedactConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationRedactConfigToProto(o.RedactConfig))\n\tp.SetCharacterMaskConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCharacterMaskConfigToProto(o.CharacterMaskConfig))\n\tp.SetCryptoReplaceFfxFpeConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoReplaceFfxFpeConfigToProto(o.CryptoReplaceFfxFpeConfig))\n\tp.SetFixedSizeBucketingConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationFixedSizeBucketingConfigToProto(o.FixedSizeBucketingConfig))\n\tp.SetBucketingConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationBucketingConfigToProto(o.BucketingConfig))\n\tp.SetReplaceWithInfoTypeConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceWithInfoTypeConfigToProto(o.ReplaceWithInfoTypeConfig))\n\tp.SetTimePartConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationTimePartConfigToProto(o.TimePartConfig))\n\tp.SetCryptoHashConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigToProto(o.CryptoHashConfig))\n\tp.SetDateShiftConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationDateShiftConfigToProto(o.DateShiftConfig))\n\tp.SetCryptoDeterministicConfig(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigToProto(o.CryptoDeterministicConfig))\n\treturn p\n}", "func ProtoToYAML(m protoreflect.ProtoMessage) ([]byte, error) {\n\tmarshalOptions := protojson.MarshalOptions{\n\t\tUseProtoNames: true,\n\t}\n\tconfigJSON, err := marshalOptions.Marshal(m)\n\tif err != nil {\n\t\tlog.Error().Err(err).Msg(\"Error marshaling proto to JSON\")\n\t\treturn nil, err\n\t}\n\n\tconfigYAML, err := yaml.JSONToYAML(configJSON)\n\tif err != nil {\n\t\tlog.Error().Err(err).Msgf(\"Error converting JSON to YAML\")\n\t\treturn nil, err\n\t}\n\treturn configYAML, err\n}", "func MapInfoToAPIModel(info *model.NodeInfo) *node.Info {\n\treturn &node.Info{\n\t\tUptime: info.Uptime,\n\t\tLabels: mapLabelsToAPIModel(info.Labels),\n\t\tHostname: info.Hostname,\n\t\tAddresses: addressesToString(info.Addresses),\n\t\tGrpcPort: int64(info.GrpcPort),\n\t\tMachineID: info.MachineID,\n\t\tSystemUUID: info.SystemUUID,\n\t\tBootID: info.BootID,\n\t\tArch: info.Arch,\n\t\tOs: info.OS,\n\t\tVersion: info.Version,\n\t\tFilesystems: mapFilesystemsToAPIModel(info.Filesystems),\n\t}\n}", "func (Mapper) ModelToEntity(m, e interface{}) error {\r\n\tobj, err := json.Marshal(m)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\treturn json.Unmarshal(obj, e)\r\n}", "func (tokens ApiToken) mapToStruct(params map[string]interface{}) *ApiToken {\n\tif _, isSet := params[\"tokenid\"]; isSet {\n\t\ttokens.TokenId = params[\"tokenid\"].(string)\n\t}\n\tif _, isSet := params[\"comment\"]; isSet {\n\t\ttokens.Comment = params[\"comment\"].(string)\n\t}\n\tif _, isSet := params[\"expire\"]; isSet {\n\t\ttokens.Expire = int64(params[\"expire\"].(float64))\n\t}\n\tif _, isSet := params[\"privsep\"]; isSet {\n\t\ttokens.Privsep = false\n\t\tif params[\"privsep\"] == 1 {\n\t\t\ttokens.Privsep = true\n\t\t}\n\t}\n\treturn &tokens\n}", "func (protocol *Protocol)Marshal() ([]byte, error) {\n buff := new(bytes.Buffer);\n binary.Write(buff, binary.BigEndian, protocol.ID);\n binary.Write(buff, binary.BigEndian, protocol.PType)\n ms, ok := protocol.Packet.(proto.Message);\n if !ok {\n\t\treturn nil, fmt.Errorf(\"Protocol error not valid protobuff\");\n\t}\n data, err := proto.Marshal(ms)\n if nil != err {\n return nil, fmt.Errorf(\"Packet Marshal Error\");\n }\n buff.Write(data);\n return buff.Bytes(), nil\n}", "func ProtoToEndpoint(p *vertexaipb.VertexaiEndpoint) *vertexai.Endpoint {\n\tobj := &vertexai.Endpoint{\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tEtag: dcl.StringOrNil(p.GetEtag()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tUpdateTime: dcl.StringOrNil(p.GetUpdateTime()),\n\t\tEncryptionSpec: ProtoToVertexaiEndpointEncryptionSpec(p.GetEncryptionSpec()),\n\t\tNetwork: dcl.StringOrNil(p.GetNetwork()),\n\t\tModelDeploymentMonitoringJob: dcl.StringOrNil(p.GetModelDeploymentMonitoringJob()),\n\t\tProject: dcl.StringOrNil(p.GetProject()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t}\n\tfor _, r := range p.GetDeployedModels() {\n\t\tobj.DeployedModels = append(obj.DeployedModels, *ProtoToVertexaiEndpointDeployedModels(r))\n\t}\n\treturn obj\n}", "func VertexaiModelEncryptionSpecToProto(o *vertexai.ModelEncryptionSpec) *vertexaipb.VertexaiModelEncryptionSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiModelEncryptionSpec{}\n\tp.SetKmsKeyName(dcl.ValueOrEmptyString(o.KmsKeyName))\n\treturn p\n}", "func (c *TypeConverter) genStructConverter(\n\tkeyPrefix string,\n\tfromPrefix string,\n\tindent string,\n\tfromFields []*compile.FieldSpec,\n\ttoFields []*compile.FieldSpec,\n\tfieldMap map[string]FieldMapperEntry,\n\tprevKeyPrefixes []string,\n) error {\n\n\tfor i := 0; i < len(toFields); i++ {\n\t\ttoField := toFields[i]\n\n\t\t// Check for same named field\n\t\tvar fromField *compile.FieldSpec\n\t\tfor j := 0; j < len(fromFields); j++ {\n\t\t\tif fromFields[j].Name == toField.Name {\n\t\t\t\tfromField = fromFields[j]\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\ttoSubIdentifier := keyPrefix + PascalCase(toField.Name)\n\t\ttoIdentifier := \"out.\" + toSubIdentifier\n\t\toverriddenIdentifier := \"\"\n\t\tfromIdentifier := \"\"\n\n\t\t// Check for mapped field\n\t\tvar overriddenField *compile.FieldSpec\n\n\t\t// check if this toField satisfies a fieldMap transform\n\t\ttransformFrom, ok := fieldMap[toSubIdentifier]\n\t\tif ok {\n\t\t\t// no existing direct fromField, just assign the transform\n\t\t\tif fromField == nil {\n\t\t\t\tfromField = transformFrom.Field\n\t\t\t\tif c.useRecurGen {\n\t\t\t\t\tfromIdentifier = \"inOriginal.\" + transformFrom.QualifiedName\n\t\t\t\t} else {\n\t\t\t\t\tfromIdentifier = \"in.\" + transformFrom.QualifiedName\n\t\t\t\t}\n\t\t\t\t// else there is a conflicting direct fromField\n\t\t\t} else {\n\t\t\t\t// depending on Override flag either the direct fromField or transformFrom is the OverrideField\n\t\t\t\tif transformFrom.Override {\n\t\t\t\t\t// check for required/optional setting\n\t\t\t\t\tif !transformFrom.Field.Required {\n\t\t\t\t\t\toverriddenField = fromField\n\t\t\t\t\t\toverriddenIdentifier = \"in.\" + fromPrefix +\n\t\t\t\t\t\t\tPascalCase(overriddenField.Name)\n\t\t\t\t\t}\n\t\t\t\t\t// If override is true and the new field is required,\n\t\t\t\t\t// there's a default instantiation value and will always\n\t\t\t\t\t// overwrite.\n\t\t\t\t\tfromField = transformFrom.Field\n\t\t\t\t\tif c.useRecurGen {\n\t\t\t\t\t\tfromIdentifier = \"inOriginal.\" + transformFrom.QualifiedName\n\t\t\t\t\t} else {\n\t\t\t\t\t\tfromIdentifier = \"in.\" + transformFrom.QualifiedName\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// If override is false and the from field is required,\n\t\t\t\t\t// From is always populated and will never be overwritten.\n\t\t\t\t\tif !fromField.Required {\n\t\t\t\t\t\toverriddenField = transformFrom.Field\n\t\t\t\t\t\tif c.useRecurGen {\n\t\t\t\t\t\t\tfromIdentifier = \"inOriginal.\" + transformFrom.QualifiedName\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\toverriddenIdentifier = \"in.\" + transformFrom.QualifiedName\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// neither direct or transform fromField was found\n\t\tif fromField == nil {\n\t\t\t// search the fieldMap toField identifiers for matching identifier prefix\n\t\t\t// e.g. the current toField is a struct and something within it has a transform\n\t\t\t// a full match identifiers for transform non-struct types would have been caught above\n\t\t\thasStructFieldMapping := false\n\t\t\tfor toID := range fieldMap {\n\t\t\t\tif strings.HasPrefix(toID, toSubIdentifier) {\n\t\t\t\t\thasStructFieldMapping = true\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// if there's no fromField and no fieldMap transform that could be applied\n\t\t\tif !hasStructFieldMapping {\n\t\t\t\tvar bypass bool\n\t\t\t\t// check if required field is filled from other resources\n\t\t\t\t// it can be used to set system default (customized tracing /auth required for clients),\n\t\t\t\t// or header propagating\n\t\t\t\tif c.optionalEntries != nil {\n\t\t\t\t\tfor toID := range c.optionalEntries {\n\t\t\t\t\t\tif strings.HasPrefix(toID, toSubIdentifier) {\n\t\t\t\t\t\t\tbypass = true\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// the toField is either covered by optionalEntries, or optional and\n\t\t\t\t// there's nothing that maps to it or its sub-fields so we should skip it\n\t\t\t\tif bypass || !toField.Required {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// unrecoverable error\n\t\t\t\treturn errors.Errorf(\n\t\t\t\t\t\"required toField %s does not have a valid fromField mapping\",\n\t\t\t\t\ttoField.Name,\n\t\t\t\t)\n\t\t\t}\n\t\t}\n\n\t\tif fromIdentifier == \"\" && fromField != nil {\n\t\t\t// should we set this if no fromField ??\n\t\t\tfromIdentifier = \"in.\" + fromPrefix + PascalCase(fromField.Name)\n\t\t}\n\n\t\tif prevKeyPrefixes == nil {\n\t\t\tprevKeyPrefixes = []string{}\n\t\t}\n\n\t\tvar overriddenFieldName string\n\t\tvar overriddenFieldType compile.TypeSpec\n\t\tif overriddenField != nil {\n\t\t\toverriddenFieldName = overriddenField.Name\n\t\t\toverriddenFieldType = overriddenField.Type\n\t\t}\n\n\t\t// Override thrift type names to avoid naming collisions between endpoint\n\t\t// and client types.\n\t\tswitch toFieldType := compile.RootTypeSpec(toField.Type).(type) {\n\t\tcase\n\t\t\t*compile.BoolSpec,\n\t\t\t*compile.I8Spec,\n\t\t\t*compile.I16Spec,\n\t\t\t*compile.I32Spec,\n\t\t\t*compile.EnumSpec,\n\t\t\t*compile.I64Spec,\n\t\t\t*compile.DoubleSpec,\n\t\t\t*compile.StringSpec:\n\n\t\t\terr := c.genConverterForPrimitive(\n\t\t\t\ttoField,\n\t\t\t\ttoIdentifier,\n\t\t\t\tfromField,\n\t\t\t\tfromIdentifier,\n\t\t\t\toverriddenField,\n\t\t\t\toverriddenIdentifier,\n\t\t\t\tindent,\n\t\t\t\tprevKeyPrefixes,\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tcase *compile.BinarySpec:\n\t\t\tfor _, line := range checkOptionalNil(indent, c.uninitialized, toIdentifier, prevKeyPrefixes, c.useRecurGen) {\n\t\t\t\tc.append(line)\n\t\t\t}\n\t\t\tc.append(toIdentifier, \" = []byte(\", fromIdentifier, \")\")\n\t\tcase *compile.StructSpec:\n\t\t\tvar (\n\t\t\t\tstFromPrefix = fromPrefix\n\t\t\t\tstFromType compile.TypeSpec\n\t\t\t\tfromTypeName string\n\t\t\t)\n\t\t\tif fromField != nil {\n\t\t\t\tstFromType = fromField.Type\n\t\t\t\tstFromPrefix = fromPrefix + PascalCase(fromField.Name)\n\n\t\t\t\tfromTypeName, _ = c.getIdentifierName(stFromType)\n\t\t\t}\n\n\t\t\ttoTypeName, err := c.getIdentifierName(toFieldType)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tif converterMethodName, ok := c.convStructMap[toFieldType.Name]; ok {\n\t\t\t\t// the converter for this struct has already been generated, so just use it\n\t\t\t\tc.append(indent, \"out.\", keyPrefix+PascalCase(toField.Name), \" = \", converterMethodName, \"(\", fromIdentifier, \")\")\n\t\t\t} else if c.useRecurGen && fromTypeName != \"\" {\n\t\t\t\t// generate a callable converter inside function literal\n\t\t\t\terr = c.genConverterForStructWrapped(\n\t\t\t\t\ttoField,\n\t\t\t\t\ttoFieldType,\n\t\t\t\t\ttoTypeName,\n\t\t\t\t\ttoSubIdentifier,\n\t\t\t\t\tfromTypeName,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t\tstFromType,\n\t\t\t\t\tfieldMap,\n\t\t\t\t\tprevKeyPrefixes,\n\t\t\t\t\tindent,\n\t\t\t\t)\n\t\t\t} else {\n\t\t\t\terr = c.genConverterForStruct(\n\t\t\t\t\ttoField.Name,\n\t\t\t\t\ttoFieldType,\n\t\t\t\t\ttoField.Required,\n\t\t\t\t\tstFromType,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t\tkeyPrefix+PascalCase(toField.Name),\n\t\t\t\t\tstFromPrefix,\n\t\t\t\t\tindent,\n\t\t\t\t\tfieldMap,\n\t\t\t\t\tprevKeyPrefixes,\n\t\t\t\t)\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tcase *compile.ListSpec:\n\t\t\terr := c.genConverterForList(\n\t\t\t\ttoFieldParam{\n\t\t\t\t\ttoFieldType,\n\t\t\t\t\ttoField.Name,\n\t\t\t\t\ttoField.Required,\n\t\t\t\t\ttoIdentifier,\n\t\t\t\t},\n\t\t\t\tfromFieldParam{\n\t\t\t\t\tfromField.Type,\n\t\t\t\t\tfromField.Name,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t},\n\t\t\t\toverriddenFieldParam{\n\t\t\t\t\toverriddenFieldType,\n\t\t\t\t\toverriddenFieldName,\n\t\t\t\t\toverriddenIdentifier,\n\t\t\t\t},\n\t\t\t\tindent,\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tcase *compile.MapSpec:\n\t\t\terr := c.genConverterForMap(\n\t\t\t\ttoFieldParam{\n\t\t\t\t\ttoFieldType,\n\t\t\t\t\ttoField.Name,\n\t\t\t\t\ttoField.Required,\n\t\t\t\t\ttoIdentifier,\n\t\t\t\t},\n\t\t\t\tfromFieldParam{\n\t\t\t\t\tfromField.Type,\n\t\t\t\t\tfromField.Name,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t\tfromIdentifier,\n\t\t\t\t},\n\t\t\t\toverriddenFieldParam{\n\t\t\t\t\toverriddenFieldType,\n\t\t\t\t\toverriddenFieldName,\n\t\t\t\t\toverriddenIdentifier,\n\t\t\t\t},\n\t\t\t\tindent,\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tdefault:\n\t\t\t// fmt.Printf(\"Unknown type %s for field %s \\n\",\n\t\t\t// \ttoField.Type.TypeCode().String(), toField.Name,\n\t\t\t// )\n\n\t\t\t// pkgName, err := h.TypePackageName(toField.Type.IDLFile())\n\t\t\t// if err != nil {\n\t\t\t// \treturn nil, err\n\t\t\t// }\n\t\t\t// typeName := pkgName + \".\" + toField.Type.ThriftName()\n\t\t\t// line := toIdentifier + \"(*\" + typeName + \")\" + postfix\n\t\t\t// c.Lines = append(c.Lines, line)\n\t\t}\n\t}\n\n\treturn nil\n}", "func ProtoToVertexaiModelSupportedExportFormats(p *vertexaipb.VertexaiModelSupportedExportFormats) *vertexai.ModelSupportedExportFormats {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelSupportedExportFormats{\n\t\tId: dcl.StringOrNil(p.GetId()),\n\t}\n\tfor _, r := range p.GetExportableContents() {\n\t\tobj.ExportableContents = append(obj.ExportableContents, *ProtoToVertexaiModelSupportedExportFormatsExportableContentsEnum(r))\n\t}\n\treturn obj\n}", "func VertexaiEndpointDeployedModelsPrivateEndpointsToProto(o *vertexai.EndpointDeployedModelsPrivateEndpoints) *vertexaipb.VertexaiEndpointDeployedModelsPrivateEndpoints {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &vertexaipb.VertexaiEndpointDeployedModelsPrivateEndpoints{}\n\tp.SetPredictHttpUri(dcl.ValueOrEmptyString(o.PredictHttpUri))\n\tp.SetExplainHttpUri(dcl.ValueOrEmptyString(o.ExplainHttpUri))\n\tp.SetHealthHttpUri(dcl.ValueOrEmptyString(o.HealthHttpUri))\n\tp.SetServiceAttachment(dcl.ValueOrEmptyString(o.ServiceAttachment))\n\treturn p\n}", "func (i *Input) ToMap() map[string]interface{} {\n\tvar keys []interface{}\n\tfor _, k := range i.StateKeys {\n\t\tkeys = append(keys, k)\n\t}\n\tvar orgs []interface{}\n\tfor _, org := range i.Organizations {\n\t\torgs = append(orgs, org)\n\t}\n\n\treturn map[string]interface{}{\n\t\t\"keys\": keys,\n\t\t\"organizations\": orgs,\n\t\t\"policy\": i.Policy,\n\t\t\"privateCollection\": i.PrivateCollection,\n\t}\n}", "func gceLabelsToModel(labels map[string]string) []string {\n\tidx := 0\n\tmLabels := make([]string, len(labels))\n\tfor k, v := range labels {\n\t\tmLabels[idx] = fmt.Sprintf(\"%s:%s\", k, v)\n\t\tidx++\n\t}\n\treturn mLabels\n}", "func MapStructureDecode(model interface{}, out interface{}) {\n\tstringToDateTimeHook := func(f reflect.Type, t reflect.Type, data interface{}) (interface{}, error) {\n\t\tif t == reflect.TypeOf(time.Time{}) {\n\t\t\treturn time.Parse(time.RFC3339, data.(string))\n\t\t}\n\n\t\treturn data, nil\n\t}\n\n\tconfig := mapstructure.DecoderConfig{\n\t\tDecodeHook: stringToDateTimeHook,\n\t\tResult: &model,\n\t}\n\n\tmpdecoder, err := mapstructure.NewDecoder(&config)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tmpdecoder.Decode(out)\n}", "func ProtoToAppengineDomainMappingSslSettings(p *appenginepb.AppengineDomainMappingSslSettings) *appengine.DomainMappingSslSettings {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &appengine.DomainMappingSslSettings{\n\t\tCertificateId: dcl.StringOrNil(p.CertificateId),\n\t\tSslManagementType: ProtoToAppengineDomainMappingSslSettingsSslManagementTypeEnum(p.GetSslManagementType()),\n\t\tPendingManagedCertificateId: dcl.StringOrNil(p.PendingManagedCertificateId),\n\t}\n\treturn obj\n}", "func (e *Server) ToModel() base.ModelInterface {\n\tm := model.Server{}\n\tbase.EntityToModel(&e.Entity, &m.Model)\n\tcreateProductInfoModel(&e.ProductInfo, &m.ProductInfo)\n\tm.OriginURIs.Chassis = e.OriginURIsChassis\n\tm.OriginURIs.System = e.OriginURIsSystem\n\tm.PhysicalUUID = e.PhysicalUUID\n\tm.Name = e.Name\n\tm.Description = e.Description\n\tm.Hostname = e.Hostname\n\tm.Type = e.Type\n\tm.Protocol = e.Protocol\n\tm.Vender = e.Vender\n\tm.PowerState = e.PowerState\n\tm.IndicatorLED = e.IndicatorLED\n\tm.Credential = e.Credential\n\tm.State = e.State\n\tm.Health = e.Health\n\t// ComputerSystem.Processors\n\tprocessors := []model.Processor{}\n\tfor i := range e.Processors {\n\t\tprocessors = append(processors, *e.Processors[i].ToModel())\n\t}\n\tm.ComputerSystem.Processors = processors\n\t// ComputerSystem.Memory\n\tmemory := []model.Memory{}\n\tfor i := range e.Memory {\n\t\tmemory = append(memory, *e.Memory[i].ToModel())\n\t}\n\tm.ComputerSystem.Memory = memory\n\n\t// ComputerSystem.EthernetInterfaces\n\tethernetInterfaces := []model.EthernetInterface{}\n\tfor i := range e.EthernetInterfaces {\n\t\tethernetInterfaces = append(ethernetInterfaces, *e.EthernetInterfaces[i].ToModel())\n\t}\n\tm.ComputerSystem.EthernetInterfaces = ethernetInterfaces\n\t// ComputerSystem.NetworkInterfaces\n\tnetworkInterfaces := []model.NetworkInterface{}\n\tfor i := range e.NetworkInterfaces {\n\t\tnetworkInterfaces = append(networkInterfaces, *e.NetworkInterfaces[i].ToModel())\n\t}\n\tm.ComputerSystem.NetworkInterfaces = networkInterfaces\n\t// ComputerSystem.Storages\n\tstorages := []model.Storage{}\n\tfor i := range e.Storages {\n\t\tstorages = append(storages, *e.Storages[i].ToModel())\n\t}\n\tm.ComputerSystem.Storages = storages\n\t// Chassis.Power\n\tcreateResourceModel(&e.Power.EmbeddedResource, &m.Chassis.Power.Resource)\n\tpowerControl := []model.PowerControl{}\n\tfor i := range e.Power.PowerControl {\n\t\tpowerControl = append(powerControl, *e.Power.PowerControl[i].ToModel())\n\t}\n\tm.Chassis.Power.PowerControl = powerControl\n\n\tpowerSupplies := []model.PowerSupply{}\n\tfor i := range e.Power.PowerSupplies {\n\t\tpowerSupplies = append(powerSupplies, *e.Power.PowerSupplies[i].ToModel())\n\t}\n\tm.Chassis.Power.PowerSupplies = powerSupplies\n\n\tredundancy := []model.Redundancy{}\n\tfor i := range e.Power.Redundancy {\n\t\tredundancy = append(redundancy, *e.Power.Redundancy[i].ToModel())\n\t}\n\tm.Chassis.Power.Redundancy = redundancy\n\t// Chassis.Thermal\n\tcreateResourceModel(&e.Thermal.EmbeddedResource, &m.Chassis.Thermal.Resource)\n\tfans := []model.Fan{}\n\tfor i := range e.Thermal.Fans {\n\t\tfans = append(fans, *e.Thermal.Fans[i].ToModel())\n\t}\n\tm.Chassis.Thermal.Fans = fans\n\t// Chassis.Boards\n\tboards := []model.Board{}\n\tfor i := range e.Boards {\n\t\tboards = append(boards, *e.Boards[i].ToModel())\n\t}\n\tm.Chassis.Boards = boards\n\t// Chassis.NetworkAdapters\n\tnetworkAdapters := []model.NetworkAdapter{}\n\tfor i := range e.NetworkAdapters {\n\t\tnetworkAdapters = append(networkAdapters, *e.NetworkAdapters[i].ToModel())\n\t}\n\tm.Chassis.NetworkAdapters = networkAdapters\n\t// Chassis.Drives\n\tdrives := []model.Drive{}\n\tfor i := range e.Drives {\n\t\tdrives = append(drives, *e.Drives[i].ToModel())\n\t}\n\tm.Chassis.Drives = drives\n\t// Chassis.PCIeDevices\n\tpcieDevices := []model.PCIeDevice{}\n\tfor i := range e.PCIeDevices {\n\t\tpcieDevices = append(pcieDevices, *e.PCIeDevices[i].ToModel())\n\t}\n\tm.Chassis.PCIeDevices = pcieDevices\n\treturn &m\n}", "func ToMap(componentName, src, root string) (map[string]interface{}, error) {\n\tobj, err := jsonnetParseFn(\"params.libsonnet\", src)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"parse jsonnet\")\n\t}\n\n\tcomponentObject, err := componentParams(obj, componentName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tm, err := convertObjectToMapFn(componentObject)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif componentName == \"\" {\n\t\treturn m[root].(map[string]interface{}), nil\n\t}\n\n\tparamsMap, ok := m[componentName].(map[string]interface{})\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"component %q params is not an object\", componentName)\n\t}\n\n\treturn paramsMap, nil\n}", "func (mapper *ChannelMapper) MapToChannels(input string) ([]string, error) {\n\tresponse := mapper.request(kMap, input)\n\treturn response.channels, response.err\n}", "func ProtoToVertexaiModelEncryptionSpec(p *vertexaipb.VertexaiModelEncryptionSpec) *vertexai.ModelEncryptionSpec {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.ModelEncryptionSpec{\n\t\tKmsKeyName: dcl.StringOrNil(p.GetKmsKeyName()),\n\t}\n\treturn obj\n}", "func VertexaiBetaModelDeploymentDedicatedResourcesMachineSpecToProto(o *beta.ModelDeploymentDedicatedResourcesMachineSpec) *betapb.VertexaiBetaModelDeploymentDedicatedResourcesMachineSpec {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.VertexaiBetaModelDeploymentDedicatedResourcesMachineSpec{}\n\tp.SetMachineType(dcl.ValueOrEmptyString(o.MachineType))\n\treturn p\n}", "func ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation(p *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation) *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformation{\n\t\tReplaceConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceConfig(p.GetReplaceConfig()),\n\t\tRedactConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationRedactConfig(p.GetRedactConfig()),\n\t\tCharacterMaskConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCharacterMaskConfig(p.GetCharacterMaskConfig()),\n\t\tCryptoReplaceFfxFpeConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoReplaceFfxFpeConfig(p.GetCryptoReplaceFfxFpeConfig()),\n\t\tFixedSizeBucketingConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationFixedSizeBucketingConfig(p.GetFixedSizeBucketingConfig()),\n\t\tBucketingConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationBucketingConfig(p.GetBucketingConfig()),\n\t\tReplaceWithInfoTypeConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationReplaceWithInfoTypeConfig(p.GetReplaceWithInfoTypeConfig()),\n\t\tTimePartConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationTimePartConfig(p.GetTimePartConfig()),\n\t\tCryptoHashConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfig(p.GetCryptoHashConfig()),\n\t\tDateShiftConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationDateShiftConfig(p.GetDateShiftConfig()),\n\t\tCryptoDeterministicConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfig(p.GetCryptoDeterministicConfig()),\n\t}\n\treturn obj\n}", "func (opts CreateOpts) ToDomainCreateMap() (map[string]interface{}, error) {\n\treturn golangsdk.BuildRequestBody(opts, \"\")\n}", "func (opts UpdateOpts) ToDomainUpdateMap() (map[string]interface{}, error) {\n\treturn golangsdk.BuildRequestBody(opts, \"\")\n}", "func (t *Translator) fromModel(req *TranslationRequest) (*JSONMessage, error) {\n\tcmpTable := t.Tables.Component\n\n\t// Get all of the components associated with that message.\n\tcomps := make([]*Component, 0)\n\terr := cmpTable.Get().Where(\"message\", req.Model.Id).All(t.Store, &comps)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Translate the components into JSON.\n\tparsedComponents := make(map[string]*JSONComponent)\n\tfor _, c := range comps {\n\t\tparsedComponents[c.Name] = &JSONComponent{\n\t\t\tBinary: c.Data,\n\t\t\tString: string(c.Data),\n\t\t}\n\t}\n\n\ttoAddrs := strings.Split(req.Model.To, \",\")\n\n\t// Download the profiles of all recipients of the message.\n\tprofiles := make([]*JSONProfile, len(toAddrs))\n\tfor i, toAddr := range toAddrs {\n\t\tprofiles[i] = req.Profiles[toAddr]\n\t}\n\n\tmyProfile := &JSONProfile{\n\t\tName: req.Me.Profile.Name,\n\t\tAvatar: req.Me.Profile.Image,\n\t\tAlias: req.Me.Alias.String(),\n\t}\n\n\t// Export the finished product.\n\treturn &JSONMessage{\n\t\tName: req.Model.Name,\n\t\tDate: time.Unix(req.Model.Date, 0),\n\n\t\t// To and From Info\n\t\tFrom: myProfile,\n\t\tTo: profiles,\n\n\t\t// Components\n\t\tComponents: parsedComponents,\n\n\t\t// Meta\n\t\tSelf: true,\n\t\tPublic: !req.Model.Alert,\n\t}, nil\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKeyToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKey) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKey {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKey{}\n\tp.SetTransient(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKeyTransientToProto(o.Transient))\n\tp.SetUnwrapped(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKeyUnwrappedToProto(o.Unwrapped))\n\tp.SetKmsWrapped(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoHashConfigCryptoKeyKmsWrappedToProto(o.KmsWrapped))\n\treturn p\n}", "func AlignStructAndMap(from interface{}, to interface{}) {\n JsonToInterface(InterfaceToJson(from), &to)\n}", "func (pr *PolicyReflector) portsToProto(ports []core_v1beta1.NetworkPolicyPort) (portsProto []*proto.Policy_Port) {\n\tfor _, port := range ports {\n\t\tportProto := &proto.Policy_Port{}\n\t\t// Protocol\n\t\tif port.Protocol != nil {\n\t\t\tswitch *port.Protocol {\n\t\t\tcase core_v1.ProtocolTCP:\n\t\t\t\tportProto.Protocol = proto.Policy_Port_TCP\n\t\t\tcase core_v1.ProtocolUDP:\n\t\t\t\tportProto.Protocol = proto.Policy_Port_UDP\n\t\t\t}\n\t\t}\n\t\t// Port number/name\n\t\tif port.Port != nil {\n\t\t\tportProto.Port = &proto.Policy_Port_PortNameOrNumber{}\n\t\t\tswitch port.Port.Type {\n\t\t\tcase intstr.Int:\n\t\t\t\tportProto.Port.Type = proto.Policy_Port_PortNameOrNumber_NUMBER\n\t\t\t\tportProto.Port.Number = port.Port.IntVal\n\t\t\tcase intstr.String:\n\t\t\t\tportProto.Port.Type = proto.Policy_Port_PortNameOrNumber_NAME\n\t\t\t\tportProto.Port.Name = port.Port.StrVal\n\t\t\t}\n\t\t}\n\t\t// append port\n\t\tportsProto = append(portsProto, portProto)\n\t}\n\treturn portsProto\n}", "func ( fq *Fq_req ) To_bwow_map( ) ( fmap map[string]string ) {\n\tfmap = make( map[string]string )\n\n\tif fq == nil {\n\t\treturn\n\t}\n\n\tif fq.Match.Smac != nil {\n\t\tfmap[\"smac\"] = *fq.Match.Smac\n\t} else {\n\t\tfmap[\"smac\"] = \"\"\n\t}\n\tif fq.Match.Dmac != nil {\t\t\t\t\t// likely nil as oneways are usually x-project when router is not a neutron device\n\t\tfmap[\"dmac\"] = *fq.Match.Dmac\n\t} else {\n\t\tfmap[\"dmac\"] = \"\"\n\t}\n\tif fq.Extip != nil {\n\t\tfmap[\"extip\"] = *fq.Extip\t\t\t\t\t\t\t\t\t\t\t\t// external IP if supplied\n\t} else {\n\t\tfmap[\"extip\"] = \"\"\n\t}\n\tif fq.Match.Vlan_id != nil {\t\t\t\t\t\t\t\t\t\t\t\t// adds a vlan number to match (should NOT be a mac)\n\t\tfmap[\"vlan_match\"] = *fq.Match.Vlan_id\n\t} else {\n\t\tfmap[\"vlan_match\"] = \"\"\n\t}\n\tif fq.Action.Vlan_id != nil {\t\t\t\t\t\t\t\t\t\t\t\t// adds a set vlan action, can be a MAC for late conversion\n\t\tfmap[\"vlan_action\"] = *fq.Action.Vlan_id\n\t} else {\n\t\tfmap[\"vlan_action\"] = \"\"\n\t}\n\n\tfmap[\"queue\"] = fmt.Sprintf( \"%d\", fq.Espq.Queuenum )\n\tfmap[\"dscp\"] = fmt.Sprintf( \"%d\", fq.Dscp << 2 )\t\t\t\t\t\t// shift left 2 bits to match what OVS wants\n\tfmap[\"ipv6\"] = fmt.Sprintf( \"%v\", fq.Ipv6 )\t\t\t\t\t\t\t// force ipv6 fmods is on\n\tfmap[\"timeout\"] = fmt.Sprintf( \"%d\", fq.Expiry - time.Now().Unix() )\n\tif fq.Tptype != nil && *fq.Tptype != \"none\" && *fq.Tptype != \"\" {\t\t\t\t\t// if transport prototype defined, turn it on\n\t\tif fq.Match.Tpsport != nil \t{\t\t\t\t\t\t\t\t\t\t\t\t\t// set src and dest ports if they are defined too\n\t\t\tfmap[\"sproto\"] = fmt.Sprintf( \"%s:%s\", *fq.Tptype, *fq.Match.Tpsport )\n\t\t}\n\t\tif fq.Match.Tpdport != nil \t{\n\t\t\tfmap[\"dproto\"] = fmt.Sprintf( \"%s:%s\", *fq.Tptype, *fq.Match.Tpdport )\n\t\t}\n\t}\n\n\tif fq_sheep.Would_baa( 2 ) {\n\t\tfor k, v := range fmap {\n\t\t\tfq_sheep.Baa( 2, \"fq_req to action id=%s %s = %s\", *fq.Id, k, v )\n\t\t}\n\t}\n\n\treturn\n}", "func ProtoToVertexaiEndpointDeployedModelsPrivateEndpoints(p *vertexaipb.VertexaiEndpointDeployedModelsPrivateEndpoints) *vertexai.EndpointDeployedModelsPrivateEndpoints {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &vertexai.EndpointDeployedModelsPrivateEndpoints{\n\t\tPredictHttpUri: dcl.StringOrNil(p.GetPredictHttpUri()),\n\t\tExplainHttpUri: dcl.StringOrNil(p.GetExplainHttpUri()),\n\t\tHealthHttpUri: dcl.StringOrNil(p.GetHealthHttpUri()),\n\t\tServiceAttachment: dcl.StringOrNil(p.GetServiceAttachment()),\n\t}\n\treturn obj\n}", "func (m *PropertyMap) ToProto() map[string]*pb.Property {\n\t// This may seem wrong, but m is a pointer to a map, which is also a\n\t// nullable reference type.\n\tif m == nil || *m == nil {\n\t\treturn nil\n\t}\n\tret := make(map[string]*pb.Property)\n\tfor k, v := range *m {\n\t\tret[k] = v.ToProto()\n\t}\n\treturn ret\n}", "func (o *ProjectWebhookPartial) ToMap() map[string]interface{} {\n\tkv := map[string]interface{}{\n\t\t\"enabled\": toProjectWebhookObject(o.Enabled, true),\n\t\t\"error_message\": toProjectWebhookObject(o.ErrorMessage, true),\n\t\t\"errored\": toProjectWebhookObject(o.Errored, true),\n\t\t\"project_id\": toProjectWebhookObject(o.ProjectID, true),\n\t\t\"url\": toProjectWebhookObject(o.URL, true),\n\t}\n\tfor k, v := range kv {\n\t\tif v == nil || reflect.ValueOf(v).IsZero() {\n\t\t\tdelete(kv, k)\n\t\t} else {\n\t\t}\n\t}\n\treturn kv\n}", "func (m *Model) Unpack(data []byte) error {\n\tif err := proto.Unmarshal(data, m.cur); err != nil {\n\t\treturn err\n\t}\n\n\tif m.cur.GetChatfriends() == nil {\n\t\tm.cur.Chatfriends = make(map[string]bool, 0)\n\t}\n\n\treturn nil\n}", "func (h *Handler) nodeConfigToProto(nodeConfig *v1.NodeConfig) *model.NodeConfig {\n\tnodeConfigProto := &model.NodeConfig{}\n\tnodeConfigProto.NodeName = nodeConfig.Name\n\tif nodeConfig.Spec.MainVPPInterface.InterfaceName != \"\" {\n\t\tnodeConfigProto.MainVppInterface = h.interfaceConfigToProto(nodeConfig.Spec.MainVPPInterface)\n\t}\n\tnodeConfigProto.Gateway = nodeConfig.Spec.Gateway\n\tnodeConfigProto.StealInterface = nodeConfig.Spec.StealInterface\n\tnodeConfigProto.NatExternalTraffic = nodeConfig.Spec.NatExternalTraffic\n\tfor _, otherNode := range nodeConfig.Spec.OtherVPPInterfaces {\n\t\tnodeConfigProto.OtherVppInterfaces = append(nodeConfigProto.OtherVppInterfaces,\n\t\t\th.interfaceConfigToProto(otherNode))\n\t}\n\n\treturn nodeConfigProto\n}", "func ProtoToFirewall(p *computepb.ComputeFirewall) *compute.Firewall {\n\tobj := &compute.Firewall{\n\t\tCreationTimestamp: dcl.StringOrNil(p.GetCreationTimestamp()),\n\t\tDescription: dcl.StringOrNil(p.Description),\n\t\tDirection: ProtoToComputeFirewallDirectionEnum(p.GetDirection()),\n\t\tDisabled: dcl.Bool(p.Disabled),\n\t\tId: dcl.StringOrNil(p.Id),\n\t\tLogConfig: ProtoToComputeFirewallLogConfig(p.GetLogConfig()),\n\t\tName: dcl.StringOrNil(p.Name),\n\t\tNetwork: dcl.StringOrNil(p.Network),\n\t\tPriority: dcl.Int64OrNil(p.Priority),\n\t\tSelfLink: dcl.StringOrNil(p.SelfLink),\n\t\tProject: dcl.StringOrNil(p.Project),\n\t}\n\tfor _, r := range p.GetAllowed() {\n\t\tobj.Allowed = append(obj.Allowed, *ProtoToComputeFirewallAllowed(r))\n\t}\n\tfor _, r := range p.GetDenied() {\n\t\tobj.Denied = append(obj.Denied, *ProtoToComputeFirewallDenied(r))\n\t}\n\tfor _, r := range p.GetDestinationRanges() {\n\t\tobj.DestinationRanges = append(obj.DestinationRanges, r)\n\t}\n\tfor _, r := range p.GetSourceRanges() {\n\t\tobj.SourceRanges = append(obj.SourceRanges, r)\n\t}\n\tfor _, r := range p.GetSourceServiceAccounts() {\n\t\tobj.SourceServiceAccounts = append(obj.SourceServiceAccounts, r)\n\t}\n\tfor _, r := range p.GetSourceTags() {\n\t\tobj.SourceTags = append(obj.SourceTags, r)\n\t}\n\tfor _, r := range p.GetTargetServiceAccounts() {\n\t\tobj.TargetServiceAccounts = append(obj.TargetServiceAccounts, r)\n\t}\n\tfor _, r := range p.GetTargetTags() {\n\t\tobj.TargetTags = append(obj.TargetTags, r)\n\t}\n\treturn obj\n}", "func FromTransmissionRecordModelToDTO(tr models.TransmissionRecord) TransmissionRecord {\n\treturn TransmissionRecord{\n\t\tStatus: string(tr.Status),\n\t\tResponse: tr.Response,\n\t\tSent: tr.Sent,\n\t}\n}", "func Map2Struct(jmap interface{}, s interface{}) error {\n tmpDataJson, err := json.Marshal(jmap)\n if err != nil {\n return err\n }\n err = json.Unmarshal(tmpDataJson, &s)\n if err != nil {\n return err\n }\n return nil\n}", "func (mb *MutableBag) ToProto(output *mixerpb.CompressedAttributes, globalDict map[string]int32, globalWordCount int) {\n\tds := newDictState(globalDict, globalWordCount)\n\tkeys := mb.Names()\n\n\tfor _, k := range keys {\n\t\tindex := ds.assignDictIndex(k)\n\t\tv, _ := mb.Get(k) // if not found, nil return will be ignored by the switch below\n\n\t\tswitch t := v.(type) {\n\t\tcase string:\n\t\t\tif output.Strings == nil {\n\t\t\t\toutput.Strings = make(map[int32]int32)\n\t\t\t}\n\t\t\toutput.Strings[index] = ds.assignDictIndex(t)\n\n\t\tcase int64:\n\t\t\tif output.Int64S == nil {\n\t\t\t\toutput.Int64S = make(map[int32]int64)\n\t\t\t}\n\t\t\toutput.Int64S[index] = t\n\n\t\tcase int:\n\t\t\tif output.Int64S == nil {\n\t\t\t\toutput.Int64S = make(map[int32]int64)\n\t\t\t}\n\t\t\toutput.Int64S[index] = int64(t)\n\n\t\tcase float64:\n\t\t\tif output.Doubles == nil {\n\t\t\t\toutput.Doubles = make(map[int32]float64)\n\t\t\t}\n\t\t\toutput.Doubles[index] = t\n\n\t\tcase bool:\n\t\t\tif output.Bools == nil {\n\t\t\t\toutput.Bools = make(map[int32]bool)\n\t\t\t}\n\t\t\toutput.Bools[index] = t\n\n\t\tcase time.Time:\n\t\t\tif output.Timestamps == nil {\n\t\t\t\toutput.Timestamps = make(map[int32]time.Time)\n\t\t\t}\n\t\t\toutput.Timestamps[index] = t\n\n\t\tcase time.Duration:\n\t\t\tif output.Durations == nil {\n\t\t\t\toutput.Durations = make(map[int32]time.Duration)\n\t\t\t}\n\t\t\toutput.Durations[index] = t\n\n\t\tcase []byte:\n\t\t\tif output.Bytes == nil {\n\t\t\t\toutput.Bytes = make(map[int32][]byte)\n\t\t\t}\n\t\t\toutput.Bytes[index] = t\n\n\t\tcase map[string]string:\n\t\t\tsm := make(map[int32]int32, len(t))\n\t\t\tfor smk, smv := range t {\n\t\t\t\tsm[ds.assignDictIndex(smk)] = ds.assignDictIndex(smv)\n\t\t\t}\n\n\t\t\tif output.StringMaps == nil {\n\t\t\t\toutput.StringMaps = make(map[int32]mixerpb.StringMap)\n\t\t\t}\n\t\t\toutput.StringMaps[index] = mixerpb.StringMap{Entries: sm}\n\n\t\tdefault:\n\t\t\tpanic(fmt.Errorf(\"cannot convert value:%v of type:%T\", v, v))\n\t\t}\n\t}\n\n\toutput.Words = ds.getMessageWordList()\n}", "func DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKeyToProto(o *beta.DeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKey) *betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKey {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &betapb.DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKey{}\n\tp.SetTransient(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKeyTransientToProto(o.Transient))\n\tp.SetUnwrapped(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKeyUnwrappedToProto(o.Unwrapped))\n\tp.SetKmsWrapped(DlpBetaDeidentifyTemplateDeidentifyConfigInfoTypeTransformationsTransformationsPrimitiveTransformationCryptoDeterministicConfigCryptoKeyKmsWrappedToProto(o.KmsWrapped))\n\treturn p\n}", "func DlpInspectTemplateInspectConfigCustomInfoTypesDictionaryToProto(o *dlp.InspectTemplateInspectConfigCustomInfoTypesDictionary) *dlppb.DlpInspectTemplateInspectConfigCustomInfoTypesDictionary {\n\tif o == nil {\n\t\treturn nil\n\t}\n\tp := &dlppb.DlpInspectTemplateInspectConfigCustomInfoTypesDictionary{}\n\tp.SetWordList(DlpInspectTemplateInspectConfigCustomInfoTypesDictionaryWordListToProto(o.WordList))\n\tp.SetCloudStoragePath(DlpInspectTemplateInspectConfigCustomInfoTypesDictionaryCloudStoragePathToProto(o.CloudStoragePath))\n\treturn p\n}", "func ProtoToDeidentifyTemplate(p *betapb.DlpBetaDeidentifyTemplate) *beta.DeidentifyTemplate {\n\tobj := &beta.DeidentifyTemplate{\n\t\tName: dcl.StringOrNil(p.GetName()),\n\t\tDisplayName: dcl.StringOrNil(p.GetDisplayName()),\n\t\tDescription: dcl.StringOrNil(p.GetDescription()),\n\t\tCreateTime: dcl.StringOrNil(p.GetCreateTime()),\n\t\tUpdateTime: dcl.StringOrNil(p.GetUpdateTime()),\n\t\tDeidentifyConfig: ProtoToDlpBetaDeidentifyTemplateDeidentifyConfig(p.GetDeidentifyConfig()),\n\t\tLocationId: dcl.StringOrNil(p.GetLocationId()),\n\t\tParent: dcl.StringOrNil(p.GetParent()),\n\t\tLocation: dcl.StringOrNil(p.GetLocation()),\n\t}\n\treturn obj\n}", "func (u OpUnion) ToProto(pb *pipelinepb.AppliedPipelineOp) error {\n\tpb.Reset()\n\tswitch u.Type {\n\tcase pipeline.TransformationOpType:\n\t\tpb.Type = pipelinepb.AppliedPipelineOp_TRANSFORMATION\n\t\treturn u.Transformation.ToProto(&pb.Transformation)\n\tcase pipeline.RollupOpType:\n\t\tpb.Type = pipelinepb.AppliedPipelineOp_ROLLUP\n\t\treturn u.Rollup.ToProto(&pb.Rollup)\n\tdefault:\n\t\treturn errUnknownOpType\n\t}\n}", "func (m *Message) ToProtobuf() *PBDHTMessage {\n\tpmes := new(PBDHTMessage)\n\tif m.Value != nil {\n\t\tpmes.Value = m.Value\n\t}\n\n\tpmes.Type = &m.Type\n\tpmes.Key = &m.Key\n\tpmes.Response = &m.Response\n\tpmes.Id = &m.ID\n\tpmes.Success = &m.Success\n\tfor _, p := range m.Peers {\n\t\tpmes.Peers = append(pmes.Peers, peerInfo(p))\n\t}\n\n\treturn pmes\n}", "func ProtoToAppengineDomainMappingResourceRecords(p *appenginepb.AppengineDomainMappingResourceRecords) *appengine.DomainMappingResourceRecords {\n\tif p == nil {\n\t\treturn nil\n\t}\n\tobj := &appengine.DomainMappingResourceRecords{\n\t\tName: dcl.StringOrNil(p.Name),\n\t\tRrdata: dcl.StringOrNil(p.Rrdata),\n\t\tType: ProtoToAppengineDomainMappingResourceRecordsTypeEnum(p.GetType()),\n\t}\n\treturn obj\n}", "func Obj2Protocol(c *Client, o *GodisObject) {\n\tif c.Cmd.Name == \"set\" {\n\t\tc.Buf = \"OK\"\n\t} else if c.Cmd.Name == \"get\" {\n\t\tc.Buf = o.Ptr.(string)\n\t} else {\n\t\tc.Buf = \"nil\"\n\t}\n}", "func (config ConfigUser) mapToStruct(params map[string]interface{}) *ConfigUser {\n\tif _, isSet := params[\"userid\"]; isSet {\n\t\tconfig.User = UserID{}.mapToStruct(params[\"userid\"].(string))\n\t}\n\tif _, isSet := params[\"comment\"]; isSet {\n\t\tconfig.Comment = params[\"comment\"].(string)\n\t}\n\tif _, isSet := params[\"email\"]; isSet {\n\t\tconfig.Email = params[\"email\"].(string)\n\t}\n\tif _, isSet := params[\"enable\"]; isSet {\n\t\tconfig.Enable = Itob(int(params[\"enable\"].(float64)))\n\t}\n\tif _, isSet := params[\"expire\"]; isSet {\n\t\tconfig.Expire = uint(params[\"expire\"].(float64))\n\t}\n\tif _, isSet := params[\"firstname\"]; isSet {\n\t\tconfig.FirstName = params[\"firstname\"].(string)\n\t}\n\tif _, isSet := params[\"keys\"]; isSet {\n\t\tconfig.Keys = params[\"keys\"].(string)\n\t}\n\tif _, isSet := params[\"lastname\"]; isSet {\n\t\tconfig.LastName = params[\"lastname\"].(string)\n\t}\n\tif _, isSet := params[\"groups\"]; isSet {\n\t\tconfig.Groups = GroupName(\"\").mapToArray(params[\"groups\"])\n\t}\n\treturn &config\n}", "func (opts PortCreateOptsExt) ToPortCreateMap() (map[string]interface{}, error) {\n\tbase, err := opts.CreateOptsBuilder.ToPortCreateMap()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tport := base[\"port\"].(map[string]interface{})\n\n\tif opts.QoSPolicyID != \"\" {\n\t\tport[\"qos_policy_id\"] = opts.QoSPolicyID\n\t}\n\n\treturn base, nil\n}" ]
[ "0.65248597", "0.6061667", "0.6005189", "0.57232744", "0.56322694", "0.5601512", "0.547078", "0.5470513", "0.54283756", "0.54171157", "0.53757435", "0.5334866", "0.53190863", "0.52980673", "0.5289407", "0.52570814", "0.52454984", "0.5202118", "0.520063", "0.5198144", "0.5189111", "0.5141389", "0.50865865", "0.50538665", "0.5040585", "0.50332946", "0.5017082", "0.5013802", "0.5003643", "0.49991983", "0.499748", "0.49701318", "0.49568555", "0.49355608", "0.49258295", "0.49164888", "0.48881087", "0.48838693", "0.48773625", "0.4874929", "0.48674455", "0.48552808", "0.4853899", "0.48498634", "0.48399386", "0.48373705", "0.48117504", "0.48114341", "0.48044437", "0.47961006", "0.478479", "0.47836363", "0.4776476", "0.4775579", "0.47731677", "0.47718498", "0.47698948", "0.47688055", "0.47663406", "0.4764255", "0.47585914", "0.47542912", "0.4743686", "0.47380537", "0.47378466", "0.47329545", "0.4731593", "0.47304088", "0.4724455", "0.47244528", "0.47193316", "0.46987256", "0.46972978", "0.46901482", "0.4688935", "0.46855468", "0.46827257", "0.46821487", "0.46714658", "0.46449858", "0.46165097", "0.46129793", "0.4612743", "0.46116802", "0.4605485", "0.45960265", "0.45898744", "0.4588033", "0.45847356", "0.4580803", "0.4577501", "0.45704487", "0.45674473", "0.4564119", "0.45627373", "0.45608056", "0.45547754", "0.45427147", "0.45422143", "0.45377922" ]
0.8126662
0
NewAutoRoller creates and returns a new AutoRoller which runs at the given frequency.
func NewAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, tickFrequency, repoFrequency time.Duration, depot_tools string, rollIntoAndroid bool, strategy string) (*AutoRoller, error) { var err error var rm repo_manager.RepoManager if rollIntoAndroid { rm, err = repo_manager.NewAndroidRepoManager(workdir, parentBranch, childPath, childBranch, repoFrequency, gerrit) } else { rm, err = repo_manager.NewDEPSRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, repoFrequency, depot_tools, gerrit) } if err != nil { return nil, err } recent, err := recent_rolls.NewRecentRolls(path.Join(workdir, "recent_rolls.db")) if err != nil { return nil, err } mh, err := autoroll_modes.NewModeHistory(path.Join(workdir, "autoroll_modes.db")) if err != nil { return nil, err } arb := &AutoRoller{ attemptCounter: util.NewAutoDecrementCounter(ROLL_ATTEMPT_THROTTLE_TIME), cqExtraTrybots: cqExtraTrybots, emails: emails, gerrit: gerrit, includeCommitLog: true, liveness: metrics2.NewLiveness("last-autoroll-landed", map[string]string{"child-path": childPath}), modeHistory: mh, recent: recent, rm: rm, status: &autoRollStatusCache{}, strategy: strategy, rollIntoAndroid: rollIntoAndroid, } // Cycle once to fill out the current status. if err := arb.doAutoRoll(); err != nil { return nil, err } go func() { for range time.Tick(tickFrequency) { util.LogErr(arb.doAutoRoll()) } }() return arb, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewAutoRoller(ctx context.Context, c *config.Config, emailer emailclient.Client, chatBotConfigReader chatbot.ConfigReader, g *gerrit.Gerrit, githubClient *github.GitHub, workdir, recipesCfgFile, serverURL string, gcsClient gcs.GCSClient, client *http.Client, rollerName string, local bool, statusDB status.DB, manualRollDB manual.DB) (*AutoRoller, error) {\n\t// Validation and setup.\n\tif err := c.Validate(); err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to validate config\")\n\t}\n\tvar cr codereview.CodeReview\n\tvar err error\n\tif c.GetGerrit() != nil {\n\t\tcr, err = codereview.NewGerrit(c.GetGerrit(), g, client)\n\t} else if c.GetGithub() != nil {\n\t\tcr, err = codereview.NewGitHub(c.GetGithub(), githubClient)\n\t} else {\n\t\treturn nil, skerr.Fmt(\"Either GitHub or Gerrit is required.\")\n\t}\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to initialize code review\")\n\t}\n\treg, err := config_vars.NewRegistry(ctx, chrome_branch.NewClient(client))\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create config var registry\")\n\t}\n\n\t// Create the RepoManager.\n\trm, err := repo_manager.New(ctx, c.GetRepoManagerConfig(), reg, workdir, rollerName, recipesCfgFile, serverURL, c.ServiceAccount, client, cr, c.IsInternal, local)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\n\tsklog.Info(\"Creating strategy history.\")\n\tsh, err := strategy.NewDatastoreStrategyHistory(ctx, rollerName, c.ValidStrategies())\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create strategy history\")\n\t}\n\tcurrentStrategy := sh.CurrentStrategy()\n\tif currentStrategy == nil {\n\t\t// If there's no history, set the initial strategy.\n\t\tsklog.Infof(\"Setting initial strategy for %s to %q\", rollerName, c.DefaultStrategy())\n\t\tif err := sh.Add(ctx, c.DefaultStrategy(), \"AutoRoll Bot\", \"Setting initial strategy.\"); err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to set initial strategy\")\n\t\t}\n\t\tcurrentStrategy = sh.CurrentStrategy()\n\t}\n\tsklog.Info(\"Setting strategy.\")\n\tstrat, err := strategy.GetNextRollStrategy(currentStrategy.Strategy)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to get next roll strategy\")\n\t}\n\n\tsklog.Info(\"Running repo_manager.Update()\")\n\tlastRollRev, tipRev, notRolledRevs, err := rm.Update(ctx)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed initial repo manager update\")\n\t}\n\tnextRollRev := strat.GetNextRollRev(notRolledRevs)\n\tif nextRollRev == nil {\n\t\tnextRollRev = lastRollRev\n\t}\n\n\tsklog.Info(\"Creating roll history\")\n\trecent, err := recent_rolls.NewRecentRolls(ctx, recent_rolls.NewDatastoreRollsDB(ctx), rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create recent rolls DB\")\n\t}\n\tsklog.Info(\"Creating mode history\")\n\tmh, err := modes.NewDatastoreModeHistory(ctx, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create mode history\")\n\t}\n\tif mh.CurrentMode() == nil {\n\t\tsklog.Info(\"Setting initial mode.\")\n\t\tif err := mh.Add(ctx, modes.ModeRunning, \"AutoRoll Bot\", \"Setting initial mode.\"); err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to set initial mode\")\n\t\t}\n\t}\n\n\t// Throttling counters.\n\tsklog.Info(\"Creating throttlers\")\n\tsafetyThrottleCfg := config.DefaultSafetyThrottleConfig\n\tif c.SafetyThrottle != nil {\n\t\tsafetyThrottleCfg = c.SafetyThrottle\n\t}\n\tsafetyThrottleDuration, err := human.ParseDuration(safetyThrottleCfg.TimeWindow)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to parse safety throttle time window\")\n\t}\n\tsafetyThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/attempt_counter\", safetyThrottleDuration, int64(safetyThrottleCfg.AttemptCount))\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create safety throttler\")\n\t}\n\n\tfailureThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/fail_counter\", time.Hour, 1)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create failure throttler\")\n\t}\n\n\tvar rollCooldown time.Duration\n\tif c.RollCooldown != \"\" {\n\t\trollCooldown, err = human.ParseDuration(c.RollCooldown)\n\t}\n\tsuccessThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/success_counter\", rollCooldown, 1)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create success throttler\")\n\t}\n\tsklog.Info(\"Getting reviewers\")\n\temails := GetReviewers(client, c.RollerName, c.Reviewer, c.ReviewerBackup)\n\tsklog.Info(\"Creating notifier\")\n\tconfigCopies := replaceReviewersPlaceholder(c.Notifiers, emails)\n\tn, err := arb_notifier.New(ctx, c.ChildDisplayName, c.ParentDisplayName, serverURL, client, emailer, chatBotConfigReader, configCopies)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create notifier\")\n\t}\n\tsklog.Info(\"Creating status cache.\")\n\tstatusCache, err := status.NewCache(ctx, statusDB, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create status cache\")\n\t}\n\tsklog.Info(\"Creating TimeWindow.\")\n\tvar tw *time_window.TimeWindow\n\tif c.TimeWindow != \"\" {\n\t\ttw, err = time_window.Parse(c.TimeWindow)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to create TimeWindow\")\n\t\t}\n\t}\n\tcommitMsgBuilder, err := commit_msg.NewBuilder(c.CommitMsg, reg, c.ChildDisplayName, c.ParentDisplayName, serverURL, c.ChildBugLink, c.ParentBugLink, c.TransitiveDeps)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tarb := &AutoRoller{\n\t\tcfg: c,\n\t\tclient: client,\n\t\tcodereview: cr,\n\t\tcommitMsgBuilder: commitMsgBuilder,\n\t\temails: emails,\n\t\tfailureThrottle: failureThrottle,\n\t\tlastRollRev: lastRollRev,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"roller\": c.RollerName}),\n\t\tmanualRollDB: manualRollDB,\n\t\tmodeHistory: mh,\n\t\tnextRollRev: nextRollRev,\n\t\tnotifier: n,\n\t\tnotRolledRevs: notRolledRevs,\n\t\trecent: recent,\n\t\treg: reg,\n\t\trm: rm,\n\t\troller: rollerName,\n\t\trollUploadAttempts: metrics2.GetCounter(\"autoroll_cl_upload_attempts\", map[string]string{\"roller\": c.RollerName}),\n\t\trollUploadFailures: metrics2.GetCounter(\"autoroll_cl_upload_failures\", map[string]string{\"roller\": c.RollerName}),\n\t\tsafetyThrottle: safetyThrottle,\n\t\tserverURL: serverURL,\n\t\treviewers: c.Reviewer,\n\t\treviewersBackup: c.ReviewerBackup,\n\t\tstatus: statusCache,\n\t\tstrategy: strat,\n\t\tstrategyHistory: sh,\n\t\tsuccessThrottle: successThrottle,\n\t\tthrottle: unthrottle.NewDatastore(ctx),\n\t\ttimeWindow: tw,\n\t\ttipRev: tipRev,\n\t}\n\tsklog.Info(\"Creating state machine\")\n\tsm, err := state_machine.New(ctx, arb, n, gcsClient, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create state machine\")\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\trollingTo, err := arb.getRevision(ctx, current.RollingTo)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\troll, err := arb.retrieveRoll(ctx, current, rollingTo)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to retrieve current roll\")\n\t\t}\n\t\tif err := roll.InsertIntoDB(ctx); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\tsklog.Info(\"Done creating autoroller\")\n\treturn arb, nil\n}", "func NewRoller() *Roller {\n\troll := &Roller{\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t}\n\treturn roll\n}", "func newAutoRoller(workdir, childPath, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, rm repo_manager.RepoManager, retrieveRoll func(*AutoRoller, int64) (RollImpl, error)) (*AutoRoller, error) {\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"child_path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\tretrieveRoll: retrieveRoll,\n\t\trm: rm,\n\t\tstatus: &AutoRollStatusCache{},\n\t}\n\tsm, err := state_machine.New(arb, workdir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\troll, err := arb.retrieveRoll(arb, current.Issue)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\treturn arb, nil\n}", "func New(redisOptions *redis.Options, namespace string, min, max uint) *Yaraus {\n\treturn &Yaraus{\n\t\tc: redis.NewClient(redisOptions),\n\t\tmin: min,\n\t\tmax: max,\n\t\tnamespace: namespace,\n\t\tInterval: 1 * time.Second,\n\t\tDelay: 2 * time.Second,\n\t\tExpire: 30 * time.Minute,\n\t}\n}", "func (r *AutoRoller) Start(ctx context.Context, tickFrequency time.Duration) {\n\tsklog.Infof(\"Starting autoroller.\")\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(tickFrequency, func(_ context.Context) {\n\t\t// Explicitly ignore the passed-in context; this allows us to\n\t\t// continue running even if the context is canceled, which helps\n\t\t// to prevent errors due to interrupted syncs, etc.\n\t\tctx := context.Background()\n\t\tif err := r.Tick(ctx); err != nil {\n\t\t\t// Hack: we frequently get failures from GoB which trigger error-rate alerts.\n\t\t\t// These alerts are noise and sometimes hide real failures. If the error is\n\t\t\t// due to a sync failure, log it as a warning instead of an error. We'll rely\n\t\t\t// on the liveness alert in the case where we see persistent sync failures.\n\t\t\tif isSyncError(err) {\n\t\t\t\tsklog.Warningf(\"Failed to run autoroll: %s\", err)\n\t\t\t} else {\n\t\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t}, nil)\n\n\t// Update the current reviewers in a loop.\n\tlvReviewers := metrics2.NewLiveness(\"last_successful_reviewers_retrieval\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(30*time.Minute, func(ctx context.Context) {\n\t\temails := GetReviewers(r.client, r.cfg.RollerName, r.cfg.Reviewer, r.cfg.ReviewerBackup)\n\t\tr.emailsMtx.Lock()\n\t\tdefer r.emailsMtx.Unlock()\n\t\tr.emails = emails\n\n\t\tconfigCopies := replaceReviewersPlaceholder(r.cfg.Notifiers, emails)\n\t\tif err := r.notifier.ReloadConfigs(ctx, configCopies); err != nil {\n\t\t\tsklog.Errorf(\"Failed to reload configs: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tlvReviewers.Reset()\n\t}, nil)\n\n\t// Handle requests for manual rolls.\n\tif r.cfg.SupportsManualRolls {\n\t\tlvManualRolls := metrics2.NewLiveness(\"last_successful_manual_roll_check\", map[string]string{\"roller\": r.roller})\n\t\tcleanup.Repeat(time.Minute, func(_ context.Context) {\n\t\t\t// Explicitly ignore the passed-in context; this allows\n\t\t\t// us to continue handling manual rolls even if the\n\t\t\t// context is canceled, which helps to prevent errors\n\t\t\t// due to interrupted syncs, etc.\n\t\t\tctx := context.Background()\n\t\t\tif err := r.handleManualRolls(ctx); err != nil {\n\t\t\t\tsklog.Error(err)\n\t\t\t} else {\n\t\t\t\tlvManualRolls.Reset()\n\t\t\t}\n\t\t}, nil)\n\t}\n}", "func (*lruFactory) New(args xreg.Args, _ *meta.Bck) xreg.Renewable {\n\treturn &lruFactory{RenewBase: xreg.RenewBase{Args: args}}\n}", "func New(fn RunFunction, rates []Rate) (*rrInstance, error) {\n\tif len(rates) == 0 {\n\t\treturn nil, errors.New(\"empty rates\")\n\t}\n\n\trateRunner := &rrInstance{\n\t\tterminateRunner: make(chan bool, 1),\n\t\trestartRates: make(chan bool, 1),\n\t\trunFunction: fn,\n\t\trates: rates,\n\t\tnextRateIndex: 0,\n\t}\n\n\treturn rateRunner, nil\n}", "func New(timeout time.Duration, cache GetSetter) Throttler {\n\tsalt, err := randomBytes(16)\n\tif err != nil {\n\t\tpanic(\"cannot initialize rate limiter\")\n\t}\n\treturn &Limiter{\n\t\tcache: cache,\n\t\ttimeout: timeout,\n\t\tsalt: salt,\n\t}\n}", "func NewDispatcher(init, max time.Duration, handler AlertHandler) *Dispatcher {\n\tif init <= 0 {\n\t\tinit = 1 * time.Second\n\t}\n\tif max < init {\n\t\tmax = init\n\t}\n\treturn &Dispatcher{\n\t\tinitInterval: init,\n\t\tmaxInterval: max,\n\t\thandler: handler,\n\t}\n}", "func NewAutoScaler(c *options.AutoScalerConfig) (*AutoScaler, error) {\n\tnewK8sClient, err := k8sclient.NewK8sClient(c.Namespace, c.Target, c.Kubeconfig, c.DryRun)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcfg := ScaleConfig{}\n\tif c.DefaultConfig != \"\" {\n\t\tif err := json.Unmarshal([]byte(c.DefaultConfig), &cfg); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"invalid default config: %v\", err)\n\t\t}\n\t}\n\treturn &AutoScaler{\n\t\tk8sClient: newK8sClient,\n\t\tdefaultConfig: cfg,\n\t\tconfigFile: c.ConfigFile,\n\t\tpollPeriod: time.Second * time.Duration(c.PollPeriodSeconds),\n\t\tclock: clock.RealClock{},\n\t\tstopCh: make(chan struct{}),\n\t\treadyCh: make(chan struct{}, 1),\n\t}, nil\n}", "func NewRates(name string, countTracker CountTracker, samples int, interval time.Duration) *Rates {\n\tif interval < 1*time.Second && interval != -1*time.Second {\n\t\tpanic(\"interval too small\")\n\t}\n\tctx, cancel := context.WithCancel(context.Background())\n\trt := &Rates{\n\t\ttimeStamps: NewRingInt64(samples + 1),\n\t\tcounts: make(map[string]*RingInt64),\n\t\tcountTracker: countTracker,\n\t\tsamples: samples + 1,\n\t\tinterval: interval,\n\t\ttimestampLastSampling: timeNow(),\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t}\n\tif name != \"\" {\n\t\tpublish(name, rt)\n\t}\n\tif interval > 0 {\n\t\tgo rt.track()\n\t}\n\treturn rt\n}", "func New(rate int, opts ...Option) Limiter {\n\treturn newAtomicBased(rate, opts...)\n}", "func newManualPollingPolicy(\n\tconfigFetcher configProvider,\n\tstore *configStore,\n\tlogger Logger) *manualPollingPolicy {\n\n\treturn &manualPollingPolicy{configRefresher: configRefresher{configFetcher: configFetcher, store: store, logger: logger}}\n}", "func New(rate int, opts ...Option) Limiter {\r\n\treturn newAtomicBased(rate, opts...)\r\n}", "func (r *AutoRoller) Tick(ctx context.Context) error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\n\t// Update the config vars.\n\tif err := r.reg.Update(ctx); err != nil {\n\t\tsklog.Errorf(\"Failed to update config registry; continuing, but config may be out of date: %s\", err)\n\t}\n\n\t// Determine if we should unthrottle.\n\tshouldUnthrottle, err := r.throttle.Get(ctx, r.roller)\n\tif err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to determine whether we should unthrottle\")\n\t}\n\tif shouldUnthrottle {\n\t\tif err := r.unthrottle(ctx); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to unthrottle\")\n\t\t}\n\t\tif err := r.throttle.Reset(ctx, r.roller); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to reset unthrottle counter\")\n\t\t}\n\t}\n\n\t// Update modes and strategies.\n\tif err := r.modeHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update mode history\")\n\t}\n\toldStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif err := r.strategyHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update strategy history\")\n\t}\n\tnewStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif oldStrategy != newStrategy {\n\t\tstrat, err := strategy.GetNextRollStrategy(newStrategy)\n\t\tif err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to get next roll strategy\")\n\t\t}\n\t\tr.strategyMtx.Lock()\n\t\tr.strategy = strat\n\t\tr.strategyMtx.Unlock()\n\t}\n\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence(ctx)\n\tlastErrStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrStr = lastErr.Error()\n\t}\n\n\t// Update the status information.\n\tif err := r.updateStatus(ctx, true, lastErrStr); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update status\")\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn skerr.Wrapf(lastErr, \"Failed state transition sequence\")\n}", "func New(callback func(types.NamespacedName), lease time.Duration) Interface {\n\treturn &impl{\n\t\tleaseDuration: lease,\n\t\tcb: callback,\n\t}\n}", "func (r *AutoRoller) Start(tickFrequency, repoFrequency time.Duration, ctx context.Context) {\n\tsklog.Infof(\"Starting autoroller.\")\n\trepo_manager.Start(r.rm, repoFrequency, ctx)\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\")\n\tgo util.RepeatCtx(tickFrequency, ctx, func() {\n\t\tif err := r.Tick(); err != nil {\n\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\tutil.LogErr(r.recent.Close())\n\t\t\t\tutil.LogErr(r.modeHistory.Close())\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\t}()\n}", "func New(backoff []time.Duration, class Classifier) *Retrier {\n\tif class == nil {\n\t\tclass = DefaultClassifier{}\n\t}\n\n\treturn &Retrier{\n\t\tbackoff: backoff,\n\t\tclass: class,\n\t\trand: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}\n}", "func New(ctx context.Context, childName, parentName, serverURL string, client *http.Client, emailer emailclient.Client, chatBotConfigReader chatbot.ConfigReader, configs []*notifier.Config) (*AutoRollNotifier, error) {\n\tn := &AutoRollNotifier{\n\t\tchildName: childName,\n\t\tclient: client,\n\t\tconfigReader: chatBotConfigReader,\n\t\temailer: emailer,\n\t\tparentName: parentName,\n\t\tserverURL: serverURL,\n\t}\n\tif err := n.ReloadConfigs(ctx, configs); err != nil {\n\t\treturn nil, err\n\t}\n\treturn n, nil\n}", "func NewWatcher(\n\tlister ListerService,\n\tinterval time.Duration,\n\topts ...func(*WatcherConfig),\n) *Watcher {\n\tcfg := WatcherConfig{\n\t\tLogger: zero.Logger(),\n\t}\n\tfor _, opt := range opts {\n\t\topt(&cfg)\n\t}\n\tw := &Watcher{\n\t\tlister: lister,\n\t\tstreamer: stream.NewPoller(\n\t\t\tfunc() (zero.Interface, error) {\n\t\t\t\treturn lister.ListGoRepos()\n\t\t\t},\n\t\t\tinterval,\n\t\t),\n\t\tlog: cfg.Logger,\n\t}\n\tgo w.run()\n\treturn w\n}", "func NewShuffler() *Shuffler { return &Shuffler{} }", "func newAlfredWatcher() *alfredWatcher {\n w, _ := inotify.NewWatcher()\n aw := &alfredWatcher{\n watcher: w,\n list: make(map[string]uint32),\n }\n return aw\n}", "func NewStrategy(\n\tmaxWaitDuration time.Duration,\n\tquiesceDuration time.Duration,\n\ttimerFrequency time.Duration) *Strategy {\n\n\ts := &Strategy{\n\t\tmaxWaitDuration: maxWaitDuration,\n\t\tquiesceDuration: quiesceDuration,\n\t\ttimerFrequency: timerFrequency,\n\t\tPublish: make(chan struct{}, 1),\n\t\tnowFn: time.Now,\n\t\tworker: util.NewWorker(\"runtime publishing strategy\", log.Scope),\n\t\tresetChan: make(chan struct{}, 1),\n\t}\n\ts.startTimerFn = s.startTimer\n\treturn s\n}", "func NewAutoDecoder(r io.Reader) Decoder {\n\tok, newR := IsNTFormat(r)\n\tif ok {\n\t\treturn NewLenientNTDecoder(newR)\n\t}\n\treturn NewBinaryDecoder(newR)\n}", "func New(pattern string, options ...Option) (*RotateLogs, error) {\n\tglobPattern := pattern\n\tfor _, re := range patternConversionRegexps {\n\t\tglobPattern = re.ReplaceAllString(globPattern, \"*\")\n\t}\n\n\tstrfobj, err := strftime.New(pattern)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, `invalid strftime pattern`)\n\t}\n\n\tvar rl RotateLogs\n\trl.clock = Local\n\trl.globPattern = globPattern\n\trl.pattern = strfobj\n\trl.rotationTime = 24 * time.Hour\n\trl.maxAge = 7 * 24 * time.Hour\n\tfor _, opt := range options {\n\t\topt.Configure(&rl)\n\t}\n\n\treturn &rl, nil\n}", "func (r *AutoRoller) Tick() error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence()\n\n\t// Update the status information.\n\tlastErrorStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrorStr = lastErr.Error()\n\t}\n\trecent := r.recent.GetRecentRolls()\n\tnumFailures := 0\n\tfor _, roll := range recent {\n\t\tif roll.Failed() {\n\t\t\tnumFailures++\n\t\t} else if roll.Succeeded() {\n\t\t\tbreak\n\t\t}\n\t}\n\tsklog.Infof(\"Updating status (%d)\", r.rm.CommitsNotRolled())\n\tif err := r.status.Set(&AutoRollStatus{\n\t\tAutoRollMiniStatus: AutoRollMiniStatus{\n\t\t\tNumFailedRolls: numFailures,\n\t\t\tNumNotRolledCommits: r.rm.CommitsNotRolled(),\n\t\t},\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tFullHistoryUrl: r.gerrit.Url(0) + \"/q/owner:\" + r.GetUser(),\n\t\tIssueUrlBase: r.gerrit.Url(0) + \"/c/\",\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t}); err != nil {\n\t\treturn err\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn lastErr\n}", "func newThrottlerWithClock(name, unit string, threadCount int, maxRate int64, maxReplicationLag int64, nowFunc func() time.Time) (*Throttler, error) {\n\treturn newThrottler(GlobalManager, name, unit, threadCount, maxRate, maxReplicationLag, nowFunc)\n}", "func NewAutoRegistryGenerator(registryWriter AutoGeneratedRegistryWriter) AutoRegistryGenerator {\n\treturn AutoRegistryGenerator{registryWriter: &registryWriter}\n}", "func New(fireAfter time.Duration, fireFunc func() ()) *Ticker {\n\treturn &Ticker{\n\t\tlastRestart: time.Now(),\n\t\tfireAfter: fireAfter,\n\t\tfireFunc: fireFunc,\n\t\tactive: false,\n\t}\n}", "func New(l log.Logger, taskInterval, taskDelay time.Duration) Timer {\n\treturn &timer{\n\t\twg: sync.WaitGroup{},\n\t\tl: l.WithModule(\"timer\"),\n\t\ttaskInterval: taskInterval,\n\t\ttaskDelay: taskDelay,\n\t}\n}", "func initKeyRenewal(registry *KeyRegistry, period time.Duration, cutoffTime time.Time) (func(), error) {\n\t// Create a new key if it's the first key,\n\t// or if it's older than cutoff time.\n\tif len(registry.keys) == 0 || registry.mostRecentKey.creationTime.Before(cutoffTime) {\n\t\tif _, err := registry.generateKey(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// wrapper function to log error thrown by generateKey function\n\tkeyGenFunc := func() {\n\t\tif _, err := registry.generateKey(); err != nil {\n\t\t\tlog.Printf(\"Failed to generate new key : %v\\n\", err)\n\t\t}\n\t}\n\tif period == 0 {\n\t\treturn keyGenFunc, nil\n\t}\n\n\t// If key rotation is enabled, we'll rotate the key when the most recent\n\t// key becomes stale (older than period).\n\tmostRecentKeyAge := time.Since(registry.mostRecentKey.creationTime)\n\tinitialDelay := period - mostRecentKeyAge\n\tif initialDelay < 0 {\n\t\tinitialDelay = 0\n\t}\n\treturn ScheduleJobWithTrigger(initialDelay, period, keyGenFunc), nil\n}", "func New(policy *Policy) *RateLimiter {\n\trl := &RateLimiter{\n\t\tpolicy: policy,\n\t\tstartTime: nowFunc(),\n\t}\n\treturn rl\n}", "func NewAutoScanner(r io.Reader) *Scanner {\n\treturn NewScanner(nil, r)\n}", "func New(connector connectors.AutoUpdateDataConnector, collectionPeriod int,\n postgresUrl string, apiConfig utils.APIDependencyConfig) *AutoUpdater {\n return &AutoUpdater{\n PostgresURL: postgresUrl,\n DataConnector: connector,\n CollectionPeriodMinutes: collectionPeriod,\n TRFApiConfig: apiConfig,\n }\n}", "func New(maxDelayBetweenHits time.Duration, factors ...Factor) Streaker {\n\tsortFactors(factors)\n\n\treturn Streaker{\n\t\tMaxDelayBetweenHits: maxDelayBetweenHits,\n\t\tFactors: factors,\n\t}\n}", "func New(cfg Config) Backoff {\n\n\t// if no interval is provided (nil), set to default value\n\tvar interval int\n\tif cfg.Interval == nil {\n\t\tinterval = DefaultInterval\n\t} else {\n\t\tinterval = *cfg.Interval\n\t}\n\n\tswitch cfg.Type {\n\tcase TypeConstant:\n\t\treturn NewConstant(interval)\n\tcase TypeLinear:\n\t\treturn NewLinear(interval)\n\tcase TypeExponential:\n\t\treturn NewExponential(interval, cfg.Multiplier)\n\tdefault:\n\t\treturn NewConstant(interval)\n\t}\n}", "func NewFrequency(v float64, s string) Frequency {\n\treturn Frequency(v) * frequency[s]\n}", "func newWatcher(loader *Loader, uri string, interval time.Duration, onStop func()) *watcher {\n\treturn &watcher{\n\t\tstate: isCreated,\n\t\tupdatedAt: 0,\n\t\tloader: loader,\n\t\turi: uri,\n\t\tupdates: make(chan Update, 1),\n\t\tinterval: interval,\n\t\tonStop: onStop,\n\t}\n}", "func NewRoll(s *S) *Roll {\n\tu := &Roll{\n\t\tS: s,\n\t\tC: NewCCap(s.Len() * 10),\n\t\tdmap: make([][]z.Lit, s.Len())}\n\treturn u\n}", "func New(datastore Datastore, restricter RestrictMiddleware, closed <-chan struct{}) *Autoupdate {\n\ta := &Autoupdate{\n\t\tdatastore: datastore,\n\t\ttopic: topic.New(topic.WithClosed(closed)),\n\t\trestricter: restricter,\n\t}\n\n\t// Update the topic when an data update is received.\n\ta.datastore.RegisterChangeListener(func(data map[string][]byte) error {\n\t\tkeys := make([]string, 0, len(data))\n\t\tfor k := range data {\n\t\t\tkeys = append(keys, k)\n\t\t}\n\n\t\ta.topic.Publish(keys...)\n\t\treturn nil\n\t})\n\n\treturn a\n}", "func New(store *storage.Store, interval int) {\n\tlog.Info().Int(\"interval\", interval).Msg(\"Starting the scheduler\")\n\n\tgo func() {\n\t\tticker := time.NewTicker(time.Minute * time.Duration(interval))\n\n\t\tfor range ticker.C {\n\t\t\tgo func() {\n\t\t\t\tnotRefreshedSince := time.Now().Add(-1 * time.Hour)\n\n\t\t\t\tfeeds, totalCount := store.FeedList(context.TODO(), &storage.FeedListOptions{\n\t\t\t\t\tNotRefreshedSince: notRefreshedSince,\n\t\t\t\t\tLimit: 100,\n\t\t\t\t})\n\n\t\t\t\tlog.Info().Int(\"feeds\", totalCount).Time(\"not_refreshed_since\", notRefreshedSince).Msg(\"Unfresh feeds found\")\n\n\t\t\t\tfor _, feed := range *feeds {\n\t\t\t\t\tif err := store.FeedRefresh(context.TODO(), feed); err != nil {\n\t\t\t\t\t\tlog.Warn().Err(err).Str(\"feed_title\", feed.Title).Msg(\"Error refreshing feed\")\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}()\n\t\t}\n\t}()\n}", "func NewFileRoller(file string, maxSize int64, logsToKeep int) (obj *FileRoller, err error) {\n\terr = nil\n\n\tif file == \"\" {\n\t\terr = ErrInvalidFile\n\t\treturn\n\t}\n\n\tobj = &FileRoller{}\n\tobj.FileName = file\n\tobj.MaxSize = maxSize\n\tobj.LogsToKeep = logsToKeep\n\n\treturn\n}", "func New(window, granularity time.Duration) (*SlidingWindow, error) {\n\tif window == 0 {\n\t\treturn nil, errors.New(\"window cannot be 0\")\n\t}\n\tif granularity == 0 {\n\t\treturn nil, errors.New(\"granularity cannot be 0\")\n\t}\n\tif window <= granularity || window%granularity != 0 {\n\t\treturn nil, errors.New(\"window size has to be a multiplier of the granularity size\")\n\t}\n\n\tsw := &SlidingWindow{\n\t\twindow: window,\n\t\tgranularity: granularity,\n\t\tsamples: make([]NamespaceMetrics, int(window/granularity)),\n\t\tstopC: make(chan struct{}),\n\t}\n\n\tgo sw.shifter()\n\treturn sw, nil\n}", "func NewRollingPolicy(window *Window, opts RollingPolicyOpts) *RollingPolicy {\n\treturn &RollingPolicy{\n\t\twindow: window,\n\t\tsize: window.Size(),\n\t\toffset: 0,\n\n\t\tbucketDuration: opts.BucketDuration,\n\t\tlastAppendTime: time.Now(),\n\t}\n}", "func NewAutoLinker(link *Link) (*AutoLinker, error) {\n\tif link == nil || len(link.Pattern) == 0 || len(link.Template) == 0 {\n\t\treturn nil, errors.New(\"Pattern or template was empty\")\n\t}\n\n\tif !link.DisableNonWordPrefix {\n\t\tlink.Pattern = \"(?P<MMDisableNonWordPrefix>^|\\\\s)\" + link.Pattern\n\t\tlink.Template = \"${MMDisableNonWordPrefix}\" + link.Template\n\t}\n\n\tif !link.DisableNonWordSuffix {\n\t\tlink.Pattern = link.Pattern + \"(?P<DisableNonWordSuffix>$|\\\\s|\\\\.|\\\\!|\\\\?|\\\\,|\\\\))\"\n\t\tlink.Template = link.Template + \"${DisableNonWordSuffix}\"\n\t}\n\n\tp, err := regexp.Compile(link.Pattern)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &AutoLinker{\n\t\tlink: link,\n\t\tpattern: p,\n\t}, nil\n}", "func NewPoller(getFunc GetFunc, period time.Duration, store Store) *Poller {\n\treturn &Poller{\n\t\tgetFunc: getFunc,\n\t\tperiod: period,\n\t\tstore: store,\n\t}\n}", "func NewFollower(database *Database, interval int) *Follower {\n\tfollower := &Follower{\n\t\tdb: database,\n\t\tstop: make(chan struct{}),\n\t\tstopped: make(chan struct{}),\n\t\tseqInterval: interval,\n\t}\n\treturn follower\n}", "func NewWatcher(filename string) Watcher {\n\treturn &watchImpl{filename: filename}\n}", "func newWatchAggregator(c, wc Client, autoWatch bool, autoWatchRetry time.Duration) *watchAggregator {\n\tif autoWatchRetry == 0 {\n\t\tautoWatchRetry = defaultAutoWatchRetry\n\t}\n\taggregator := &watchAggregator{\n\t\tClient: c,\n\t\tpassiveClient: wc,\n\t\tautoWatch: autoWatch,\n\t\tautoWatchRetry: autoWatchRetry,\n\t\tlog: log.DefaultLogger(),\n\t\tsubscribers: make([]subscriber, 0),\n\t}\n\treturn aggregator\n}", "func NewRetrier(maxTries int, initialDelay, maxDelay time.Duration) *Retrier {\n\tif maxTries <= 0 {\n\t\tmaxTries = DefaultMaxTries\n\t}\n\tif initialDelay <= 0 {\n\t\tinitialDelay = DefaultInitialDelay\n\t}\n\tif maxDelay <= 0 {\n\t\tmaxDelay = DefaultMaxDelay\n\t}\n\treturn &Retrier{maxTries, initialDelay, maxDelay}\n}", "func NewManifestAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, depot_tools string, strategy repo_manager.NextRollStrategy, preUploadSteps []string, serverURL string) (*AutoRoller, error) {\n\trm, err := repo_manager.NewManifestRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, depot_tools, gerrit, strategy, preUploadSteps, serverURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tretrieveRoll := func(arb *AutoRoller, issue int64) (RollImpl, error) {\n\t\treturn newGerritRoll(arb.gerrit, arb.rm, arb.recent, issue)\n\t}\n\treturn newAutoRoller(workdir, childPath, cqExtraTrybots, emails, gerrit, rm, retrieveRoll)\n}", "func NewHook(executions int, length string, interval int, effect func(), effectOff func()) *Hook {\n\tlengthTime, _ := strconv.Atoi(length)\n\tparseLength := time.Duration(lengthTime) * time.Second\n\treturn &Hook{executions,\n\t\ttime.Now(),\n\t\tparseLength,\n\t\ttime.Now(),\n\t\tinterval,\n\t\teffect,\n\t\teffectOff}\n}", "func New(patterns []string, words string, nameStrategy string) *Namer {\n\tps := []Pattern{}\n\tfor _, p := range patterns {\n\t\tps = append(ps, Pattern(p))\n\t}\n\tw := &Words{Name: words}\n\tif err := w.Read(); err != nil {\n\t\tpanic(err)\n\t}\n\treturn &Namer{Patterns: ps, WordsName: words, Words: w, NameStrategy: nameStrategy}\n}", "func NewWatcher(filePath string, interval time.Duration, checks ...WatcherCheck) (*Watcher, error) {\n\treturn NewWatcherWithOpts(filePath, DefaultWatcherOpts, interval, checks...)\n}", "func newAssigner(period time.Duration, newBucket bucketFactory, clock clock) *assigner {\n\ta := &assigner{\n\t\tperiod: period,\n\t\tclock: clock,\n\t\tnewBucket: newBucket,\n\t}\n\ttc := a.nextT(clock.Now())\n\ttp := a.prevT(tc)\n\ttn := a.nextT(tc)\n\ta.buckets[prev] = newBucket(tp)\n\ta.buckets[curr] = newBucket(tc)\n\ta.buckets[next] = newBucket(tn)\n\ta.tmin = tp.Add(-a.period)\n\treturn a\n}", "func NewThrottler(elements uint, period time.Duration, bufferSize uint, mode ThrottleMode) *Throttler {\n\tthrottler := &Throttler{\n\t\tmaxElements: uint64(elements),\n\t\tperiod: period,\n\t\tmode: mode,\n\t\tin: make(chan interface{}),\n\t\tout: make(chan interface{}, bufferSize),\n\t\tnotify: make(chan struct{}),\n\t\tdone: make(chan struct{}),\n\t\tcounter: 0,\n\t}\n\tgo throttler.resetCounterLoop(period)\n\tgo throttler.bufferize()\n\n\treturn throttler\n}", "func New() *PollerRegistry {\n\treturn &PollerRegistry{\n\t\tRegistry: make(map[string]*poller.Poller),\n\t\tToDB: make(chan interface{}),\n\t\tUpdateStatus: make(chan string),\n\t}\n}", "func (*EventNotificationsV1) NewRules(eventTypeFilter string) (_model *Rules, err error) {\n\t_model = &Rules{\n\t\tEventTypeFilter: core.StringPtr(eventTypeFilter),\n\t}\n\terr = core.ValidateStruct(_model, \"required parameters\")\n\treturn\n}", "func New(\n\tmetrics []string,\n\tconsumer func(string),\n\trpm uint,\n) (*Generator, error) {\n\tc := cron.New()\n\tp := make(chan string, 1000)\n\tg := &Generator{\n\t\tcron: c,\n\t\tmetrics: metrics,\n\t\tprovide: p,\n\t\tlastProvide: 0,\n\t\trpm: rpm,\n\t}\n\n\t_, err := c.AddFunc(\"@every 10s\", func() { g.feed() })\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t_, err = c.AddFunc(\"@every 1m\", func() { g.consume(consumer) })\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn g, nil\n}", "func (rb *redisBackend) newExponentialBackoffStrategy() backoff.BackOff {\n\tbackoffStrat := backoff.NewExponentialBackOff()\n\tbackoffStrat.InitialInterval = rb.cfg.GetDuration(\"backoff.initialInterval\")\n\tbackoffStrat.RandomizationFactor = rb.cfg.GetFloat64(\"backoff.randFactor\")\n\tbackoffStrat.Multiplier = rb.cfg.GetFloat64(\"backoff.multiplier\")\n\tbackoffStrat.MaxInterval = rb.cfg.GetDuration(\"backoff.maxInterval\")\n\tbackoffStrat.MaxElapsedTime = rb.cfg.GetDuration(\"backoff.maxElapsedTime\")\n\treturn backoff.BackOff(backoffStrat)\n}", "func (rb *redisBackend) newExponentialBackoffStrategy() backoff.BackOff {\n\tbackoffStrat := backoff.NewExponentialBackOff()\n\tbackoffStrat.InitialInterval = rb.cfg.GetDuration(\"backoff.initialInterval\")\n\tbackoffStrat.RandomizationFactor = rb.cfg.GetFloat64(\"backoff.randFactor\")\n\tbackoffStrat.Multiplier = rb.cfg.GetFloat64(\"backoff.multiplier\")\n\tbackoffStrat.MaxInterval = rb.cfg.GetDuration(\"backoff.maxInterval\")\n\tbackoffStrat.MaxElapsedTime = rb.cfg.GetDuration(\"backoff.maxElapsedTime\")\n\treturn backoff.BackOff(backoffStrat)\n}", "func New(appName string) (Manager, error) {\n\tconn, err := dbus.SessionBus()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmanager := &UnixManager{\n\t\tappName: appName,\n\t\tbusObj: conn.Object(prefix, notifPath),\n\t\tactives: make(map[uint32]*Notification, 8),\n\t}\n\n\t// Listen to notifications signals\n\tdbusSignal := make(chan *dbus.Signal)\n\n\terr = conn.AddMatchSignal(dbus.WithMatchSender(prefix))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tconn.Signal(dbusSignal)\n\n\tgo manager.listenTo(dbusSignal)\n\n\treturn manager, nil\n}", "func NewWheel(interval time.Duration, count int) *Wheel {\n\tw := &Wheel{\n\t\tr: ring.New(count),\n\t\tinterval: interval,\n\t\tstopper: make(data.Chan, 1),\n\t}\n\tgo w.start()\n\treturn w\n}", "func NewThrottler(opts Options) *Throttler {\n\tt := &Throttler{\n\t\tmode: opts.Mode,\n\t\tevaluators: make(map[string]LoadEvaluator),\n\t\tthresholds: make(map[string]float64),\n\t}\n\n\tif t.mode == Disabled {\n\t\t// don't bother to create Evaluators if throttling is not\n\t\t// enabled.\n\t\treturn t\n\t}\n\n\tif opts.AverageLatencyThreshold > 0 {\n\t\te := NewGRPCLatencyEvaluatorWithThreshold(opts.SamplesPerSecond, opts.SampleHalfLife, opts.LatencyEnforcementThreshold)\n\t\tt.evaluators[e.Name()] = e\n\t\tt.thresholds[e.Name()] = opts.AverageLatencyThreshold.Seconds()\n\t}\n\n\tif opts.MaxRequestsPerSecond > 0 {\n\t\te := NewRateLimitEvaluator(opts.MaxRequestsPerSecond, opts.BurstSize)\n\t\tt.evaluators[e.Name()] = e\n\t\tt.thresholds[e.Name()] = float64(opts.MaxRequestsPerSecond)\n\t}\n\n\tscope.Debugf(\"Built Throttler(%#v) from opts(%#v)\", t, opts)\n\treturn t\n}", "func NewCounter(ttl time.Duration, callback func(float64)) (counter *Counter) {\n\tcounter = &Counter{\n\t\tentries: make(map[[sha256.Size]byte]time.Time),\n\t\tttl: ttl,\n\t\tcallback: callback,\n\t}\n\n\tgo func() {\n\t\tfor now := range time.Tick(time.Second) {\n\t\t\tcounter.mutex.Lock()\n\t\t\tfor k, v := range counter.entries {\n\t\t\t\tif delta := now.Sub(v); delta > counter.ttl {\n\t\t\t\t\tdelete(counter.entries, k)\n\t\t\t\t}\n\t\t\t}\n\t\t\tcounter.callback(float64(len(counter.entries)))\n\t\t\tcounter.mutex.Unlock()\n\t\t}\n\t}()\n\n\treturn\n}", "func NewFactory(buckets int) xkit.Factory {\n\treturn factory{\n\t\tbuckets: buckets,\n\t}\n}", "func NewRulesSubscriber(nc *nats.Conn, service engine.Service, logger *zap.Logger) *rulesSubscriber {\n\treturn &rulesSubscriber{nc, service, logger}\n}", "func (r *EndpointsRolloutV1RestClient) AutoAddRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func RollingFileHandler(conf string, fmtr log15.Format) (log15.Handler, error) {\n\tvar (\n\t\tlj lumberjack.Logger\n\t\terr error\n\t)\n\tif _, err = toml.Decode(conf, &lj); err != nil {\n\t\treturn nil, err\n\t} else if _, err = lj.Write([]byte(\"init test\")); err != nil {\n\t\treturn nil, err\n\t}\n\treturn log15.StreamHandler(&lj, fmtr), nil\n}", "func NewAutoGrant() GrantHandler {\n\treturn &autoGrant{}\n}", "func newAgeStrategy(c *StrategyConfig, dir *directory, action performer, log logger) *ageStrategy {\n\treturn &ageStrategy{Strategy{c, dir, action, log}, 0}\n}", "func newRule(defaultExpiration time.Duration, allowed int, estimated ...int) *singleRule {\n\tif allowed <= 0 {\n\t\tallowed = 1\n\t}\n\tuserEstimated := 0\n\tif len(estimated) > 0 {\n\t\tuserEstimated = estimated[0]\n\t}\n\tif userEstimated <= 0 {\n\t\tuserEstimated = allowed\n\t}\n\tcleanupInterval := defaultExpiration / 100\n\tif cleanupInterval < time.Second*1 {\n\t\tcleanupInterval = time.Second * 1\n\t}\n\tif cleanupInterval > time.Second*60 {\n\t\tcleanupInterval = time.Second * 60\n\t}\n\tvc := createRule(defaultExpiration, cleanupInterval, allowed, userEstimated)\n\tgo vc.deleteExpired()\n\treturn vc\n}", "func NewPoller(url string, interval time.Duration, out chan PollMsg, shutdown chan *sync.WaitGroup) *Poller {\n\treturn &Poller{\n\t\tURL: url,\n\t\tInterval: interval,\n\t\tOut: out,\n\t\tShutdown: shutdown,\n\t}\n}", "func New(numConcurrent int32) *dispatcher {\n\td := new(dispatcher)\n\td.queue = new(queue)\n\td.cap = numConcurrent\n\td.active = new(int32)\n\td.poke = make(chan struct{})\n\td.bot() // starts a daemon that will schedule pending funcs\n\treturn d\n}", "func New() *BaseTradingRules {\n\tvar m BaseTradingRules\n\treturn &m\n}", "func NewDispatcher(count int) *Dispatcher {\n pool := make(chan chan Job, count)\n return &Dispatcher{WorkerPool: pool, maxWorkers: count}\n}", "func NewWatcher() *Watcher {\n\treturn &Watcher{\n\t\tLogFunc: noop,\n\t\tInterval: 500,\n\t}\n}", "func NewPoller(poll PollerFunc, interval time.Duration) *Poller {\n\treturn &Poller{\n\t\tChannel: make(chan interface{}),\n\t\tPoll: poll,\n\t\tWaitInterval: interval,\n\t\tisStopped: false,\n\t\tisFinished: false,\n\t\tgroup: &sync.WaitGroup{},\n\t\tstopMutex: &sync.Mutex{},\n\t}\n}", "func NewAutoTransport(maxIdle, maxIdlePerHost int) http.RoundTripper {\n\treturn newAutoTransport(\n\t\tnewHTTPTransport(false /*disable keep-alives*/, false /*disable auto-compression*/, maxIdle, maxIdlePerHost),\n\t\tnewH2CTransport(false /*disable auto-compression*/))\n}", "func NewDispatcher(handler Handler, waiter Waiter, config *DispatcherConfig) *Dispatcher {\n\tif config == nil {\n\t\tconfig = defaultDispatcherConfig\n\t}\n\n\tpool := make(chan chan RetryJob, config.MaxWorkers)\n\treturn &Dispatcher{\n\t\tWorkerPool: pool,\n\t\tConfig: config,\n\t\thandler: handler,\n\t\twaiter: waiter,\n\t}\n}", "func New(min, max int) *Dice {\n\tif max < min {\n\t\tpanic(\"max must equal or greater than min\")\n\t}\n\treturn &Dice{\n\t\tm: &sync.Mutex{},\n\t\tr: newNowRand(),\n\t\tmin: min,\n\t\tmax: max,\n\t}\n}", "func newTyphaAutoscaler(client client.Client, options ...typhaAutoscalerOption) *typhaAutoscaler {\n\tta := new(typhaAutoscaler)\n\tta.client = client\n\tta.syncPeriod = defaultTyphaAutoscalerSyncPeriod\n\n\tfor _, option := range options {\n\t\toption(ta)\n\t}\n\treturn ta\n}", "func NewRate(base string, reference string, rate decimal.Decimal) Rate {\n\treturn Rate{\n\t\tbase: base,\n\t\treference: reference,\n\t\trate: rate,\n\t}\n}", "func New(propVal int64) Client {\n\treturn &subs{Prop: propVal}\n}", "func New(c *Config) (Poller, error) {\n\tcfg := c.withDefaults()\n\n\tkq, err := KqueueCreate(&KqueueConfig{\n\t\tOnWaitError: cfg.OnWaitError,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn poller{kq}, nil\n}", "func NewRateLimiter(redisURL string, prefix string, dynamicConfig bool, dynamicConfigCacheTTL time.Duration) RateLimiter {\n\treturn &redisRateLimiter{\n\t\tpool: &redis.Pool{\n\t\t\tDial: func() (redis.Conn, error) {\n\t\t\t\treturn redis.DialURL(redisURL)\n\t\t\t},\n\t\t\tTestOnBorrow: func(c redis.Conn, _ time.Time) error {\n\t\t\t\t_, err := c.Do(\"PING\")\n\t\t\t\treturn err\n\t\t\t},\n\t\t\tMaxIdle: redisRateLimiterPoolMaxIdle,\n\t\t\tMaxActive: redisRateLimiterPoolMaxActive,\n\t\t\tIdleTimeout: redisRateLimiterPoolIdleTimeout,\n\t\t\tWait: true,\n\t\t},\n\t\tprefix: prefix,\n\n\t\tdynamicConfig: dynamicConfig,\n\t\tdynamicConfigCacheTTL: dynamicConfigCacheTTL,\n\t}\n}", "func New(max time.Duration, interval time.Duration) *Backoff {\n\tif max < 0 || interval < 0 {\n\t\tpanic(\"backoff: max or interval is negative\")\n\t}\n\n\tb := &Backoff{\n\t\tmaxDuration: max,\n\t\tinterval: interval,\n\t}\n\tb.setup()\n\treturn b\n}", "func newRuleHandler(d discovery.Discoverer, cfg RunConfig) discovery.RuleHandler {\n\trh := &ruleHandler{\n\t\td: d.(*discoverer),\n\t\tph: cfg.Handler,\n\t\tdaemon: cfg.Daemon,\n\t\tkubeClient: cfg.KubeClient,\n\t\tlister: cfg.Lister,\n\t\trulesCount: gm.GetOrRegisterGauge(\"discovery.rules.count\", gm.DefaultRegistry),\n\t}\n\tcount := int64(len(rh.d.delegates))\n\trh.rulesCount.Update(count)\n\treturn rh\n}", "func New() ResourceWatcher {\n\treturn ResourceWatcher{\n\t\twatched: make(map[types.NamespacedName][]types.NamespacedName),\n\t}\n}", "func New(qps int64) *limiter {\n\tif qps <= 0 {\n\t\treturn nil\n\t}\n\n\trl := &limiter{\n\t\tqps: qps,\n\t}\n\trl.current = make(map[string]int64, 0)\n\n\t// launch a goroutine to reset the counter every second\n\tgo rl.reset()\n\n\treturn rl\n}", "func newFreqCount() freqCount {\n\treturn make(freqCount, 3000000)\n}", "func NewRules(nextReceiver events.DataEventReceiver, opts *options.RuleOptions) *rules.Rules {\n\treturn rules.NewRules(nextReceiver, opts)\n}", "func NewFsWatcher(root string, interval int) (*FsWatcher, error) {\r\n\r\n\tp := filepath.Clean(root)\r\n\r\n\tinfo, err := os.Stat(p)\r\n\tif err != nil || !info.IsDir() {\r\n\t\treturn nil, errors.New(\"The specified path either doesn't exists or leads to a file \" + p)\r\n\t}\r\n\r\n\ti := 500\r\n\tif interval > 0 {\r\n\t\ti = interval\r\n\t}\r\n\treturn &FsWatcher{\r\n\t\tmake(map[string]FsEventHandler),\r\n\t\tmake(map[string]string),\r\n\t\tp,\r\n\t\ti,\r\n\t\tfalse,\r\n\t\tfalse,\r\n\t}, nil\r\n}", "func NewDispatcher(alertsConfig *config.Alerter,\n\tstatusChan <-chan StatusUpdate) (*Dispatcher, error) {\n\tvar alerters []alerter.Alerter\n\n\tif alertsConfig.Email != nil {\n\t\tlog.Debugf(\"setting up email alerter ...\")\n\t\talerter, err := alerter.NewEmailAlerter(alertsConfig.Email)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"dispatcher: failed to initialize email alerter: %s\", err)\n\t\t}\n\t\talerters = append(alerters, alerter)\n\t}\n\n\talertHistory := make(map[string]time.Time)\n\tbaseURL := fmt.Sprintf(\"https://%s:%d\", alertsConfig.AdvertisedIP, alertsConfig.AdvertisedPort)\n\treturn &Dispatcher{statusChan, alerters, alertHistory, alertsConfig.ReminderDelay.Duration, baseURL}, nil\n}", "func NewWatcher(resource CustomResource, namespace string, handlers cache.ResourceEventHandlerFuncs, client rest.Interface) *ResourceWatcher {\n\treturn &ResourceWatcher{\n\t\tresource: resource,\n\t\tnamespace: namespace,\n\t\tresourceEventHandlers: handlers,\n\t\tclient: client,\n\t}\n}", "func (o ReservedInstanceOutput) AutoRenewPeriod() pulumi.IntOutput {\n\treturn o.ApplyT(func(v *ReservedInstance) pulumi.IntOutput { return v.AutoRenewPeriod }).(pulumi.IntOutput)\n}", "func NewRateLimitWatcher(c Client, logger log.Logger, threshold int) AroundFunctionCreator {\n\treturn &rateLimitWatcher{client: c, logger: logger, threshold: threshold}\n}", "func NewAutoscalersController(client kubernetes.Interface, eventRecorder record.EventRecorder, isLeaderFunc func() bool, dogCl autoscalers.DatadogClient) (*AutoscalersController, error) {\n\tvar err error\n\th := &AutoscalersController{\n\t\tclientSet: client,\n\t\tisLeaderFunc: isLeaderFunc, // only trigger GC and updateExternalMetrics by the Leader.\n\t\tHPAqueue: workqueue.NewNamedRateLimitingQueue(workqueue.DefaultItemBasedRateLimiter(), \"autoscalers\"),\n\t\tEventRecorder: eventRecorder,\n\t}\n\n\th.toStore.data = make(map[string]custommetrics.ExternalMetricValue)\n\n\tgcPeriodSeconds := config.Datadog.GetInt(\"hpa_watcher_gc_period\")\n\trefreshPeriod := config.Datadog.GetInt(\"external_metrics_provider.refresh_period\")\n\n\tif gcPeriodSeconds <= 0 || refreshPeriod <= 0 {\n\t\treturn nil, fmt.Errorf(\"tickers must be strictly positive in the AutoscalersController\"+\n\t\t\t\" [GC: %d s, Refresh: %d s]\", gcPeriodSeconds, refreshPeriod)\n\t}\n\n\th.poller = PollerConfig{\n\t\tgcPeriodSeconds: gcPeriodSeconds,\n\t\trefreshPeriod: refreshPeriod,\n\t}\n\n\t// Setup the client to process the Ref and metrics\n\th.hpaProc = autoscalers.NewProcessor(dogCl)\n\tdatadogHPAConfigMap := custommetrics.GetConfigmapName()\n\th.store, err = custommetrics.NewConfigMapStore(client, common.GetResourcesNamespace(), datadogHPAConfigMap)\n\tif err != nil {\n\t\tlog.Errorf(\"Could not instantiate the local store for the External Metrics %v\", err)\n\t\treturn nil, err\n\t}\n\treturn h, nil\n}", "func newKaosRules(c *KaosV1Client, namespace string) *kaosRules {\n\treturn &kaosRules{\n\t\tclient: c.RESTClient(),\n\t\tns: namespace,\n\t}\n}", "func New() *Subscriptions {\n\tsubscriptions := &Subscriptions{}\n\n\tcgtRepos := persistence.CommunitygoaltrackerRepos\n\tCreateAchiever := usecase.NewCreateAchieverUsecase(&cgtRepos.Achiever)\n\tsubscriptions.AfterUserCreated = NewAfterUserCreated(CreateAchiever)\n\n\treturn subscriptions\n}" ]
[ "0.64369714", "0.6415797", "0.62879425", "0.5383594", "0.52797127", "0.51585424", "0.5117648", "0.49690348", "0.4956778", "0.48977634", "0.48946106", "0.4839171", "0.48256236", "0.48152933", "0.47980905", "0.479752", "0.478508", "0.47703877", "0.4768817", "0.4767869", "0.4767482", "0.47662973", "0.47570905", "0.47505602", "0.4707021", "0.47038847", "0.46899113", "0.46593603", "0.46485135", "0.46415573", "0.46311116", "0.46291026", "0.4626871", "0.46117425", "0.45937568", "0.4581228", "0.45670092", "0.45640793", "0.45552686", "0.45339996", "0.45297587", "0.45208195", "0.45167774", "0.45052427", "0.4499404", "0.44984764", "0.4491575", "0.44851702", "0.44843313", "0.44787395", "0.4470779", "0.44683447", "0.4465318", "0.44547525", "0.4443844", "0.4431312", "0.44140905", "0.44043565", "0.4403399", "0.43919805", "0.43919805", "0.43691462", "0.43620148", "0.43610948", "0.43531716", "0.4350281", "0.43406528", "0.43338025", "0.4332532", "0.4323888", "0.43229845", "0.43215054", "0.43135056", "0.43058294", "0.43036258", "0.43034053", "0.42919764", "0.42891312", "0.42890105", "0.4283896", "0.42832816", "0.42806417", "0.42800823", "0.42743692", "0.42723018", "0.42707133", "0.4257695", "0.42552668", "0.42440286", "0.42421347", "0.4241673", "0.42413637", "0.42409432", "0.42401537", "0.42350975", "0.42336324", "0.42313936", "0.42309833", "0.42285466", "0.4227301" ]
0.6057423
3
Close closes all substructs of the AutoRoller.
func (r *AutoRoller) Close() error { err1 := r.recent.Close() err2 := r.modeHistory.Close() if err1 != nil { return err1 } if err2 != nil { return err2 } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (it *FlytrapACLChangeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AccessControlRoleAdminChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (manager *Manager) Close() {\n\tfor _, level := range manager.levels {\n\t\tfor _, dimension := range level.GetDimensions() {\n\t\t\tdimension.Close(false)\n\t\t}\n\t}\n}", "func (lsm *lsm) Close() {\n\tfor _, level := range lsm.levels {\n\t\tlevel.Close()\n\t}\n}", "func (it *AuditableRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MonsterAccessControlContractUpgradeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ConsortiumManagementVoteIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OnesplitauditOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MonsterOwnershipContractUpgradeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *PollVoteIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AuditableRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *UpgradeableAddedOwnerIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ERC20HecoManagerOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *RBACRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *RegistryOperatorContractUpgraderUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *RBACRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BaseAccessControlGroupRightsChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CrowdsaleRoyaltyCrowdsaleUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SingleAutoAddIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *TellorMesosphereRoleAdminChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CakevaultOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (lo *LuaObject) Close() {\n lo.L.Unref(lua.LUA_REGISTRYINDEX, lo.Ref)\n}", "func (it *UpgradeableRemovedOwnerIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AccessControlRoleGrantedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LoggerRollbackFinalisationIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *KeepRegistryOperatorContractUpgraderUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BaseAccessWalletRightsChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SmartchefOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderRunFinalizeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *UserableRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CrowdsaleFrameUsdUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SuperuserableRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *EthCrossChainChangeBookKeeperEventIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SingleAutoOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ZKOnacciApprovalForAllIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderLogIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *PlasmaFrameworkVaultRegisteredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MinterRoleMinterAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (*Root) Close() error { return nil }", "func (it *SuperuserableRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AccessIndexorRightsChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CrowdsaleMaxFramesUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OracleMgrOwnershipRenouncedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BREMRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MinterRoleMinterRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *TTFT20AddedOwnerIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ERC20HecoManagerBurnedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CrowdsaleOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SingleAutoDepositIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ContractAdminChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *UsersDataOnSetOrgUuidIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *DelegationControllerRoleGrantedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BREMFactoryRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OracleMgrOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BREMFactoryRoleAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MonsterOwnershipTransferIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *UserableRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OnesplitauditImplementationUpdatedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderRunStatusChangeIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderLogInt256Iterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CakevaultUnpausedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ERC20HecoManagerMintedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *GatekeeperRootAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *GameJamVoteCastIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ERC725OwnerChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *TellorMesosphereRoleGrantedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *GameJamWinnerDeclaredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CrTokenNewAdminIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *GameJamGameJamAdminAddedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *McapscontrollerOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (b *Balancer) Close() (err error) {\n\tfor _, b := range b.selector {\n\t\tif e := b.Close(); e != nil {\n\t\t\terr = e\n\t\t}\n\t}\n\treturn\n}", "func (it *BREMRoleRemovedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderRunCreateIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *RegistryApprovalForAllIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *MonsterOwnershipApprovalIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AuthContractTriggeredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (writer *FileLogWriter) Close() {\n\tfor l := writer.level; l <= _LEVEL_MAX; l++ {\n\t\twriter.files[l].close()\n\t\twriter.files[l] = nil\n\t}\n}", "func (it *GatekeeperOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ElvTradableOwnershipTransferredIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *AccessControlRoleRevokedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LvStreamRightsHolderLogUint256Iterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *ElvTradableApprovalForAllIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func Close() {\n\tVac.close()\n}", "func (it *LvStreamRightsHolderRunAccessIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *Erc1820RegistryManagerChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LoggerBatchRollbackIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *FlopperKickIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func(this*Window)Close()error{\n\n\n/*23:*/\n\n\n//line goacme.w:312\n\nfor _,v:=range this.files{\nv.Close()\n}\n\n\n\n/*:23*/\n\n\n\n/*31:*/\n\n\n//line goacme.w:396\n\nif this.next!=nil{\nthis.next.prev= this.prev\n}\nif this.prev!=nil{\nthis.prev.next= this.next\n}\nif fwin==this{\nfwin= this.next\n}\nif lwin==this{\nlwin= this.prev\n}\n\n\n\n/*:31*/\n\n\n//line goacme.w:190\n\nreturn nil\n}", "func (it *ContractsNewVoterIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *XStakingOwnerChangedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BaseAccessControlGroupUnauthorizedOperationIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OwnableOwnershipRenouncedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *OwnableOwnershipRenouncedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CakevaultUnpauseIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *LoggerDepositSubTreeReadyIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *CraftingIApprovalForAllIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *TTFT20RemovedOwnerIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *BaseAccessControlGroupManagerAccessGrantedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *DistributorRoleGrantedIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}", "func (it *SmartchefDepositIterator) Close() error {\n\tit.sub.Unsubscribe()\n\treturn nil\n}" ]
[ "0.5826136", "0.57758576", "0.572692", "0.5679884", "0.56690097", "0.5659184", "0.5650129", "0.56471294", "0.5644417", "0.56347626", "0.5625366", "0.5616441", "0.56142664", "0.55810994", "0.55592614", "0.5556965", "0.55303437", "0.5521246", "0.5513769", "0.5507827", "0.55045825", "0.55028987", "0.55016863", "0.5498721", "0.5495024", "0.5488974", "0.54861176", "0.54817206", "0.5481073", "0.54758704", "0.54742527", "0.54741925", "0.54728174", "0.546514", "0.5462742", "0.545839", "0.54534817", "0.54516864", "0.5447794", "0.544741", "0.54369617", "0.5423421", "0.5421388", "0.54168904", "0.5415633", "0.54137564", "0.541311", "0.54115075", "0.5411232", "0.54032856", "0.5403008", "0.54013586", "0.5394213", "0.53929085", "0.5392027", "0.539067", "0.53884494", "0.53863484", "0.53825647", "0.5363712", "0.53632504", "0.53631186", "0.536214", "0.5350184", "0.5349822", "0.534581", "0.5343595", "0.5340632", "0.5340549", "0.5339682", "0.5338875", "0.5335007", "0.53302366", "0.5328761", "0.532165", "0.53173804", "0.53115445", "0.53114223", "0.5308893", "0.5305807", "0.5304813", "0.53026867", "0.52995", "0.5294209", "0.529374", "0.52934515", "0.5291937", "0.52918947", "0.5291845", "0.529132", "0.52864563", "0.52864516", "0.52864516", "0.528554", "0.5281962", "0.5280227", "0.5279016", "0.5278553", "0.52783", "0.5275996" ]
0.6832475
0
Get returns the current status information.
func (c *autoRollStatusCache) Get(includeError bool) *AutoRollStatus { c.mtx.RLock() defer c.mtx.RUnlock() recent := make([]*autoroll.AutoRollIssue, 0, len(c.recent)) for _, r := range c.recent { recent = append(recent, r.Copy()) } validModes := make([]string, len(autoroll_modes.VALID_MODES)) copy(validModes, autoroll_modes.VALID_MODES) s := &AutoRollStatus{ GerritUrl: c.gerritUrl, LastRollRev: c.lastRollRev, Mode: c.mode.Copy(), Recent: recent, Status: c.status, ValidModes: validModes, } if c.currentRoll != nil { s.CurrentRoll = c.currentRoll.Copy() } if c.lastRoll != nil { s.LastRoll = c.lastRoll.Copy() } if includeError && c.lastError != "" { s.Error = c.lastError } return s }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) StatusGet() (status *Status, err error) {\n\terr = client.get(client.buildPath(\"/status/\"), &status)\n\treturn\n}", "func (s *statusClient) Get() (*Status, error) {\n\tlogger := s.logger().WithField(\"method\", \"Get\")\n\tvar status *Status\n\tif err := s.c.getInto(endpointGetStatus, &status); err != nil {\n\t\tlogger.Debug(err)\n\t\treturn nil, err\n\t}\n\treturn status, nil\n}", "func (app *Application) GetStatus() *Status {\n if app.status == nil {\n app.status = app.Get(\"status\").(*Status)\n }\n\n return app.status\n}", "func GetStatus(w http.ResponseWriter, r *http.Request, argv map[string]string) error {\n\tresTemplate := `{\"MetricsService\":\"STARTED\",\"Implementation-Version\":\"%s\",\"MohawkVersion\":\"%s\",\"MohawkStorage\":\"%s\"}`\n\tres := fmt.Sprintf(resTemplate, defaultAPI, VER, BackendName)\n\n\tfmt.Fprintln(w, res)\n\treturn nil\n}", "func (c *Client) Status() (*Status, error) {\n\tstatus := &Status{}\n\terr := c.Get().UsePath(\"/status\").Do().Unmarshal(status)\n\treturn status, err\n}", "func GetStatus(c *gin.Context) {\n\tc.String(http.StatusOK, \"OK\")\n}", "func GetStatus() (string, error) {\n\tbody, err := getRequest(GetURL()+getStatusPath, \"\", false)\n\tif err != nil {\n\t\treturn \"Connection failed\", err\n\t}\n\tif string(body) != `\"any-compute\"` {\n\t\terr = util.ErrServerUnavailable\n\t}\n\treturn string(body), err\n}", "func (f *FortiWebClient) GetStatus() string {\n\n\tclient := &http.Client{}\n\n\treq, err := http.NewRequest(\"GET\", strings.Join([]string{f.URL, \"api/v1.0/System/Status/Status\"}, \"\"), nil)\n\treq.Header.Add(\"Authorization\", encodeBase64(f.Username, f.Password))\n\tresponse, error := client.Do(req)\n\n\tif error != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t\treturn strings.Join([]string{\"Error: The HTTP request failed with error \", error.Error()}, \"\")\n\t}\n\n\tdefer response.Body.Close()\n\tbody, _ := ioutil.ReadAll(response.Body)\n\n\treturn string(body[:])\n}", "func (b *Base) GetStatus() string {\n\treturn `\n\tGoCryptoTrader Service: Online\n\tService Started: ` + ServiceStarted.String()\n}", "func (m *ThreatAssessmentRequest) GetStatus()(*ThreatAssessmentStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*ThreatAssessmentStatus)\n }\n return nil\n}", "func (m *ProgramControl) GetStatus()(*string) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (r *StatusREST) Get(ctx context.Context, name string, options *metav1.GetOptions) (runtime.Object, error) {\n\treturn r.store.Get(ctx, name, options)\n}", "func (w *walletStatus) Get() models.Status {\n\treturn w.text\n}", "func (m *IndustryDataRunActivity) GetStatus()(*IndustryDataActivityStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*IndustryDataActivityStatus)\n }\n return nil\n}", "func GetStatus() (string, error) {\n\n\tvar url bytes.Buffer\n\taddr := GetURL()\n\tif addr == \"\" {\n\t\treturn \"\", util.MAPIADDRNotSet\n\t}\n\turl.WriteString(addr + \"/latest/meta-data/instance-id\")\n\tresp, err := http.Get(url.String())\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn string(body[:]), err\n}", "func (m *SecurityActionState) GetStatus()(*OperationStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*OperationStatus)\n }\n return nil\n}", "func (s *StatusController) GetStatus() {\n\tlogs.Debug(\"GetStatus\")\n//\tvar hPatients []models.HospitalPatientInfo\n\tvar data models.DataTable\n\t_ = checkAccount(s.Ctx)\n\tdata.Data,_ = Calculatepara()\n//\tlogs.Debug(\"Status Patients:\", data.Data)\n\ts.Data[\"json\"] = &data\n\ts.ServeJSON()\n}", "func (c Client) GetStatus() (*StatusResponse, error) {\n\turl := fmt.Sprint(DefaultBaseURL, APIVersion, \"/status\")\n\tfmt.Println(\"URL:>\", url)\n\ts := new(StatusResponse)\n\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\treq.Header.Set(\"token\", c.Token)\n\n\tclient := &http.Client{}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\treturn s, err\n\t}\n\tif resp.StatusCode != http.StatusOK {\n\t\tlog.Fatal(resp.Status)\n\t\treturn s, err\n\t}\n\tdefer resp.Body.Close() //resp.Body.Close() will run when we're finished.\n\n\t// Copy the body to Stdout\n\t//_, err = io.Copy(os.Stdout, resp.Body)\n\n\terr = json.NewDecoder(resp.Body).Decode(s)\n\tif err != nil {\n\t\tfmt.Println(\"Error parsing status response\")\n\t\tlog.Fatal(err)\n\t\treturn s, err\n\t}\n\t//fmt.Println(\"Current Status is: \")\n\t//fmt.Println(s.SystemStatus[0].Status)\n\treturn s, nil\n}", "func (o *SmartstackBackend) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func GetStatus() string {\n\tif lastStatus == nil {\n\t\tlastStatus = &status{\n\t\t\tstate: \"unknown\",\n\t\t\treason: \"no known last status\",\n\t\t}\n\t}\n\treturn lastStatus.state\n}", "func (s *Service) GetStatus(ctx context.Context, req *request.Status) (*response.Status, error) {\n\treturn &response.Status{}, nil\n}", "func (s *Source) GetStatus() interface{} {\n\treturn s.Status\n}", "func (m *ConnectorStatusDetails) GetStatus()(*ConnectorHealthState) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*ConnectorHealthState)\n }\n return nil\n}", "func (m *ThreatAssessmentRequest) GetStatus()(*ThreatAssessmentStatus) {\n return m.status\n}", "func GetStatus(cfg *Config) Status {\n\tnow := time.Now().Unix()\n\n\ts := Status{\n\t\tPID: os.Getpid(),\n\t\tService: \"list-service\",\n\t}\n\n\ts.Status = \"ok\"\n\ts.Version = Version()\n\ts.CPUs = runtime.NumCPU()\n\ts.GoVers = runtime.Version()\n\ts.TimeStamp = now\n\ts.UpTime = now - InstanceStartTime\n\ts.LogLevel = log.GetLevel()\n\n\tif host, err := os.Hostname(); err == nil {\n\t\ts.Hostname = host\n\t}\n\n\treturn s\n}", "func (c *Client) getStatus() Status {\n\n\treturn c.status\n}", "func (app *ApplicationStatus) Get() int {\n\tapp.Lock()\n\tdefer app.Unlock()\n\n\treturn app.code\n}", "func (m *AccessPackageAssignment) GetStatus()(*string) {\n return m.status\n}", "func GetStatus(c echo.Context) error {\n\tserver := c.(*Server)\n\n\tcount, err := server.Repository.Games()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.JSON(http.StatusOK, Status{\n\t\tName: server.Name,\n\t\tGames: count,\n\t})\n}", "func (*StatusResource) Get() restful.Exchanger { return &StatusHandler{} }", "func (m *LongRunningOperation) GetStatus()(*LongRunningOperationStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*LongRunningOperationStatus)\n }\n return nil\n}", "func (m *SynchronizationJob) GetStatus()(SynchronizationStatusable) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(SynchronizationStatusable)\n }\n return nil\n}", "func (s *StatusController) Get() {\n\ts.Data[\"IsSconfig\"] = true\n\ts.TplName = \"patient_status.html\"\n\tflag := checkAccount(s.Ctx)\n\ts.Data[\"ISLogin\"] = flag\n//\ts.Data[\"Hospital\"] = Hospital\n\tif !flag {\n\t\ts.Redirect(\"/login\", 302)\n\t\treturn\n\t}\n\t_,_ = Calculatepara()\n}", "func (a *Client) GetStatus(ctx context.Context, params *GetStatusParams) (*GetStatusOK, error) {\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getStatus\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/status/{id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetStatusReader{formats: a.formats},\n\t\tAuthInfo: a.authInfo,\n\t\tContext: ctx,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetStatusOK), nil\n\n}", "func GetStatus(w http.ResponseWriter, r *http.Request) {\n\tmiddleware.EnableCors(&w)\n\n\tstatus := core.GetStatus()\n\tresponse := webStatusResponse{\n\t\tOnline: status.Online,\n\t\tViewerCount: status.ViewerCount,\n\t\tLastConnectTime: status.LastConnectTime,\n\t\tLastDisconnectTime: status.LastDisconnectTime,\n\t\tVersionNumber: status.VersionNumber,\n\t\tStreamTitle: status.StreamTitle,\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tif err := json.NewEncoder(w).Encode(response); err != nil {\n\t\tInternalErrorHandler(w, err)\n\t}\n}", "func (s *Status) GetStatus(rw http.ResponseWriter, r *http.Request) {\n\tstatus, err := s.client.Status()\n\tif err != nil {\n\t\ts.l.Printf(\"failed to query status: %s\\n\", err)\n\t}\n\n\tvalidatorSet, err := s.client.ValidatorSet(status.SyncInfo.LatestBlockHeight)\n\tif err != nil {\n\t\ts.l.Printf(\"failed to query validators et: %s\\n\", err)\n\t}\n\n\tblock, err := s.client.Block(status.SyncInfo.LatestBlockHeight)\n\tif err != nil {\n\t\ts.l.Printf(\"failed to query block: %s\\n\", err)\n\t}\n\n\tprevBlock, err := s.client.Block(status.SyncInfo.LatestBlockHeight - 1)\n\tif err != nil {\n\t\ts.l.Printf(\"failed to query previous block: %s\\n\", err)\n\t}\n\n\tvar blockTime float64\n\tif block.Block != nil {\n\t\tblockTime = block.Block.Time.UTC().Sub(prevBlock.Block.Time.UTC()).Seconds()\n\t} else {\n\t\ts.l.Printf(\"block.Block is nil\")\n\t}\n\n\tresult := &models.Status{\n\t\tChainID: status.NodeInfo.Network,\n\t\tBlockTime: blockTime,\n\t\tLatestBlockHeight: status.SyncInfo.LatestBlockHeight,\n\t\tTotalValidatorNum: len(validatorSet.Validators),\n\t\tTimestamp: status.SyncInfo.LatestBlockTime,\n\t}\n\n\tutils.Respond(rw, result)\n\treturn\n}", "func (m *LongRunningOperation) GetStatusDetail()(*string) {\n val, err := m.GetBackingStore().Get(\"statusDetail\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (m *TeamsAsyncOperation) GetStatus()(*TeamsAsyncOperationStatus) {\n return m.status\n}", "func GetStatus(c *db.Cocoon, _ []byte) (interface{}, error) {\n\tvar err error\n\n\tstatuses, err := c.QueryBuildStatusesWithMemcache()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tagentStatuses, err := c.QueryAgentStatuses()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &GetStatusResult{\n\t\tStatuses: statuses,\n\t\tAgentStatuses: agentStatuses,\n\t}, nil\n}", "func (client *Client) GetStatus() (*Response, error) {\n\tpath := \"/status\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\treq, err := http.NewRequest(\"GET\", uri, bytes.NewBuffer([]byte(\"\")))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := client.performRequest(req, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, err\n}", "func (sc *StreamClient) GetStatus() int {\n\treturn sc.status\n}", "func (d *Docker) GetStatus(ctx context.Context, name string) error {\n\treturn nil\n}", "func GetStatus() (status AuditStatus, err error) {\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn AuditStatus{}, errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tkstatus, err := client.GetStatus()\n\tif err != nil {\n\t\treturn AuditStatus{}, errors.Wrap(err, \"Failed to get audit status\")\n\t}\n\tstatus.Enabled = kstatus.Enabled\n\tstatus.Failure = kstatus.Failure\n\tstatus.PID = kstatus.PID\n\tstatus.RateLimit = kstatus.RateLimit\n\tstatus.BacklogLimit = kstatus.BacklogLimit\n\tstatus.Lost = kstatus.Lost\n\tstatus.Backlog = kstatus.Backlog\n\tstatus.BacklogWaitTime = kstatus.BacklogWaitTime\n\treturn status, nil\n}", "func GetStatus(key string) Status {\n\n\tvar netClient = &http.Client{\n\t\tTimeout: time.Second * 10,\n\t}\n\n\treq, err := http.NewRequest(\"GET\", \"https://oslobysykkel.no/api/v1/status\", nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Add(\"Client-Identifier\", key)\n\tresp, err := netClient.Do(req)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tvar status Status\n\terr = json.Unmarshal(body, &status)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn status\n\n}", "func GetStatus() (*Status, error) {\n\tvar status *Status\n\tif err := sendRequest(\"/api/status.json\", &status); err != nil {\n\t\treturn nil, err\n\t}\n\treturn status, nil\n}", "func (m *ApplicationSignInDetailedSummary) GetStatus()(SignInStatusable) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(SignInStatusable)\n }\n return nil\n}", "func (a *Action) GetStatus(action *models.Action) (*models.ActionStatus, error) {\n\traw, err := a.GetClient().Get(fmt.Sprintf(STREAM_LAST_UPDATE, action.GetDevice().ID, action.Name), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres := &models.ActionStatus{}\n\terr = a.GetClient().FromJSON(raw, res)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres.SetAction(action)\n\treturn res, nil\n}", "func (j *Juju) GetStatus() (string, error) {\n\ttmp := \"JUJU_DATA=\" + JujuDataPrefix + j.Name\n\tcmd := exec.Command(\"juju\", \"status\")\n\tcmd.Env = append(os.Environ(), tmp)\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"GetStatus error: %v: %s\", err, err.(*exec.ExitError).Stderr)\n\t}\n\tlog.Debug(string(out))\n\treturn string(out), nil\n}", "func (m *RetentionEventStatus) GetStatus()(*EventStatusType) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*EventStatusType)\n }\n return nil\n}", "func (c Cas) GetStatus(scaleID uint32) (Status, error) {\n\tvar status Status\n\n\topcode := [2]byte{'R', 'N'}\n\tbuf := encodePacket(scaleID, opcode, []byte{})\n\n\tif _, err := c.conn.Write(buf); err != nil {\n\t\treturn status, err\n\t}\n\n\ttmp := make([]byte, 512)\n\n\tif _, err := c.conn.Read(tmp); err != nil {\n\t\treturn status, err\n\t}\n\n\tif tmp[0] != 'W' || tmp[1] != 'N' {\n\t\treturn status, fmt.Errorf(\"GetStatus %s %x\", string(tmp[0:2]), tmp)\n\t}\n\n\treturn status, nil\n}", "func (v *Kounta) GetStatus(token string, company string) error {\n\tclient := &http.Client{}\n\tclient.CheckRedirect = checkRedirectFunc\n\n\tu, _ := url.ParseRequestURI(baseURL)\n\tu.Path = fmt.Sprintf(companyStatus, company)\n\turlStr := fmt.Sprintf(\"%v\", u)\n\n\tfmt.Println(urlStr)\n\n\tr, err := http.NewRequest(\"GET\", urlStr, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.Header = http.Header(make(map[string][]string))\n\tr.Header.Set(\"Accept\", \"application/json\")\n\tr.Header.Set(\"Authorization\", \"Bearer \"+token)\n\n\tres, err := client.Do(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trawResBody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif res.StatusCode == 200 {\n\n\t\tfmt.Println(string(rawResBody))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\tfmt.Println(string(rawResBody))\n\treturn fmt.Errorf(\"Failed to get Kounta Sale %s\", res.Status)\n\n}", "func (puppetDb *PuppetDb) GetStatus() (*operations.GetStatusOK, error) {\n\tstringToken, err := puppetDb.Token.Read()\n\tif err != nil {\n\t\tlog.Debug(err.Error())\n\t}\n\n\tclient, err := puppetDb.Client.GetClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tapiKeyHeaderAuth := httptransport.APIKeyAuth(\"X-Authentication\", \"header\", stringToken)\n\tgetStatusParameters := operations.NewGetStatusParamsWithContext(context.Background())\n\treturn client.Operations.GetStatus(getStatusParameters, apiKeyHeaderAuth)\n}", "func (c *Client) GetStatus(projectID string, jobID string) (*JobStatus, error) {\n\n\tjob := &JobStatus{}\n\tpath := fmt.Sprintf(DataProcessingStatus, url.QueryEscape(projectID), url.QueryEscape(jobID))\n\treturn job, c.OVH.Get(path, job)\n}", "func (self *client) GetStatus() {\n\n}", "func (server *FlexibleServer) GetStatus() genruntime.ConvertibleStatus {\n\treturn &server.Status\n}", "func (s *Service) GetStatus(ctx context.Context, req *request.Status) (*response.Status, error) {\n\tres := &response.Status{Callback: req.Callback}\n\tif req.Callback == \"I don't like launch pad\" {\n\t\treturn res, errors.New(\"launch pad is the best and you know it\")\n\t}\n\treturn res, nil\n}", "func (c *StatusRClient) Get(ctx context.Context, id int) (*StatusR, error) {\n\treturn c.Query().Where(statusr.ID(id)).Only(ctx)\n}", "func (m *BrowserSiteList) GetStatus()(*BrowserSiteListStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*BrowserSiteListStatus)\n }\n return nil\n}", "func (site *Site) GetStatus() genruntime.ConvertibleStatus {\n\treturn &site.Status\n}", "func (sr *StatusResource) Get(*restful.RouteMap, http.ResponseWriter, *http.Request, httprouter.Params) restful.Exchanger {\n\treturn &StatusHandler{\n\t\tAutoResolver: sr.context.AutoResolver,\n\t\tResolveFilter: sr.context.ResolveFilter,\n\t}\n}", "func (o *StorageSpaceAllOf) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (cl *Client) Get_New_Status(do_read_lock bool) (s string, err error) {\n\tif do_read_lock {\n\t\tread_lock, xerr := cl.RLockNamed(\"Get_New_Status\")\n\t\tif xerr != nil {\n\t\t\terr = xerr\n\t\t\treturn\n\t\t}\n\t\tdefer cl.RUnlockNamed(read_lock)\n\t}\n\ts = cl.Status\n\treturn\n}", "func (c *CryptohomeBinary) GetStatusString(ctx context.Context) (string, error) {\n\tout, err := c.call(ctx, \"--action=status\")\n\treturn string(out), err\n}", "func (m *ScheduleItem) GetStatus()(*FreeBusyStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*FreeBusyStatus)\n }\n return nil\n}", "func (e *entry) getStatus() int {\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\n\treturn e.status\n}", "func (m *SchemaExtension) GetStatus()(*string) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (s *Session) getStatus() (map[string]interface{}, error) {\n\tvar dummyMap = make(map[string]interface{})\n\terr := s.checkTank()\n\tif err != nil {\n\t\treturn dummyMap, err\n\t}\n\terr = s.checkName()\n\tif err != nil {\n\t\treturn dummyMap, err\n\t}\n\n\tresp, err := netClient.Get(fmt.Sprintf(\"%v/status?session=%v\", s.Tank.Url, s.Name))\n\tif err != nil {\n\t\terr = fmt.Errorf(\"http.GET failed: %w\", err)\n\t\ts.Status = \"disconnect\"\n\t\treturn dummyMap, err\n\t}\n\tdefer resp.Body.Close()\n\n\trespBody, err := checkResponseCode(*resp)\n\tif err != nil {\n\t\ts.setFailed([]string{err.Error()})\n\t\treturn dummyMap, err\n\t}\n\n\tvar respJson map[string]interface{}\n\terr = json.Unmarshal(respBody, &respJson)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"fail to unmarshal get status response: %w\", err)\n\t\treturn dummyMap, err\n\t}\n\n\tswitch stage := respJson[\"current_stage\"].(type) {\n\tcase string:\n\t\ts.Stage = stage\n\t}\n\tswitch status := respJson[\"status\"].(type) {\n\tcase string:\n\t\ts.Status = status\n\t}\n\treturn respJson, nil\n}", "func (t *Service) GetStatus() *duckv1.Status {\n\treturn &t.Status.Status\n}", "func (t *Task) GetStatus() int {\n\treturn t.Status\n}", "func (clt *client) Status() Status {\n\tclt.statusLock.Lock()\n\tstatus := clt.status\n\tclt.statusLock.Unlock()\n\treturn status\n}", "func (m *WorkbookOperation) GetStatus()(*WorkbookOperationStatus) {\n return m.status\n}", "func (s *ApiService) GetStatus(ctx context.Context) (ordf.ImplResponse, error) {\n\tstatus := ordf.Status{\n\t\tStatus: \"UP\",\n\t}\n\n\t// check NowPayments API\n\tnowPaymentsStatus, err := s.nowPaymentsService.Status(ctx)\n\tif err != nil {\n\t\tstatus.Status = fmt.Sprintf(\"Failed to check NowPayments status: %+v\", err)\n\t} else if nowPaymentsStatus != \"OK\" {\n\t\tstatus.Status = fmt.Sprintf(\"NowPayments is down: %s\", nowPaymentsStatus)\n\t}\n\n\treturn ordf.Response(200, status), nil\n}", "func (wAPI WalletAPI) Status() (*Status, error) {\n\tvar stat Status\n\t_, raw, err := wAPI.sendRequest(\n\t\t\"GET\",\n\t\twAPI.Host+\":\"+wAPI.Port+\"/status\",\n\t\t\"\",\n\t)\n\n\tif err == nil {\n\t\terr = json.Unmarshal(*raw, &stat)\n\t}\n\n\treturn &stat, err\n}", "func (h *HealthCheck) GetStatus() int {\n\treturn h.status.getStatus()\n}", "func (c *Client) Status() string {\n\n\treturn c.status.String()\n}", "func (m *SimulationAutomationRun) GetStatus()(*SimulationAutomationRunStatus) {\n val, err := m.GetBackingStore().Get(\"status\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*SimulationAutomationRunStatus)\n }\n return nil\n}", "func (candidate *Candidate) GetStatus() byte {\n\tcandidate.lock.RLock()\n\tdefer candidate.lock.RUnlock()\n\n\treturn candidate.Status\n}", "func (c *Client) Status() error {\n\tclient, err := c.client(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trel, err := url.Parse(\"status\")\n\tif err != nil {\n\t\t// This indicates a programming error.\n\t\tpanic(err)\n\t}\n\n\tresp, err := client.Get(c.baseURL.ResolveReference(rel).String())\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusNoContent {\n\t\treturn ErrInvalidServiceBehavior\n\t}\n\treturn nil\n}", "func (machine *VirtualMachine) GetStatus() genruntime.ConvertibleStatus {\n\treturn &machine.Status\n}", "func (machine *VirtualMachine) GetStatus() genruntime.ConvertibleStatus {\n\treturn &machine.Status\n}", "func (t *TaskBox[T, U, C, CT, TF]) GetStatus() int32 {\n\treturn t.status.Load()\n}", "func (r *Random) GetStatus() interface{} {\n\treturn SourceStatusRunning\n}", "func GetStatus() (*status.S, error) {\n\tsr, e := StatusRetriever(amqpURI)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\tr, e := sr.ReceiveProduce(nil)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\ts := status.S{}\n\tif e = json.Unmarshal(r, &s); e != nil {\n\t\treturn nil, e\n\t}\n\treturn &s, nil\n}", "func (dateService) Status(ctx context.Context) (string, error) {\n\treturn \"ok\", nil\n}", "func (k *NatssSource) GetStatus() *duckv1.Status {\n\treturn &k.Status.Status\n}", "func (p *Init) Status(ctx context.Context) (string, error) {\n\tp.mu.Lock()\n\tdefer p.mu.Unlock()\n\n\treturn p.initState.State(ctx)\n}", "func (s *AccessPolicy) GetStatus() resv1.Status {\n\treturn &s.Status\n}", "func (client *Client) Status() *Status {\n\treturn client.status\n}", "func (enterprise *RedisEnterprise) GetStatus() genruntime.ConvertibleStatus {\n\treturn &enterprise.Status\n}", "func (a *Agent) Status() string {\n\ta.mutex.RLock()\n\tdefer a.mutex.RUnlock()\n\n\treturn a.status\n}", "func (p *Porthole) GetStatus() *status.Status {\n\t// copy current status values into a new Status instance\n\treturn &status.Status{\n\t\tGitCommit: p.coordinator.Status.GitCommit,\n\t\tLastRequest: p.coordinator.Status.LastRequest,\n\t\tLastFetch: p.coordinator.Status.LastFetch,\n\t\tLatestAdditions: p.coordinator.Status.LatestAdditions,\n\t}\n}", "func (audit Audit) GetStatus() string {\n\treturn audit.status\n}", "func (o *VirtualizationIweVirtualMachine) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (c *RuntimeSecurityClient) GetStatus() (*api.Status, error) {\n\tapiClient := api.NewSecurityModuleClient(c.conn)\n\treturn apiClient.GetStatus(context.Background(), &api.GetStatusParams{})\n}", "func (this ResponseDetailsViewV5) GetStatus() int { return this.Status }", "func (r *RpmOstreeClient) GetStatus() (string, error) {\n\toutput, err := RunGetOut(\"rpm-ostree\", \"status\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(output), nil\n}", "func (c *Client) GetStatus() (*StatusRequest, error) {\n\tvar sq StatusRequest\n\tresp, err := c.client.Get(c.formURI(\"api/v1/status\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tif err = json.NewDecoder(resp.Body).Decode(&sq); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &sq, nil\n}", "func (c *Client) Status() *StatusInfo {\n\tc.slock.Lock()\n\tdefer c.slock.Unlock()\n\n\tsi := &StatusInfo{\n\t\tWriters: make(map[string]*WriterStatus),\n\t\tReaders: make(map[string]*ReaderStatus),\n\t\tRPCs: make(map[time.Duration]MethodList),\n\t}\n\n\tfor name, w := range c.sWriters {\n\t\tsi.Writers[name] = w.status()\n\t}\n\n\tfor name, r := range c.sReaders {\n\t\tsi.Readers[name] = r.status()\n\t}\n\n\tfor _, c := range c.sMethods {\n\t\tsi.RPCs[c.d] = c.retrieve()\n\t}\n\n\treturn si\n}", "func (s Strategy) GetStatus(coralName string, foreign string) string {\n\treturn s.Map.Entities[coralName].Status[foreign]\n}", "func (n *RocketmqChannel) GetStatus() *duckv1.Status {\n\treturn &n.Status.Status\n}", "func getStatus(n Status) string {\n\treturn status[n]\n}" ]
[ "0.8026335", "0.7606744", "0.73826325", "0.7238587", "0.7139378", "0.70920545", "0.70797706", "0.70747554", "0.7072484", "0.702497", "0.7019969", "0.69774234", "0.6936601", "0.690858", "0.69042635", "0.6895668", "0.6861055", "0.6858933", "0.6851915", "0.6842642", "0.68219066", "0.6806521", "0.67994153", "0.6799081", "0.6791105", "0.6779214", "0.6735871", "0.67247343", "0.6716745", "0.6711738", "0.6705159", "0.6702835", "0.67003286", "0.6697672", "0.66975117", "0.6694961", "0.66911185", "0.6688484", "0.6687688", "0.66863716", "0.66752845", "0.6668017", "0.664924", "0.664681", "0.6646019", "0.66214126", "0.6618933", "0.66132903", "0.6585878", "0.6583966", "0.658319", "0.6582561", "0.658153", "0.6577774", "0.6571326", "0.65588826", "0.6554037", "0.6536579", "0.65362906", "0.65112233", "0.6506463", "0.6502275", "0.6500366", "0.64917946", "0.64884585", "0.6477195", "0.64769953", "0.6468776", "0.6467511", "0.64605546", "0.6460165", "0.64569753", "0.6456296", "0.6452113", "0.6444243", "0.6442782", "0.6436652", "0.64266294", "0.6425506", "0.6425506", "0.64151156", "0.64063394", "0.6403065", "0.6397244", "0.6394918", "0.63940305", "0.6389341", "0.63846266", "0.63793665", "0.63679874", "0.6363418", "0.6359462", "0.63594157", "0.63581216", "0.63576293", "0.6349637", "0.6348754", "0.6346555", "0.6344798", "0.6343329", "0.6339364" ]
0.0
-1
set sets the current status information.
func (c *autoRollStatusCache) set(s *AutoRollStatus) error { if !util.In(string(s.Status), VALID_STATUSES) { return fmt.Errorf("Invalid status: %s", s.Status) } if s.Status == STATUS_ERROR { if s.Error == "" { return fmt.Errorf("Cannot set error status without an error!") } } else if s.Error != "" { return fmt.Errorf("Cannot be in any status other than error when an error occurred.") } c.mtx.Lock() defer c.mtx.Unlock() recent := make([]*autoroll.AutoRollIssue, 0, len(s.Recent)) for _, r := range s.Recent { recent = append(recent, r.Copy()) } c.currentRoll = nil if s.CurrentRoll != nil { c.currentRoll = s.CurrentRoll.Copy() } c.lastRoll = nil if s.LastRoll != nil { c.lastRoll = s.LastRoll.Copy() } c.gerritUrl = s.GerritUrl c.lastRollRev = s.LastRollRev c.mode = s.Mode.Copy() c.recent = recent c.status = s.Status return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *Myself) setStatus(stat int) {\n\tm.mutex.Lock()\n\tm.status = stat\n\tm.mutex.Unlock()\n}", "func SetStatus(stat int) {\n\tmutex.Lock()\n\tstatus = stat\n\tmutex.Unlock()\n}", "func (clt *client) setStatus(newStatus Status) {\n\tclt.statusLock.Lock()\n\tclt.status = newStatus\n\tclt.statusLock.Unlock()\n}", "func (m *TeamsAsyncOperation) SetStatus(value *TeamsAsyncOperationStatus)() {\n m.status = value\n}", "func (m *ProgramControl) SetStatus(value *string)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *SecurityActionState) SetStatus(value *OperationStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (e *entry) setStatus(status int) {\n\te.mu.Lock()\n\tdefer e.mu.Unlock()\n\te.status = status\n}", "func (m *AccessPackageAssignment) SetStatus(value *string)() {\n m.status = value\n}", "func (m *IndustryDataRunActivity) SetStatus(value *IndustryDataActivityStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (t *TaskBox[T, U, C, CT, TF]) SetStatus(s int32) {\n\tt.status.Store(s)\n}", "func (m *LongRunningOperation) SetStatus(value *LongRunningOperationStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *ThreatAssessmentRequest) SetStatus(value *ThreatAssessmentStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *SynchronizationJob) SetStatus(value SynchronizationStatusable)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *ThreatAssessmentRequest) SetStatus(value *ThreatAssessmentStatus)() {\n m.status = value\n}", "func (m *ConnectorStatusDetails) SetStatus(value *ConnectorHealthState)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *WorkbookOperation) SetStatus(value *WorkbookOperationStatus)() {\n m.status = value\n}", "func (m *SimulationAutomationRun) SetStatus(value *SimulationAutomationRunStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *Machine) SetStatus(status params.Status, info string, data params.StatusData) error {\n\tvar result params.ErrorResults\n\targs := params.SetStatus{\n\t\tEntities: []params.EntityStatus{\n\t\t\t{Tag: m.tag.String(), Status: status, Info: info, Data: data},\n\t\t},\n\t}\n\terr := m.st.call(\"SetStatus\", args, &result)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn result.OneError()\n}", "func (me *TpubStatusInt) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (m *ScheduleItem) SetStatus(value *FreeBusyStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (self *Response) SetStatus(isfail bool, errormsg string) {\n\tself.isfail = isfail\n\tself.errormsg = errormsg\n}", "func (m *ApplicationSignInDetailedSummary) SetStatus(value SignInStatusable)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *RetentionEventStatus) SetStatus(value *EventStatusType)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (a *Action) SetStatus(status *models.ActionStatus) error {\n\taction := status.GetAction()\n\tif action == nil {\n\t\treturn errors.New(\"Action is missing, use Action.CreateStatus to initialize an ActionStatus\")\n\t}\n\t_, err := a.GetClient().Post(fmt.Sprintf(STREAM_LAST_UPDATE, action.GetDevice().ID, action.Name), status, nil)\n\treturn err\n}", "func (ts *TechStoryService) setStatus (w http.ResponseWriter, r *http.Request) {\n\tvar techStory model.TechStory\n\tmodel.ReadJsonBody(r, &techStory)\n\ttechStory.Key = mux.Vars(r)[\"id\"]\n\n\tWithTechStoryDao(func(dao techStoryDao) {\n\t\tuser := model.GetUserFromRequest(w, r)\n\t\tentity, err := dao.SetStatus(&techStory, techStory.Status, user)\n\t\tmodel.CheckErr(err)\n\t\tmodel.WriteResponse(true, nil, entity, w)\n\t})\n}", "func (me *THITStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (s *ServerStatus) Set(v ServerStatusValue) {\n\ts.mu.Lock()\n\ts.val = v\n\ts.mu.Unlock()\n}", "func (r *Reconciler) setStatus(\n\tctx context.Context,\n\tinstance *v1alpha1.ServiceBindingRequest,\n\tstatus string,\n) error {\n\tinstance.Status.BindingStatus = status\n\treturn r.client.Status().Update(ctx, instance)\n}", "func (a *Agent) SetStatus(status string) {\n\ta.mutex.Lock()\n\tdefer a.mutex.Unlock()\n\n\ta.status = status\n}", "func (h *HealthCheck) SetStatus(s int) {\n\th.status.setStatus(s)\n}", "func SetStatus(status Status) {\n\thandler.status = status\n}", "func (r *QuotaResult) SetStatus(s rpc.Status) { r.Status = s }", "func (m *EducationAssignment) SetStatus(value *EducationAssignmentStatus)() {\n m.status = value\n}", "func SetStatus(metaObj genruntime.MetaObject, status interface{}) error {\n\tptr := reflect.ValueOf(metaObj)\n\tval := ptr.Elem()\n\n\tif val.Kind() != reflect.Struct {\n\t\treturn errors.Errorf(\"metaObj kind was not struct\")\n\t}\n\n\tfield := val.FieldByName(\"Status\")\n\tstatusVal := reflect.ValueOf(status).Elem()\n\tfield.Set(statusVal)\n\n\treturn nil\n}", "func (p *PackageInfo) SetStatus(new string) {\n\tp.Status = new\n\tp.StatusDirty = true\n\t// Override in DEB 822 document used to write the status file\n\told := p.Paragraph.Values[\"Status\"]\n\tparts := strings.Split(old, \" \")\n\tnewStatus := fmt.Sprintf(\"%s %s %s\", parts[0], parts[1], new)\n\tp.Paragraph.Values[\"Status\"] = newStatus\n}", "func (s *severityValue) set(val severity.Severity) {\n\tatomic.StoreInt32((*int32)(&s.Severity), int32(val))\n}", "func (m *SchemaExtension) SetStatus(value *string)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (em *EvaluationManager) SetStatus(status string, avsData *internal.AvsLifecycleData, logger logrus.FieldLogger) error {\n\t// do internal monitor status update\n\terr := em.delegator.SetStatus(logger, avsData, em.internalAssistant, status)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// do external monitor status update\n\terr = em.delegator.SetStatus(logger, avsData, em.externalAssistant, status)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (me *TpubStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (r *TransferRecord) SetStatus(status string) {\n\tr.mutex.Lock()\n\tr.Status = status\n\tr.mutex.Unlock()\n}", "func (m *BrowserSiteList) SetStatus(value *BrowserSiteListStatus)() {\n err := m.GetBackingStore().Set(\"status\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *LongRunningOperation) SetStatusDetail(value *string)() {\n err := m.GetBackingStore().Set(\"statusDetail\", value)\n if err != nil {\n panic(err)\n }\n}", "func SetStatus(ctx context.Context, db gorp.SqlExecutor, workerID string, status string) error {\n\tw, err := LoadByID(ctx, db, workerID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tw.Status = status\n\tif status == sdk.StatusBuilding || status == sdk.StatusWaiting {\n\t\tw.JobRunID = nil\n\t}\n\tdbData := &dbWorker{Worker: *w}\n\tif err := gorpmapping.UpdateAndSign(ctx, db, dbData); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *GormClient) SetStatus(j *jobinator.Job, status int) error {\n\terr := c.db.Model(j).Update(\"status\", status).Error\n\treturn err\n}", "func (scaleSet *VirtualMachineScaleSet) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*VirtualMachineScaleSet_STATUS); ok {\n\t\tscaleSet.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st VirtualMachineScaleSet_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tscaleSet.Status = st\n\treturn nil\n}", "func (server *FlexibleServer) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*FlexibleServer_STATUS); ok {\n\t\tserver.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st FlexibleServer_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tserver.Status = st\n\treturn nil\n}", "func set(c client.Client, opCond *operators.OperatorCondition, conditionType string, status meta.ConditionStatus,\n\treason, message string) error {\n\tnewCond := meta.Condition{\n\t\tType: conditionType,\n\t\tStatus: status,\n\t\tReason: reason,\n\t\tMessage: message,\n\t}\n\tif err := patch(c, opCond, newCond); err != nil {\n\t\treturn err\n\t}\n\t// Wait to ensure change is actually picked up\n\treturn wait(c, opCond.GetNamespace(), newCond.Type, newCond.Status)\n}", "func (r *Response) SetStatus(s int) {\n\tr.StatusCode = s\n\tr.Message = http.StatusText(s)\n}", "func (b *block) setBlockStatus(ns, alarmVersion, hostname, tagString string, status, statusTTL int) error {\n\treturn b.c.SetWithTTL(\n\t\tBlockStatusKey(ns, alarmVersion, hostname, tagString),\n\t\tstrconv.Itoa(int(status)),\n\t\ttime.Duration(statusTTL)*time.Minute-5*time.Second)\n}", "func (builder *Builder) SetStatus(task string, status bool) {\n\tbuilder.LinuxBuild.Status[task] = status\n}", "func (obj *MsgStatus) Set() (code uint64, err error) {\n\tif obj == nil {\n\t\tcode = errc.NOTINIT\n\t\terr = fmt.Errorf(\"No initialization structure: MsgStatus\")\n\t\treturn\n\t}\n\tif obj.MsgID == \"\" {\n\t\tcode = errc.ParameterError\n\t\terr = fmt.Errorf(\"The parameter is incorrect:MsgID Can not be empty\")\n\t\treturn\n\t}\n\n\t_t := new(MsgStatus)\n\t_t.MsgID = obj.MsgID\n\tcode, err = _t.Get()\n\tif code == errc.NOTEXISTRECORD {\n\t\tcode, err = obj.Add()\n\t} else if code == 0 {\n\t\tcode, err = obj.Update()\n\t}\n\treturn\n}", "func (machine *VirtualMachine) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*VirtualMachine_STATUS); ok {\n\t\tmachine.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st VirtualMachine_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tmachine.Status = st\n\treturn nil\n}", "func (machine *VirtualMachine) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*VirtualMachine_STATUS); ok {\n\t\tmachine.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st VirtualMachine_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tmachine.Status = st\n\treturn nil\n}", "func (oiuo *OrderInfoUpdateOne) SetStatus(i int8) *OrderInfoUpdateOne {\n\toiuo.mutation.ResetStatus()\n\toiuo.mutation.SetStatus(i)\n\treturn oiuo\n}", "func (r *ReconcileHumioCluster) setState(ctx context.Context, state string, hc *corev1alpha1.HumioCluster) error {\n\tr.logger.Infof(\"setting cluster state to %s\", state)\n\thc.Status.State = state\n\terr := r.client.Status().Update(ctx, hc)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (am *appManager) setStatus(ctx context.Context, app *v1alpha1.FLApp) error {\n\t_, err := am.appStatusUpdater.UpdateStatusWithRetry(ctx, app, func(mutatingApp *v1alpha1.FLApp) bool {\n\t\tif apiequality.Semantic.DeepEqual(app.Status, mutatingApp.Status) {\n\t\t\treturn false\n\t\t}\n\t\tapp.Status.DeepCopyInto(&mutatingApp.Status)\n\t\treturn true\n\t})\n\treturn err\n}", "func (xdc *XxxDemoCreate) SetStatus(i int16) *XxxDemoCreate {\n\txdc.mutation.SetStatus(i)\n\treturn xdc\n}", "func setStatus(w http.ResponseWriter, req *http.Request) {\n\n\tkey, _ := req.URL.Query()[\"key\"]\n\n\tos.Setenv(\"APP_STATUS\", key[0])\n\n\tstatus := Status{Code: 200, Message: \"APP_STATUS SET\"}\n\n\tresponse, err := json.Marshal(status)\n\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Write(response)\n\n}", "func (pu *PostUpdate) SetStatus(i int8) *PostUpdate {\n\tpu.mutation.ResetStatus()\n\tpu.mutation.SetStatus(i)\n\treturn pu\n}", "func (s *DevStat) SetStatus(active, connected bool) {\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\ts.DeviceActive = active\n\ts.DeviceConnected = connected\n\n}", "func (me *TReviewActionStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (service *ServiceObject) setServiceStatus(status uint32) {\n\tif service.serviceStatusHandle == 0 {\n\t\treturn\n\t}\n\n\tservice.currentServiceStatus.currentState = status\n\tsetServiceStatusFunction(service.serviceStatusHandle, &service.currentServiceStatus)\n}", "func (ktuo *KqiTargetUpdateOne) SetStatus(b bool) *KqiTargetUpdateOne {\n\tktuo.mutation.SetStatus(b)\n\treturn ktuo\n}", "func (me *TReviewableHITStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (image *Image) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*Image_STATUS); ok {\n\t\timage.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st Image_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\timage.Status = st\n\treturn nil\n}", "func (puo *PostUpdateOne) SetStatus(i int8) *PostUpdateOne {\n\tpuo.mutation.ResetStatus()\n\tpuo.mutation.SetStatus(i)\n\treturn puo\n}", "func (me *TAssignmentStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (oiu *OrderInfoUpdate) SetStatus(i int8) *OrderInfoUpdate {\n\toiu.mutation.ResetStatus()\n\toiu.mutation.SetStatus(i)\n\treturn oiu\n}", "func (group *ResourceGroup) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*ResourceGroup_STATUS); ok {\n\t\tgroup.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st ResourceGroup_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tgroup.Status = st\n\treturn nil\n}", "func (group *ResourceGroup) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*ResourceGroup_STATUS); ok {\n\t\tgroup.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st ResourceGroup_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tgroup.Status = st\n\treturn nil\n}", "func (f *Friend) SetStatus(status int) {\n\tf.status = status\n}", "func (c *Collector) SetRunStatus(status lib.RunStatus) {}", "func (me *TQualificationStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (network *VirtualNetwork) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*VirtualNetwork_STATUS); ok {\n\t\tnetwork.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st VirtualNetwork_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tnetwork.Status = st\n\treturn nil\n}", "func (loop *Device) SetStatus(info *Info64) error {\n\t_, _, err := syscall.Syscall(syscall.SYS_IOCTL, loop.file.Fd(), CmdSetStatus64, uintptr(unsafe.Pointer(info)))\n\tif err != 0 {\n\t\treturn fmt.Errorf(\"Failed to set loop flags on loop device: %s\", syscall.Errno(err))\n\t}\n\treturn nil\n}", "func SetStatus(ctx context.Context, id string, status int) error {\n\t// language=SQL\n\t_, err := pgctx.Exec(ctx, `\n\t\tupdate payments\n\t\tset status = $2,\n\t\t updated_at = now(),\n\t\t at = now()\n\t\twhere id = $1\n\t`, id, status)\n\treturn err\n}", "func (r *reflectedStatusAccessor) SetStatus(status interface{}) {\n\tif r != nil && r.status.IsValid() && r.status.CanSet() {\n\t\tr.status.Set(reflect.ValueOf(status))\n\t}\n}", "func (o *Cause) SetStatus(v int32) {\n\to.Status.Set(&v)\n}", "func (me *THITReviewStatus) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (enterprise *RedisEnterprise) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*RedisEnterprise_STATUS); ok {\n\t\tenterprise.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st RedisEnterprise_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tenterprise.Status = st\n\treturn nil\n}", "func (ruleset *DnsForwardingRuleset) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*DnsForwardingRuleset_STATUS); ok {\n\t\truleset.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st DnsForwardingRuleset_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\truleset.Status = st\n\treturn nil\n}", "func (ktu *KqiTargetUpdate) SetStatus(b bool) *KqiTargetUpdate {\n\tktu.mutation.SetStatus(b)\n\treturn ktu\n}", "func set_response_status(w http.ResponseWriter, s int) {\n\t// Argument s should be one of the RFC2616 constants defined here:\n\t// https://golang.org/src/net/http/status.go\n\tw.WriteHeader(s)\n}", "func (r *responseWriter) SetStatus(status ATTError) {\n\tr.status = status\n}", "func (r *Response) SetStatus(code int, message string) *Response {\n\tr.SetStatusCode(code)\n\tr.SetStatusMessage(message)\n\treturn r\n}", "func (r *ResultsProxy) SetStatus(status Status) {\n\tr.Atomic(func(results Results) { results.SetStatus(status) })\n}", "func (database *SqlDatabase) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*DatabaseAccounts_SqlDatabase_STATUS); ok {\n\t\tdatabase.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st DatabaseAccounts_SqlDatabase_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tdatabase.Status = st\n\treturn nil\n}", "func (bar *StatusBar) Set(inString string, pos int) {\n\tif pos > len(bar.Messages)-1 || pos < 0 {\n\t\tinString = \"Statusbar error: Invalid range to setting this message -> \" + inString\n\t\tpos = len(bar.Messages) - 1\n\t}\n\tbar.Messages[pos] = inString\n\tbar.Disp()\n}", "func (m *PrintJobStatus) SetState(value *PrintJobProcessingState)() {\n m.state = value\n}", "func Set(ctx context.Context, rollerName string, st *AutoRollStatus) error {\n\tbuf := bytes.NewBuffer(nil)\n\tif err := gob.NewEncoder(buf).Encode(st); err != nil {\n\t\treturn err\n\t}\n\tw := &DsStatusWrapper{\n\t\tData: buf.Bytes(),\n\t\tRoller: rollerName,\n\t}\n\t_, err := ds.DS.RunInTransaction(ctx, func(tx *datastore.Transaction) error {\n\t\t_, err := tx.Put(key(rollerName), w)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Optionally export the mini version of the internal roller's status\n\t// to the external datastore.\n\tif util.In(rollerName, EXPORT_WHITELIST) {\n\t\texportStatus := &AutoRollStatus{\n\t\t\tAutoRollMiniStatus: st.AutoRollMiniStatus,\n\t\t}\n\t\tbuf := bytes.NewBuffer(nil)\n\t\tif err := gob.NewEncoder(buf).Encode(exportStatus); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tw := &DsStatusWrapper{\n\t\t\tData: buf.Bytes(),\n\t\t\tRoller: rollerName,\n\t\t}\n\t\t_, err := ds.DS.RunInTransaction(ctx, func(tx *datastore.Transaction) error {\n\t\t\tk := key(rollerName)\n\t\t\tk.Namespace = ds.AUTOROLL_NS\n\t\t\tk.Parent.Namespace = ds.AUTOROLL_NS\n\t\t\t_, err := tx.Put(k, w)\n\t\t\treturn err\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (m *CloudPcConnection) SetHealthCheckStatus(value *string)() {\n err := m.GetBackingStore().Set(\"healthCheckStatus\", value)\n if err != nil {\n panic(err)\n }\n}", "func (v *NullableSyntheticsTestPauseStatus) Set(val *SyntheticsTestPauseStatus) {\n\tv.value = val\n\tv.isSet = true\n}", "func (pool *WorkspacesBigDataPool) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*Workspaces_BigDataPool_STATUS); ok {\n\t\tpool.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st Workspaces_BigDataPool_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tpool.Status = st\n\treturn nil\n}", "func (a *API) setWidgetStatus( /*ctx context.Context,*/ id int, value string) error {\n\t// - [ ] `<id>[|<value>]`: Sets given `<value>` of widget with given `<id>`\n\t// - `Button`: `<value>` equals to `1` for press and `0` for release\n\t// - `Slider`: `<value>` is in range of slider\n\t// - `AudioTrigger`: `<value>` equals to `1` for active and `0` for inactive\n\t// - `CueList`: `<value>` is `<command>[|<index>]`:\n\t// - `PLAY`: Plays cue\n\t// - `STOP`: Stops cue\n\t// - `PREV`: Selects previous cue\n\t// - `NEXT`: Selects next cue\n\t// - `STEP|<index>`: Selects cue with given `<index>`\n\t// - `Frame`, `SoloFrame`: `<value>` is one of `NEXT_PG` (next page) or `PREV_PG` (previous page)\n\n\t// lock API\n\ta.lock.Lock()\n\tdefer a.lock.Unlock()\n\n\t// Send message\n\terr := a.writeText([]byte(fmt.Sprintf(\"QLC+API|setFunctionStatus|%d|%s\", id, value)))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (peering *VirtualNetworksVirtualNetworkPeering) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*VirtualNetworks_VirtualNetworkPeering_STATUS); ok {\n\t\tpeering.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st VirtualNetworks_VirtualNetworkPeering_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tpeering.Status = st\n\treturn nil\n}", "func (e *Engine) setHAStatus(status seesaw.HAStatus) error {\n\tselect {\n\tcase e.haManager.statusChan <- status:\n\tdefault:\n\t\treturn fmt.Errorf(\"status channel if full\")\n\t}\n\treturn nil\n}", "func (ftd *FakeTiCDCControl) SetStatus(status *CaptureStatus) {\n\tftd.status = status\n}", "func (etc *ExportTaskCreate) SetStatus(e exporttask.Status) *ExportTaskCreate {\n\tetc.mutation.SetStatus(e)\n\treturn etc\n}", "func (service *StorageAccountsTableService) SetStatus(status genruntime.ConvertibleStatus) error {\n\t// If we have exactly the right type of status, assign it\n\tif st, ok := status.(*StorageAccounts_TableService_STATUS); ok {\n\t\tservice.Status = *st\n\t\treturn nil\n\t}\n\n\t// Convert status to required version\n\tvar st StorageAccounts_TableService_STATUS\n\terr := status.ConvertStatusTo(&st)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to convert status\")\n\t}\n\n\tservice.Status = st\n\treturn nil\n}", "func setStatus(wpa *datadoghqv1alpha1.WatermarkPodAutoscaler, currentReplicas, desiredReplicas int32, metricStatuses []autoscalingv2.MetricStatus, rescale bool) {\n\twpa.Status.CurrentReplicas = currentReplicas\n\twpa.Status.DesiredReplicas = desiredReplicas\n\twpa.Status.CurrentMetrics = metricStatuses\n\n\tif rescale {\n\t\tnow := metav1.NewTime(time.Now())\n\t\twpa.Status.LastScaleTime = &now\n\t}\n}" ]
[ "0.73551786", "0.7231414", "0.70108753", "0.69346637", "0.68965626", "0.68820643", "0.68415093", "0.6834977", "0.67909706", "0.6782313", "0.6776894", "0.6771386", "0.67292285", "0.6720781", "0.66878325", "0.6678704", "0.6673506", "0.66666865", "0.6638011", "0.6560599", "0.6559855", "0.6529623", "0.6470555", "0.64622104", "0.64084333", "0.6393523", "0.6359125", "0.6323644", "0.63138455", "0.6307357", "0.6294869", "0.6293322", "0.62809783", "0.6242291", "0.6229815", "0.6224772", "0.6221573", "0.6184331", "0.61816597", "0.6180488", "0.61796385", "0.61652523", "0.6149392", "0.61421424", "0.61171657", "0.6093434", "0.6084109", "0.60830784", "0.60826766", "0.6051309", "0.6046568", "0.60319185", "0.60319185", "0.6029438", "0.60270053", "0.60209566", "0.60202247", "0.6018428", "0.6014103", "0.5980961", "0.59711075", "0.5969833", "0.59496236", "0.59470415", "0.5944282", "0.59385973", "0.593609", "0.5935625", "0.59337103", "0.59337103", "0.59211266", "0.5919058", "0.5883052", "0.5881345", "0.5878406", "0.5862858", "0.58579886", "0.58550596", "0.58539927", "0.5847153", "0.58449346", "0.584415", "0.5840265", "0.5839299", "0.5817933", "0.58098173", "0.5801152", "0.57992125", "0.5799132", "0.5798961", "0.5786918", "0.578486", "0.57788324", "0.5777014", "0.5767373", "0.57495034", "0.5743534", "0.57414854", "0.57405597", "0.57336795" ]
0.65698147
19
GetStatus returns the rollup status of the bot.
func (r *AutoRoller) GetStatus(includeError bool) *AutoRollStatus { return r.status.Get(includeError) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) GetStatus(includeError bool) *AutoRollStatus {\n\treturn r.status.Get(includeError, nil)\n}", "func (b *Base) GetStatus() string {\n\treturn `\n\tGoCryptoTrader Service: Online\n\tService Started: ` + ServiceStarted.String()\n}", "func (s *ApiService) GetStatus(ctx context.Context) (ordf.ImplResponse, error) {\n\tstatus := ordf.Status{\n\t\tStatus: \"UP\",\n\t}\n\n\t// check NowPayments API\n\tnowPaymentsStatus, err := s.nowPaymentsService.Status(ctx)\n\tif err != nil {\n\t\tstatus.Status = fmt.Sprintf(\"Failed to check NowPayments status: %+v\", err)\n\t} else if nowPaymentsStatus != \"OK\" {\n\t\tstatus.Status = fmt.Sprintf(\"NowPayments is down: %s\", nowPaymentsStatus)\n\t}\n\n\treturn ordf.Response(200, status), nil\n}", "func GetStatus(c echo.Context) error {\n\tserver := c.(*Server)\n\n\tcount, err := server.Repository.Games()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.JSON(http.StatusOK, Status{\n\t\tName: server.Name,\n\t\tGames: count,\n\t})\n}", "func (r *RpmOstreeClient) GetStatus() (string, error) {\n\toutput, err := RunGetOut(\"rpm-ostree\", \"status\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(output), nil\n}", "func (f *FortiWebClient) GetStatus() string {\n\n\tclient := &http.Client{}\n\n\treq, err := http.NewRequest(\"GET\", strings.Join([]string{f.URL, \"api/v1.0/System/Status/Status\"}, \"\"), nil)\n\treq.Header.Add(\"Authorization\", encodeBase64(f.Username, f.Password))\n\tresponse, error := client.Do(req)\n\n\tif error != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t\treturn strings.Join([]string{\"Error: The HTTP request failed with error \", error.Error()}, \"\")\n\t}\n\n\tdefer response.Body.Close()\n\tbody, _ := ioutil.ReadAll(response.Body)\n\n\treturn string(body[:])\n}", "func (client *Client) StatusGet() (status *Status, err error) {\n\terr = client.get(client.buildPath(\"/status/\"), &status)\n\treturn\n}", "func (h *HealthCheck) GetStatus() int {\n\treturn h.status.getStatus()\n}", "func (sc *StreamClient) GetStatus() int {\n\treturn sc.status\n}", "func (o *WebhooksJsonWebhook) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (o *OnpremUpgradePhase) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (o *ApplianceClusterInstallPhase) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (o *LinkSessionFinishedWebhook) GetStatus() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Status\n}", "func (g *AzureSqlActionManager) GetStatus(obj runtime.Object) (*v1alpha1.ASOStatus, error) {\n\tinstance, err := g.convert(obj)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &instance.Status, nil\n}", "func (o *Run) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func GetStatus(cfg *Config) Status {\n\tnow := time.Now().Unix()\n\n\ts := Status{\n\t\tPID: os.Getpid(),\n\t\tService: \"list-service\",\n\t}\n\n\ts.Status = \"ok\"\n\ts.Version = Version()\n\ts.CPUs = runtime.NumCPU()\n\ts.GoVers = runtime.Version()\n\ts.TimeStamp = now\n\ts.UpTime = now - InstanceStartTime\n\ts.LogLevel = log.GetLevel()\n\n\tif host, err := os.Hostname(); err == nil {\n\t\ts.Hostname = host\n\t}\n\n\treturn s\n}", "func (r *Composition) GetStatus() resv1.Status {\n\treturn &r.Status\n}", "func (_TrialRulesAbstract *TrialRulesAbstractTransactor) GetStatus(opts *bind.TransactOpts, witnessStatus uint8, trialStatus uint8) (*types.Transaction, error) {\n\treturn _TrialRulesAbstract.contract.Transact(opts, \"getStatus\", witnessStatus, trialStatus)\n}", "func (s *Service) GetStatus(ctx context.Context, req *request.Status) (*response.Status, error) {\n\tres := &response.Status{Callback: req.Callback}\n\tif req.Callback == \"I don't like launch pad\" {\n\t\treturn res, errors.New(\"launch pad is the best and you know it\")\n\t}\n\treturn res, nil\n}", "func GetStatus() string {\n\tif lastStatus == nil {\n\t\tlastStatus = &status{\n\t\t\tstate: \"unknown\",\n\t\t\treason: \"no known last status\",\n\t\t}\n\t}\n\treturn lastStatus.state\n}", "func (j *Juju) GetStatus() (string, error) {\n\ttmp := \"JUJU_DATA=\" + JujuDataPrefix + j.Name\n\tcmd := exec.Command(\"juju\", \"status\")\n\tcmd.Env = append(os.Environ(), tmp)\n\tout, err := cmd.Output()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"GetStatus error: %v: %s\", err, err.(*exec.ExitError).Stderr)\n\t}\n\tlog.Debug(string(out))\n\treturn string(out), nil\n}", "func GetStatus() (string, error) {\n\tbody, err := getRequest(GetURL()+getStatusPath, \"\", false)\n\tif err != nil {\n\t\treturn \"Connection failed\", err\n\t}\n\tif string(body) != `\"any-compute\"` {\n\t\terr = util.ErrServerUnavailable\n\t}\n\treturn string(body), err\n}", "func GetStatus() (status AuditStatus, err error) {\n\tclient, err := libaudit.NewAuditClient(nil)\n\tdefer client.Close()\n\tif err != nil {\n\t\treturn AuditStatus{}, errors.Wrap(err, \"Failed to initialize client\")\n\t}\n\tkstatus, err := client.GetStatus()\n\tif err != nil {\n\t\treturn AuditStatus{}, errors.Wrap(err, \"Failed to get audit status\")\n\t}\n\tstatus.Enabled = kstatus.Enabled\n\tstatus.Failure = kstatus.Failure\n\tstatus.PID = kstatus.PID\n\tstatus.RateLimit = kstatus.RateLimit\n\tstatus.BacklogLimit = kstatus.BacklogLimit\n\tstatus.Lost = kstatus.Lost\n\tstatus.Backlog = kstatus.Backlog\n\tstatus.BacklogWaitTime = kstatus.BacklogWaitTime\n\treturn status, nil\n}", "func (c *Client) getStatus() Status {\n\n\treturn c.status\n}", "func (puppetDb *PuppetDb) GetStatus() (*operations.GetStatusOK, error) {\n\tstringToken, err := puppetDb.Token.Read()\n\tif err != nil {\n\t\tlog.Debug(err.Error())\n\t}\n\n\tclient, err := puppetDb.Client.GetClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tapiKeyHeaderAuth := httptransport.APIKeyAuth(\"X-Authentication\", \"header\", stringToken)\n\tgetStatusParameters := operations.NewGetStatusParamsWithContext(context.Background())\n\treturn client.Operations.GetStatus(getStatusParameters, apiKeyHeaderAuth)\n}", "func (profile *Profile) GetStatus() genruntime.ConvertibleStatus {\n\treturn &profile.Status\n}", "func (group *ResourceGroup) GetStatus() genruntime.ConvertibleStatus {\n\treturn &group.Status\n}", "func (group *ResourceGroup) GetStatus() genruntime.ConvertibleStatus {\n\treturn &group.Status\n}", "func (c *CheckRun) GetStatus() string {\n\tif c == nil || c.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *c.Status\n}", "func (o *CheckoutResponse) GetStatus() string {\n\tif o == nil || IsNil(o.Status) {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (t *SelfTester) GetStatus() *api.SelfTestsStatus {\n\treturn &api.SelfTestsStatus{\n\t\tLastTimestamp: t.lastTimestamp.Format(time.RFC822),\n\t\tSuccess: t.success,\n\t\tFails: t.fails,\n\t}\n}", "func (d *Docker) GetStatus(ctx context.Context, name string) error {\n\treturn nil\n}", "func (rule *NamespacesEventhubsAuthorizationRule) GetStatus() genruntime.ConvertibleStatus {\n\treturn &rule.Status\n}", "func GetStatus(w http.ResponseWriter, r *http.Request, argv map[string]string) error {\n\tresTemplate := `{\"MetricsService\":\"STARTED\",\"Implementation-Version\":\"%s\",\"MohawkVersion\":\"%s\",\"MohawkStorage\":\"%s\"}`\n\tres := fmt.Sprintf(resTemplate, defaultAPI, VER, BackendName)\n\n\tfmt.Fprintln(w, res)\n\treturn nil\n}", "func GetStatus(cfg *config.Config, c client.Client) string {\n\tvar status string\n\n\tvalStatus := GetValStatusFromDB(cfg, c)\n\tif valStatus == \"1\" {\n\t\tvalStatus = \"voting\"\n\t} else {\n\t\tvalStatus = \"jailed\"\n\t}\n\tstatus = fmt.Sprintf(\"Heimdall Node Status:\\n- Your validator is currently %s \\n\", valStatus)\n\n\tvalHeight := GetValidatorBlock(cfg, c) // get heimdall validator block height\n\tstatus = status + fmt.Sprintf(\"- Validator current block height %s \\n\", valHeight)\n\n\tnetworkHeight := GetNetworkBlock(cfg, c) // get heimdall network block height\n\tstatus = status + fmt.Sprintf(\"- Network current block height %s \\n\", networkHeight)\n\n\tvotingPower := GetVotingPowerFromDb(cfg, c) // get heimdall validator voting power\n\tstatus = status + fmt.Sprintf(\"- Voting power of your validator is %s \\n\", votingPower)\n\n\tborHeight := GetBorCurrentBlokHeight(cfg, c) // get bor validator block height\n\tstatus = status + fmt.Sprintf(\"\\nBor Node :\\n- Validator current block height %s \\n\", borHeight)\n\n\tspanID := GetBorSpanIDFromDb(cfg, c) // get bor latest span ID\n\tstatus = status + fmt.Sprintf(\"- Current span id is %s \\n\", spanID)\n\n\treturn status\n}", "func (image *Image) GetStatus() genruntime.ConvertibleStatus {\n\treturn &image.Status\n}", "func (c *CheckSuite) GetStatus() string {\n\tif c == nil || c.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *c.Status\n}", "func (s *State) GetStatus(sett *Setting) Status {\n\ts.m.Lock()\n\tdefer s.m.Unlock()\n\n\tif sett.FailureThreshold > s.currentFailureCount {\n\t\treturn Closed\n\t}\n\n\tretry := s.lastFailureTimestamp.Add(sett.RetryTimeout)\n\tnow := time.Now().UTC()\n\n\tif retry.Before(now) || retry.Equal(now) {\n\n\t\tif s.retrySuccessCount >= sett.RetrySuccessThreshold {\n\t\t\ts.innerReset()\n\t\t\treturn Closed\n\t\t}\n\n\t\tif s.currentExecutions > sett.MaxRetryExecutionThreshold {\n\t\t\treturn Open\n\t\t}\n\n\t\treturn HalfOpen\n\t}\n\n\treturn Open\n}", "func (c *CryptohomeBinary) GetStatusString(ctx context.Context) (string, error) {\n\tout, err := c.call(ctx, \"--action=status\")\n\treturn string(out), err\n}", "func (o *ImageImportOperation) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func GetStatus(c *gin.Context) {\n\tc.String(http.StatusOK, \"OK\")\n}", "func (client *Client) GetStatus() (*Response, error) {\n\tpath := \"/status\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\treq, err := http.NewRequest(\"GET\", uri, bytes.NewBuffer([]byte(\"\")))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := client.performRequest(req, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp, err\n}", "func (c Cas) GetStatus(scaleID uint32) (Status, error) {\n\tvar status Status\n\n\topcode := [2]byte{'R', 'N'}\n\tbuf := encodePacket(scaleID, opcode, []byte{})\n\n\tif _, err := c.conn.Write(buf); err != nil {\n\t\treturn status, err\n\t}\n\n\ttmp := make([]byte, 512)\n\n\tif _, err := c.conn.Read(tmp); err != nil {\n\t\treturn status, err\n\t}\n\n\tif tmp[0] != 'W' || tmp[1] != 'N' {\n\t\treturn status, fmt.Errorf(\"GetStatus %s %x\", string(tmp[0:2]), tmp)\n\t}\n\n\treturn status, nil\n}", "func (audit Audit) GetStatus() string {\n\treturn audit.status\n}", "func (enterprise *RedisEnterprise) GetStatus() genruntime.ConvertibleStatus {\n\treturn &enterprise.Status\n}", "func (o *UcsdBackupInfoAllOf) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func GetStatus(c *db.Cocoon, _ []byte) (interface{}, error) {\n\tvar err error\n\n\tstatuses, err := c.QueryBuildStatusesWithMemcache()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tagentStatuses, err := c.QueryAgentStatuses()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &GetStatusResult{\n\t\tStatuses: statuses,\n\t\tAgentStatuses: agentStatuses,\n\t}, nil\n}", "func (o *SmartstackBackend) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (o *InlineResponse20030) GetSTATUS() string {\n\tif o == nil || o.STATUS == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.STATUS\n}", "func (o *Authorization) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (p *PagesBuild) GetStatus() string {\n\tif p == nil || p.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *p.Status\n}", "func (h *Handler) GetStatus(object interface{}) (string, error) {\n\tswitch val := object.(type) {\n\tcase string:\n\t\tpod, err := h.Get(val)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn string(pod.Status.Phase), nil\n\tcase *corev1.Pod:\n\t\treturn string(val.Status.Phase), nil\n\tcase corev1.Pod:\n\t\treturn string(val.Status.Phase), nil\n\tdefault:\n\t\treturn \"\", ErrInvalidToolsType\n\t}\n}", "func (gh *groupHandler) GetStatus() (bool, string, error) {\n\tfor {\n\t\tselect {\n\t\tcase err := <-gh.errCh:\n\t\t\treturn false, \"\", err\n\t\tdefault:\n\t\t\tgh.RLock()\n\t\t\tcurrentLeader := gh.currentLeader\n\t\t\tgh.RUnlock()\n\t\t\tif currentLeader == \"\" {\n\t\t\t\t// wait until a leader is elected\n\t\t\t\ttime.Sleep(time.Second)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tgh.RLock()\n\t\t\tdefer gh.RUnlock()\n\t\t\treturn gh.areWeLeader, gh.currentLeader, nil\n\t\t}\n\t}\n}", "func (network *VirtualNetwork) GetStatus() genruntime.ConvertibleStatus {\n\treturn &network.Status\n}", "func (o *HyperflexSnapshotStatus) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (n *RocketmqChannel) GetStatus() *duckv1.Status {\n\treturn &n.Status.Status\n}", "func (o *PaymentInitiationPayment) GetStatus() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Status\n}", "func (s *AccessPolicy) GetStatus() resv1.Status {\n\treturn &s.Status\n}", "func (t *Task) GetStatus() int {\n\treturn t.Status\n}", "func (s *Service) GetStatus(ctx context.Context, req *request.Status) (*response.Status, error) {\n\treturn &response.Status{}, nil\n}", "func (p *Porthole) GetStatus() *status.Status {\n\t// copy current status values into a new Status instance\n\treturn &status.Status{\n\t\tGitCommit: p.coordinator.Status.GitCommit,\n\t\tLastRequest: p.coordinator.Status.LastRequest,\n\t\tLastFetch: p.coordinator.Status.LastFetch,\n\t\tLatestAdditions: p.coordinator.Status.LatestAdditions,\n\t}\n}", "func GetStatus(w http.ResponseWriter, r *http.Request) {\n\tmiddleware.EnableCors(&w)\n\n\tstatus := core.GetStatus()\n\tresponse := webStatusResponse{\n\t\tOnline: status.Online,\n\t\tViewerCount: status.ViewerCount,\n\t\tLastConnectTime: status.LastConnectTime,\n\t\tLastDisconnectTime: status.LastDisconnectTime,\n\t\tVersionNumber: status.VersionNumber,\n\t\tStreamTitle: status.StreamTitle,\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tif err := json.NewEncoder(w).Encode(response); err != nil {\n\t\tInternalErrorHandler(w, err)\n\t}\n}", "func (o *InlineResponse20083) GetSTATUS() string {\n\tif o == nil || o.STATUS == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.STATUS\n}", "func (t *Topic) GetStatus() *duckv1.Status {\n\treturn &t.Status.Status\n}", "func (rule *NamespacesTopicsSubscriptionsRule) GetStatus() genruntime.ConvertibleStatus {\n\treturn &rule.Status\n}", "func (subscription *NamespacesTopicsSubscription) GetStatus() genruntime.ConvertibleStatus {\n\treturn &subscription.Status\n}", "func (r *Random) GetStatus() interface{} {\n\treturn SourceStatusRunning\n}", "func (v *Kounta) GetStatus(token string, company string) error {\n\tclient := &http.Client{}\n\tclient.CheckRedirect = checkRedirectFunc\n\n\tu, _ := url.ParseRequestURI(baseURL)\n\tu.Path = fmt.Sprintf(companyStatus, company)\n\turlStr := fmt.Sprintf(\"%v\", u)\n\n\tfmt.Println(urlStr)\n\n\tr, err := http.NewRequest(\"GET\", urlStr, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.Header = http.Header(make(map[string][]string))\n\tr.Header.Set(\"Accept\", \"application/json\")\n\tr.Header.Set(\"Authorization\", \"Bearer \"+token)\n\n\tres, err := client.Do(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trawResBody, err := ioutil.ReadAll(res.Body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif res.StatusCode == 200 {\n\n\t\tfmt.Println(string(rawResBody))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\tfmt.Println(string(rawResBody))\n\treturn fmt.Errorf(\"Failed to get Kounta Sale %s\", res.Status)\n\n}", "func (account *DatabaseAccount) GetStatus() genruntime.ConvertibleStatus {\n\treturn &account.Status\n}", "func (ruleset *DnsForwardingRuleset) GetStatus() genruntime.ConvertibleStatus {\n\treturn &ruleset.Status\n}", "func getStatus(cmd *cobra.Command, args []string) {\n\tconfig, err := statusKubeconfig.Get()\n\tif err != nil {\n\t\terrlog.LogError(errors.Wrap(err, \"couldn't get kubernetes config\"))\n\t\tos.Exit(1)\n\t}\n\tclient, err := kubernetes.NewForConfig(config)\n\tif err != nil {\n\t\terrlog.LogError(errors.Wrap(err, \"couldn't initialise kubernete client\"))\n\t\tos.Exit(1)\n\t}\n\n\tstatus, err := ops.NewSonobuoyClient().GetStatus(statusNamespace, client)\n\tif err != nil {\n\t\terrlog.LogError(errors.Wrap(err, \"error attempting to run sonobuoy\"))\n\t\tos.Exit(1)\n\t}\n\n\ttw := tabwriter.NewWriter(os.Stdout, 1, 8, 1, '\\t', tabwriter.AlignRight)\n\n\tfmt.Fprintf(tw, \"PLUGIN\\tNODE\\tSTATUS\\n\")\n\tfor _, pluginStatus := range status.Plugins {\n\t\tfmt.Fprintf(tw, \"%s\\t%s\\t%s\\n\", pluginStatus.Plugin, pluginStatus.Node, pluginStatus.Status)\n\t}\n\n\tif err := tw.Flush(); err != nil {\n\t\terrlog.LogError(errors.Wrap(err, \"couldn't write status out\"))\n\t\tos.Exit(1)\n\t}\n\tfmt.Printf(\"\\n%s\\n\", humanReadableStatus(status.Status))\n}", "func (t *Service) GetStatus() *duckv1.Status {\n\treturn &t.Status.Status\n}", "func (o *InlineResponse20082) GetSTATUS() string {\n\tif o == nil || o.STATUS == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.STATUS\n}", "func (topic *Topic) GetStatus() genruntime.ConvertibleStatus {\n\treturn &topic.Status\n}", "func (u *UpdateCheckRunOptions) GetStatus() string {\n\tif u == nil || u.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *u.Status\n}", "func (p *Pages) GetStatus() string {\n\tif p == nil || p.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *p.Status\n}", "func (o *InlineResponse20085) GetSTATUS() string {\n\tif o == nil || o.STATUS == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.STATUS\n}", "func (subnet *VirtualNetworksSubnet) GetStatus() genruntime.ConvertibleStatus {\n\treturn &subnet.Status\n}", "func (o *AuthenticationResponse) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (p Project) GetStatus() string {\n\tstatus := getFirstMatch(\"^\\\\[([R|Y|G|!])\\\\] .*$\", p.Name)\n\n\tif status == \"\" {\n\t\treturn \"Unknown\"\n\t}\n\n\tif status == \"!\" {\n\t\treturn \"Done\"\n\t}\n\n\treturn status\n}", "func (m *MaintainerManager) GetStatus(pr *gh.PullRequest) (gh.CombinedStatus, error) {\n\to := &gh.Options{}\n\to.QueryParams = map[string]string{}\n\treturn m.client.CombinedStatus(m.repo, pr.Head.Sha, o)\n}", "func (o *VirtualizationIweVirtualMachine) GetStatus() string {\n\tif o == nil || o.Status == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Status\n}", "func (app *Application) GetStatus() *Status {\n if app.status == nil {\n app.status = app.Get(\"status\").(*Status)\n }\n\n return app.status\n}", "func (o *RequestTarget) GetStatus() *string {\n\tif o == nil {\n\t\treturn nil\n\t}\n\n\treturn o.Status\n}", "func (machine *VirtualMachine) GetStatus() genruntime.ConvertibleStatus {\n\treturn &machine.Status\n}", "func (machine *VirtualMachine) GetStatus() genruntime.ConvertibleStatus {\n\treturn &machine.Status\n}", "func (s Strategy) GetStatus(coralName string, foreign string) string {\n\treturn s.Map.Entities[coralName].Status[foreign]\n}", "func (policy *ServersConnectionPolicy) GetStatus() genruntime.ConvertibleStatus {\n\treturn &policy.Status\n}", "func (k *NatssSource) GetStatus() *duckv1.Status {\n\treturn &k.Status.Status\n}", "func (i *Import) GetStatus() string {\n\tif i == nil || i.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *i.Status\n}", "func (database *SqlDatabase) GetStatus() genruntime.ConvertibleStatus {\n\treturn &database.Status\n}", "func GetStatus(address string, obj interface{}) (error, int) {\n\treplica, err := NewReplicaClient(address)\n\tif err != nil {\n\t\treturn err, -1\n\t}\n\turl := replica.address + \"/stats\"\n\tresp, err := replica.httpClient.Get(url)\n\tif resp != nil {\n\t\tif resp.StatusCode == 500 {\n\t\t\treturn err, 500\n\t\t} else if resp.StatusCode == 503 {\n\t\t\treturn err, 503\n\t\t}\n\t} else {\n\t\treturn err, -1\n\t}\n\tif err != nil {\n\t\treturn err, -1\n\t}\n\tdefer resp.Body.Close()\n\n\treturn json.NewDecoder(resp.Body).Decode(obj), 0\n}", "func (s *Source) GetStatus() interface{} {\n\treturn s.Status\n}", "func (o *SubscriptionRegistration) GetStatus() (value string, ok bool) {\n\tok = o != nil && o.bitmap_&16 != 0\n\tif ok {\n\t\tvalue = o.status\n\t}\n\treturn\n}", "func (d *Docker) GetStatus(ctx context.Context, ID string, checkHealth bool) (*types.WorkloadStatus, error) {\n\tlogger := log.WithFunc(\"GetStatus\").WithField(\"ID\", ID)\n\tcontainer, err := d.detectWorkload(ctx, ID)\n\tif err != nil {\n\t\tlogger.Error(ctx, err, \"failed to detect workload\")\n\t\treturn nil, err\n\t}\n\n\tbytes, err := json.Marshal(container.Labels)\n\tif err != nil {\n\t\tlogger.Error(ctx, err, \"failed to marshal labels\")\n\t\treturn nil, err\n\t}\n\n\tstatus := &types.WorkloadStatus{\n\t\tID: container.ID,\n\t\tRunning: container.Running,\n\t\tNetworks: container.Networks,\n\t\tExtension: bytes,\n\t\tAppname: container.Name,\n\t\tNodename: d.config.HostName,\n\t\tEntrypoint: container.Entrypoint,\n\t\tHealthy: container.Running && container.HealthCheck == nil,\n\t}\n\n\t// only check the running containers\n\tif checkHealth && container.Running {\n\t\tfree, acquired := d.cas.Acquire(container.ID)\n\t\tif !acquired {\n\t\t\treturn nil, common.ErrGetLockFailed\n\t\t}\n\t\tdefer free()\n\t\tstatus.Healthy = container.CheckHealth(ctx, time.Duration(d.config.HealthCheck.Timeout)*time.Second)\n\t}\n\n\treturn status, nil\n}", "func (candidate *Candidate) GetStatus() byte {\n\tcandidate.lock.RLock()\n\tdefer candidate.lock.RUnlock()\n\n\treturn candidate.Status\n}", "func (c *CreateCheckRunOptions) GetStatus() string {\n\tif c == nil || c.Status == nil {\n\t\treturn \"\"\n\t}\n\treturn *c.Status\n}", "func (o *InlineResponse20026) GetSTATUS() string {\n\tif o == nil || o.STATUS == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.STATUS\n}", "func (o *Transfer) GetStatus() TransferStatus {\n\tif o == nil {\n\t\tvar ret TransferStatus\n\t\treturn ret\n\t}\n\n\treturn o.Status\n}", "func (o GetReposRepoTagOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetReposRepoTag) string { return v.Status }).(pulumi.StringOutput)\n}" ]
[ "0.6807277", "0.65729356", "0.64715016", "0.6376956", "0.6359956", "0.6329804", "0.63049203", "0.6242687", "0.6237769", "0.6228739", "0.6201623", "0.6189542", "0.61757827", "0.61606675", "0.61528236", "0.61432916", "0.61419344", "0.6122873", "0.61217576", "0.6087857", "0.6086998", "0.6085552", "0.607469", "0.6064227", "0.6062816", "0.6061582", "0.60591143", "0.60591143", "0.60532266", "0.6047967", "0.6025664", "0.6019174", "0.6015318", "0.6011342", "0.6005811", "0.60043836", "0.59884185", "0.59848833", "0.59843427", "0.597708", "0.59747267", "0.596191", "0.5961503", "0.5961399", "0.5960866", "0.5955751", "0.595566", "0.5951439", "0.59471494", "0.5933466", "0.59284055", "0.5928256", "0.5922757", "0.58972156", "0.5890357", "0.58896935", "0.5887634", "0.5886708", "0.58825684", "0.5881119", "0.5880627", "0.5879805", "0.5877447", "0.5866725", "0.58617365", "0.5860017", "0.58574045", "0.5854143", "0.5850491", "0.5840824", "0.5837269", "0.58330786", "0.58325243", "0.58269686", "0.582477", "0.5817987", "0.5808612", "0.58064646", "0.58007723", "0.5796345", "0.5795617", "0.5794969", "0.57909167", "0.5790392", "0.578584", "0.578584", "0.578252", "0.5778288", "0.5774434", "0.5770707", "0.5762911", "0.5758687", "0.57538784", "0.5753513", "0.57444495", "0.57438385", "0.5739818", "0.5737203", "0.5736394", "0.573396" ]
0.68261963
0
SetMode sets the desired mode of the bot. This forces the bot to run and blocks until it finishes.
func (r *AutoRoller) SetMode(m, user, message string) error { r.modeMtx.Lock() defer r.modeMtx.Unlock() if err := r.modeHistory.Add(m, user, message); err != nil { return err } return r.doAutoRoll() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) SetMode(m, user, message string) error {\n\tif err := r.modeHistory.Add(m, user, message); err != nil {\n\t\treturn err\n\t}\n\treturn r.Tick()\n}", "func SetMode(flag int) (reset func()) {\n\tMustTestMode()\n\told := mode\n\treset = func() {\n\t\tmode = old\n\t}\n\tmode = flag\n\treturn\n}", "func (s *Server) SetMode(mode Mode) {\n\tif s != nil {\n\t\ts.mode = mode\n\t}\n}", "func (m *Modem) SetMode(mode uint8) error {\n\treturn m.sendAndWaitForACK([]byte{\n\t\tSetMode,\n\t\tmode,\n\t}, m.Timeout)\n}", "func SetMode(m TunnelMode) {\n\tmode = m\n}", "func SetMode(value string) {\n\trgin.SetMode(value)\n}", "func (c *Config) SetMode(mode Mode) {\n\tc.general.Mode = mode\n\tc.event <- &Event{Type: \"mode\", Payload: mode}\n}", "func (p *LegoPort) SetMode(m string) *LegoPort {\n\tif p.err != nil {\n\t\treturn p\n\t}\n\tp.err = setAttributeOf(p, mode, m)\n\treturn p\n}", "func (m *msg) SetMode(md mode) {\n\tm.LiVnMode = (m.LiVnMode & 0xf8) | byte(md)\n}", "func (r *Receiver) SetMode(m ChannelMode, sideKey trinary.Trytes) error {\n\tif m != ChannelModePublic && m != ChannelModePrivate && m != ChannelModeRestricted {\n\t\treturn ErrUnknownChannelMode\n\t}\n\tif m == ChannelModeRestricted {\n\t\tif sideKey == \"\" {\n\t\t\treturn ErrNoSideKey\n\t\t}\n\t\tr.sideKey = sideKey\n\t}\n\tr.mode = m\n\treturn nil\n}", "func (s *Sensor) SetMode(m string) *Sensor {\n\tif s.err != nil {\n\t\treturn s\n\t}\n\ts.err = setAttributeOf(s, mode, m)\n\treturn s\n}", "func (b *Base) setMode(mode proto.Mode) {\n\tb.messageChan <- &proto.Update{\n\t\tMode: mode,\n\t}\n}", "func (w *WarpState) SetMode(\n\tuser string,\n\tmode warp.Mode,\n) error {\n\tuserState, ok := w.users[user]\n\tif !ok {\n\t\treturn errors.Trace(\n\t\t\terrors.Newf(\"Unknown user: %s\", user),\n\t\t)\n\t}\n\n\tuserState.mode = mode\n\tw.users[user] = userState\n\n\treturn nil\n}", "func (option *SetAttribute) SetMode(value wire.Mode) {\n\toption.Mode = &value\n}", "func (lp *LoadPoint) SetMode(mode api.ChargeMode) {\n\tlp.Lock()\n\tdefer lp.Unlock()\n\n\tlp.log.INFO.Printf(\"set charge mode: %s\", string(mode))\n\n\t// apply immediately\n\tif lp.Mode != mode {\n\t\tlp.Mode = mode\n\t\tlp.publish(\"mode\", mode)\n\n\t\t// immediately allow pv mode activity\n\t\tlp.elapsePVTimer()\n\n\t\tlp.requestUpdate()\n\t}\n}", "func (q *Queue) SetMode(mode uint8) error {\n\tif q.cH == nil {\n\t\treturn ErrNotInitialized\n\t}\n\n\tif q.cQh == nil {\n\t\treturn ErrNotInitialized\n\t}\n\n\tC.nfq_set_mode(q.cQh, C.u_int8_t(mode), 0xffff)\n\n\treturn nil\n}", "func SetMode(mode Mode) error {\n\tmodeString := fmt.Sprintf(\"%d\", mode)\n\treturn writeSysfsValue(\"mode\", modeString)\n}", "func (device *LaserRangeFinderBricklet) SetMode(mode Mode) (err error) {\n\tvar buf bytes.Buffer\n\tbinary.Write(&buf, binary.LittleEndian, mode)\n\n\tresultBytes, err := device.device.Set(uint8(FunctionSetMode), buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 8 {\n\t\t\treturn fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 8)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tbytes.NewBuffer(resultBytes[8:])\n\n\t}\n\n\treturn nil\n}", "func (i *CommitInfo) SetMode(mode WriteMode, rev string) {\n\tif mode == WriteModeUpdate {\n\t\ti.Mode = newWriteModeUpdate(rev)\n\t} else {\n\t\ti.Mode = mode\n\t}\n}", "func (e *Encoder) SetMode(mode Mode) {\n\te.mode = mode\n}", "func (d *Detector) SetMode(x int) error {\n\terrno := C.fvad_set_mode(d.fvad, C.int(x))\n\tif errno != 0 {\n\t\treturn fmt.Errorf(\"invalid mode: %v\", x)\n\t}\n\treturn nil\n}", "func (_options *CreateSecretLocksBulkOptions) SetMode(mode string) *CreateSecretLocksBulkOptions {\n\t_options.Mode = core.StringPtr(mode)\n\treturn _options\n}", "func (_options *CreateSecretVersionLocksBulkOptions) SetMode(mode string) *CreateSecretVersionLocksBulkOptions {\n\t_options.Mode = core.StringPtr(mode)\n\treturn _options\n}", "func (irc *IrcCon) ChMode(user, channel, mode string) {\n\tirc.Send(\"MODE \" + channel + \" \" + mode + \" \" + user)\n}", "func (o *UserDisco) SetMode(v UserModeEnum) {\n\to.Mode = &v\n}", "func (r *Recorder) SetMode(newMode Mode) {\n\tif r.mode == newMode {\n\t\treturn\n\t}\n\tr.mode = newMode\n\tr.logger.Logf(\"recorder %s from/to %v\", r.mode.toHumanString(), r.recordsDir)\n}", "func (o *ProjectDeploymentRuleResponse) SetMode(v string) {\n\to.Mode = v\n}", "func (c *Client) Mode(target, mode, arg string) error {\n\tif len(arg) > 0 {\n\t\treturn c.Raw(\"MODE %s %s %s\", target, mode, arg)\n\t}\n\treturn c.Raw(\"MODE %s %s\", target, mode)\n}", "func (o *EquipmentFanControl) SetMode(v string) {\n\to.Mode = &v\n}", "func (s *Autotune) SetMode(v string) *Autotune {\n\ts.Mode = &v\n\treturn s\n}", "func (s *SessionSpecification) SetMode(v string) *SessionSpecification {\n\ts.Mode = &v\n\treturn s\n}", "func (o *HyperflexVmSnapshotInfoAllOf) SetMode(v string) {\n\to.Mode = &v\n}", "func (o *QtreeCreateRequest) SetMode(newValue string) *QtreeCreateRequest {\n\to.ModePtr = &newValue\n\treturn o\n}", "func (key Key) SetMode(mode C.DWORD) error {\n\tif C.CryptSetKeyParam(key.hKey, C.KP_MODE, C.LPBYTE(unsafe.Pointer(&mode)), 0) == 0 {\n\t\treturn getErr(\"Error setting mode for key\")\n\t}\n\treturn nil\n}", "func (s *AutoMLJobConfig) SetMode(v string) *AutoMLJobConfig {\n\ts.Mode = &v\n\treturn s\n}", "func (c *Client) SetTestMode(testMode bool) {\n\tc.testMode = testMode\n}", "func (s *Encryption) SetMode(v string) *Encryption {\n\ts.Mode = &v\n\treturn s\n}", "func (d *ModeDiff) setMode(mode rune) {\n\td.pos.setMode(mode)\n\td.neg.unsetMode(mode)\n}", "func (modes *Modes) SwitchMode(mode Mode) {\n\tif modes.Mode != nil {\n\t\tmodes.Mode.Hide()\n\t}\n\tmodes.Mode = mode\n\tmodes.Mode.Show()\n}", "func (s *Swarm64) ChangeMode(mode Mode) {\n\ts.mode = mode\n}", "func (s *UtteranceSpecification) SetMode(v string) *UtteranceSpecification {\n\ts.Mode = &v\n\treturn s\n}", "func (s *TabularJobConfig) SetMode(v string) *TabularJobConfig {\n\ts.Mode = &v\n\treturn s\n}", "func (s *ContainerDefinition) SetMode(v string) *ContainerDefinition {\n\ts.Mode = &v\n\treturn s\n}", "func SetMode(mode uint8) {\n\t_, _, errno := syscall.Syscall(syscall.SYS_IOCTL, spiFile.Fd(), spiIOCWrMode, uintptr(unsafe.Pointer(&mode)))\n\tif errno != 0 {\n\t\terr := syscall.Errno(errno)\n\t\tlog.Fatal(err)\n\t}\n}", "func (option *CreateDirectory) SetMode(value wire.Mode) {\n\toption.Mode = &value\n}", "func (o *JourneyJourneyResultsParams) SetMode(mode []string) {\n\to.Mode = mode\n}", "func (s *InferenceExecutionConfig) SetMode(v string) *InferenceExecutionConfig {\n\ts.Mode = &v\n\treturn s\n}", "func (a *AutoRollNotifier) SendModeChange(ctx context.Context, user, mode, message string) {\n\ta.send(ctx, &tmplVars{\n\t\tMessage: message,\n\t\tMode: mode,\n\t\tUser: user,\n\t}, subjectTmplModeChange, bodyTmplModeChange, notifier.SEVERITY_WARNING, MSG_TYPE_MODE_CHANGE, nil)\n}", "func (plan Handle) SetCompatibilityMode(mode CompatibilityMode) {\n\terr := Result(C.cufftSetCompatibilityMode(\n\t\tC.cufftHandle(plan),\n\t\tC.cufftCompatibility(mode)))\n\tif err != SUCCESS {\n\t\tpanic(err)\n\t}\n}", "func (v *ToggleButton) SetMode(drawIndicator bool) {\n\tC.gtk_toggle_button_set_mode(v.native(), gbool(drawIndicator))\n}", "func (enc *C14NEncoder) SetMode(mode C14NMode) {\n\tenc.mode = mode\n}", "func (s *EvaluationFormScoringStrategy) SetMode(v string) *EvaluationFormScoringStrategy {\n\ts.Mode = &v\n\treturn s\n}", "func (a *Client) UpdateMode(params *UpdateModeParams) (*UpdateModeOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewUpdateModeParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"updateMode\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/mode/{subject}\",\n\t\tProducesMediaTypes: []string{\"application/json; qs=0.5\", \"application/vnd.schemaregistry+json; qs=0.9\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\", \"application/vnd.schemaregistry+json\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &UpdateModeReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*UpdateModeOK), nil\n\n}", "func (ch *Channel) SetOperationMode(mode fgen.OperationMode) error {\n\tswitch mode {\n\tcase fgen.BurstMode:\n\t\treturn ch.Set(\"MENA1;MTYP5\\n\")\n\tcase fgen.ContinuousMode:\n\t\treturn ch.Set(\"MENA0\\n\")\n\t}\n\treturn errors.New(\"bad fgen operation mode\")\n}", "func (a *API) SetTestMode(t bool) {\n\ta.Sandbox = t\n}", "func (c *CursesConfig) SetCommandMode(mode tileslib.CommandModeType) {\n\tc.base.CommandMode = mode\n}", "func (s *AvailSuppression) SetMode(v string) *AvailSuppression {\n\ts.Mode = &v\n\treturn s\n}", "func SetConsoleMode(hConsoleHandle HANDLE, dwMode DWORD) bool {\n\tret1 := syscall3(setConsoleMode, 2,\n\t\tuintptr(hConsoleHandle),\n\t\tuintptr(dwMode),\n\t\t0)\n\treturn ret1 != 0\n}", "func (a *Client) UpdateMode(params *UpdateModeParams) (*UpdateModeOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewUpdateModeParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"updateMode\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/mode/{subject}\",\n\t\tProducesMediaTypes: []string{\"application/json; qs=0.5\", \"application/vnd.schemaregistry+json; qs=0.9\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\", \"application/vnd.schemaregistry+json\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &UpdateModeReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*UpdateModeOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for updateMode: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func SetFlightMode(iC *InterfaceConfig, mode string) {\n if mode==\"gps\" {\n\t\t\t// pca9685.Write(CHANNEL(iC.(\"U\",\"channel\")), VALUE(iC.(\"U\", \"gps\")));\n iC.pca.SetChannel(Uchannel, 0, iC.GpsModeFlipSwitchDutyCycle)\n\t\t\tprintln(\"Setting flight mode: \"+ mode)\n\t\t} else if mode==\"failsafe\" {\n\t\t\t// pca9685.Write(CHANNEL(iC.(\"U\",\"channel\")), VALUE(iC.(\"U\", \"failsafe\")));\n iC.pca.SetChannel(Uchannel, 0, iC.FailsafeModeFlipSwitchDutyCycle)\n\t\t\tprintln(\"Setting flight mode: \"+ mode)\n\t\t} else if mode==\"selectable\" {\n\t\t\t// pca9685.Write(CHANNEL(iC.(\"U\",\"channel\")), VALUE(iC.(\"U\", \"selectable\")));\n iC.pca.SetChannel(Uchannel, 0, iC.SelectableModeFlipSwitchDutyCycle)\n\t\t\tprintln(\"Setting flight mode: \"+ mode)\n\t\t}\n}", "func (a *DeviceAPI) SetDeviceMode(ctx context.Context, req *api.SetDeviceModeRequest) (*api.SetDeviceModeResponse, error) {\n\tlogInfo := \"api/appserver_serves_ui/SetDeviceMode org=\" + strconv.FormatInt(req.OrgId, 10)\n\n\t// verify if user is global admin\n\tu, err := devmod.NewValidator(a.st).GetUser(ctx)\n\tif err != nil {\n\t\tlog.WithError(err).Error(logInfo)\n\t\treturn &api.SetDeviceModeResponse{}, status.Errorf(codes.Internal, \"unable to verify user: %s\", err.Error())\n\t}\n\t// is user is not global admin, user must have accesss to this organization\n\tif !u.IsGlobalAdmin {\n\t\tif valid, err := organization.NewValidator(a.st).ValidateOrganizationAccess(ctx, authcus.Read, req.OrgId); !valid || err != nil {\n\t\t\treturn &api.SetDeviceModeResponse{}, status.Errorf(codes.Unauthenticated, \"authentication failed: %s\", err)\n\t\t}\n\t}\n\n\tdevClient := mxpcli.Global.GetM2MDeviceServiceClient()\n\n\tresp, err := devClient.SetDeviceMode(ctx, &pb.SetDeviceModeRequest{\n\t\tOrgId: req.OrgId,\n\t\tDevId: req.DevId,\n\t\tDevMode: pb.DeviceMode(req.DevMode),\n\t})\n\tif err != nil {\n\t\tlog.WithError(err).Error(logInfo)\n\t\treturn &api.SetDeviceModeResponse{}, status.Errorf(codes.Unavailable, err.Error())\n\t}\n\n\treturn &api.SetDeviceModeResponse{\n\t\tStatus: resp.Status,\n\t}, status.Error(codes.OK, \"\")\n}", "func (c *Context) SetTestMode(on int) (err int) {\n\treturn int(C.rtlsdr_set_testmode((*C.rtlsdr_dev_t)(c.dev),\n\t\tC.int(on)))\n}", "func SetEnforceMode(mode int) error {\n\treturn setEnforceMode(mode)\n}", "func Bcm2835_pwm_set_mode(Channel byte, Markspace byte, Enabled byte) {\n\tcChannel, _ := (C.uint8_t)(Channel), cgoAllocsUnknown\n\tcMarkspace, _ := (C.uint8_t)(Markspace), cgoAllocsUnknown\n\tcEnabled, _ := (C.uint8_t)(Enabled), cgoAllocsUnknown\n\tC.bcm2835_pwm_set_mode(cChannel, cMarkspace, cEnabled)\n}", "func (conn *Conn) Mode(t string, modestring ...string) {\n\tmode := strings.Join(modestring, \" \")\n\tif mode != \"\" {\n\t\tmode = \" \" + mode\n\t}\n\tconn.Raw(MODE + \" \" + t + mode)\n}", "func (x *XBee) SetAPIMode(mode api.EscapeMode) error {\n\tif mode != api.EscapeModeInactive && mode != api.EscapeModeActive {\n\t\treturn api.ErrInvalidAPIEscapeMode\n\t}\n\tx.apiMode = mode\n\treturn nil\n}", "func (m *Model) SetCursorMode(mode CursorMode) tea.Cmd {\n\tm.cursorMode = mode\n\tm.blink = m.cursorMode == CursorHide || !m.focus\n\tif mode == CursorBlink {\n\t\treturn Blink\n\t}\n\treturn nil\n}", "func (i Mode) Set(v string) error {\n\tswitch v {\n\tcase string(SimpleMode):\n\t\t// nolint:ineffassign\n\t\ti = SimpleMode\n\tcase string(RingMode):\n\t\t// nolint:ineffassign\n\t\ti = RingMode\n\tdefault:\n\t\treturn fmt.Errorf(\"mode %s not supported. list of supported modes: simple (default), ring\", v)\n\t}\n\treturn nil\n}", "func (h *header) setMode(md mode) {\n\th.LiVnMode = (h.LiVnMode & 0xf8) | uint8(md)\n}", "func (_options *UpdateBotManagementOptions) SetFightMode(fightMode bool) *UpdateBotManagementOptions {\n\t_options.FightMode = core.BoolPtr(fightMode)\n\treturn _options\n}", "func (d *Device) SetMux(mode int) error {\n\tval := strconv.Itoa(mode)\n\td.Set(TCPMultiple, val)\n\t_, err := d.Response(pause)\n\treturn err\n}", "func (iface *Iface) SetEdgeMode() error {\n\tiface.mode = EdgeMode\n\treturn nil\n}", "func (dev *Device) SetDepthMode(mode FrameMode) int {\n\treturn int(C.freenect_set_depth_mode(dev.ptr(), *mode.ptr()))\n}", "func (dev *Device) SetVideoMode(mode FrameMode) int {\n\treturn int(C.freenect_set_video_mode(dev.ptr(), *mode.ptr()))\n}", "func (s *Swarm32) ChangeMode(mode Mode) {\n\ts.mode = mode\n\tvar m Mode\n\tif s.mode == m.Constriction() {\n\t\tif math.IsNaN(float64(s.constriction)) {\n\t\t\tpanic(\"Constriction is nan: Cognative + Social mus be > 4\")\n\t\t}\n\t}\n}", "func (d *Dev) SetInputMode(inputMode InputMode) error {\n\td.mu.Lock()\n\tdefer d.mu.Unlock()\n\td.inputMode = inputMode\n\t_, err := d.readRaw()\n\treturn err\n}", "func WithMode(m Mode) Option {\n\treturn func(lp *Longpoll) error {\n\t\tlp.Mode = m\n\n\t\treturn nil\n\t}\n}", "func (s Sequence) Mode(m Mode) Sequence {\n\ts.mode = m\n\treturn s\n}", "func Mode(m mode) PinningOption {\n\treturn func(o *Pinning) {\n\t\to.Mode = m\n\t}\n}", "func (device *DCV2Bricklet) SetDriveMode(mode DriveMode) (err error) {\n\tvar buf bytes.Buffer\n\tbinary.Write(&buf, binary.LittleEndian, mode)\n\n\tresultBytes, err := device.device.Set(uint8(FunctionSetDriveMode), buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 8 {\n\t\t\treturn fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 8)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tbytes.NewBuffer(resultBytes[8:])\n\n\t}\n\n\treturn nil\n}", "func (y *YeeLight) SetPower(power PowerValue, effect Effect, duration int, mode TurnOnValue) (*Answer, error) {\n\tif !power.isValid() {\n\t\treturn nil, errors.Wrapf(ErrInvalidType, \"invalid power value: %v\", power)\n\t}\n\tif !isValidDuration(duration) {\n\t\treturn nil, errors.Wrapf(ErrInvalidType, \"invalid duration value: %d\", duration)\n\t}\n\tcmd, err := y.newCommand(\"set_power\", []interface{}{power, effect, duration, mode})\n\tif err != nil {\n\t\treturn nil, errors.WithStack(err)\n\t}\n\treturn y.sendCommand(cmd)\n}", "func SetAPIMode(ctx context.Context) context.Context {\n\treturn context.WithValue(ctx, ResultModeContextKey, ResultModeAPI)\n}", "func Bcm2835_spi_setDataMode(Mode byte) {\n\tcMode, _ := (C.uint8_t)(Mode), cgoAllocsUnknown\n\tC.bcm2835_spi_setDataMode(cMode)\n}", "func (gpio *RpiGpio) Mode(m Mode) error {\n\tif m != GPIO && m != PI {\n\t\treturn fmt.Errorf(\"Mode must be GPIO or PI\")\n\t}\n\tgpio.mode = m\n\treturn nil\n}", "func SetLogMode(value string) {\n\tswitch value {\n\tcase DebugMode, \"\":\n\t\tvglogModeCode = debugCode\n\tcase ReleaseMode:\n\t\tvglogModeCode = releaseCode\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"vglog mode unknown: %s\", value))\n\t}\n\tif value == \"\" {\n\t\tvalue = DebugMode\n\t}\n\tvglogModeName = value\n}", "func (_this *DigitalPin) SetPinMode(mode uint8) error {\n\t_this.lock.Lock()\n\tdefer _this.lock.Unlock()\n\n\tif !isPinExported(_this.realPin) || !_this.useable {\n\t\texportPin(_this.realPin)\n\t\tif !isPinExported(_this.realPin) {\n\t\t\treturn ErrPinNotExported\n\t\t}\n\t}\n\treturn setPinMode(_this.realPin, mode)\n}", "func (v Account) SetSilenceMode(params AccountSetSilenceModeParams) (bool, error) {\n\tr, err := v.API.Request(\"account.setSilenceMode\", params)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn decodeBoolIntResponse(r)\n}", "func (b *OGame) SetVacationMode() error {\n\treturn b.WithPriority(taskRunner.Normal).SetVacationMode()\n}", "func (fb *FlowBox) SetSelectionMode(mode SelectionMode) {\n\tC.gtk_flow_box_set_selection_mode(fb.native(), C.GtkSelectionMode(mode))\n}", "func (void *VoidResponse) SetParseMode(mode string) *VoidResponse {\n\tbody := JSON{\n\t\t\"parse_mode\": mode,\n\t}\n\tvoid.Request = void.Request.Send(body)\n\n\treturn void\n}", "func (e *Input) SetEchoMode(m EchoMode) {\n\te.echoMode = m\n}", "func (t *Tortoise) Mode() Mode {\n\tt.mu.Lock()\n\tdefer t.mu.Unlock()\n\tif t.trtl.isFull {\n\t\treturn Full\n\t}\n\treturn Verifying\n}", "func (htpc *HttpProcessorConfig) Mode(deliveryMode HttpClientMode) *HttpProcessorConfig {\n\thtpc.mode = deliveryMode\n\treturn htpc\n}", "func SetEnvironmentMode(state string) {\n\tenvState = state\n\n\tif IsInDevMode() {\n\t\tlog.Println(\"Application is running in development mode\")\n\t} else {\n\t\tlog.Println(\"Application is running in release mode\")\n\t}\n}", "func SetFirewallMode(v string) { firewallMode.Store(v) }", "func (f *Factory) WithMode(mode sdk.BroadcastMode) *Factory {\n\tf.mode = mode\n\treturn f\n}", "func (f *Fridge) SetEcoMode(useEcoMode bool) {\n\tlog.Warnf(\"SetEcoMode: %v\", useEcoMode)\n\ts := f.GetStatusReport().Settings\n\tif s.EcoMode != useEcoMode {\n\t\ts.EcoMode = useEcoMode\n\t\tf.settingsC <- s\n\t}\n}", "func (self *PhysicsP2) SetSleepModeA(member int) {\n self.Object.Set(\"sleepMode\", member)\n}", "func (dbm *DBManager) SetBatchMode(mode bool) {\n\tdbm.batchMode = mode\n\t//if the batch mode is turned off, close DB directly\n\tif !mode {\n\t\tdbm.closeDB()\n\t}\n}", "func (mc *MockContiv) SetSTNMode(stnMode bool) {\n\tmc.stnMode = stnMode\n}" ]
[ "0.75422305", "0.7333513", "0.72548175", "0.71374756", "0.7098122", "0.7036112", "0.7028319", "0.6999698", "0.6992441", "0.6967338", "0.68850285", "0.6764373", "0.67014676", "0.66776246", "0.6617654", "0.65172714", "0.64836884", "0.64371777", "0.64287746", "0.6417797", "0.6388564", "0.63228995", "0.63065624", "0.62954485", "0.6291088", "0.6275549", "0.62453234", "0.62200886", "0.61925066", "0.61309385", "0.6126139", "0.6108641", "0.6083284", "0.6082428", "0.606437", "0.60530216", "0.6052523", "0.60477597", "0.6005569", "0.5991353", "0.5982294", "0.5961046", "0.59419346", "0.5939898", "0.59374046", "0.59135294", "0.59126663", "0.5911811", "0.58869797", "0.5867836", "0.58594203", "0.5850929", "0.5776328", "0.5774187", "0.57428396", "0.57315916", "0.55931556", "0.55630034", "0.55285364", "0.5478626", "0.54746395", "0.54594326", "0.541321", "0.538077", "0.5344816", "0.5333703", "0.53283477", "0.532352", "0.5316781", "0.5297905", "0.52918166", "0.5263948", "0.52601236", "0.5246655", "0.5242927", "0.5226065", "0.5216471", "0.5203621", "0.51780325", "0.51454985", "0.51387465", "0.51301676", "0.50923127", "0.5077569", "0.5068614", "0.5067996", "0.50375736", "0.5011774", "0.49762842", "0.4966058", "0.49455884", "0.49440017", "0.49424627", "0.49352607", "0.49290138", "0.49270886", "0.4891504", "0.4872688", "0.48565137", "0.4855729" ]
0.7240473
3
isMode determines whether the bot is in the given mode.
func (r *AutoRoller) isMode(s string) bool { return r.modeHistory.CurrentMode().Mode == s }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func IsValidMode(mode *string) bool {\n\tmodes := []string{Broker, Hybrid, Receiver, Rest, TokenServer, HTTPOnly, HTTPWithNoRest}\n\treturn StrContains(modes, *mode)\n}", "func (o *UserDisco) HasMode() bool {\n\tif o != nil && o.Mode != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c Conf) IsModeAllowed(mode Mode) bool {\n\tfor _, m := range c.AllowedModes {\n\t\tif m == mode {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (o *EquipmentFanControl) HasMode() bool {\n\tif o != nil && o.Mode != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (d ModeDiff) isUserMode(mode rune) (is bool) {\n\tif d.userPrefixes != nil {\n\t\tis = d.modeBit(mode) > 0\n\t}\n\treturn\n}", "func TestMode() bool {\n\treturn mode&TestModeFlag == TestModeFlag\n}", "func (t *Tortoise) Mode() Mode {\n\tt.mu.Lock()\n\tdefer t.mu.Unlock()\n\tif t.trtl.isFull {\n\t\treturn Full\n\t}\n\treturn Verifying\n}", "func (app *Application) GetMode() int {\n return app.mode\n}", "func (s *Server) GetMode() Mode {\n\treturn s.mode\n}", "func IsStatsMode() bool {\n\tc := GetConfig()\n\treturn c.FederatedPromInterval != \"\"\n}", "func (s *UtteranceSpecification) SetMode(v string) *UtteranceSpecification {\n\ts.Mode = &v\n\treturn s\n}", "func (config *Config) IsAgentMode() bool {\n\treturn config.IsHost || config.BrownfieldToken != \"\" || config.CommandPollingToken != \"\"\n}", "func (m *Monitor) SupportsMode(mode VideoMode) bool {\n\treturn m.internal.supportsMode(mode)\n}", "func (s *SessionSpecification) SetMode(v string) *SessionSpecification {\n\ts.Mode = &v\n\treturn s\n}", "func (o *EquipmentFanControl) GetMode() string {\n\tif o == nil || o.Mode == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Mode\n}", "func (s *Autotune) SetMode(v string) *Autotune {\n\ts.Mode = &v\n\treturn s\n}", "func Mode(s string) (core.Mode, error) {\n\tfor _, mode := range core.Modes {\n\t\tif string(mode) == s {\n\t\t\treturn mode, nil\n\t\t}\n\t}\n\treturn core.NoMode, fmt.Errorf(\"%q is not a supported mode\", s)\n}", "func (e *Exporter) GetMode() (string, bool) {\n\tmodeInt := e.sched.getMode()\n\tif modeInt == 0 {\n\t\treturn \"default\", true\n\t}\n\n\tfor modeName, modeNum := range e.modes {\n\t\tif modeNum == modeInt {\n\t\t\treturn modeName, false\n\t\t}\n\t}\n\n\tpanic(\"Unknown mode found\")\n}", "func (o *ProjectDeploymentRuleResponse) GetMode() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Mode\n}", "func (o *ProjectDeploymentRuleResponse) SetMode(v string) {\n\to.Mode = v\n}", "func (o *HyperflexVmSnapshotInfoAllOf) HasMode() bool {\n\tif o != nil && o.Mode != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (s *AvailSuppression) SetMode(v string) *AvailSuppression {\n\ts.Mode = &v\n\treturn s\n}", "func (s *Encryption) SetMode(v string) *Encryption {\n\ts.Mode = &v\n\treturn s\n}", "func (irc *IrcCon) ChMode(user, channel, mode string) {\n\tirc.Send(\"MODE \" + channel + \" \" + mode + \" \" + user)\n}", "func (o *HyperflexVmSnapshotInfoAllOf) GetMode() string {\n\tif o == nil || o.Mode == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Mode\n}", "func (o *UserDisco) GetMode() UserModeEnum {\n\tif o == nil || o.Mode == nil {\n\t\tvar ret UserModeEnum\n\t\treturn ret\n\t}\n\treturn *o.Mode\n}", "func (v *ToggleButton) GetMode() bool {\n\tc := C.gtk_toggle_button_get_mode(v.native())\n\treturn gobool(c)\n}", "func (s *ContainerDefinition) SetMode(v string) *ContainerDefinition {\n\ts.Mode = &v\n\treturn s\n}", "func (s *AutoMLJobConfig) SetMode(v string) *AutoMLJobConfig {\n\ts.Mode = &v\n\treturn s\n}", "func (o *UserDisco) SetMode(v UserModeEnum) {\n\to.Mode = &v\n}", "func (c *Client) Mode(target, mode, arg string) error {\n\tif len(arg) > 0 {\n\t\treturn c.Raw(\"MODE %s %s %s\", target, mode, arg)\n\t}\n\treturn c.Raw(\"MODE %s %s\", target, mode)\n}", "func (server *Server) Mode() string {\n\treturn server.mode\n}", "func (ch *Channel) OperationMode() (fgen.OperationMode, error) {\n\tvar mode fgen.OperationMode\n\ts, err := ch.QueryString(\"MENA?\\n\")\n\tif err != nil {\n\t\treturn mode, fmt.Errorf(\"error getting operation mode: %s\", err)\n\t}\n\tswitch s {\n\tcase \"0\":\n\t\treturn fgen.ContinuousMode, nil\n\tcase \"1\":\n\t\tmod, err := ch.QueryString(\"MTYP?\\n\")\n\t\tif err != nil {\n\t\t\treturn mode, fmt.Errorf(\"error determining modulation type: %s\", err)\n\t\t}\n\t\tswitch mod {\n\t\tcase \"5\":\n\t\t\treturn fgen.BurstMode, nil\n\t\tdefault:\n\t\t\treturn mode, fmt.Errorf(\"error determining operation mode, mtyp = %s\", mod)\n\t\t}\n\tdefault:\n\t\treturn mode, fmt.Errorf(\"error determining operation mode from fgen: %s\", s)\n\t}\n}", "func (o *EquipmentFanControl) SetMode(v string) {\n\to.Mode = &v\n}", "func (option *SetAttribute) SetMode(value wire.Mode) {\n\toption.Mode = &value\n}", "func (s *TabularJobConfig) SetMode(v string) *TabularJobConfig {\n\ts.Mode = &v\n\treturn s\n}", "func (s *EvaluationFormScoringStrategy) SetMode(v string) *EvaluationFormScoringStrategy {\n\ts.Mode = &v\n\treturn s\n}", "func Mode() TunnelMode {\n\treturn mode\n}", "func (bp *BusPirate) GetMode() (int, int) {\n\treturn bp.mode, bp.modeversion\n}", "func (o *UserDisco) GetModeOk() (*UserModeEnum, bool) {\n\tif o == nil || o.Mode == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Mode, true\n}", "func (r *AutoRoller) GetMode() string {\n\treturn r.modeHistory.CurrentMode().Mode\n}", "func (r *Receiver) Mode() ChannelMode {\n\treturn r.mode\n}", "func (o *HyperflexVmSnapshotInfoAllOf) SetMode(v string) {\n\to.Mode = &v\n}", "func (m *Mask) GetMode() string {\n\tif m == nil || m.Mode == nil {\n\t\treturn \"\"\n\t}\n\treturn *m.Mode\n}", "func (o *Content) HasAppMode() bool {\n\tif o != nil && o.AppMode != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *QtreeCreateRequest) SetMode(newValue string) *QtreeCreateRequest {\n\to.ModePtr = &newValue\n\treturn o\n}", "func (s *Server) SetMode(mode Mode) {\n\tif s != nil {\n\t\ts.mode = mode\n\t}\n}", "func (list *List) AddMode(nick string, mode rune) (ok bool) {\n\tif !list.isupport.IsPermissionMode(mode) {\n\t\treturn false\n\t}\n\n\tlist.mutex.RLock()\n\tdefer list.mutex.RUnlock()\n\n\tuser := list.index[strings.ToLower(nick)]\n\tif user == nil {\n\t\treturn false\n\t}\n\tif strings.ContainsRune(user.Modes, mode) {\n\t\treturn true\n\t}\n\n\tprevHighest := user.HighestMode()\n\tuser.Modes = list.isupport.SortModes(user.Modes + string(mode))\n\tuser.Prefixes = list.isupport.Prefixes(user.Modes)\n\tuser.updatePrefixedNick()\n\n\t// Only sort if the new mode changed the highest mode.\n\tif list.autosort && prevHighest != user.HighestMode() {\n\t\tlist.sort()\n\t}\n\n\treturn true\n}", "func (conn *Conn) Mode(t string, modestring ...string) {\n\tmode := strings.Join(modestring, \" \")\n\tif mode != \"\" {\n\t\tmode = \" \" + mode\n\t}\n\tconn.Raw(MODE + \" \" + t + mode)\n}", "func (c *AppConfig) GetMode() string {\n\tmode := c.DefaultString(\"GO_ENV\", \"development\")\n\treturn mode\n}", "func (u *UserModeKinds) GetModeBit(mode rune) byte {\n\tfor i := uint(0); i < uint(len(u.modeInfo)); i++ {\n\t\tif u.modeInfo[i][0] == mode {\n\t\t\treturn 1 << i\n\t\t}\n\t}\n\treturn 0\n}", "func (option *CreateDirectory) SetMode(value wire.Mode) {\n\toption.Mode = &value\n}", "func (a *Client) GetMode(params *GetModeParams) (*GetModeOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetModeParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getMode\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/mode/{subject}\",\n\t\tProducesMediaTypes: []string{\"application/json; qs=0.5\", \"application/vnd.schemaregistry+json; qs=0.9\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\", \"application/vnd.schemaregistry+json\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetModeReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*GetModeOK), nil\n\n}", "func (s *InferenceExecutionConfig) SetMode(v string) *InferenceExecutionConfig {\n\ts.Mode = &v\n\treturn s\n}", "func SetMode(m TunnelMode) {\n\tmode = m\n}", "func (i DbInfo) Mode() (ModeFlag, error) {\n\t_, _, mode, err := i.Parse()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn ParseModeFlag(mode)\n}", "func (obj *Global) IsDesktopMode(ctx context.Context) (bool, error) {\n\tresult := &struct {\n\t\tReturn bool `json:\"qReturn\"`\n\t}{}\n\terr := obj.RPC(ctx, \"IsDesktopMode\", result)\n\treturn result.Return, err\n}", "func requireMode(t testing.TB, modes ...mode) {\n\ttestMode := getTestMode()\n\tfor _, mode := range modes {\n\t\tif testMode == string(mode) {\n\t\t\treturn\n\t\t}\n\t}\n\n\tt.Skipf(\"test is in \\\"%s\\\" mode, but it requires one of \\\"%s\\\"\", testMode, modes)\n}", "func (act *ActionTrace) Mode() ResponseMode {\n\treturn act.mode\n}", "func (t *Task) Mode() ModeType {\n\treturn t.mode\n}", "func (o *Config) getMode() (os.FileMode, bool, error) {\n\tmodeOverride := o.OutMode != \"\"\n\tm, err := strconv.ParseUint(\"0\"+o.OutMode, 8, 32)\n\tif err != nil {\n\t\treturn 0, false, err\n\t}\n\tmode := os.FileMode(m)\n\treturn mode, modeOverride, nil\n}", "func (o *Content) GetAppMode() string {\n\tif o == nil || o.AppMode == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.AppMode\n}", "func ReplayMode() bool {\n\treturn mode&ReplayModeFlag == ReplayModeFlag\n}", "func (o SplitTunnelOutput) Mode() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *SplitTunnel) pulumi.StringOutput { return v.Mode }).(pulumi.StringOutput)\n}", "func (m Mode) IsGood() bool {\n\treturn modeGood[m]\n}", "func (p *LegoPort) Mode() (string, error) {\n\treturn stringFrom(attributeOf(p, mode))\n}", "func GetConsoleMode(hConsoleHandle HANDLE, lpMode *uint32) bool {\n\tret1 := syscall3(getConsoleMode, 2,\n\t\tuintptr(hConsoleHandle),\n\t\tuintptr(unsafe.Pointer(lpMode)),\n\t\t0)\n\treturn ret1 != 0\n}", "func (h *header) getMode() mode {\n\treturn mode(h.LiVnMode & 0x07)\n}", "func (o *EquipmentFanControl) GetModeOk() (*string, bool) {\n\tif o == nil || o.Mode == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Mode, true\n}", "func (s *AppServer) Mode() string {\n\treturn string(s.mode)\n}", "func (e *Entry) HasMode() tree.Mode {\n\treturn tree.Mode(e.root.fsInfo.GetReal(e.realPath(), e.st).Filesystem().Permissions)\n}", "func (ds DiscoverStandalone) Mode() string {\n\treturn \"hardware\"\n}", "func (lp *LoadPoint) GetMode() api.ChargeMode {\n\tlp.Lock()\n\tdefer lp.Unlock()\n\treturn lp.Mode\n}", "func IsValidDeviceMode(mode string) bool {\n\tvar legalDeviceMode = map[rune]bool{\n\t\t'r': true,\n\t\t'w': true,\n\t\t'm': true,\n\t}\n\tif mode == \"\" {\n\t\treturn false\n\t}\n\tfor _, c := range mode {\n\t\tif !legalDeviceMode[c] {\n\t\t\treturn false\n\t\t}\n\t\tlegalDeviceMode[c] = false\n\t}\n\treturn true\n}", "func NewMode(cf cipher.CipherFactory) *Mode {\n\treturn &Mode{\n\t\tCf: cf,\n\t\tErrorLog: log.New(ioutil.Discard, \"\", 0),\n\t\tInfoLog: log.New(ioutil.Discard, \"\", 0),\n\t\tDebugLog: log.New(ioutil.Discard, \"\", 0),\n\t}\n}", "func (device *LaserRangeFinderBricklet) GetMode() (mode Mode, err error) {\n\tvar buf bytes.Buffer\n\n\tresultBytes, err := device.device.Get(uint8(FunctionGetMode), buf.Bytes())\n\tif err != nil {\n\t\treturn mode, err\n\t}\n\tif len(resultBytes) > 0 {\n\t\tvar header PacketHeader\n\n\t\theader.FillFromBytes(resultBytes)\n\n\t\tif header.Length != 9 {\n\t\t\treturn mode, fmt.Errorf(\"Received packet of unexpected size %d, instead of %d\", header.Length, 9)\n\t\t}\n\n\t\tif header.ErrorCode != 0 {\n\t\t\treturn mode, DeviceError(header.ErrorCode)\n\t\t}\n\n\t\tresultBuf := bytes.NewBuffer(resultBytes[8:])\n\t\tbinary.Read(resultBuf, binary.LittleEndian, &mode)\n\n\t}\n\n\treturn mode, nil\n}", "func (me TxsdFeBlendTypeMode) IsScreen() bool { return me.String() == \"screen\" }", "func (m *External_Postgresql_SSL) GetMode() *wrappers.StringValue {\n\tif m != nil {\n\t\treturn m.Mode\n\t}\n\treturn nil\n}", "func (c *Client) IsOn() (bool, error) {\n\treq, err := http.NewRequest(http.MethodGet, \"/state/device/power_mode\", nil)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tvar resp respPowerMode\n\tif err = c.do(req, &resp); err != nil {\n\t\treturn false, err\n\t}\n\n\treturn resp.on(), nil\n}", "func (a *Client) UpdateMode(params *UpdateModeParams) (*UpdateModeOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewUpdateModeParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"updateMode\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/mode/{subject}\",\n\t\tProducesMediaTypes: []string{\"application/json; qs=0.5\", \"application/vnd.schemaregistry+json; qs=0.9\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\", \"application/vnd.schemaregistry+json\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &UpdateModeReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*UpdateModeOK), nil\n\n}", "func (m modeKinds) modeBit(mode rune) byte {\n\tm.RLock()\n\tdefer m.RUnlock()\n\n\tfor i := uint(0); i < uint(len(m.userPrefixes)); i++ {\n\t\tif m.userPrefixes[i][0] == mode {\n\t\t\treturn 1 << i\n\t\t}\n\t}\n\treturn 0\n}", "func isLineModeVar(mode lineMode) bool {\n\tif mode&lineModeValue > 0 {\n\t\treturn true\n\t}\n\tif mode&lineModeMulti > 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func (t *Table) Mode() Mode {\n\treturn t.parent.Config.Mode\n}", "func (a *Client) GetMode(params *GetModeParams) (*GetModeOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetModeParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"getMode\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/mode/{subject}\",\n\t\tProducesMediaTypes: []string{\"application/json; qs=0.5\", \"application/vnd.schemaregistry+json; qs=0.9\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\", \"application/octet-stream\", \"application/vnd.schemaregistry+json\", \"application/vnd.schemaregistry.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &GetModeReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetModeOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for getMode: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (o *JourneyJourneyResultsParams) SetMode(mode []string) {\n\to.Mode = mode\n}", "func modeFromString(s string) (Mode, error) {\n\tswitch s {\n\tcase \"replay\":\n\t\treturn Replay, nil\n\tcase \"overwrite\":\n\t\treturn Overwrite, nil\n\tcase \"append\":\n\t\treturn Append, nil\n\t}\n\treturn invalidMode, fmt.Errorf(`invalid mode: \"%s\"`, s)\n}", "func (mc MultiCursor) NavModeIsColumn() bool {\n\treturn mc.navMode == Column\n}", "func (s *Sensor) Mode() (string, error) {\n\treturn stringFrom(attributeOf(s, mode))\n}", "func (c *CmdBuff) InCmdMode() bool {\n\tc.mx.RLock()\n\tdefer c.mx.RUnlock()\n\n\tif !c.active {\n\t\treturn false\n\t}\n\n\treturn len(c.buff) > 0\n}", "func (i *CommitInfo) GetMode() (mode WriteMode, rev string) {\n\t// Update Mode if empty\n\ti.checkMode()\n\n\tswitch m := i.Mode.(type) {\n\tcase *writeModeUpdate:\n\t\treturn WriteModeUpdate, m.Rev\n\tcase WriteMode:\n\t\treturn m, \"\"\n\t}\n\n\treturn \"\", \"\"\n}", "func (p *Panorama) ViewMode() core.ViewMode {\n\treturn p.viewMode\n}", "func (o *ProjectDeploymentRuleResponse) GetModeOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Mode, true\n}", "func (obj *Global) IsPersonalMode(ctx context.Context) (bool, error) {\n\tresult := &struct {\n\t\tReturn bool `json:\"qReturn\"`\n\t}{}\n\terr := obj.RPC(ctx, \"IsPersonalMode\", result)\n\treturn result.Return, err\n}", "func (t *TreeEntry) GetMode() string {\n\tif t == nil || t.Mode == nil {\n\t\treturn \"\"\n\t}\n\treturn *t.Mode\n}", "func (e *Set) AppMode() string {\n\tv, ex := os.LookupEnv(\"APP_MODE\")\n\tif ex {\n\t\treturn v\n\t}\n\tif e.DefaultMode != \"\" {\n\t\treturn e.DefaultMode\n\t}\n\treturn \"development\"\n}", "func NormalMode() bool {\n\treturn mode == 0\n}", "func (m MotionState) Modes(mode string) FieldRestricter {\n\tswitch mode {\n\tcase \"A\":\n\t\treturn m.see\n\t}\n\treturn nil\n}", "func ModeToString(mode Mode) (string, bool) {\n\tm := map[Mode]string{\n\t\tRecursive: linkRecursive,\n\t\tDirect: linkDirect,\n\t\tIndirect: linkIndirect,\n\t\tInternal: linkInternal,\n\t\tNotPinned: linkNotPinned,\n\t\tAny: linkAny,\n\t}\n\ts, ok := m[mode]\n\treturn s, ok\n}", "func IsBroker(mode *string) bool {\n\treturn *mode == Broker\n}", "func (t *TokenFilterKeepTypes) Mode(mode string) *TokenFilterKeepTypes {\n\tt.mode = mode\n\treturn t\n}" ]
[ "0.6802548", "0.6668311", "0.62792194", "0.6217034", "0.62001127", "0.6193424", "0.6187222", "0.6110261", "0.60738635", "0.5964042", "0.59610176", "0.59582126", "0.584919", "0.5847806", "0.58149815", "0.5808729", "0.5804469", "0.57858014", "0.5785639", "0.5766421", "0.57568014", "0.571062", "0.5699275", "0.5695387", "0.56828105", "0.568218", "0.5625645", "0.55942875", "0.5579997", "0.5579965", "0.55756485", "0.55728877", "0.55639637", "0.5558044", "0.55575573", "0.55437034", "0.5521671", "0.5518031", "0.55150306", "0.54707444", "0.5459591", "0.5458345", "0.5456563", "0.54310393", "0.54285103", "0.54179376", "0.5410558", "0.5387256", "0.5379578", "0.53753304", "0.536885", "0.5364881", "0.5363512", "0.53560555", "0.5334171", "0.53319067", "0.53292376", "0.5318325", "0.53158295", "0.53099054", "0.5306687", "0.5303079", "0.5302491", "0.53016144", "0.52981794", "0.5285821", "0.5278761", "0.5274947", "0.5261183", "0.5253308", "0.52417153", "0.5238676", "0.5219543", "0.52100724", "0.52086717", "0.52045256", "0.5197103", "0.51952356", "0.51916796", "0.51909137", "0.51718485", "0.51594794", "0.5156063", "0.5154061", "0.51537025", "0.51381713", "0.5136127", "0.5134757", "0.5130156", "0.511896", "0.51166064", "0.51139605", "0.51074916", "0.50995374", "0.5096136", "0.5080173", "0.5055633", "0.50531685", "0.50515324", "0.50505614" ]
0.75790805
0
GetEmails returns the list of email addresses which are copied on DEPS rolls.
func (r *AutoRoller) GetEmails() []string { r.emailMtx.RLock() defer r.emailMtx.RUnlock() rv := make([]string, len(r.emails)) copy(rv, r.emails) return rv }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) GetEmails() []string {\n\tr.emailsMtx.RLock()\n\tdefer r.emailsMtx.RUnlock()\n\trv := make([]string, len(r.emails))\n\tcopy(rv, r.emails)\n\treturn rv\n}", "func (r *AutoRoller) GetEmails() []string {\n\tr.emailsMtx.RLock()\n\tdefer r.emailsMtx.RUnlock()\n\trv := make([]string, len(r.emails))\n\tcopy(rv, r.emails)\n\treturn rv\n}", "func (o *DataExportQuery) GetEmails() []string {\n\tif o == nil || o.Emails == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn *o.Emails\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v VulnerabilityAssessmentRecurringScansProperties) []string { return v.Emails }).(pulumi.StringArrayOutput)\n}", "func GetEmailList() []Result {\n\tclearSchedule()\n\tupdateSchedule()\n\n\tquery := `SELECT u.pid,\n\t\tu.email, \n\t\ts.subject, \n\t\ts.catalog, \n\t\ts.section, \n\t\ts.title, \n\t\ts.instructor,\n\t\ts.class\n\tFROM USER_INFO u INNER JOIN SECTION_INFO s\n\tON u.level = s.level \n\t\tAND u.term = s.term \n\t\tAND u.subject = s.subject \n\t\tAND u.catalog = s.catalog \n\t\tAND u.section = s.section\n\tWHERE s.enrollment < s.capacity;`\n\n\tretVal := make([]Result, 0, 1)\n\n\trows, err := database.Query(query)\n\tdefer rows.Close()\n\tif err != nil {\n\t\tprintln(err.Error())\n\t\treturn retVal\n\t}\n\n\tfor rows.Next() {\n\t\tresult := Result{}\n\t\trows.Scan(&result.Pid,\n\t\t\t&result.Email,\n\t\t\t&result.Subject,\n\t\t\t&result.Catalog,\n\t\t\t&result.Section,\n\t\t\t&result.Title,\n\t\t\t&result.Instructor,\n\t\t\t&result.Class)\n\n\t\tretVal = append(retVal, result)\n\t}\n\n\treturn retVal\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesResponseOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v VulnerabilityAssessmentRecurringScansPropertiesResponse) []string { return v.Emails }).(pulumi.StringArrayOutput)\n}", "func (out Outlooky) GetMails(arg ...interface{}) (int, []MailItem) {\n\tdefer util.TimeTrack(time.Now(), \"GetMails\")\n\n\t//get inbox\n\tvar (\n\t\tinbox *ole.IDispatch\n\t\tlength = len(arg)\n\t)\n\n\tif length == 0 {\n\t\treturn 0, []MailItem{}\n\t}\n\n\tif length == 1 {\n\t\tswitch arg[0].(type) {\n\t\tcase int:\n\t\t\tinbox = out.GetDefaultFolder(arg[0].(int))\n\t\tcase string:\n\t\t\tinbox = out.GetCustomFolder(arg[0].(string))\n\t\t}\n\t} else {\n\t\tinbox = out.GetCustomFolder(arg[0].(string), arg[1:]...)\n\t}\n\n\tinterfaces := out.GetLeaf(inbox, MailItem{}, true) //Returns []MailItem\n\tmails := make([]MailItem, len(interfaces))\n\n\t//Transfer\n\tfor i, v := range interfaces {\n\t\tmails[i] = v.(MailItem)\n\t}\n\n\tutil.Logger(\"GetMail Fetched: \", len(mails))\n\n\treturn len(mails), mails\n}", "func (o *DataExportQuery) SetEmails(v []string) {\n\to.Emails = &v\n}", "func (e *EmailGetter) PrintEmails() {\n\tfor _, email := range e.Addresses {\n\t\tfmt.Println(email)\n\t}\n}", "func (m *MockedAcknowledgementMailContext) GetUnsentMails() ([]mailData.Mail, error) {\n\treturn []mailData.Mail{}, nil\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesResponsePtrOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *VulnerabilityAssessmentRecurringScansPropertiesResponse) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Emails\n\t}).(pulumi.StringArrayOutput)\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesPtrOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *VulnerabilityAssessmentRecurringScansProperties) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Emails\n\t}).(pulumi.StringArrayOutput)\n}", "func GetAllMail() []Email {\n\tdb, err := dbOpen()\n\tdefer db.Close()\n\tcheckError(err)\n\tstatement, err := db.Prepare(\"SELECT id, sender, receiver, subject, SUBSTR(text, 0, 100) FROM mail ORDER BY id DESC\")\n\tdefer statement.Close()\n\tcheckError(err)\n\trows, err := statement.Query()\n\tcheckError(err)\n\tdefer rows.Close()\n\tmailList := make([]Email, 0)\n\tfor rows.Next() {\n\t\tmail := Email{}\n\t\terr := rows.Scan(&mail.ID, &mail.To, &mail.From, &mail.Subject, &mail.Text)\n\t\tcheckError(err)\n\t\tmailList = append(mailList, mail)\n\t}\n\treturn mailList\n}", "func TestGetEmails(t *testing.T) {\n\n\tdbsql, err := sql.Open(\"postgres\", \"user=postgres dbname=gorm password=simsim sslmode=disable\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tdb, err := InitDB(dbsql)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\taddresses, err := db.GetEmails(344178872)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif addresses != nil {\n\t\tt.Errorf(\"Addresses must be empty %d!\", len(addresses))\n\t}\n\n}", "func (s *Author) Emails() []string {\n\treturn s.email\n}", "func (email *Email) GetAll(client *redis.Client) []string {\n\tkeys, _ := client.ZRangeByScore(emailListKey, &redis.ZRangeBy{\n\t\tMin: \"-inf\",\n\t\tMax: \"+inf\",\n\t}).Result()\n\n\tdata := []string{}\n\tfor _, emailKey := range keys {\n\t\tval, _ := client.HGetAll(emailKey).Result()\n\t\tm, _ := json.Marshal(val)\n\t\tdata = append(data, string(m))\n\t}\n\n\treturn data\n}", "func (o *DataExportQuery) GetEmailsOk() (*[]string, bool) {\n\tif o == nil || o.Emails == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Emails, true\n}", "func (appConfig *AppConfiguration) EmailRecipients() []string {\n\n\tkey := \"email.recipients\"\n\n\tif appConfig.viper.IsSet(key) {\n\t\treturn appConfig.viper.GetStringSlice(key)\n\t}\n\n\treturn nil\n}", "func getOnCallEmails(ctx context.Context, rotationProxyName string) ([]string, error) {\n\tclient := GetClient(ctx)\n\tdata, err := client.sendRequest(ctx, rotationProxyName)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err,\n\t\t\t\"error when querying for on-call rotation\").Err()\n\t}\n\n\tres := &rotationResponse{}\n\tif err = json.Unmarshal([]byte(data), res); err != nil {\n\t\treturn nil, errors.Annotate(err,\n\t\t\t\"failed to unmarshal rotation response (data = %s)\", data).Err()\n\t}\n\n\treturn res.Emails, nil\n}", "func DefaultListEmail(ctx context.Context, db *gorm.DB) ([]*Email, error) {\n\tormResponse := []EmailORM{}\n\tdb, err := ops.ApplyCollectionOperators(db, ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := db.Set(\"gorm:auto_preload\", true).Find(&ormResponse).Error; err != nil {\n\t\treturn nil, err\n\t}\n\tpbResponse := []*Email{}\n\tfor _, responseEntry := range ormResponse {\n\t\ttemp, err := responseEntry.ToPB(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpbResponse = append(pbResponse, &temp)\n\t}\n\treturn pbResponse, nil\n}", "func Emails(emails ...string) string {\n\treturn strings.Join(emails, \", \")\n}", "func GetEmailAddress(s string) ([]string, error) {\n\tvar address []string\n\tfor _, v := range strings.Split(s, \" \") {\n\t\te, err := mail.ParseAddress(v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\taddress = append(address, e.Address)\n\t}\n\treturn address, nil\n}", "func (c *Client) GetMailboxes() []Mailbox {\n\turi := fmt.Sprintf(\"mailboxes.json\")\n\tpage := c.getPage(uri)\n\tvar mailboxList []Mailbox\n\tif err := json.Unmarshal([]byte(page.Items), &mailboxList); err != nil {\n\t\tpanic(err)\n\t}\n\treturn mailboxList\n}", "func (m *Mailer) Deliveries() []Mail {\n\treturn m.deliveries\n}", "func (s *Client) GetMailingList(domain string) ([]uint8, error) {\n\tdata := url.Values{}\n\tdata.Set(\"domain\", domain)\n\turl := fmt.Sprint(baseURL, \"email/ml/list?\", data.Encode())\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tresp, err := s.doRequest(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn resp, nil\n}", "func GetServiceEmailsOnPush(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *ServiceEmailsOnPushState, opts ...pulumi.ResourceOption) (*ServiceEmailsOnPush, error) {\n\tvar resource ServiceEmailsOnPush\n\terr := ctx.ReadResource(\"gitlab:index/serviceEmailsOnPush:ServiceEmailsOnPush\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func Emails(generator *Generator) Criterion {\n\treturn emailsCriterion{g: generator}\n}", "func (a *SendEmailApiService) GetEmailLogs(ctx _context.Context) ApiGetEmailLogsRequest {\n\treturn ApiGetEmailLogsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (o *User) GetOtherMails() []string {\n\tif o == nil || o.OtherMails == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn *o.OtherMails\n}", "func GetOnCallEmails(ctx context.Context, project string) ([]string, error) {\n\tswitch project {\n\tcase \"chromium/src\":\n\t\treturn getOnCallEmails(ctx, \"oncallator:chrome-build-sheriff\")\n\tdefault:\n\t\t// getting on-call rotation not supported\n\t}\n\n\treturn nil, fmt.Errorf(\"could not get on-call rotation for project %s\", project)\n}", "func (k PublicKey) EmailAddresses() []string {\n\treturn nil\n}", "func (o ServiceEmailsOnPushOutput) Recipients() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ServiceEmailsOnPush) pulumi.StringOutput { return v.Recipients }).(pulumi.StringOutput)\n}", "func (a *EmailControllerApiService) GetEmailsPaginated(ctx _context.Context, localVarOptionals *GetEmailsPaginatedOpts) (PageEmailProjection, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue PageEmailProjection\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/emails\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif localVarOptionals != nil && localVarOptionals.InboxId.IsSet() {\n\t\tt:=localVarOptionals.InboxId.Value()\n\t\tif reflect.TypeOf(t).Kind() == reflect.Slice {\n\t\t\ts := reflect.ValueOf(t)\n\t\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\t\tlocalVarQueryParams.Add(\"inboxId\", parameterToString(s.Index(i), \"multi\"))\n\t\t\t}\n\t\t} else {\n\t\t\tlocalVarQueryParams.Add(\"inboxId\", parameterToString(t, \"multi\"))\n\t\t}\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Page.IsSet() {\n\t\tlocalVarQueryParams.Add(\"page\", parameterToString(localVarOptionals.Page.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Size.IsSet() {\n\t\tlocalVarQueryParams.Add(\"size\", parameterToString(localVarOptionals.Size.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.Sort.IsSet() {\n\t\tlocalVarQueryParams.Add(\"sort\", parameterToString(localVarOptionals.Sort.Value(), \"\"))\n\t}\n\tif localVarOptionals != nil && localVarOptionals.UnreadOnly.IsSet() {\n\t\tlocalVarQueryParams.Add(\"unreadOnly\", parameterToString(localVarOptionals.UnreadOnly.Value(), \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"x-api-key\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 200 {\n\t\t\tvar v PageEmailProjection\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func MergeEmailLists(lists ...[]string) []string {\n\tconst (\n\t\tmaxEmailLen = 1000\n\t\tmaxEmails = 50\n\t)\n\tmerged := make(map[string]bool)\n\tfor _, list := range lists {\n\t\tfor _, email := range list {\n\t\t\taddr, err := mail.ParseAddress(email)\n\t\t\tif err != nil || len(addr.Address) > maxEmailLen {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tmerged[addr.Address] = true\n\t\t}\n\t}\n\tvar result []string\n\tfor e := range merged {\n\t\tresult = append(result, e)\n\t}\n\tsort.Strings(result)\n\tif len(result) > maxEmails {\n\t\tresult = result[:maxEmails]\n\t}\n\treturn result\n}", "func (e *EmailTemplatesService) GetTemplates(opt *GetEmailTemplatesOptions, options ...OptionFunc) (*[]EmailTemplate, *Response, error) {\n\treq, err := e.client.newRequest(IDM, \"GET\", \"authorize/identity/EmailTemplate\", opt, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\treq.Header.Set(\"api-version\", emailTemplateAPIVersion)\n\n\tvar bundleResponse struct {\n\t\tTotal int `json:\"total\"`\n\t\tEntry []struct {\n\t\t\tID string `json:\"id\"`\n\t\t} `json:\"entry\"`\n\t}\n\tvar templates []EmailTemplate\n\n\tresp, err := e.client.do(req, &bundleResponse)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\tif bundleResponse.Total == 0 {\n\t\treturn nil, resp, ErrNotFound\n\t}\n\tfor _, t := range bundleResponse.Entry {\n\t\ttemplate, _, err := e.GetTemplateByID(t.ID)\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\t\ttemplates = append(templates, *template)\n\t}\n\treturn &templates, resp, nil\n}", "func (s *AccountsService) ListAccountEmails(accountID string) (*[]EmailInfo, *Response, error) {\n\tu := fmt.Sprintf(\"accounts/%s/emails\", accountID)\n\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tv := new([]EmailInfo)\n\tresp, err := s.client.Do(req, v)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn v, resp, err\n}", "func (out Outlooky) ListMails(mail []MailItem, unread bool) (int, []MailItem) {\n\tdefer util.TimeTrack(time.Now(), \"ListMails\")\n\n\tnewList := make([]MailItem, 0)\n\n\tfor _, item := range mail {\n\t\tif item.UnRead == unread {\n\t\t\tnewList = append(newList, item)\n\t\t}\n\t}\n\n\treturn len(newList), newList\n}", "func (o IntegrationEmailsOnPushOutput) Recipients() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *IntegrationEmailsOnPush) pulumi.StringOutput { return v.Recipients }).(pulumi.StringOutput)\n}", "func (m *MailRepoImpl) List() ([]model.History, error) {\n\tvar results []model.History\n\n\tfindOptions := options.Find()\n\tfindOptions.SetLimit(100)\n\n\t// select all\n\tcur, err := m.Db.Collection(\"emails\").Find(context.Background(), bson.D{}, findOptions)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn results, err\n\t}\n\n\tfor cur.Next(context.Background()) {\n\t\tvar elem model.History\n\t\terr := cur.Decode(&elem)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tresults = append(results, elem)\n\t}\n\n\tif err := cur.Err(); err != nil {\n\t\tlog.Error(err)\n\t\treturn results, err\n\t}\n\n\tcur.Close(context.Background())\n\n\treturn results, nil\n}", "func (c *V3Client) GetAuthenticatedUserEmails(ctx context.Context) ([]*UserEmail, error) {\n\tif MockGetAuthenticatedUserEmails != nil {\n\t\treturn MockGetAuthenticatedUserEmails(ctx)\n\t}\n\n\tvar emails []*UserEmail\n\terr := c.requestGet(ctx, \"/user/emails?per_page=100\", &emails)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn emails, nil\n}", "func (sched *Scheduler) getEmailForSending() ([]*mail.EmailContent, error) {\n\tresp, err := sched.scanFromDB()\n\tif err != nil {\n\t\treturn resp, err\n\t}\n\t// fill FromUser\n\t// why we can set FromUser here?\n\tfor _, emailContent := range resp {\n\t\temailContent.FromUser = &mail.EmailUser{\n\t\t\tName: DefaultFromName,\n\t\t\tEmail: DefaultFromEmail,\n\t\t}\n\t}\n\n\treturn resp, err\n}", "func (m *BrandResource) ListEmailTemplates(ctx context.Context, brandId string, qp *query.Params) ([]*EmailTemplate, *Response, error) {\n\turl := fmt.Sprintf(\"/api/v1/brands/%v/templates/email\", brandId)\n\tif qp != nil {\n\t\turl = url + qp.String()\n\t}\n\n\trq := m.client.CloneRequestExecutor()\n\n\treq, err := rq.WithAccept(\"application/json\").WithContentType(\"application/json\").NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar emailTemplate []*EmailTemplate\n\n\tresp, err := rq.Do(ctx, req, &emailTemplate)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn emailTemplate, resp, nil\n}", "func (s *Client) GetEmailCounters(domain, login, uid string) ([]uint8, error) {\n\tdata := url.Values{}\n\tdata.Set(\"domain\", domain)\n\tdata.Set(\"login\", login)\n\tdata.Set(\"uid\", uid)\n\turl := fmt.Sprint(baseURL, \"email/counters?\", data.Encode())\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tresp, err := s.doRequest(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn resp, nil\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponseOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponse) []string {\n\t\treturn v.ContributorEmails\n\t}).(pulumi.StringArrayOutput)\n}", "func (m *MeetingParticipants) GetAttendees()([]MeetingParticipantInfoable) {\n return m.attendees\n}", "func (c *Client) GetUsers(ctx context.Context, emails []models.Email) ([]*models.User, error) {\n\tvar resp struct {\n\t\tUsers []*models.User `json:\"identities\"`\n\t}\n\n\tvariables := make(map[string]interface{})\n\tvariables[\"emails\"] = emails\n\n\terr := c.transport.Raw(ctx, `\n\t\tquery GetUsers($emails: [Email!]) {\n\t\t\tidentities(emails: $emails) {\n\t\t\t\tid\n\t\t\t\temail\n\t\t\t}\n\t\t}\n\t`, variables, &resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resp.Users, nil\n}", "func GetEmail(ctx *pulumi.Context) string {\n\treturn config.Get(ctx, \"cloudflare:email\")\n}", "func (e *eventsBatcher) getEvents() []*evtsapi.Event {\n\treturn append(e.evts, e.expiredEvts...)\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudDatalabelingV1beta1HumanAnnotationConfig) []string { return v.ContributorEmails }).(pulumi.StringArrayOutput)\n}", "func (m *User) GetMailFolders()([]MailFolderable) {\n return m.mailFolders\n}", "func (s *Store) GetRecipients(ctx context.Context, name string) (*recipients.Recipients, error) {\n\treturn s.getRecipients(ctx, s.idFile(ctx, name))\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponsePtrOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponse) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContributorEmails\n\t}).(pulumi.StringArrayOutput)\n}", "func (a *SendEmailApiService) GetEmailDeliveryReports(ctx _context.Context) ApiGetEmailDeliveryReportsRequest {\n\treturn ApiGetEmailDeliveryReportsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigPtrOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *GoogleCloudDatalabelingV1beta1HumanAnnotationConfig) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ContributorEmails\n\t}).(pulumi.StringArrayOutput)\n}", "func GetIntegrationEmailsOnPush(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *IntegrationEmailsOnPushState, opts ...pulumi.ResourceOption) (*IntegrationEmailsOnPush, error) {\n\tvar resource IntegrationEmailsOnPush\n\terr := ctx.ReadResource(\"gitlab:index/integrationEmailsOnPush:IntegrationEmailsOnPush\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (msg *Message) GetRecipients() ([]string, error) {\n\trecipientLength := 0\n\taddrHeaderList := []string{\"To\", \"Cc\", \"Bcc\"}\n\n\tfor _, field := range addrHeaderList {\n\t\tif addresses, ok := msg.header[field]; ok {\n\t\t\trecipientLength += len(addresses)\n\t\t}\n\t}\n\trecipients := make([]string, recipientLength)\n\tindex := 0\n\n\tfor _, field := range addrHeaderList {\n\t\tif addresses, ok := msg.header[field]; ok {\n\t\t\tfor _, addr := range addresses {\n\t\t\t\tif addr, err := common.ParseAddress(addr); err != nil {\n\t\t\t\t\treturn nil, fmt.Errorf(\n\t\t\t\t\t\t\"m-mail: Unable to parse address. Address: %s, Error: %v\", addr, err)\n\t\t\t\t} else {\n\t\t\t\t\trecipients[index] = addr\n\t\t\t\t\tindex++\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn recipients, nil\n}", "func RequestEmails(clientType, domain string, page *uint32, perPage *uint32) (*http.Request, error) {\n\tparams := NewParams().domain(domain)\n\tif page != nil {\n\t\tif *page == 0 {\n\t\t\ttmp := uint32(1)\n\t\t\tpage = &tmp\n\t\t}\n\t\tparams = params.Page(*page)\n\t\tif perPage != nil {\n\t\t\tif *perPage == 0 {\n\t\t\t\ttmp := uint32(20)\n\t\t\t\tperPage = &tmp\n\t\t\t}\n\t\t\tparams = params.OnPage(*perPage)\n\t\t}\n\t}\n\treturn http.NewRequest(\n\t\thttp.MethodGet,\n\t\tutils.URL(clientType, \"email\", \"list\", url.Values(*params)),\n\t\tnil,\n\t)\n}", "func (r *AutoRoller) SetEmails(e []string) {\n\tr.emailsMtx.Lock()\n\tdefer r.emailsMtx.Unlock()\n\temails := make([]string, len(e))\n\tcopy(emails, e)\n\tr.emails = emails\n}", "func MailRecipients() (string, error) {\n\tvar templist []string\n\tfor _, rec := range os.Args[2:] {\n\t\ttemplist = append(templist, rec)\n\t}\n\trecipients := strings.Join(templist, \" \")\n\treturn recipients, nil\n}", "func GetRecordsByEmail(email string) ([]Password, error) {\n\treturn fetchRecordsForQuery(&Password{Email: email})\n}", "func (r *AutoRoller) SetEmails(e []string) {\n\tr.emailMtx.Lock()\n\tdefer r.emailMtx.Unlock()\n\temails := make([]string, len(e))\n\tcopy(emails, e)\n\tr.emails = emails\n}", "func (test *Test) GetIPs(projectName string) ([]models.IP, error) {\n\treturn tests.NormalIPs, nil\n}", "func (o *CheckoutResponse) GetEmailsSent() int32 {\n\tif o == nil || IsNil(o.EmailsSent) {\n\t\tvar ret int32\n\t\treturn ret\n\t}\n\treturn *o.EmailsSent\n}", "func (a *API) GetListsByEmail(email string) (*Lists, error) {\n\tep, err := url.ParseRequestURI(a.endPoint.String() + \"/api/lists/\" + email)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(http.MethodGet, ep.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := a.Request(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar lists Lists\n\terr = json.Unmarshal(res, &lists)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &lists, nil\n}", "func GetDistributionAddresses() []string {\n\taddrs := make([]string, len(distributionAddresses))\n\tfor i := range distributionAddresses {\n\t\taddrs[i] = distributionAddresses[i]\n\t}\n\treturn addrs\n}", "func (o *InlineResponse2004People) GetEmail() string {\n\tif o == nil || o.Email == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Email\n}", "func (o *SLOCorrectionResponseAttributesModifier) GetEmail() string {\n\tif o == nil || o.Email == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Email\n}", "func (q emailQuery) All(ctx context.Context, exec boil.ContextExecutor) (EmailSlice, error) {\n\tvar o []*Email\n\n\terr := q.Bind(ctx, exec, &o)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"mysql: failed to assign all query results to Email slice\")\n\t}\n\n\tif len(emailAfterSelectHooks) != 0 {\n\t\tfor _, obj := range o {\n\t\t\tif err := obj.doAfterSelectHooks(ctx, exec); err != nil {\n\t\t\t\treturn o, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn o, nil\n}", "func (db *DB) GetPubs() ([]feed.Pub, error) {\n\treturn db.e.GetPubs()\n}", "func (c *Client) List() ([]*Email, error) {\n\terr := c.writeMsg(\"LIST\\r\\n\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmsg, err := c.readMsg(multiLineMessageTerminator)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfmt.Print(\"Listing messages\\n\")\n\n\tvar emails []*Email\n\tlines := strings.Split(msg, \"\\r\\n\")\n\n\t// remove the first item (expecting +OK) and last item (expecing terminator)\n\tlines = lines[1 : len(lines)-2]\n\tfor _, line := range lines {\n\t\temail := NewEmail()\n\t\terr := email.ParseLine(line)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\temails = append(emails, email)\n\t}\n\n\treturn emails, nil\n}", "func (me *XsdGoPkgHasElems_Email) Walk() (err error) {\n\tif fn := WalkHandlers.XsdGoPkgHasElems_Email; me != nil {\n\t\tif fn != nil {\n\t\t\tif err = fn(me, true); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tfor _, x := range me.Emails {\n\t\t\tif err = x.Walk(); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif fn != nil {\n\t\t\tif err = fn(me, false); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (c *BaseMail) GetTo() []*Email {\n\treturn c.To\n}", "func (o ServicePrincipalOutput) NotificationEmailAddresses() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *ServicePrincipal) pulumi.StringArrayOutput { return v.NotificationEmailAddresses }).(pulumi.StringArrayOutput)\n}", "func (l *EmailsList) ToProto() []*emailService.Email {\n\tmails := make([]*emailService.Email, 0, len(l.Emails))\n\tfor _, e := range l.Emails {\n\t\tmails = append(mails, e.ToProto())\n\t}\n\treturn mails\n}", "func (_m *RelationRepo) GetRetrivableEmails(id string) ([]string, error) {\n\tret := _m.Called(id)\n\n\tvar r0 []string\n\tif rf, ok := ret.Get(0).(func(string) []string); ok {\n\t\tr0 = rf(id)\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).([]string)\n\t\t}\n\t}\n\n\tvar r1 error\n\tif rf, ok := ret.Get(1).(func(string) error); ok {\n\t\tr1 = rf(id)\n\t} else {\n\t\tr1 = ret.Error(1)\n\t}\n\n\treturn r0, r1\n}", "func (m *ChatMessage) GetAttachments()([]ChatMessageAttachmentable) {\n return m.attachments\n}", "func getWebhooks(context context.Context, email string) []Webhook {\n\tquery := datastore.NewQuery(\"Webhook\").Filter(\"User =\", email).Limit(50)\n\twebhooks := make([]Webhook, 0, 50)\n\tquery.GetAll(context, &webhooks)\n\treturn webhooks\n}", "func (s *Client) GetListSubscribers(domain, mailList string) ([]uint8, error) {\n\tdata := url.Values{}\n\tdata.Set(\"domain\", domain)\n\tdata.Set(\"maillist\", mailList)\n\turl := fmt.Sprint(baseURL, \"email/ml/subscribers?\", data.Encode())\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tresp, err := s.doRequest(req)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn resp, nil\n}", "func (mw loggingMiddleware) GetAccounts(ctx context.Context) (accounts []Account, err error) {\n\tdefer func(begin time.Time) {\n\t\tmw.logger.Log(\"method\", \"GetAddresses\", \"took\", time.Since(begin), \"err\", err)\n\t}(time.Now())\n\treturn mw.next.GetAccounts(ctx)\n}", "func getTo(e *mail.Envelope) []string {\n\tvar ret []string\n\tfor i := range e.RcptTo {\n\t\tret = append(ret, e.RcptTo[i].String())\n\t}\n\treturn ret\n}", "func (m *InvitedUserMessageInfo) GetCcRecipients()([]Recipientable) {\n val, err := m.GetBackingStore().Get(\"ccRecipients\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]Recipientable)\n }\n return nil\n}", "func (o LookupServiceAccountResultOutput) Email() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServiceAccountResult) string { return v.Email }).(pulumi.StringOutput)\n}", "func (o LookupServiceAccountResultOutput) Email() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupServiceAccountResult) string { return v.Email }).(pulumi.StringOutput)\n}", "func (ms *ManagerService) GetMailboxesHandler(w http.ResponseWriter, r *http.Request) {\n\tvar files []string\n\tvar boxes []string\n\n\troot := ms.GetSectionPropertyOrDefault(\"0box\", \"mboxroot\", \"/var/mail\")\n\n\tfiles, err := ms.GetFilesInPath(root)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tfor _, file := range files {\n\t\tboxes = append(boxes, filepath.Base(file))\n\t}\n\n\tms.WebService.JsonStatusResponse(w, strings.Join(boxes, \",\"), http.StatusOK)\n}", "func getTeams(c *gin.Context) {\n\tvar tms []Team\n\t//Reads from database\n\tif err := db.Preload(\"VolunteerEmails\").Find(&tms).Error; err != nil {\n\t\tcreateNotFoundResponse(c)\n\t\treturn\n\t}\n\t//Checks if user is reporter\n\tif !reporterAuth(c) {\n\t\treturn\n\t}\n\tc.JSON(200, tms)\n}", "func (c *ConfigurationData) GetIgnoreEmailInProd() string {\n\treturn c.v.GetString(varIgnoreEmailInProd)\n}", "func GetEmail(c *gin.Context, auth *oauth2.Config, apiToken *oauth2.Token) (email string, err error) {\n\tvar profile Profile\n\tprofile, err = GetProfile(c, auth, apiToken)\n\temail = profile.Email\n\tif len(profile.Email) == 0 {\n\t\terr = errors.New(\"Empty Email\")\n\t}\n\treturn\n}", "func (b *batch) GetEvents() []*evtsapi.Event {\n\treturn b.evts\n}", "func (s *ServerConnection) MailingListsGetSuffixes() (StringList, error) {\n\tdata, err := s.CallRaw(\"MailingLists.getSuffixes\", nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuffixes := struct {\n\t\tResult struct {\n\t\t\tSuffixes StringList `json:\"suffixes\"`\n\t\t} `json:\"result\"`\n\t}{}\n\terr = json.Unmarshal(data, &suffixes)\n\treturn suffixes.Result.Suffixes, err\n}", "func (a *EmailControllerApiService) DeleteAllEmails(ctx _context.Context) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/emails\"\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarHeaderParams[\"x-api-key\"] = key\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (policy *AccountCreationPolicy) GetVerificationEmailTemplates(ctx context.Context) (*EmailTemplates, error) {\n\terr := getClient(ctx).get(policy.VerificationEmailTemplates.Href, emptyPayload(), policy.VerificationEmailTemplates)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn policy.VerificationEmailTemplates, nil\n}", "func (r *RedisFailoverChecker) GetSentinelsIPs(rf *redisfailoverv1.RedisFailover) ([]string, error) {\n\tsentinels := []string{}\n\trps, err := r.k8sService.GetDeploymentPods(rf.Namespace, GetSentinelName(rf))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, sp := range rps.Items {\n\t\tif sp.Status.Phase == corev1.PodRunning && sp.DeletionTimestamp == nil { // Only work with running pods\n\t\t\tsentinels = append(sentinels, sp.Status.PodIP)\n\t\t}\n\t}\n\treturn sentinels, nil\n}", "func (c *Client) GetAssignees(user, repo string) ([]*User, *Response, error) {\n\tif err := c.checkServerVersionGreaterThanOrEqual(version1_15_0); err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif err := escapeValidatePathSegments(&user, &repo); err != nil {\n\t\treturn nil, nil, err\n\t}\n\tassignees := make([]*User, 0, 5)\n\tresp, err := c.getParsedResponse(\"GET\", fmt.Sprintf(\"/repos/%s/%s/assignees\", user, repo), nil, nil, &assignees)\n\treturn assignees, resp, err\n}", "func (msg MsgLeave) GetSigners() []sdk.AccAddress {\n\treturn []sdk.AccAddress{msg.Signer}\n}", "func (s *ServerConnection) MailingListsGet(query SearchQuery, domainId KId) (MlList, int, error) {\n\tquery = addMissedParametersToSearchQuery(query)\n\tparams := struct {\n\t\tQuery SearchQuery `json:\"query\"`\n\t\tDomainId KId `json:\"domainId\"`\n\t}{query, domainId}\n\tdata, err := s.CallRaw(\"MailingLists.get\", params)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tlist := struct {\n\t\tResult struct {\n\t\t\tList MlList `json:\"list\"`\n\t\t\tTotalItems int `json:\"totalItems\"`\n\t\t} `json:\"result\"`\n\t}{}\n\terr = json.Unmarshal(data, &list)\n\treturn list.Result.List, list.Result.TotalItems, err\n}", "func (o *Post) GetAttachments() []MicrosoftGraphAttachment {\n\tif o == nil || o.Attachments == nil {\n\t\tvar ret []MicrosoftGraphAttachment\n\t\treturn ret\n\t}\n\treturn *o.Attachments\n}", "func (mock *MailgunMock) GetDomainsCalls() []struct {\n\tLimit int\n\tSkip int\n} {\n\tvar calls []struct {\n\t\tLimit int\n\t\tSkip int\n\t}\n\tlockMailgunMockGetDomains.RLock()\n\tcalls = mock.calls.GetDomains\n\tlockMailgunMockGetDomains.RUnlock()\n\treturn calls\n}", "func getTo(e *mail.Envelope) []string {\n\tvar ret []string\n\tfor _, addy := range e.RcptTo {\n\t\tret = append(ret, addy.String())\n\t}\n\treturn ret\n}", "func (s *SourceImportAuthor) GetEmail() string {\n\tif s == nil || s.Email == nil {\n\t\treturn \"\"\n\t}\n\treturn *s.Email\n}", "func fetchAllEmailTemplates(c context.Context, configService configInterface.Interface, projectID string) (map[string]*EmailTemplate, error) {\n\tconfigSet := configInterface.ProjectSet(projectID)\n\tfiles, err := configService.ListFiles(c, configSet)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tret := map[string]*EmailTemplate{}\n\n\t// This runs in a cron job. It is not performance critical, so we don't have\n\t// to fetch files concurrently.\n\tfilenameRegexp, err := emailTemplateFilenameRegexp(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, f := range files {\n\t\tm := filenameRegexp.FindStringSubmatch(f)\n\t\tif m == nil {\n\t\t\t// Not a template file or a template of another instance of luci-notify.\n\t\t\tcontinue\n\t\t}\n\t\ttemplateName := m[1]\n\n\t\tlogging.Infof(c, \"fetching email template from %s:%s\", configSet, f)\n\t\tconfig, err := configService.GetConfig(c, configSet, f, false)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Annotate(err, \"failed to fetch %q\", f).Err()\n\t\t}\n\n\t\tsubject, body, err := mailtmpl.SplitTemplateFile(config.Content)\n\t\tif err != nil {\n\t\t\t// Should not happen. luci-config should not have passed this commit in\n\t\t\t// because luci-notify exposes its validation code to luci-conifg.\n\t\t\tlogging.Warningf(c, \"invalid email template content in %q: %s\", f, err)\n\t\t\tcontinue\n\t\t}\n\n\t\tret[templateName] = &EmailTemplate{\n\t\t\tName: templateName,\n\t\t\tSubjectTextTemplate: subject,\n\t\t\tBodyHTMLTemplate: body,\n\t\t\tDefinitionURL: config.ViewURL,\n\t\t}\n\t}\n\treturn ret, nil\n}" ]
[ "0.7715624", "0.7715624", "0.74551654", "0.6335076", "0.6334469", "0.6235819", "0.6171455", "0.6099361", "0.607236", "0.60697114", "0.60478425", "0.6023765", "0.5858898", "0.58205324", "0.579142", "0.579061", "0.5731278", "0.5712766", "0.5696087", "0.5695946", "0.5594052", "0.55629766", "0.5544716", "0.5475195", "0.5394716", "0.5384075", "0.5349314", "0.531611", "0.52919734", "0.52880406", "0.52509815", "0.5233902", "0.52123344", "0.52072394", "0.5190933", "0.5165197", "0.5110923", "0.51069677", "0.5106531", "0.5086584", "0.5078241", "0.5044253", "0.5027112", "0.50135493", "0.49810943", "0.4978855", "0.49753553", "0.49482536", "0.49285403", "0.49267343", "0.49224585", "0.49210623", "0.49045464", "0.48970836", "0.48907477", "0.48902255", "0.48756334", "0.48719802", "0.48595864", "0.48500416", "0.48374307", "0.48368028", "0.48220524", "0.481038", "0.48019415", "0.47946346", "0.47744325", "0.47740257", "0.47644106", "0.47373194", "0.47185916", "0.4712629", "0.47059175", "0.46816224", "0.46800655", "0.4669129", "0.46562478", "0.46476248", "0.46364933", "0.46260825", "0.4621843", "0.46182284", "0.46182284", "0.46114466", "0.4610975", "0.46086967", "0.4601717", "0.45911813", "0.45907164", "0.459038", "0.45848176", "0.45772305", "0.45545426", "0.45457536", "0.4544399", "0.45411366", "0.45408332", "0.45403314", "0.45373568", "0.4533403" ]
0.76409173
2
SetEmails sets the list of email addresses which are copied on DEPS rolls.
func (r *AutoRoller) SetEmails(e []string) { r.emailMtx.Lock() defer r.emailMtx.Unlock() emails := make([]string, len(e)) copy(emails, e) r.emails = emails }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) SetEmails(e []string) {\n\tr.emailsMtx.Lock()\n\tdefer r.emailsMtx.Unlock()\n\temails := make([]string, len(e))\n\tcopy(emails, e)\n\tr.emails = emails\n}", "func (o *DataExportQuery) SetEmails(v []string) {\n\to.Emails = &v\n}", "func (_m *Repository) EmailUsers(userID uuid.UUID, name string, email string, payloadType string) {\n\t_m.Called(userID, name, email, payloadType)\n}", "func Emails(emails ...string) string {\n\treturn strings.Join(emails, \", \")\n}", "func (s UserSet) SetEmail(value string) {\n\ts.RecordCollection.Set(models.NewFieldName(\"Email\", \"email\"), value)\n}", "func (p *politeiawww) setUserEmailsCache(email string, id uuid.UUID) {\n\tp.Lock()\n\tdefer p.Unlock()\n\tp.userEmails[email] = id\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v VulnerabilityAssessmentRecurringScansProperties) []string { return v.Emails }).(pulumi.StringArrayOutput)\n}", "func Emails(generator *Generator) Criterion {\n\treturn emailsCriterion{g: generator}\n}", "func (s *ServerConnection) MailingListsSet(mlIds KIdList, pattern Ml) (ErrorList, error) {\n\tparams := struct {\n\t\tMlIds KIdList `json:\"mlIds\"`\n\t\tPattern Ml `json:\"pattern\"`\n\t}{mlIds, pattern}\n\tdata, err := s.CallRaw(\"MailingLists.set\", params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terrors := struct {\n\t\tResult struct {\n\t\t\tErrors ErrorList `json:\"errors\"`\n\t\t} `json:\"result\"`\n\t}{}\n\terr = json.Unmarshal(data, &errors)\n\treturn errors.Result.Errors, err\n}", "func (m *MeetingParticipants) SetAttendees(value []MeetingParticipantInfoable)() {\n m.attendees = value\n}", "func (m *User) SetMailFolders(value []MailFolderable)() {\n m.mailFolders = value\n}", "func (g *GitLocal) SetEmail(dir string, email string) error {\n\t// Use GitFake as this is a global setting\n\treturn g.GitFake.SetEmail(dir, email)\n}", "func (b *LDAPAttributesBuilder) Email(values ...string) *LDAPAttributesBuilder {\n\tb.email = make([]string, len(values))\n\tcopy(b.email, values)\n\tb.bitmap_ |= 2\n\treturn b\n}", "func WithEmails(t *template.Template) Config {\n\treturn func(r *router) {\n\t\tr.emails = t\n\t}\n}", "func (e *EmailGetter) PrintEmails() {\n\tfor _, email := range e.Addresses {\n\t\tfmt.Println(email)\n\t}\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesResponseOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v VulnerabilityAssessmentRecurringScansPropertiesResponse) []string { return v.Emails }).(pulumi.StringArrayOutput)\n}", "func (m *MailTips) SetEmailAddress(value EmailAddressable)() {\n err := m.GetBackingStore().Set(\"emailAddress\", value)\n if err != nil {\n panic(err)\n }\n}", "func (s *Store) SetRecipients(ctx context.Context, rs *recipients.Recipients) error {\n\treturn s.saveRecipients(ctx, rs, \"Set Recipients\")\n}", "func (m *UnifiedRoleManagementPolicyNotificationRule) SetNotificationRecipients(value []string)() {\n m.notificationRecipients = value\n}", "func (user *User) SetEmail(newEmail string, currentUser IUser) error {\n\terr := verifyAuthorization(user, currentUser)\n\tif err != nil {\n\t\treturn err\n\t}\n\tuser.Email = newEmail\n\treturn nil\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesPtrOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *VulnerabilityAssessmentRecurringScansProperties) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Emails\n\t}).(pulumi.StringArrayOutput)\n}", "func (mc *ManagerCreate) SetEmail(s string) *ManagerCreate {\n\tmc.mutation.SetEmail(s)\n\treturn mc\n}", "func (s *Author) Emails() []string {\n\treturn s.email\n}", "func (j *Jail) SetIpAddrs(addrs []net.IP) error {\n\tjpps := jailParamList{}\n\tdefer jpps.release()\n\n\tip4addrs := []net.IP{}\n\tip6addrs := []net.IP{}\n\n\tfor _, addr := range addrs {\n\t\tif ip4addr := addr.To4(); ip4addr != nil {\n\t\t\tip4addrs = append(ip4addrs, ip4addr)\n\n\t\t} else {\n\t\t\tip6addrs = append(ip6addrs, addr)\n\t\t}\n\t}\n\n\tparams := map[string]interface{}{\n\t\t\"jid\": &j.jid,\n\t\t\"ip4.addr\": ip4addrs,\n\t\t\"ip6.addr\": ip6addrs,\n\t}\n\n\tif er := jpps.bindParameters(params); er != nil {\n\t\treturn er\n\t}\n\n\tif _, er := C.jailparam_set(&jpps.params[0], jpps.numParams(), C.JAIL_UPDATE); er != nil {\n\t\treturn er\n\t}\n\n\tj.addrs = addrs\n\n\treturn nil\n}", "func (m *Group) SetSites(value []Siteable)() {\n m.sites = value\n}", "func (appConfig *AppConfiguration) EmailRecipients() []string {\n\n\tkey := \"email.recipients\"\n\n\tif appConfig.viper.IsSet(key) {\n\t\treturn appConfig.viper.GetStringSlice(key)\n\t}\n\n\treturn nil\n}", "func (dc *DentistCreate) SetEmail(s string) *DentistCreate {\n\tdc.mutation.SetEmail(s)\n\treturn dc\n}", "func (o VulnerabilityAssessmentRecurringScansPropertiesResponsePtrOutput) Emails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *VulnerabilityAssessmentRecurringScansPropertiesResponse) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Emails\n\t}).(pulumi.StringArrayOutput)\n}", "func (o *DataExportQuery) GetEmails() []string {\n\tif o == nil || o.Emails == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn *o.Emails\n}", "func EmailNotifier(emails []string, emailer emailclient.Client, markup string) (Notifier, error) {\n\treturn &emailNotifier{\n\t\tfrom: emailFromAddress,\n\t\temailer: emailer,\n\t\tmarkup: markup,\n\t\tto: emails,\n\t}, nil\n}", "func (m *AadUserConversationMember) SetEmail(value *string)() {\n m.email = value\n}", "func (m *BrowserSiteList) SetSites(value []BrowserSiteable)() {\n err := m.GetBackingStore().Set(\"sites\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *SLOCorrectionResponseAttributesModifier) SetEmail(v string) {\n\to.Email = &v\n}", "func (pu *PharmacistUpdate) SetEmail(s string) *PharmacistUpdate {\n\tpu.mutation.SetEmail(s)\n\treturn pu\n}", "func (r *AutoRoller) GetEmails() []string {\n\tr.emailsMtx.RLock()\n\tdefer r.emailsMtx.RUnlock()\n\trv := make([]string, len(r.emails))\n\tcopy(rv, r.emails)\n\treturn rv\n}", "func (r *AutoRoller) GetEmails() []string {\n\tr.emailsMtx.RLock()\n\tdefer r.emailsMtx.RUnlock()\n\trv := make([]string, len(r.emails))\n\tcopy(rv, r.emails)\n\treturn rv\n}", "func (puo *PharmacistUpdateOne) SetEmail(s string) *PharmacistUpdateOne {\n\tpuo.mutation.SetEmail(s)\n\treturn puo\n}", "func (op *ListRecipientNamesByEmailOp) Email(val string) *ListRecipientNamesByEmailOp {\n\tif op != nil {\n\t\top.QueryOpts.Set(\"email\", val)\n\t}\n\treturn op\n}", "func (s UserSet) SetEmailFormatted(value string) {\n\ts.RecordCollection.Set(models.NewFieldName(\"EmailFormatted\", \"email_formatted\"), value)\n}", "func (m *ManagerMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (obj *MessengerUser) SetEmail(email string) {\n\temailChar := C.CString(email)\n\tdefer C.free(unsafe.Pointer(emailChar))\n\temailStr := C.vsc_str_from_str(emailChar)\n\n\tC.vssq_messenger_user_set_email(obj.cCtx, emailStr)\n\n\truntime.KeepAlive(obj)\n\n\truntime.KeepAlive(email)\n\n\treturn\n}", "func (pu *PatientrecordUpdate) SetEmail(s string) *PatientrecordUpdate {\n\tpu.mutation.SetEmail(s)\n\treturn pu\n}", "func (puo *PatientrecordUpdateOne) SetEmail(s string) *PatientrecordUpdateOne {\n\tpuo.mutation.SetEmail(s)\n\treturn puo\n}", "func (m *MailTips) SetRecipientSuggestions(value []Recipientable)() {\n err := m.GetBackingStore().Set(\"recipientSuggestions\", value)\n if err != nil {\n panic(err)\n }\n}", "func (b *Basket) SetEggs(newEggs []string) {\n\tb.eggs = newEggs\n}", "func replaceReviewersPlaceholder(configs []*config.NotifierConfig, emails []string) []*notifier.Config {\n\tconfigCopies := []*notifier.Config{}\n\tfor _, n := range configs {\n\t\tconfigCopy := arb_notifier.ProtoToConfig(n)\n\t\tif configCopy.Email != nil {\n\t\t\tnewEmails := []string{}\n\t\t\tfor _, e := range configCopy.Email.Emails {\n\t\t\t\tif e == \"$REVIEWERS\" {\n\t\t\t\t\tnewEmails = append(newEmails, emails...)\n\t\t\t\t} else {\n\t\t\t\t\tnewEmails = append(newEmails, e)\n\t\t\t\t}\n\t\t\t}\n\t\t\tconfigCopy.Email.Emails = newEmails\n\t\t}\n\t\tconfigCopies = append(configCopies, configCopy)\n\t}\n\treturn configCopies\n}", "func (r *AutoRoller) GetEmails() []string {\n\tr.emailMtx.RLock()\n\tdefer r.emailMtx.RUnlock()\n\trv := make([]string, len(r.emails))\n\tcopy(rv, r.emails)\n\treturn rv\n}", "func (aruo *AuthRequestUpdateOne) SetClaimsEmail(s string) *AuthRequestUpdateOne {\n\taruo.mutation.SetClaimsEmail(s)\n\treturn aruo\n}", "func (aru *AuthRequestUpdate) SetClaimsEmail(s string) *AuthRequestUpdate {\n\taru.mutation.SetClaimsEmail(s)\n\treturn aru\n}", "func SendVerificationEmails(\n\ttxn *sql.Tx, publicKey *pgpkey.PgpKey, meta VerificationMetadata) error {\n\n\tfor _, email := range publicKey.Emails(true) {\n\t\tshouldSend, err := shouldSendVerificationEmail(txn, email)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t} else if shouldSend {\n\t\t\tif err := sendVerificationEmail(txn, email, publicKey, meta); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (am *ACMEIssuer) setEmail(ctx context.Context, allowPrompts bool) error {\n\tleEmail := am.Email\n\n\t// First try package default email, or a discovered email address\n\tif leEmail == \"\" {\n\t\tleEmail = DefaultACME.Email\n\t}\n\tif leEmail == \"\" {\n\t\tdiscoveredEmailMu.Lock()\n\t\tleEmail = discoveredEmail\n\t\tdiscoveredEmailMu.Unlock()\n\t}\n\n\t// Then try to get most recent user email from storage\n\tvar gotRecentEmail bool\n\tif leEmail == \"\" {\n\t\tleEmail, gotRecentEmail = am.mostRecentAccountEmail(ctx, am.CA)\n\t}\n\tif !gotRecentEmail && leEmail == \"\" && allowPrompts {\n\t\t// Looks like there is no email address readily available,\n\t\t// so we will have to ask the user if we can.\n\t\tvar err error\n\t\tleEmail, err = am.promptUserForEmail()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// User might have just signified their agreement\n\t\tam.mu.Lock()\n\t\tam.agreed = DefaultACME.Agreed\n\t\tam.mu.Unlock()\n\t}\n\n\t// Save the email for later and ensure it is consistent\n\t// for repeated use; then update cfg with the email\n\tleEmail = strings.TrimSpace(strings.ToLower(leEmail))\n\tdiscoveredEmailMu.Lock()\n\tif discoveredEmail == \"\" {\n\t\tdiscoveredEmail = leEmail\n\t}\n\tdiscoveredEmailMu.Unlock()\n\n\t// The unexported email field is the one we use\n\t// because we have thread-safe control over it\n\tam.mu.Lock()\n\tam.email = leEmail\n\tam.mu.Unlock()\n\n\treturn nil\n}", "func (m *InvitedUserMessageInfo) SetCcRecipients(value []Recipientable)() {\n err := m.GetBackingStore().Set(\"ccRecipients\", value)\n if err != nil {\n panic(err)\n }\n}", "func (du *DoctorUpdate) SetEmail(s string) *DoctorUpdate {\n\tdu.mutation.SetEmail(s)\n\treturn du\n}", "func (m *User) SetImAddresses(value []string)() {\n m.imAddresses = value\n}", "func Emails(mods ...qm.QueryMod) emailQuery {\n\tmods = append(mods, qm.From(\"`email`\"))\n\treturn emailQuery{NewQuery(mods...)}\n}", "func (m *Printer) SetTaskTriggers(value []PrintTaskTriggerable)() {\n err := m.GetBackingStore().Set(\"taskTriggers\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *EmployeeMutation) SetEmail(s string) {\n\tm._Email = &s\n}", "func (m *EmployeeMutation) SetEmail(s string) {\n\tm._Email = &s\n}", "func (uuo *UserUpdateOne) SetEmail(s string) *UserUpdateOne {\n\tuuo.mutation.SetEmail(s)\n\treturn uuo\n}", "func (uuo *UserUpdateOne) SetEmail(s string) *UserUpdateOne {\n\tuuo.mutation.SetEmail(s)\n\treturn uuo\n}", "func (euo *EmployeeUpdateOne) SetEmail(s string) *EmployeeUpdateOne {\n\teuo.mutation.SetEmail(s)\n\treturn euo\n}", "func (duo *DoctorUpdateOne) SetEmail(s string) *DoctorUpdateOne {\n\tduo.mutation.SetEmail(s)\n\treturn duo\n}", "func (uu *UserUpdate) SetEmail(s string) *UserUpdate {\n\tuu.mutation.SetEmail(s)\n\treturn uu\n}", "func (uu *UserUpdate) SetEmail(s string) *UserUpdate {\n\tuu.mutation.SetEmail(s)\n\treturn uu\n}", "func (ifc *Interface) SetAddresses(addresses []*net.IPNet) error {\n\n\terr := ifc.FlushAddresses()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ifc.AddAddresses(addresses)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *BackupUnitProperties) SetEmail(v string) {\n\n\to.Email = &v\n\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponseOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudDatalabelingV1beta1HumanAnnotationConfigResponse) []string {\n\t\treturn v.ContributorEmails\n\t}).(pulumi.StringArrayOutput)\n}", "func (o GoogleCloudDatalabelingV1beta1HumanAnnotationConfigOutput) ContributorEmails() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v GoogleCloudDatalabelingV1beta1HumanAnnotationConfig) []string { return v.ContributorEmails }).(pulumi.StringArrayOutput)\n}", "func (in *ActionUserRequestRegistrationCreateInput) SetEmail(value string) *ActionUserRequestRegistrationCreateInput {\n\tin.Email = value\n\n\tif in._selectedParameters == nil {\n\t\tin._selectedParameters = make(map[string]interface{})\n\t}\n\n\tin._selectedParameters[\"Email\"] = nil\n\treturn in\n}", "func (b *Builder) Email(s string) *Builder {\n\treturn b.appendMessage(s, func(offset, limit int) tg.MessageEntityClass {\n\t\treturn &tg.MessageEntityEmail{Offset: offset, Length: limit}\n\t})\n}", "func SetMail(Config MailConfig) {\n\n\t// Get the Auth\n\tmauth := smtp.PlainAuth(\n\t\t\"\",\n\t\tConfig.MailSender, // Sender e-mail\n\t\tConfig.MailPassword, // Sender password\n\t\tConfig.Server, // Server. Eg: smtp.google.com\n\t)\n\n\temailFrom := []string{Config.MailReceiver}\n\temailHeader := \"To:\" + Config.MailReceiver + \"\\r\\nSubject:\" + Config.Subject + \"\\r\\nMIME-Version: 1.0\\r\\nContent-Type: text/html; charset=ISO-8891-1\\r\\n\\r\\n\"\n\n\temailBody := []byte(emailHeader + Config.Message)\n\t// Send the e-mail\n\terr := smtp.SendMail(\n\t\tConfig.Server+\":\"+strconv.Itoa(Config.Port), // Server + Port\n\t\tmauth, // Get the Auth setup\n\t\tConfig.MailSender, // Get who is sending the e-mail\n\t\temailFrom, // Get who will receive the e-mail\n\t\temailBody, // Get the message\n\t)\n\n\tif err != nil {\n\t\tlog.Fatal(err) // Log if error\n\t} else {\n\t\tlog.Println(\"E-Mail send to: \" + Config.MailReceiver) // Log if succeful and display who receive the e-mail\n\t}\n\n}", "func (m *EducationAssignment) SetSubmissions(value []EducationSubmissionable)() {\n m.submissions = value\n}", "func (d *Dispatcher) SetEmailer(emailer Emailer) {\n\td.emailer = emailer\n}", "func (e *Account) SetEmailIfDifferent(v string) bool {\n\tif e.email == v {\n\t\treturn false\n\t}\n\te.SetEmail(v)\n\treturn true\n}", "func (e *Account) SetEmailIfDifferent(v string) bool {\n\tif e.email == v {\n\t\treturn false\n\t}\n\te.SetEmail(v)\n\treturn true\n}", "func (m *DeviceManagementConfigurationPolicy) SetAssignments(value []DeviceManagementConfigurationPolicyAssignmentable)() {\n err := m.GetBackingStore().Set(\"assignments\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *EmployeeMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (m *EmployeeMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (o *GetSearchEmployeesParams) SetEmail(email *string) {\n\to.Email = email\n}", "func (m *PaymentMutation) SetEMAIL(s string) {\n\tm._EMAIL = &s\n}", "func (b *GroupsEditBuilder) Email(v string) *GroupsEditBuilder {\n\tb.Params[\"email\"] = v\n\treturn b\n}", "func DefaultListEmail(ctx context.Context, db *gorm.DB) ([]*Email, error) {\n\tormResponse := []EmailORM{}\n\tdb, err := ops.ApplyCollectionOperators(db, ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := db.Set(\"gorm:auto_preload\", true).Find(&ormResponse).Error; err != nil {\n\t\treturn nil, err\n\t}\n\tpbResponse := []*Email{}\n\tfor _, responseEntry := range ormResponse {\n\t\ttemp, err := responseEntry.ToPB(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpbResponse = append(pbResponse, &temp)\n\t}\n\treturn pbResponse, nil\n}", "func (ec *EmployeeCreate) SetEmail(s string) *EmployeeCreate {\n\tec.mutation.SetEmail(s)\n\treturn ec\n}", "func (k PublicKey) EmailAddresses() []string {\n\treturn nil\n}", "func (m *ProfileMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (o *CheckoutResponse) SetEmailsSent(v int32) {\n\to.EmailsSent = &v\n}", "func (eu *EmployeeUpdate) SetEmail(s string) *EmployeeUpdate {\n\teu.mutation.SetEmail(s)\n\treturn eu\n}", "func (m *CustomerMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (m *User) SetFollowedSites(value []Siteable)() {\n m.followedSites = value\n}", "func (m *BookingBusiness) SetEmail(value *string)() {\n err := m.GetBackingStore().Set(\"email\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *SMTPOptions) ResetRecipients() {\n\to.mutex.Lock()\n\tdefer o.mutex.Unlock()\n\n\to.toAddrs = []*mail.Address{}\n}", "func (m *AudioRoutingGroup) SetReceivers(value []string)() {\n err := m.GetBackingStore().Set(\"receivers\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *RelatedContact) SetEmailAddress(value *string)() {\n err := m.GetBackingStore().Set(\"emailAddress\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *AgreementAcceptance) SetUserEmail(value *string)() {\n m.userEmail = value\n}", "func (m *PolicySetsItemUpdatePostRequestBody) SetAssignments(value []ie233ee762e29b4ba6970aa2a2efce4b7fde11697ca9ea81099d0f8269309c1be.PolicySetAssignmentable)() {\n err := m.GetBackingStore().Set(\"assignments\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *RouterChain) SetInvokers(invokers []protocol.Invoker) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\tc.invokers = invokers\n\tfor _, v := range c.routers {\n\t\tv.Notify(c.invokers)\n\t}\n}", "func (tr *TestRunner) Set(targets []*net.UDPAddr) {\n\ttr.mutex.Lock()\n\ttr.targets = targets\n\tdefer tr.mutex.Unlock()\n}", "func (m *InviteeMutation) SetEmail(s string) {\n\tm.email = &s\n}", "func (o *VerificationTraits) SetEmail(v string) {\n\to.Email = &v\n}", "func (m *User) SetAppRoleAssignments(value []AppRoleAssignmentable)() {\n m.appRoleAssignments = value\n}" ]
[ "0.82714236", "0.74483263", "0.6024663", "0.57086015", "0.57007587", "0.5677544", "0.564232", "0.559006", "0.55650765", "0.55586106", "0.5557214", "0.5537994", "0.5531404", "0.545678", "0.5394481", "0.53895247", "0.53594536", "0.53209436", "0.52935517", "0.5258642", "0.51996505", "0.5166886", "0.5151537", "0.5147885", "0.5147463", "0.5107439", "0.5099635", "0.50641274", "0.5047429", "0.5038694", "0.5036904", "0.50282246", "0.5018853", "0.5003119", "0.5000451", "0.5000451", "0.49978504", "0.49913058", "0.49877056", "0.496166", "0.49476275", "0.49464467", "0.4944898", "0.4942702", "0.49376005", "0.49360916", "0.49358043", "0.49326554", "0.49108335", "0.48945785", "0.48900852", "0.4870944", "0.48698375", "0.48664632", "0.48656607", "0.485665", "0.48543146", "0.48543146", "0.48542702", "0.48542702", "0.4848793", "0.48347917", "0.48197988", "0.48197988", "0.48108637", "0.48039928", "0.48011574", "0.4793094", "0.47855023", "0.4785466", "0.47809535", "0.477962", "0.4778315", "0.47773844", "0.47773844", "0.47743866", "0.4770375", "0.4770375", "0.47702295", "0.47700182", "0.47687903", "0.47663295", "0.4762036", "0.47615424", "0.47596574", "0.47590795", "0.47588423", "0.47583652", "0.4755003", "0.4752587", "0.47523484", "0.47456157", "0.47454727", "0.47369164", "0.47315666", "0.47276706", "0.47248527", "0.47155184", "0.47091126", "0.47080317" ]
0.8253865
1
closeIssue closes the given issue with the given message.
func (r *AutoRoller) closeIssue(issue *autoroll.AutoRollIssue, result, msg string) error { sklog.Infof("Closing issue %d (result %q) with message: %s", issue.Issue, result, msg) info, err := issue.ToGerritChangeInfo() if err != nil { return fmt.Errorf("Failed to convert issue to Gerrit ChangeInfo: %s", err) } if err := r.gerrit.Abandon(info, msg); err != nil { return err } issue.Result = result issue.Closed = true issue.CommitQueue = false issue.CommitQueueDryRun = false return r.recent.Update(issue) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *RepoAPI) closeIssue(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tname := m.Get(\"name\").Str()\n\treference := m.Get(\"reference\").Str()\n\treturn rpc.Success(util.Map{\n\t\t\"data\": a.mods.Repo.CloseIssue(name, reference),\n\t})\n}", "func (c *client) CloseIssue(org, repo string, number int) error {\n\tdurationLogger := c.log(\"CloseIssue\", org, repo, number)\n\tdefer durationLogger()\n\n\treturn c.closeIssue(org, repo, number, \"completed\")\n}", "func IssueClose(pid interface{}, id int) error {\n\t_, _, err := lab.Issues.UpdateIssue(pid, id, &gitlab.UpdateIssueOptions{\n\t\tStateEvent: gitlab.String(\"close\"),\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (g Gitlab) CloseIssue(ctx context.Context, projectID, issueIID IntID, resolution Resolution) error {\n\tparams := IssueParams{\n\t\tStateEvent: \"close\",\n\t\tRemoveLabels: g.labels[\"pending\"],\n\t\tAddLabels: g.labels[string(resolution.Tag)],\n\t}\n\tresp, err := g.NewRequest(ctx).\n\t\tSetBody(params).\n\t\tPut(g.APIV4URL(\"projects\", projectID, \"issues\", issueIID))\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif resp.IsError() {\n\t\treturn trace.Wrap(responseError(resp))\n\t}\n\n\tlogger.Get(ctx).Debug(\"Successfully closed the issue\")\n\treturn nil\n}", "func (r *Rietveld) Close(issue int64, message string) error {\n\tif err := r.AddComment(issue, message); err != nil {\n\t\treturn err\n\t}\n\treturn r.post(fmt.Sprintf(\"/%d/close\", issue), nil)\n}", "func (m *MockClient) CloseIssue(org, repo string, number int) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CloseIssue\", org, repo, number)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (m *MockUCCodeHubI) CloseIssue(issueID, userID int64) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CloseIssue\", issueID, userID)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (c *client) ReopenIssue(org, repo string, number int) error {\n\tdurationLogger := c.log(\"ReopenIssue\", org, repo, number)\n\tdefer durationLogger()\n\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodPatch,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/issues/%d\", org, repo, number),\n\t\torg: org,\n\t\trequestBody: map[string]string{\"state\": \"open\"},\n\t\texitCodes: []int{200},\n\t}, nil)\n\treturn stateCannotBeChangedOrOriginalError(err)\n}", "func (m *MockIssueClient) CloseIssue(org, repo string, number int) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CloseIssue\", org, repo, number)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func closeRelatedIssues(ctx context.Context, client *github.Client, owner string, repositoryName string, pr *github.PullRequest, dryRun bool) error {\n\tissueNumbers := parseIssueFixes(pr.GetBody())\n\n\trepo, _, err := client.Repositories.Get(ctx, owner, repositoryName)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to access repository %s/%s: %w\", owner, repositoryName, err)\n\t}\n\n\tfor _, issueNumber := range issueNumbers {\n\t\tlog.Printf(\"PR #%d: closes issue #%d, add milestones %s\", pr.GetNumber(), issueNumber, pr.Milestone.GetTitle())\n\t\tif !dryRun {\n\t\t\terr := closeIssue(ctx, client, owner, repositoryName, pr, issueNumber)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"unable to close issue #%d: %w\", issueNumber, err)\n\t\t\t}\n\t\t}\n\n\t\t// Add comment if needed\n\t\tif pr.Base.GetRef() != repo.GetDefaultBranch() {\n\t\t\tmessage := fmt.Sprintf(\"Closed by #%d.\", pr.GetNumber())\n\n\t\t\tlog.Printf(\"PR #%d: issue #%d, add comment: %s\", pr.GetNumber(), issueNumber, message)\n\n\t\t\tif !dryRun {\n\t\t\t\terr := addComment(ctx, client, owner, repositoryName, issueNumber, message)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn fmt.Errorf(\"unable to add comment on issue #%d: %w\", issueNumber, err)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (mr *MockClientMockRecorder) CloseIssue(org, repo, number interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CloseIssue\", reflect.TypeOf((*MockClient)(nil).CloseIssue), org, repo, number)\n}", "func (a *RepoAPI) reopenIssue(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tname := m.Get(\"name\").Str()\n\treference := m.Get(\"reference\").Str()\n\treturn rpc.Success(util.Map{\n\t\t\"data\": a.mods.Repo.ReopenIssue(name, reference),\n\t})\n}", "func (g Gitlab) ResolveIssue(ctx context.Context, projectID, issueIID IntID, resolution Resolution) error {\n\t// Try to add a comment.\n\terr1 := trace.Wrap(g.PostResolutionComment(ctx, projectID, issueIID, resolution))\n\n\t// Try to close the issue.\n\terr2 := trace.Wrap(g.CloseIssue(ctx, projectID, issueIID, resolution))\n\n\treturn trace.NewAggregate(err1, err2)\n}", "func (mr *MockUCCodeHubIMockRecorder) CloseIssue(issueID, userID interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CloseIssue\", reflect.TypeOf((*MockUCCodeHubI)(nil).CloseIssue), issueID, userID)\n}", "func (mr *MockIssueClientMockRecorder) CloseIssue(org, repo, number interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"CloseIssue\", reflect.TypeOf((*MockIssueClient)(nil).CloseIssue), org, repo, number)\n}", "func Issue(command string, message string) {\n\tIssueCommand(command, nil, message)\n}", "func MRClose(pid interface{}, id int) error {\n\tmr, _, err := lab.MergeRequests.GetMergeRequest(pid, id, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif mr.State == \"closed\" {\n\t\treturn fmt.Errorf(\"mr already closed\")\n\t}\n\t_, _, err = lab.MergeRequests.UpdateMergeRequest(pid, int(id), &gitlab.UpdateMergeRequestOptions{\n\t\tStateEvent: gitlab.String(\"close\"),\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *client) CloseIssueAsNotPlanned(org, repo string, number int) error {\n\tdurationLogger := c.log(\"CloseIssueAsNotPlanned\", org, repo, number)\n\tdefer durationLogger()\n\n\treturn c.closeIssue(org, repo, number, \"not_planned\")\n}", "func (c *client) ClosePR(org, repo string, number int) error {\n\tdurationLogger := c.log(\"ClosePR\", org, repo, number)\n\tdefer durationLogger()\n\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodPatch,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/pulls/%d\", org, repo, number),\n\t\torg: org,\n\t\trequestBody: map[string]string{\"state\": \"closed\"},\n\t\texitCodes: []int{200},\n\t}, nil)\n\treturn err\n}", "func (a *ProblemsApiService) CloseProblemExecute(r ApiCloseProblemRequest) (ProblemCloseResult, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue ProblemCloseResult\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"ProblemsApiService.CloseProblem\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/problems/{problemId}/close\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"problemId\"+\"}\", _neturl.PathEscape(parameterToString(r.problemId, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json; charset=utf-8\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json; charset=utf-8\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = r.closingComment\n\tif r.ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := r.ctx.Value(ContextAPIKeys).(map[string]APIKey); ok {\n\t\t\tif apiKey, ok := auth[\"Api-Token\"]; ok {\n\t\t\t\tvar key string\n\t\t\t\tif apiKey.Prefix != \"\" {\n\t\t\t\t\tkey = apiKey.Prefix + \" \" + apiKey.Key\n\t\t\t\t} else {\n\t\t\t\t\tkey = apiKey.Key\n\t\t\t\t}\n\t\t\t\tlocalVarHeaderParams[\"Authorization\"] = key\n\t\t\t}\n\t\t}\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func (ec *executionContext) field_Mutation_closeIssue_args(ctx context.Context, rawArgs map[string]interface{}) (map[string]interface{}, error) {\n\tvar err error\n\targs := map[string]interface{}{}\n\tvar arg0 CloseIssue\n\tif tmp, ok := rawArgs[\"input\"]; ok {\n\t\targ0, err = ec.unmarshalNcloseIssue2githubᚗcomᚋwakashiyoᚋtraphᚑgoᚐCloseIssue(ctx, tmp)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\targs[\"input\"] = arg0\n\treturn args, nil\n}", "func (g *Github) ClosePullRequest(ctx context.Context, pullReq scm.PullRequest) error {\n\tpr := pullReq.(pullRequest)\n\n\tg.modLock()\n\tdefer g.modUnlock()\n\n\t_, _, err := retry(ctx, func() (*github.PullRequest, *github.Response, error) {\n\t\treturn g.ghClient.PullRequests.Edit(ctx, pr.ownerName, pr.repoName, pr.number, &github.PullRequest{\n\t\t\tState: &[]string{\"closed\"}[0],\n\t\t})\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = retryWithoutReturn(ctx, func() (*github.Response, error) {\n\t\treturn g.ghClient.Git.DeleteRef(ctx, pr.prOwnerName, pr.prRepoName, fmt.Sprintf(\"heads/%s\", pr.branchName))\n\t})\n\treturn err\n}", "func (g *Gitlab) ClosePullRequest(ctx context.Context, pullReq scm.PullRequest) error {\n\tpr := pullReq.(pullRequest)\n\n\tstateEvent := \"close\"\n\t_, _, err := g.glClient.MergeRequests.UpdateMergeRequest(pr.targetPID, pr.iid, &gitlab.UpdateMergeRequestOptions{\n\t\tStateEvent: &stateEvent,\n\t}, gitlab.WithContext(ctx))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = g.glClient.Branches.DeleteBranch(pr.sourcePID, pr.branchName, gitlab.WithContext(ctx))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *Message) Close() error {\n\t// Do some processing here\n\tm.parse(&m.buffer)\n\n\t// Display what we received\n\tlog.Printf(\"Received message from %s: [%s] %s\\n\", m.From, m.Subject, m.Body)\n\n\t// Send to application for processing\n\ta, err := NewAction(*m)\n\tif err != nil {\n\t\tlog.Printf(\"Error: Failed to parse action. %v\", err)\n\t\treturn fmt.Errorf(\"Error: Failed to parse action. %v\", err)\n\t}\n\terr = a.Do()\n\tif err != nil {\n\t\tlog.Printf(\"Error: Failed to run action. %v\", err)\n\t\treturn fmt.Errorf(\"Error: Failed to run action. %v\", err)\n\t}\n\n\treturn nil\n}", "func close(object ibmmq.MQObject) error {\n\terr := object.Close(0)\n\tif err == nil {\n\t\tfmt.Println(\"Closed queue\")\n\t} else {\n\t\tfmt.Println(err)\n\t}\n\treturn err\n}", "func (i *IssuesService) StopWatchingIssue(owner, repoSlug string, id int64) (*simpleresty.Response, error) {\n\turlStr := i.client.http.RequestURL(\"/repositories/%s/%s/issues/%v/watch\", owner, repoSlug, id)\n\tresponse, err := i.client.http.Delete(urlStr, nil, nil)\n\n\treturn response, err\n}", "func qr_decoder_close(p _QrDecoderHandle) {\n\tC.qr_decoder_close(C.QrDecoderHandle(p))\n}", "func (_Registry *RegistryTransactor) Issue(opts *bind.TransactOpts, _to common.Address, _validityData []byte, _topic *big.Int, _value *big.Int, _data []byte) (*types.Transaction, error) {\n\treturn _Registry.contract.Transact(opts, \"issue\", _to, _validityData, _topic, _value, _data)\n}", "func (i *IssuesService) StopWatchingIssue(owner, repoSlug string, id int64) (*Response, error) {\n\turlStr := i.client.requestURL(\"/repositories/%s/%s/issues/%v/watch\", owner, repoSlug, id)\n\tresponse, err := i.client.execute(\"DELETE\", urlStr, nil, nil)\n\n\treturn response, err\n}", "func Close(cmdInfo CommandInfo) {\n\t// Arg at cmdInfo.CmdOps[1] should be specifiying event or trade\n\tif len(cmdInfo.CmdOps) != 3 {\n\t\t// wrong arguments - error\n\t\tmsg := cmdInfo.createMsgEmbed(\n\t\t\t\"Error: Incorrect Arguments Supplied\", errThumbURL, \"Please check your syntax.\", errColor,\n\t\t\tformat(\n\t\t\t\tcreateFields(\"EXAMPLE\", cmdInfo.Prefix+\"close event 1234\", true),\n\t\t\t\tcreateFields(\"EXAMPLE\", cmdInfo.Prefix+\"close trade 1234\", true),\n\t\t\t))\n\t\tcmdInfo.Ses.ChannelMessageSendEmbed(cmdInfo.BotChID, msg)\n\t\treturn\n\t}\n\n\tt := strings.ToLower(cmdInfo.CmdOps[1])\n\tswitch t {\n\tcase \"event\":\n\t\tcloseEvent(cmdInfo.CmdOps[2], cmdInfo)\n\tcase \"trade\":\n\t\tcloseTrade(cmdInfo.CmdOps[2], cmdInfo)\n\t}\n}", "func ClosedIssues(ctx context.Context, c *client.Client, org string, project string, since time.Time, until time.Time, users []string) ([]*IssueSummary, error) {\n\tclosed, err := issues(ctx, c, org, project, since, until, users, \"closed\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresult := make([]*IssueSummary, 0, len(closed))\n\tfor _, i := range closed {\n\t\tresult = append(result, &IssueSummary{\n\t\t\tURL: i.GetHTMLURL(),\n\t\t\tDate: i.GetClosedAt().Format(dateForm),\n\t\t\tAuthor: i.GetUser().GetLogin(),\n\t\t\tCloser: i.GetClosedBy().GetLogin(),\n\t\t\tProject: project,\n\t\t\tTitle: i.GetTitle(),\n\t\t})\n\t}\n\n\treturn result, nil\n}", "func (m *MockClient) ReopenIssue(org, repo string, number int) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ReopenIssue\", org, repo, number)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func onClose(code int, reason string) {\n\tfmt.Println(\"Close: \", code, reason)\n}", "func (a *ProblemsApiService) CloseProblem(ctx _context.Context, problemId string) ApiCloseProblemRequest {\n\treturn ApiCloseProblemRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tproblemId: problemId,\n\t}\n}", "func IFBClose() error {\n\treturn execModProbe(ifbModuleName, \"-r\")\n}", "func (m *MockIssueClient) ReopenIssue(org, repo string, number int) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ReopenIssue\", org, repo, number)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func close(object ibmmq.MQObject) error {\n\terr := object.Close(0)\n\tif err == nil {\n\t\tfmt.Println(\"Closed topic\")\n\t} else {\n\t\tfmt.Println(err)\n\t}\n\treturn err\n}", "func (g *GH) DeleteProjectIssueCard(contentType string, issue github.Issue, repoName string, projectName string) {\n\tlog.Println(\"Deleting Project Card\")\n\tctx := context.Background()\n\tprojectID := g.GetProjectID(projectName)\n\tcard := g.GetProjectCardByIssue(issue, repoName, *projectID)\n\tif card == nil {\n\t\tlog.Print(\"There is no card to delete for issue #\", issue.ID)\n\t\treturn\n\t}\n\tg.c.Projects.DeleteProjectCard(ctx, *card.ID)\n}", "func (c *client) DeleteIssue(\n\tid interface{},\n\tissue int,\n\toptions ...glab.RequestOptionFunc,\n) (*glab.Response, error) {\n\treturn c.c.Issues.DeleteIssue(id, issue, options...)\n}", "func (p *MessagePartition) releaseMessagefile(fileId uint64, file *os.File) {\n\tfile.Close()\n}", "func (s *IssueService) Issue(id int64) (*domain.Issue, error) {\n\treturn s.IssueRepository.GetById(id)\n}", "func (a *RepoAPI) readIssue(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tname := m.Get(\"name\").Str()\n\treference := m.Get(\"reference\").Str()\n\treturn rpc.Success(util.Map{\n\t\t\"data\": a.mods.Repo.ReadIssue(name, reference),\n\t})\n}", "func CloseWithMessage(conn *websocket.Conn, msg string) {\n\terr := conn.WriteControl(websocket.CloseMessage,\n\t\twebsocket.FormatCloseMessage(websocket.ClosePolicyViolation, msg), time.Now().Add(time.Second))\n\tif err != nil {\n\t\treturn\n\t}\n\tconn.Close()\n}", "func (mq *LinuxMessageQueue) Close() error {\n\tif mq.cancelSocket >= 0 {\n\t\tif err := mq.NotifyCancel(); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to cancel notifications\")\n\t\t}\n\t}\n\terr := unix.Close(mq.ID())\n\t*mq = LinuxMessageQueue{cancelSocket: -1}\n\treturn err\n}", "func MoveIssueCommand(ctx *cli.Context) error {\n\tif ctx.Args().Len() != 2 {\n\t\treturn fmt.Errorf(\"expected exactly two argument, the issue ID and the pipeline ID. Received %d\", ctx.Args().Len())\n\t}\n\n\tissueID, err := strconv.Atoi(ctx.Args().First())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"expected issue ID to be an int, got %s\", ctx.Args().First())\n\t}\n\n\tpipelineID := ctx.Args().Get(1)\n\n\ttoken, err := GetZenHubToken()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient := http.Client{\n\t\tTransport: &AuthenticationTransport{\n\t\t\ttransport: http.DefaultTransport,\n\t\t\tauthenticationToken: token,\n\t\t},\n\t}\n\n\tworkspaceID := ctx.String(\"workspace-id\")\n\tif workspaceID == \"\" {\n\t\treturn fmt.Errorf(\"invalid workpace-id value of %s\", workspaceID)\n\t}\n\n\trepositoryID := ctx.Uint(\"repository-id\")\n\tif repositoryID == 0 {\n\t\treturn fmt.Errorf(\"invalid repository-id value of %d\", repositoryID)\n\t}\n\n\turl := fmt.Sprintf(\"%s/p2/workspaces/%s/repositories/%d/issues/%d/moves\",\n\t\tctx.String(\"base-url\"),\n\t\tworkspaceID,\n\t\trepositoryID,\n\t\tissueID,\n\t)\n\trequest := MoveIssueRequest{\n\t\tPipelineID: pipelineID,\n\t\tPosition: \"bottom\",\n\t}\n\tbody, err := json.Marshal(request)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to convert move issue request %v to JSON: %w\", request, err)\n\t}\n\n\tlogrus.WithFields(logrus.Fields{\n\t\t\"url\": url,\n\t\t\"body\": string(body),\n\t}).Debug(\"Sending move issue request\")\n\tresp, err := client.Post(url, \"application/json\", bytes.NewReader(body))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to move issue between pipelines: %w\", err)\n\t}\n\n\tif err := ErrorFromStatusCode(resp.StatusCode); err != nil {\n\t\treturn fmt.Errorf(\"failed to move issue between pipelines: %w\", err)\n\t}\n\n\tfmt.Printf(\"Successfully moved issue %d to pipeline %s\\n\", issueID, pipelineID)\n\n\treturn nil\n}", "func (c *Circuit) close(now time.Time, forceClosed bool) {\n\tif !c.IsOpen() {\n\t\t// Not open. Don't need to close it\n\t\treturn\n\t}\n\tif c.threadSafeConfig.CircuitBreaker.ForceOpen.Get() {\n\t\treturn\n\t}\n\tif forceClosed || c.OpenToClose.ShouldClose(now) {\n\t\tc.CircuitMetricsCollector.Closed(now)\n\t\tc.isOpen.Set(false)\n\t}\n}", "func (s *sqlReadTx) DeleteIssue(issueID string) (*model.Issue, error) {\n\tselectQuery := s.statementBuilder.Select(\n\t\t\"*\",\n\t).From(\n\t\t\"issues\",\n\t).Where(\n\t\tsquirrel.Eq{\n\t\t\t\"id\": issueID,\n\t\t},\n\t)\n\tissue, err := s.querySingleIssue(selectQuery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdeleteQuery := s.statementBuilder.Delete(\n\t\t\"issues\",\n\t).Where(\n\t\tsquirrel.Eq{\n\t\t\t\"id\": issueID,\n\t\t},\n\t)\n\tif _, err := deleteQuery.QueryContext(s.ctx); err != nil {\n\t\treturn nil, err\n\t}\n\treturn issue, nil\n}", "func (_Registry *RegistrySession) Issue(_to common.Address, _validityData []byte, _topic *big.Int, _value *big.Int, _data []byte) (*types.Transaction, error) {\n\treturn _Registry.Contract.Issue(&_Registry.TransactOpts, _to, _validityData, _topic, _value, _data)\n}", "func (c *Client) CloseStory(ctx context.Context, request *CloseStoryRequest) error {\n\tvar ok Ok\n\n\tif err := c.rpc.Invoke(ctx, request, &ok); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (mr *MockClientMockRecorder) ReopenIssue(org, repo, number interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ReopenIssue\", reflect.TypeOf((*MockClient)(nil).ReopenIssue), org, repo, number)\n}", "func (_Registry *RegistryTransactorSession) Issue(_to common.Address, _validityData []byte, _topic *big.Int, _value *big.Int, _data []byte) (*types.Transaction, error) {\n\treturn _Registry.Contract.Issue(&_Registry.TransactOpts, _to, _validityData, _topic, _value, _data)\n}", "func IsErrIssueWasClosed(err error) bool {\n\t_, ok := err.(ErrIssueWasClosed)\n\treturn ok\n}", "func (m *MockIssueClient) UnassignIssue(org, repo string, number int, logins []string) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"UnassignIssue\", org, repo, number, logins)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func DeleteIssueAttachment(ctx *context.APIContext) {\n\t// swagger:operation DELETE /repos/{owner}/{repo}/issues/{index}/assets/{attachment_id} issue issueDeleteIssueAttachment\n\t// ---\n\t// summary: Delete an issue attachment\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: index\n\t// in: path\n\t// description: index of the issue\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// - name: attachment_id\n\t// in: path\n\t// description: id of the attachment to delete\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// responses:\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/error\"\n\n\tattachment := getIssueAttachmentSafeWrite(ctx)\n\tif attachment == nil {\n\t\treturn\n\t}\n\n\tif err := repo_model.DeleteAttachment(attachment, true); err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"DeleteAttachment\", err)\n\t\treturn\n\t}\n\n\tctx.Status(http.StatusNoContent)\n}", "func (j *Jira) Issue(id string) (i Issue, err error) {\n\turi := j.ApiPath + \"/issue/\" + id\n\n\tcontents, err := j.getRequest(uri)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tvar issue Issue\n\terr = json.Unmarshal(contents, &issue)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn issue, nil\n}", "func (m *MockClient) UnassignIssue(org, repo string, number int, logins []string) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"UnassignIssue\", org, repo, number, logins)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func NewMsgIssue(auth types.AccAddress, creator, symbol types.Name, amount Coin) MsgIssueCoin {\n\treturn MsgIssueCoin{\n\t\t*msg.MustNewKuMsg(\n\t\t\tRouterKeyName,\n\t\t\tmsg.WithAuth(auth),\n\t\t\tmsg.WithData(Cdc(), &MsgIssueCoinData{\n\t\t\t\tCreator: creator,\n\t\t\t\tSymbol: symbol,\n\t\t\t\tAmount: amount,\n\t\t\t}),\n\t\t),\n\t}\n}", "func (r *Reitveld) Issue(issue int64) (*Issue, error) {\n\treq, err := http.NewRequest(\"GET\", fmt.Sprintf(\"https://codereview.chromium.org/api/%d/\", issue), nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to build request: %s\", err)\n\t}\n\tresp, err := r.client.Do(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to retrieve issue info: %s\", err)\n\t}\n\tif resp.StatusCode != 200 {\n\t\treturn nil, fmt.Errorf(\"No such issue %d\", issue)\n\t}\n\tdec := json.NewDecoder(resp.Body)\n\tinfo := &Issue{}\n\tif err := dec.Decode(info); err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to read Reitveld issue info: %s\", err)\n\n\t}\n\treturn info, nil\n}", "func UnpinIssue(ctx *context.APIContext) {\n\t// swagger:operation DELETE /repos/{owner}/{repo}/issues/{index}/pin issue unpinIssue\n\t// ---\n\t// summary: Unpin an Issue\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: index\n\t// in: path\n\t// description: index of issue to unpin\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// responses:\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/notFound\"\n\tissue, err := issues_model.GetIssueByIndex(ctx, ctx.Repo.Repository.ID, ctx.ParamsInt64(\":index\"))\n\tif err != nil {\n\t\tif issues_model.IsErrIssueNotExist(err) {\n\t\t\tctx.NotFound()\n\t\t} else {\n\t\t\tctx.Error(http.StatusInternalServerError, \"GetIssueByIndex\", err)\n\t\t}\n\t\treturn\n\t}\n\n\t// If we don't do this, it will crash when trying to add the unpin event to the comment history\n\terr = issue.LoadRepo(ctx)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"LoadRepo\", err)\n\t\treturn\n\t}\n\n\terr = issue.Unpin(ctx, ctx.Doer)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"UnpinIssue\", err)\n\t\treturn\n\t}\n\n\tctx.Status(http.StatusNoContent)\n}", "func (c *ConnectionMock) CloseWithReason(code int, reason string) {\n\tc.Called(code, reason)\n}", "func (m *DataHistoryManager) CheckCandleIssue(job *DataHistoryJob, multiplier int64, apiData, dbData float64, candleField string) (issue string, replace bool) {\n\tif m == nil {\n\t\treturn ErrNilSubsystem.Error(), false\n\t}\n\tif atomic.LoadInt32(&m.started) == 0 {\n\t\treturn ErrSubSystemNotStarted.Error(), false\n\t}\n\tif job == nil {\n\t\treturn errNilJob.Error(), false\n\t}\n\n\tfloatiplier := float64(multiplier)\n\tif floatiplier > 0 {\n\t\tapiData = math.Round(apiData*floatiplier) / floatiplier\n\t\tdbData = math.Round(dbData*floatiplier) / floatiplier\n\t}\n\tif apiData != dbData {\n\t\tvar diff float64\n\t\tif apiData > dbData {\n\t\t\tdiff = gctmath.CalculatePercentageGainOrLoss(apiData, dbData)\n\t\t} else {\n\t\t\tdiff = gctmath.CalculatePercentageGainOrLoss(dbData, apiData)\n\t\t}\n\t\tif diff > job.IssueTolerancePercentage {\n\t\t\tissue = fmt.Sprintf(\"%s api: %v db: %v diff: %v %%\", candleField, apiData, dbData, diff)\n\t\t}\n\t\tif job.ReplaceOnIssue &&\n\t\t\tjob.IssueTolerancePercentage != 0 &&\n\t\t\tdiff > job.IssueTolerancePercentage &&\n\t\t\tjob.SecondaryExchangeSource == \"\" {\n\t\t\treplace = true\n\t\t}\n\t}\n\treturn issue, replace\n}", "func RemoveIssueDependency(user *User, issue *Issue, dep *Issue, depType DependencyType) (err error) {\n\tsess := x.NewSession()\n\tdefer sess.Close()\n\tif err = sess.Begin(); err != nil {\n\t\treturn err\n\t}\n\n\tvar issueDepToDelete IssueDependency\n\n\tswitch depType {\n\tcase DependencyTypeBlockedBy:\n\t\tissueDepToDelete = IssueDependency{IssueID: issue.ID, DependencyID: dep.ID}\n\tcase DependencyTypeBlocking:\n\t\tissueDepToDelete = IssueDependency{IssueID: dep.ID, DependencyID: issue.ID}\n\tdefault:\n\t\treturn ErrUnknownDependencyType{depType}\n\t}\n\n\taffected, err := sess.Delete(&issueDepToDelete)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If we deleted nothing, the dependency did not exist\n\tif affected <= 0 {\n\t\treturn ErrDependencyNotExists{issue.ID, dep.ID}\n\t}\n\n\t// Add comment referencing the removed dependency\n\tif err = createIssueDependencyComment(sess, user, issue, dep, false); err != nil {\n\t\treturn err\n\t}\n\treturn sess.Commit()\n}", "func (ch *Channel) Close() {}", "func (a *RepoAPI) createIssue(params interface{}) (resp *rpc.Response) {\n\tm := objx.New(cast.ToStringMap(params))\n\tname := m.Get(\"name\").Str()\n\tcallParams := m.Get(\"params\").MSI()\n\treturn rpc.Success(util.Map{\n\t\t\"data\": a.mods.Repo.CreateIssue(name, callParams),\n\t})\n}", "func (mr *MockIssueClientMockRecorder) ReopenIssue(org, repo, number interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"ReopenIssue\", reflect.TypeOf((*MockIssueClient)(nil).ReopenIssue), org, repo, number)\n}", "func createIssueComment(\n\tctx context.Context,\n\tpr *github.PullRequest,\n\tclient *github.Client,\n\tmessage string,\n) error {\n\tcomment := &github.IssueComment{Body: &message}\n\t_, _, err := client.Issues.CreateComment(\n\t\tctx,\n\t\tpr.Base.Repo.Owner.GetLogin(),\n\t\tpr.Base.Repo.GetName(),\n\t\tpr.GetNumber(),\n\t\tcomment,\n\t)\n\treturn err\n}", "func (c *Client) CommentAndOpenIssue(ctx context.Context, owner, repo string, issueNumber int, body string) error {\n\t_, _, err := c.GHClient.Issues.CreateComment(ctx, owner, repo, issueNumber, &gh.IssueComment{\n\t\tBody: &body,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, _, err = c.GHClient.Issues.Edit(ctx, owner, repo, issueNumber, &gh.IssueRequest{\n\t\tState: func() *string { a := \"open\"; return &a }(),\n\t})\n\n\treturn err\n}", "func (q *jobQueue) close() error {\n\t// queue is closing\n\tclose(q.msgq)\n\t// wait until go-routines of consumption was exited\n\treturn q.eg.Wait()\n}", "func IsErrIssueIsClosed(err error) bool {\n\t_, ok := err.(ErrIssueIsClosed)\n\treturn ok\n}", "func closeAlert(alertClient *alert.Client, event *types.Event, alertid string) error {\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tdefer cancel()\n\tnotes := fmt.Sprintf(\"Closed Automatically\\n %s\", event.Check.Output)\n\tcloseResult, err := alertClient.Close(ctx, &alert.CloseAlertRequest{\n\t\tIdentifierType: alert.ALERTID,\n\t\tIdentifierValue: alertid,\n\t\tSource: source,\n\t\tNote: notes,\n\t})\n\tif err != nil {\n\t\tfmt.Printf(\"[ERROR] Not Closed: %s \\n\", err)\n\t}\n\tfmt.Printf(\"RequestID %s to Close %s \\n\", alertid, closeResult.RequestId)\n\n\treturn nil\n}", "func (p *Note) Close() {}", "func (r *Replayer) Close() error {\n\treturn errors.New(\"implement me\")\n}", "func (rg *reportGenerator) createIssue() *github.Issue {\n\ttitle := rg.getIssueTitle()\n\tbody := os.Expand(issueBodyTemplate, rg.templateHelper)\n\n\tissue, response, err := rg.client.Issues.Create(\n\t\trg.ctx,\n\t\trg.envVariables[projectUsernameKey],\n\t\trg.envVariables[projectRepoNameKey],\n\t\t&github.IssueRequest{\n\t\t\tTitle: &title,\n\t\t\tBody: &body,\n\t\t\t// TODO: Set Assignees and labels\n\t\t})\n\tif err != nil {\n\t\trg.logger.Fatal(\"Failed to create GitHub Issue\", zap.Error(err))\n\t}\n\n\tif response.StatusCode != http.StatusCreated {\n\t\trg.handleBadResponses(response)\n\t}\n\n\treturn issue\n}", "func (c *client) GetIssue(\n\tpid interface{},\n\tid int,\n\toptions ...glab.RequestOptionFunc,\n) (*glab.Issue, *glab.Response, error) {\n\treturn c.c.Issues.GetIssue(pid, id, options...)\n}", "func IssueCommand(command string, properties CommandProperties, message string) {\n\tcmd := NewCommand(command, properties, message)\n\tfmt.Fprintln(os.Stdout, cmd.String())\n}", "func GetIssue(queryRoute string, cdc *codec.Codec) *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"issue [issueID]\",\n\t\tShort: \"Get issue by ID\",\n\t\tExample: \"issue issue1\",\n\t\tArgs: cobra.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\tcliCtx := context.NewCLIContext().WithCodec(cdc)\n\n\t\t\t// prepare request\n\t\t\treq := types.IssueReq{ID: args[0]}\n\n\t\t\tbz, err := cliCtx.Codec.MarshalJSON(req)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// query and parse the result\n\t\t\tres, _, err := cliCtx.QueryWithData(fmt.Sprintf(\"custom/%s/%s\", queryRoute, types.QueryIssue), bz)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tvar out types.Issue\n\t\t\tcdc.MustUnmarshalJSON(res, &out)\n\n\t\t\treturn cliCtx.PrintOutput(out)\n\t\t},\n\t}\n\thelpers.BuildCmdHelp(cmd, []string{\n\t\t\"unique issue ID\",\n\t})\n\n\treturn cmd\n}", "func TestIssue(t *testing.T) {\n\tts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tcheckAuth(t, r)\n\t\tval := `{\"expand\":\"renderedFields,names,schema,transitions,operations,editmeta,changelog\",\"id\":\"1234\",\"self\":\"https://test.atlassian.net/rest/api/2/issue/1234\",\"key\":\"ABC-01\",\"fields\":{\"summary\":\"This is a test\"}}`\n\t\tfmt.Fprintln(w, val)\n\t}))\n\tdefer ts.Close()\n\n\tc := NewJiraClient(ts.URL, \"foo\", \"bar\", 100)\n\ti, err := c.Issue(\"abc\", nil)\n\tif err != nil {\n\t\tt.Error(\"Error loading issue:\", err)\n\t}\n\n\tif i == nil {\n\t\tt.Error(\"issue is nil\")\n\t}\n\n\tif i.ID != \"1234\" {\n\t\tt.Error(\"Error expected issue ID to be 1234 got:\", i.ID)\n\t}\n\n\tif i.Key != \"ABC-01\" {\n\t\tt.Error(\"Error expected issue Key to be ABC-01 got:\", i.Key)\n\t}\n}", "func FormatCloseMessage(closeCode int, text string) []byte {\n\tif closeCode == CloseNoStatusReceived {\n\t\t// Return empty message because it's illegal to send\n\t\t// CloseNoStatusReceived. Return non-nil value in case application\n\t\t// checks for nil.\n\t\treturn []byte{}\n\t}\n\tbuf := make([]byte, 2+len(text))\n\tbinary.BigEndian.PutUint16(buf, uint16(closeCode))\n\tcopy(buf[2:], text)\n\treturn buf\n}", "func (c *channel) close() {\n\tc.heartbeatLck.Lock()\n\t// nil heartbeatStop indicates that heartbeat is disabled\n\t// due to lost remote and channel is basically in\n\t// \"dangling\" state.\n\tif c.heartbeatInterval != 0 && c.heartbeatStop != nil {\n\t\tc.heartbeatStop <- struct{}{}\n\t\tc.heartbeatTicker.Stop()\n\t\t<-c.heartbeatStop\n\t}\n\n\tc.heartbeatLck.Unlock()\n\n\tc.stateLck.Lock()\n\tdefer c.stateLck.Unlock()\n\n\tc.stopping = true\n\n\tclose(c.queue)\n\tc.errorCh = nil\n}", "func (b *BcBotAction) Issue(_ *bot.Bot) (bool, error) {\n\n\tadminBot := b.GetAdminBot()\n\n\tbalance := strconv.FormatFloat(adminBot.Balance, 'f', -1, 64)\n\tchargeAmount, err := strconv.ParseFloat(b.jobs.Config.AdminChargeAmount, 64)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tnewIssue := &issue{\n\t\tTxID: uuid.Must(uuid.NewUUID()).String(),\n\t\tWalletAddress: adminBot.Id,\n\t\tAmount: b.jobs.Config.AdminChargeAmount,\n\t\tBalance: balance,\n\t\tTxTime: time.Now().Format(\"2006-01-02T15:04:05.000Z\"),\n\t}\n\n\tisDone, err := b.requestHandler(\"/issue\", newIssue)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tb.mutex.Lock()\n\tdefer b.mutex.Unlock()\n\tb.jobs.AdminBot.Balance -= chargeAmount\n\n\treturn isDone, nil\n}", "func MergeIssue(parentIssue *ParentIssue, i jiraworklog.Issue) {\r\n\tresolvedDate := time.Time{}\r\n\tif i.Fields.ResolutionDate != nil {\r\n\t\tresolvedDate, _ = time.Parse(\"2006-01-02T15:04:05.000-0700\", *i.Fields.ResolutionDate)\r\n\t} else {\r\n\t\tresolvedDate, _ = time.Parse(\"2006-01-02T15:04:05.000-0700\", *i.Fields.StatusCategoryChangeDate)\r\n\t}\r\n\tdaysToResolve := int(math.Ceil(resolvedDate.Sub(parentIssue.CreateDate).Hours() / 24))\r\n\r\n\tparentIssue.IsResolved = true\r\n\tparentIssue.ResolvedDate = &resolvedDate\r\n\tparentIssue.UpdateDate = resolvedDate\r\n\tparentIssue.DaysToResolve = daysToResolve\r\n\tparentIssue.Type = i.Fields.Issuetype.Name\r\n\tparentIssue.Priority = i.Fields.Priority.Name\r\n\tparentIssue.Status = i.Fields.Status.Name\r\n\tparentIssue.Summary = i.Fields.Summary\r\n\tparentIssue.Project = strings.Split(i.Key, \"-\")[0]\r\n\tparentIssue.AggregateTimeOriginalEstimate = i.Fields.Aggregatetimeoriginalestimate\r\n\tparentIssue.AggregateTimeSpent = i.Fields.Aggregatetimespent\r\n}", "func (s *CertificatesService) Reissue(id string, body *CertificateReissue) error {\n\tenc, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = s.client.Post(\"/v1/certificates/\"+id+\"/reissue\", enc)\n\n\treturn err\n}", "func (c *coordinator) CloseRoom(id string) { c.Notify(api.CloseRoom, id) }", "func (cb *GSCCloseBranch) Close() error {\n\tcurrentBranch := cb.git.GetCurrentBranch()\n\tdefaultBranch := cb.git.GetDefaultBranch()\n\tif currentBranch == defaultBranch {\n\t\terr := fmt.Errorf(\"Cannot close main branch\")\n\t\tcb.GetLogger().Error(err.Error())\n\t\treturn err\n\t}\n\tcb.GetLogger().Infof(\"Closing %s branch and discarding all the changes.\", currentBranch)\n\tif err := wzlib_subprocess.ExecCommand(\"git\", \"checkout\", cb.git.GetDefaultBranch()).Run(); err != nil {\n\t\treturn err\n\t}\n\treturn wzlib_subprocess.ExecCommand(\"git\", \"branch\", \"--delete\", \"--force\", currentBranch).Run()\n}", "func (c CommitMessage) issuesList() string {\n\tvar issues []string\n\tfor _, item := range strings.Split(c.Close, \",\") {\n\t\titem = strings.TrimSpace(item)\n\t\tif item != \"\" {\n\t\t\tissues = append(issues, \"Closes #\"+item)\n\t\t}\n\t}\n\treturn strings.Join(issues, \"\\n\")\n}", "func NewMsgCloseBid(id BidID) *MsgCloseBid {\n\treturn &MsgCloseBid{\n\t\tBidID: id,\n\t}\n}", "func (c *client) UnassignIssue(org, repo string, number int, logins []string) error {\n\tdurationLogger := c.log(\"UnassignIssue\", org, repo, number, logins)\n\tdefer durationLogger()\n\n\tassigned := make(map[string]bool)\n\tvar i Issue\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodDelete,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/issues/%d/assignees\", org, repo, number),\n\t\torg: org,\n\t\trequestBody: map[string][]string{\"assignees\": logins},\n\t\texitCodes: []int{200},\n\t}, &i)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, assignee := range i.Assignees {\n\t\tassigned[NormLogin(assignee.Login)] = true\n\t}\n\textra := ExtraUsers{action: \"unassign\"}\n\tfor _, login := range logins {\n\t\tif assigned[NormLogin(login)] {\n\t\t\textra.Users = append(extra.Users, login)\n\t\t}\n\t}\n\tif len(extra.Users) > 0 {\n\t\treturn extra\n\t}\n\treturn nil\n}", "func formatIssueAsGithub(issue *result.Issue) string {\n\tseverity := defaultGithubSeverity\n\tif issue.Severity != \"\" {\n\t\tseverity = issue.Severity\n\t}\n\n\tret := fmt.Sprintf(\"::%s file=%s,line=%d\", severity, issue.FilePath(), issue.Line())\n\tif issue.Pos.Column != 0 {\n\t\tret += fmt.Sprintf(\",col=%d\", issue.Pos.Column)\n\t}\n\n\tret += fmt.Sprintf(\"::%s (%s)\", issue.Text, issue.FromLinter)\n\treturn ret\n}", "func (c *Client) Issue9(ctx context.Context, params *Issue9Params, body *Issue9RequestBody) (*http.Response, error) {\n\treq, err := NewIssue9Request(c.Server, params, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\treturn c.Client.Do(req)\n}", "func (e *PythonOp) closeCmd() {\n\te.stdout.Close()\n\te.stdin.Close()\n\tif e.cmd != nil {\n\t\terr := e.cmd.Wait()\n\t\te.cmd = nil\n\t\tif err != nil && e.err == nil {\n\t\t\te.err = err\n\t\t}\n\t}\n}", "func (c *Client) close(reason string) (err error) {\n\tc.closer.Do(func() {\n\t\tif reason != \"\" {\n\t\t\tc.Logln(LogConn, \"Close reason:\", reason)\n\t\t}\n\t\tif err = c.t.Close(false); err != nil {\n\t\t\tc.Logln(LogConn, \"Close error:\", err)\n\t\t}\n\t})\n\treturn\n}", "func (s *SmartContract) CloseAuction(ctx contractapi.TransactionContextInterface, auctionID string) error {\n\n\tauctionBytes, err := ctx.GetStub().GetState(auctionID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get auction %v: %v\", auctionID, err)\n\t}\n\n\tif auctionBytes == nil {\n\t\treturn fmt.Errorf(\"Auction interest object %v not found\", auctionID)\n\t}\n\n\tvar auctionJSON Auction\n\terr = json.Unmarshal(auctionBytes, &auctionJSON)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to create auction object JSON: %v\", err)\n\t}\n\n\t// the auction can only be closed by the seller\n\n\t// get ID of submitting client\n\tclientID, err := ctx.GetClientIdentity().GetID()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to get client identity %v\", err)\n\t}\n\n\tSeller := auctionJSON.Seller\n\tif Seller != clientID {\n\t\treturn fmt.Errorf(\"auction can only be closed by seller: %v\", err)\n\t}\n\n\tStatus := auctionJSON.Status\n\tif Status != \"open\" {\n\t\treturn fmt.Errorf(\"cannot close auction that is not open\")\n\t}\n\n\tauctionJSON.Status = string(\"closed\")\n\n\tclosedAuction, _ := json.Marshal(auctionJSON)\n\n\terr = ctx.GetStub().PutState(auctionID, closedAuction)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to close auction: %v\", err)\n\t}\n\n\treturn nil\n}", "func (c *client) GetIssue(org, repo string, number int) (*Issue, error) {\n\tdurationLogger := c.log(\"GetIssue\", org, repo, number)\n\tdefer durationLogger()\n\n\tvar i Issue\n\t_, err := c.request(&request{\n\t\t// allow emoji\n\t\t// https://developer.github.com/changes/2018-02-22-label-description-search-preview/\n\t\taccept: \"application/vnd.github.symmetra-preview+json\",\n\t\tmethod: http.MethodGet,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/issues/%d\", org, repo, number),\n\t\torg: org,\n\t\texitCodes: []int{200},\n\t}, &i)\n\treturn &i, err\n}", "func (c *client) EditIssue(org, repo string, number int, issue *Issue) (*Issue, error) {\n\tdurationLogger := c.log(\"EditIssue\", org, repo, number)\n\tdefer durationLogger()\n\n\tif c.dry {\n\t\treturn issue, nil\n\t}\n\tedit := struct {\n\t\tTitle string `json:\"title,omitempty\"`\n\t\tBody string `json:\"body,omitempty\"`\n\t\tState string `json:\"state,omitempty\"`\n\t}{\n\t\tTitle: issue.Title,\n\t\tBody: issue.Body,\n\t\tState: issue.State,\n\t}\n\tvar ret Issue\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodPatch,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/issues/%d\", org, repo, number),\n\t\torg: org,\n\t\texitCodes: []int{200},\n\t\trequestBody: &edit,\n\t}, &ret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &ret, nil\n}", "func (c CloseTester) Close() (err error) {\n\tif c.errorMsg != \"\" {\n\t\treturn fmt.Errorf(c.errorMsg)\n\t}\n\n\treturn nil\n}", "func (e *agentEndpoint) close() {\n\te.mutex.Lock()\n\tdefer func() {\n\t\t_ = recover()\n\t\te.mutex.Unlock()\n\t}()\n\tif e.isClosed {\n\t\treturn\n\t}\n\tif err := e.conn.WriteMessage(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseGoingAway, \"bye bye\")); err != nil {\n\t\tlogger.WithError(err).Errorf(\"error sending closing message to agent with id == %s\", e.id)\n\t}\n\te.isClosed = true\n}", "func ConvertIssue(repostitory string, in github.Issue) Issue {\n\n\tissue := Issue{\n\t\tNumber: *in.Number,\n\t\tState: *in.State,\n\t\tTitle: *in.Title,\n\t\tLabels: []Label{},\n\t\tClosedAt: in.ClosedAt,\n\t\tCreatedAt: in.CreatedAt,\n\t\tUpdatedAt: in.UpdatedAt,\n\t\tRepository: repostitory,\n\t}\n\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tfmt.Fprintln(os.Stderr, in)\n\t\t\tpanic(\"unknown data\")\n\t\t}\n\t}()\n\n\tif in.Body != nil {\n\t\tissue.Body = *in.Body\n\t}\n\n\tif in.User != nil {\n\t\tissue.User = *in.User.Login\n\t}\n\n\tif in.Assignee != nil {\n\t\tissue.Assignee = *in.Assignee.Login\n\t}\n\n\tif in.Milestone != nil {\n\t\tissue.Milestone = *in.Milestone.Title\n\t}\n\n\tfor _, ghLabel := range in.Labels {\n\t\tissue.Labels = append(issue.Labels, ConvertLabel(ghLabel))\n\t}\n\n\treturn issue\n}", "func ResetIssueTime(ctx *context.APIContext) {\n\t// swagger:operation Delete /repos/{owner}/{repo}/issues/{index}/times issue issueResetTime\n\t// ---\n\t// summary: Reset a tracked time of an issue\n\t// consumes:\n\t// - application/json\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: index\n\t// in: path\n\t// description: index of the issue to add tracked time to\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// responses:\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"400\":\n\t// \"$ref\": \"#/responses/error\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\n\tissue, err := issues_model.GetIssueByIndex(ctx, ctx.Repo.Repository.ID, ctx.ParamsInt64(\":index\"))\n\tif err != nil {\n\t\tif issues_model.IsErrIssueNotExist(err) {\n\t\t\tctx.NotFound(err)\n\t\t} else {\n\t\t\tctx.Error(http.StatusInternalServerError, \"GetIssueByIndex\", err)\n\t\t}\n\t\treturn\n\t}\n\n\tif !ctx.Repo.CanUseTimetracker(issue, ctx.Doer) {\n\t\tif !ctx.Repo.Repository.IsTimetrackerEnabled(ctx) {\n\t\t\tctx.JSON(http.StatusBadRequest, struct{ Message string }{Message: \"time tracking disabled\"})\n\t\t\treturn\n\t\t}\n\t\tctx.Status(http.StatusForbidden)\n\t\treturn\n\t}\n\n\terr = issues_model.DeleteIssueUserTimes(issue, ctx.Doer)\n\tif err != nil {\n\t\tif db.IsErrNotExist(err) {\n\t\t\tctx.Error(http.StatusNotFound, \"DeleteIssueUserTimes\", err)\n\t\t} else {\n\t\t\tctx.Error(http.StatusInternalServerError, \"DeleteIssueUserTimes\", err)\n\t\t}\n\t\treturn\n\t}\n\tctx.Status(http.StatusNoContent)\n}", "func CloseCmd() *cobra.Command {\n\tcmd := &cobra.Command{\n\t\tUse: \"close\",\n\t\tShort: \"close the specified peer\",\n\t\tRun: closePeer,\n\t}\n\tcmd.Flags().StringP(\"pid\", \"\", \"\", \"peer id,support dht network\")\n\treturn cmd\n}", "func closeRecv(route wire.Route, ctx common.Context, in <-chan wire.Packet, out chan<- wire.Packet, challenge uint64) error {\n\n\t// Send: close verify\n\toffset := uint64(rand.Uint32())\n\n\tif err := sendOrTimeout(ctx, out, wire.BuildPacket(route).SetClose(offset).SetVerify(challenge).Build()); err != nil {\n\t\treturn NewClosingError(fmt.Sprintf(\"Failed: send(close,verify): %v\", err.Error()))\n\t}\n\n\t// Receive: verify\n\tp, err := recvOrTimeout(ctx, in)\n\tif err != nil {\n\t\treturn NewClosingError(fmt.Sprintf(\"Failed: receive(verify): %v\", err.Error()))\n\t}\n\n\tif verify := p.Verify(); verify == nil || verify.Val() != offset {\n\t\treturn NewClosingError(\"Failed: receive(verify). Incorrect value\")\n\t}\n\n\treturn nil\n}" ]
[ "0.73282236", "0.7246137", "0.72274476", "0.6945461", "0.68043506", "0.5832222", "0.57497317", "0.57276416", "0.56256056", "0.5598965", "0.55777115", "0.5551263", "0.5520711", "0.5398505", "0.5302305", "0.5202957", "0.51250786", "0.50397885", "0.49553296", "0.49533334", "0.49102956", "0.4908899", "0.4906045", "0.48721108", "0.4859826", "0.4791794", "0.4781017", "0.47390512", "0.47323102", "0.46723056", "0.4658327", "0.46500865", "0.46381354", "0.46377167", "0.46315363", "0.45652935", "0.45458642", "0.4533149", "0.45329142", "0.4506478", "0.4483633", "0.44816947", "0.44615856", "0.4442329", "0.44347754", "0.44199762", "0.44162673", "0.44018543", "0.4392392", "0.43890694", "0.43705657", "0.43678534", "0.43659985", "0.43656844", "0.43633214", "0.4357856", "0.43469122", "0.43394768", "0.43391708", "0.43350506", "0.4322366", "0.43214524", "0.4316305", "0.43040317", "0.42992488", "0.4298688", "0.42820427", "0.4275749", "0.42738685", "0.4272571", "0.42725113", "0.4267987", "0.42494982", "0.4233093", "0.42309344", "0.4230348", "0.42271185", "0.42260772", "0.42259574", "0.42188314", "0.42080224", "0.42037052", "0.42017046", "0.4194758", "0.4191451", "0.41903198", "0.41873813", "0.4182653", "0.41819105", "0.41800517", "0.41787252", "0.41708392", "0.41686574", "0.41625512", "0.41616866", "0.41538665", "0.41446388", "0.41440755", "0.41428912", "0.41401288" ]
0.69842666
3
addIssueComment adds a comment to the given issue.
func (r *AutoRoller) addIssueComment(issue *autoroll.AutoRollIssue, msg string) error { sklog.Infof("Adding comment to issue: %q", msg) info, err := issue.ToGerritChangeInfo() if err != nil { return fmt.Errorf("Failed to convert issue to Gerrit ChangeInfo: %s", err) } if err := r.gerrit.AddComment(info, msg); err != nil { return err } updated, err := r.retrieveRoll(issue.Issue) if err != nil { return err } return r.recent.Update(updated) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (j *Jira) AddComment(issue *Issue, comment string) error {\n\tvar cMap = make(map[string]string)\n\tcMap[\"body\"] = comment\n\n\tcJson, err := json.Marshal(cMap)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\turi := j.BaseUrl + j.ApiPath + \"/issue/\" + issue.Key + \"/comment\"\n\tbody := bytes.NewBuffer(cJson)\n\n\t_, err = j.postJson(uri, body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *Rietveld) AddComment(issue int64, message string) error {\n\tdata := url.Values{}\n\tdata.Add(\"message\", message)\n\tdata.Add(\"message_only\", \"True\")\n\tdata.Add(\"add_as_reviewer\", \"False\")\n\tdata.Add(\"send_mail\", \"True\")\n\tdata.Add(\"no_redirect\", \"True\")\n\treturn r.post(fmt.Sprintf(\"/%d/publish\", issue), data)\n}", "func (c Client) AddComment(ctx context.Context, ID it.IssueID, comment it.Comment) (it.CommentID, error) {\n\tissueID, err := strconv.Atoi(string(ID))\n\tif err != nil {\n\t\treturn it.CommentID(\"\"), err\n\t}\n\tid, err := c.Client.IssueNoteAdd(ctx, issueID, mantis.IssueNoteData{\n\t\t//Reporter:\n\t\tDateSubmitted: mantis.Time(comment.CreatedAt),\n\t\tText: comment.Body,\n\t})\n\treturn it.CommentID(strconv.Itoa(id)), err\n}", "func (r *AutoRoller) AddComment(ctx context.Context, issueNum int64, message, user string, timestamp time.Time) error {\n\troll, err := r.recent.Get(ctx, issueNum)\n\tif err != nil {\n\t\treturn skerr.Fmt(\"No such issue %d\", issueNum)\n\t}\n\tid := fmt.Sprintf(\"%d_%d\", issueNum, len(roll.Comments))\n\troll.Comments = append(roll.Comments, comment.New(id, message, user))\n\treturn r.recent.Update(ctx, roll)\n}", "func (c *Client) AddComment(ctx context.Context, change *gerritpb.ChangeInfo, message string) (*gerritpb.ReviewResult, error) {\n\treq := c.createSetReviewRequest(ctx, change, message)\n\tres, err := c.setReview(ctx, req)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"error adding comment\").Err()\n\t}\n\n\treturn res, nil\n}", "func (r *AutoRoller) AddComment(issueNum int64, message, user string, timestamp time.Time) error {\n\troll, err := r.recent.Get(issueNum)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"No such issue %d\", issueNum)\n\t}\n\tid := fmt.Sprintf(\"%d_%d\", issueNum, len(roll.Comments))\n\troll.Comments = append(roll.Comments, comment.New(id, message, user))\n\treturn r.recent.Update(roll)\n}", "func (p PRMirror) AddComment(id int, comment string) bool {\n\tissueComment := github.IssueComment{}\n\tissueComment.Body = &comment\n\n\t_, _, err := p.GitHubClient.Issues.CreateComment(*p.Context, p.Configuration.DownstreamOwner, p.Configuration.DownstreamRepo, id, &issueComment)\n\tif err != nil {\n\t\tlog.Errorf(\"Error while adding a comment to issue#:%d - %s\", id, err.Error())\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (rg *reportGenerator) commentOnIssue(issue *github.Issue) *github.IssueComment {\n\tbody := os.Expand(issueCommentTemplate, rg.templateHelper)\n\n\tissueComment, response, err := rg.client.Issues.CreateComment(\n\t\trg.ctx,\n\t\trg.envVariables[projectUsernameKey],\n\t\trg.envVariables[projectRepoNameKey],\n\t\t*issue.Number,\n\t\t&github.IssueComment{\n\t\t\tBody: &body,\n\t\t},\n\t)\n\tif err != nil {\n\t\trg.logger.Fatal(\"Failed to search GitHub Issues\", zap.Error(err))\n\t}\n\n\tif response.StatusCode != http.StatusCreated {\n\t\trg.handleBadResponses(response)\n\t}\n\n\treturn issueComment\n}", "func addComment(gh *octokat.Client, repo octokat.Repo, prNum, comment, commentType string) error {\n\t// get the comments\n\tcomments, err := gh.Comments(repo, prNum, &octokat.Options{})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// check if we already made the comment\n\tfor _, c := range comments {\n\t\t// if we already made the comment return nil\n\t\tif strings.ToLower(c.User.Login) == \"gordontheturtle\" && strings.Contains(c.Body, commentType) {\n\t\t\tlogrus.Debugf(\"Already made comment about %q on PR %s\", commentType, prNum)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// add the comment because we must not have already made it\n\tif _, err := gh.AddComment(repo, prNum, comment); err != nil {\n\t\treturn err\n\t}\n\n\tlogrus.Infof(\"Would have added comment about %q PR %s\", commentType, prNum)\n\treturn nil\n}", "func (r *Review) AddComment(c comment.Comment) error {\n\tcommentNote, err := c.Write()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.Repo.AppendNote(comment.Ref, r.Revision, commentNote)\n\treturn nil\n}", "func (m *MaintainerManager) AddComment(number, comment string) (gh.Comment, error) {\n\treturn m.client.AddComment(m.repo, number, comment)\n}", "func createIssueComment(\n\tctx context.Context,\n\tpr *github.PullRequest,\n\tclient *github.Client,\n\tmessage string,\n) error {\n\tcomment := &github.IssueComment{Body: &message}\n\t_, _, err := client.Issues.CreateComment(\n\t\tctx,\n\t\tpr.Base.Repo.Owner.GetLogin(),\n\t\tpr.Base.Repo.GetName(),\n\t\tpr.GetNumber(),\n\t\tcomment,\n\t)\n\treturn err\n}", "func (db *Database) AddComment(userID types.UserID, objectID types.ObjectID, content string) (err error) {\n\tif err = userID.Validate(); err != nil {\n\t\treturn\n\t}\n\tif err = objectID.Validate(); err != nil {\n\t\treturn\n\t}\n\tif len(content) > 1024 {\n\t\treturn errors.New(\"content too large\")\n\t}\n\n\terr = db.comments.Insert(types.Comment{\n\t\tUserID: userID,\n\t\tObjectID: objectID,\n\t\tContent: content,\n\t\tDate: time.Now(),\n\t})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to insert new comment\")\n\t}\n\n\treturn\n}", "func (repo *commentRepository) AddComment(c *comment.Comment) (*comment.Comment, error) {\n\tc, err := (*repo.secondaryRepo).AddComment(c)\n\tif err == nil {\n\t\trepo.cache[c.ID] = *c\n\t}\n\treturn c, err\n}", "func CreateIssueComment(id int64, login, owner, repo string) error {\n\tmessage := fmt.Sprintf(\"Thank you for opening an issue @%s. Your contributions are welcome.\", login)\n\n\tissueComment := github.IssueComment{\n\t\tID: &id,\n\t\tBody: &message,\n\t}\n\n\tgithubClient := New()\n\tcomment, _, err := githubClient.Issues.CreateComment(context.Background(), owner, repo, int(id), &issueComment)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.SetOutput(os.Stdout)\n\tlog.Print(comment)\n\n\treturn nil\n}", "func CreateIssueComment(ctx *context.APIContext) {\n\t// swagger:operation POST /repos/{owner}/{repo}/issues/{index}/comments issue issueCreateComment\n\t// ---\n\t// summary: Add a comment to an issue\n\t// consumes:\n\t// - application/json\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: index\n\t// in: path\n\t// description: index of the issue\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// - name: body\n\t// in: body\n\t// schema:\n\t// \"$ref\": \"#/definitions/CreateIssueCommentOption\"\n\t// responses:\n\t// \"201\":\n\t// \"$ref\": \"#/responses/Comment\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\tform := web.GetForm(ctx).(*api.CreateIssueCommentOption)\n\tissue, err := issues_model.GetIssueByIndex(ctx, ctx.Repo.Repository.ID, ctx.ParamsInt64(\":index\"))\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"GetIssueByIndex\", err)\n\t\treturn\n\t}\n\n\tif issue.IsLocked && !ctx.Repo.CanWriteIssuesOrPulls(issue.IsPull) && !ctx.Doer.IsAdmin {\n\t\tctx.Error(http.StatusForbidden, \"CreateIssueComment\", errors.New(ctx.Tr(\"repo.issues.comment_on_locked\")))\n\t\treturn\n\t}\n\n\tcomment, err := issue_service.CreateIssueComment(ctx, ctx.Doer, ctx.Repo.Repository, issue, form.Body, nil)\n\tif err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"CreateIssueComment\", err)\n\t\treturn\n\t}\n\n\tctx.JSON(http.StatusCreated, convert.ToAPIComment(ctx, ctx.Repo.Repository, comment))\n}", "func (s *DiscussionsService) AddIssueDiscussionNote(pid interface{}, issue int, discussion string, opt *AddIssueDiscussionNoteOptions, options ...RequestOptionFunc) (*Note, *Response, error) {\n\tproject, err := parseID(pid)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tu := fmt.Sprintf(\"projects/%s/issues/%d/discussions/%s/notes\",\n\t\tPathEscape(project),\n\t\tissue,\n\t\tdiscussion,\n\t)\n\n\treq, err := s.client.NewRequest(http.MethodPost, u, opt, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tn := new(Note)\n\tresp, err := s.client.Do(req, n)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn n, resp, nil\n}", "func (c *Client) CreateIssueComment(issueIDOrKey string, input *CreateIssueCommentInput) (*IssueComment, error) {\n\treturn c.CreateIssueCommentContext(context.Background(), issueIDOrKey, input)\n}", "func (c *Client) CommentAndOpenIssue(ctx context.Context, owner, repo string, issueNumber int, body string) error {\n\t_, _, err := c.GHClient.Issues.CreateComment(ctx, owner, repo, issueNumber, &gh.IssueComment{\n\t\tBody: &body,\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, _, err = c.GHClient.Issues.Edit(ctx, owner, repo, issueNumber, &gh.IssueRequest{\n\t\tState: func() *string { a := \"open\"; return &a }(),\n\t})\n\n\treturn err\n}", "func (s *IdeaStorage) AddComment(number int, content string, userID int) (int, error) {\n\treturn 0, nil\n}", "func (c *Client) CreateComment(owner, repo string, number int, comment string) error {\n\tif c.dry {\n\t\treturn nil\n\t}\n\n\tic := IssueComment{\n\t\tBody: comment,\n\t}\n\tresp, err := c.request(http.MethodPost, fmt.Sprintf(\"%s/repos/%s/%s/issues/%d/comments\", c.base, owner, repo, number), ic)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 201 {\n\t\treturn fmt.Errorf(\"response not 201: %s\", resp.Status)\n\t}\n\treturn nil\n}", "func (issue *Issue) SetComment(comment io.Reader) (*Comment, error) {\n\turl := fmt.Sprintf(\"%s/issue/%s/comment\", BaseUrl, issue.Key)\n\tcode, body := execRequest(\"POST\", url, comment)\n\tif code == http.StatusCreated {\n\t\tvar jiraComment Comment\n\t\terr := json.Unmarshal(body, &jiraComment)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn &jiraComment, nil\n\t} else {\n\t\treturn nil, handleJiraError(body)\n\t}\n}", "func (issue *Issue) SetComment(comment *Comment) (*Comment, error) {\n\turl := fmt.Sprintf(\"%s/issue/%s/comment\", BaseURL, issue.Key)\n\tencodedParams, err := json.Marshal(comment)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcode, body := execRequest(\"POST\", url, bytes.NewBuffer(encodedParams))\n\tif code == http.StatusCreated {\n\t\tvar jiraComment Comment\n\t\terr := json.Unmarshal(body, &jiraComment)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn &jiraComment, nil\n\t}\n\treturn nil, handleJiraError(body)\n}", "func (g *baseGithub) PRComment(\n\tctx context.Context,\n\towner,\n\trepository string,\n\tpullRequestNumber int64,\n\tcomment string,\n) error {\n\t_, _, err := g.c.Issues.CreateComment(ctx, owner, repository, int(pullRequestNumber), &github.IssueComment{\n\t\tBody: github.String(comment),\n\t})\n\treturn err\n}", "func (s *Store) AddIssue(issue Issue) {\n\ts.Issues = append(s.Issues, &issue)\n}", "func AddComment(commentValue, threadId, collectionName, projectName, repositoryId, pullRequestId string) error {\n\tusername, pass, _, _, _, domain := GetConfigDatas()\n\turl := domain + collectionName + `/` + projectName + `/_apis/git/repositories/` + repositoryId + `/pullRequests/` + pullRequestId + `/threads/` + threadId + `/comments?api-version=4.1`\n\n\tfmt.Println(\"url: \", url)\n\tbody := `{\n\t\t\t \"content\": \"` + commentValue + `\",\n\t\t\t \"parentCommentId\": 1,\n\t\t\t \"commentType\": 1\n\t\t\t}`\n\n\tfmt.Println(\"body: \", body)\n\n\tclient := &http.Client{}\n\treq, _ := http.NewRequest(\"POST\", url, bytes.NewBuffer([]byte(body)))\n\n\treq.SetBasicAuth(username, pass)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := client.Do(req)\n\n\tif err != nil {\n\t\tCreateLogJson(\"Error\", \"AzureOps/AddComment\", \"Error while requesting to update tasks on Azure DevOps for adding comment on thread.\", err.Error())\n\t\treturn err\n\t}\n\n\tioutil.ReadAll(resp.Body)\n\n\t//CreateLogJson(\"Info\",\"AddComment\",\"Adding comment on pull request thread.\",\"AzureDevops thread comment is added. =>\"+bodyString)\n\treturn nil\n}", "func (ts *TechStoryService) addComment (w http.ResponseWriter, r *http.Request) {\n\tvar techStory model.TechStory\n\ttechStory.Key = mux.Vars(r)[\"id\"]\n\n\tvar com model.Comment\n\tmodel.ReadJsonBody(r, &com)\n\n\tWithTechStoryDao(func(dao techStoryDao) {\n\t\tuser := model.GetUserFromRequest(w, r)\n\t\tcomment := dao.AddComment(techStory, com.Title, com.Text, user.GetCollection()+\"/\"+user.Key)\n\n\t\tmodel.WriteResponse(true, nil, comment, w)\n\t})\n}", "func (c *Client) CreateIssueCommentContext(ctx context.Context, issueIDOrKey string, input *CreateIssueCommentInput) (*IssueComment, error) {\n\tu := fmt.Sprintf(\"/api/v2/issues/%v/comments\", issueIDOrKey)\n\n\treq, err := c.NewRequest(\"POST\", u, input)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tissueComment := new(IssueComment)\n\tif err := c.Do(ctx, req, &issueComment); err != nil {\n\t\treturn nil, err\n\t}\n\treturn issueComment, nil\n}", "func (s Surface) AddComment(key, value string) error {\n\treturn s.AddComments(Comments{Comment(key, value)})\n}", "func AddComment(tid, nickname, content string) error {\n\ttidNum, err := strconv.ParseInt(tid, 10, 64)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcomment := new(Comment)\n\tcomment.Tid = tidNum\n\tcomment.Name = nickname\n\tcomment.Content = content\n\tcomment.Created = time.Now()\n\n\to := orm.NewOrm()\n\t/*insert a reply*/\n\t_, err = o.Insert(comment)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* update topic reply count */\n\ttopic := new(Topic)\n\tqs := o.QueryTable(\"topic\")\n\terr = qs.Filter(\"Id\", tid).One(topic)\n\tif err != nil {\n\t\treturn err\n\t}\n\ttopic.ReplyCount++\n\ttopic.ReplyTime = time.Now()\n\n\t_, err = o.Update(topic)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (self *CourseService)AddComment(content, courseId, userId string) (commentVo *course.CourseCommentVo, error *err.HttpError) {\n\tcommentVo = new(course.CourseCommentVo)\n\tcommentTable := new(table.CourseCommentTable)\n\tcommentTable.UUID = uuid.New()\n\tcommentTable.Content = content\n\tcommentTable.CourseId = courseId\n\tcommentTable.CreateUser = userId\n\tcommentTable.CreateTime = time.Now()\n\tcommentTable.FrozenStatus = value.STATUS_ENABLED\n\tinsertNum, insertErr := self.Session.InsertOne(commentTable)\n\tif insertNum == 0 {\n\t\tif insertErr != nil {\n\t\t\tself.Log.Println(insertErr)\n\t\t}\n\t\terror = err.COURSE_COMMENT_INSERT_ERR\n\t\treturn\n\t}\n\tcommentVo = course.NewCommentVo(commentTable, self.Session, self.Log)\n\terror = nil\n\treturn\n}", "func (s *DiscussionsService) AddCommitDiscussionNote(pid interface{}, commit string, discussion string, opt *AddCommitDiscussionNoteOptions, options ...RequestOptionFunc) (*Note, *Response, error) {\n\tproject, err := parseID(pid)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tu := fmt.Sprintf(\"projects/%s/repository/commits/%s/discussions/%s/notes\",\n\t\tPathEscape(project),\n\t\tcommit,\n\t\tdiscussion,\n\t)\n\n\treq, err := s.client.NewRequest(http.MethodPost, u, opt, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tn := new(Note)\n\tresp, err := s.client.Do(req, n)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn n, resp, nil\n}", "func (res *Result) AddComment(c string) {\n\tres.comment += \"\\n\" + c\n}", "func AddTime(ctx context.Context, user *user_model.User, issue *Issue, amount int64, created time.Time) (*TrackedTime, error) {\n\tctx, committer, err := db.TxContext(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer committer.Close()\n\n\tt, err := addTime(ctx, user, issue, amount, created)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := issue.LoadRepo(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif _, err := CreateComment(ctx, &CreateCommentOptions{\n\t\tIssue: issue,\n\t\tRepo: issue.Repo,\n\t\tDoer: user,\n\t\t// Content before v1.21 did store the formated string instead of seconds,\n\t\t// so use \"|\" as delimeter to mark the new format\n\t\tContent: fmt.Sprintf(\"|%d\", amount),\n\t\tType: CommentTypeAddTimeManual,\n\t\tTimeID: t.ID,\n\t}); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn t, committer.Commit()\n}", "func (f CommentsFilter) ApplyIssue(context operations.Context, issue *github.Issue) bool {\n\treturn f.predicate(*issue.Comments)\n}", "func (b *Client) CreateComment(repo models.Repo, pullNum int, comment string, command string) error {\n\t// NOTE: I tried to find the maximum size of a comment for bitbucket.org but\n\t// I got up to 200k chars without issue so for now I'm not going to bother\n\t// to detect this.\n\tbodyBytes, err := json.Marshal(map[string]map[string]string{\"content\": {\n\t\t\"raw\": comment,\n\t}})\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"json encoding\")\n\t}\n\tpath := fmt.Sprintf(\"%s/2.0/repositories/%s/pullrequests/%d/comments\", b.BaseURL, repo.FullName, pullNum)\n\t_, err = b.makeRequest(\"POST\", path, bytes.NewBuffer(bodyBytes))\n\treturn err\n}", "func AddComment(taskID string, comment string) (*Task, error) {\n\tcl, err := workflow.New(user, appkey, authtoken)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\twfTask, err := cl.AddComment(taskID, comment)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// XXX: refactor\n\ttask := Task{\n\t\tID: wfTask.ID,\n\t\tTitle: wfTask.Title,\n\t\tCreatedDate: wfTask.CreatedDate,\n\t\tURL: wfTask.URL,\n\t\tDue: wfTask.Due,\n\t\tPeriod: wfTask.Period,\n\t\tDateLastActivity: wfTask.DateLastActivity,\n\t\tDesc: wfTask.Desc,\n\t\tChecklistItems: wfTask.ChecklistItems,\n\t\tList: &BoardList{\n\t\t\twfTask.List.Board,\n\t\t\twfTask.List.List,\n\t\t},\n\t}\n\treturn &task, nil\n}", "func EditIssueComment(ctx *context.APIContext) {\n\t// swagger:operation PATCH /repos/{owner}/{repo}/issues/comments/{id} issue issueEditComment\n\t// ---\n\t// summary: Edit a comment\n\t// consumes:\n\t// - application/json\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: id\n\t// in: path\n\t// description: id of the comment to edit\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// - name: body\n\t// in: body\n\t// schema:\n\t// \"$ref\": \"#/definitions/EditIssueCommentOption\"\n\t// responses:\n\t// \"200\":\n\t// \"$ref\": \"#/responses/Comment\"\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/notFound\"\n\n\tform := web.GetForm(ctx).(*api.EditIssueCommentOption)\n\teditIssueComment(ctx, *form)\n}", "func DeleteIssueComment(ctx *context.APIContext) {\n\t// swagger:operation DELETE /repos/{owner}/{repo}/issues/comments/{id} issue issueDeleteComment\n\t// ---\n\t// summary: Delete a comment\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: id\n\t// in: path\n\t// description: id of comment to delete\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// responses:\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/notFound\"\n\n\tdeleteIssueComment(ctx)\n}", "func GetIssueComment(ctx *context.APIContext) {\n\t// swagger:operation GET /repos/{owner}/{repo}/issues/comments/{id} issue issueGetComment\n\t// ---\n\t// summary: Get a comment\n\t// consumes:\n\t// - application/json\n\t// produces:\n\t// - application/json\n\t// parameters:\n\t// - name: owner\n\t// in: path\n\t// description: owner of the repo\n\t// type: string\n\t// required: true\n\t// - name: repo\n\t// in: path\n\t// description: name of the repo\n\t// type: string\n\t// required: true\n\t// - name: id\n\t// in: path\n\t// description: id of the comment\n\t// type: integer\n\t// format: int64\n\t// required: true\n\t// responses:\n\t// \"200\":\n\t// \"$ref\": \"#/responses/Comment\"\n\t// \"204\":\n\t// \"$ref\": \"#/responses/empty\"\n\t// \"403\":\n\t// \"$ref\": \"#/responses/forbidden\"\n\t// \"404\":\n\t// \"$ref\": \"#/responses/notFound\"\n\n\tcomment, err := issues_model.GetCommentByID(ctx, ctx.ParamsInt64(\":id\"))\n\tif err != nil {\n\t\tif issues_model.IsErrCommentNotExist(err) {\n\t\t\tctx.NotFound(err)\n\t\t} else {\n\t\t\tctx.Error(http.StatusInternalServerError, \"GetCommentByID\", err)\n\t\t}\n\t\treturn\n\t}\n\n\tif err = comment.LoadIssue(ctx); err != nil {\n\t\tctx.InternalServerError(err)\n\t\treturn\n\t}\n\tif comment.Issue.RepoID != ctx.Repo.Repository.ID {\n\t\tctx.Status(http.StatusNotFound)\n\t\treturn\n\t}\n\n\tif comment.Type != issues_model.CommentTypeComment {\n\t\tctx.Status(http.StatusNoContent)\n\t\treturn\n\t}\n\n\tif err := comment.LoadPoster(ctx); err != nil {\n\t\tctx.Error(http.StatusInternalServerError, \"comment.LoadPoster\", err)\n\t\treturn\n\t}\n\n\tctx.JSON(http.StatusOK, convert.ToAPIComment(ctx, ctx.Repo.Repository, comment))\n}", "func Comment(ctx context.Context, cfg *v1.Config, pr int, contents []byte) error {\n\tc := newClient(ctx, cfg.Github)\n\treturn c.CommentOnPR(pr, string(contents))\n}", "func (a *ProblemsApiService) CreateComment(ctx _context.Context, problemId string) ApiCreateCommentRequest {\n\treturn ApiCreateCommentRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tproblemId: problemId,\n\t}\n}", "func (j *DSGitHub) EnrichIssueComments(ctx *Ctx, issue map[string]interface{}, comments []map[string]interface{}, affs bool) (richItems []interface{}, err error) {\n\t// type: category, type(_), item_type( ), issue_comment=true\n\t// copy issue: github_repo, repo_name, repository\n\t// copy comment: created_at, updated_at, body, body_analyzed, author_association, url, html_url\n\t// identify: id, id_in_repo, issue_comment_id, url_id\n\t// standard: metadata..., origin, project, project_slug, uuid\n\t// parent: issue_id, issue_number\n\t// calc: n_reactions\n\t// identity: author_... -> commenter_...,\n\t// common: is_github_issue=1, is_github_issue_comment=1\n\tiID, _ := issue[\"id\"]\n\tid, _ := iID.(string)\n\tiIssueID, _ := issue[\"issue_id\"]\n\tissueID := int(iIssueID.(float64))\n\tissueNumber, _ := issue[\"id_in_repo\"]\n\tiNumber, _ := issueNumber.(int)\n\tiGithubRepo, _ := issue[\"github_repo\"]\n\tgithubRepo, _ := iGithubRepo.(string)\n\tcopyIssueFields := []string{\"category\", \"github_repo\", \"repo_name\", \"repository\", \"repo_short_name\", \"pull_request\"}\n\tcopyCommentFields := []string{\"created_at\", \"updated_at\", \"body\", \"body_analyzed\", \"author_association\", \"url\", \"html_url\"}\n\tfor _, comment := range comments {\n\t\trich := make(map[string]interface{})\n\t\tfor _, field := range RawFields {\n\t\t\tv, _ := issue[field]\n\t\t\trich[field] = v\n\t\t}\n\t\tfor _, field := range copyIssueFields {\n\t\t\trich[field], _ = issue[field]\n\t\t}\n\t\tfor _, field := range copyCommentFields {\n\t\t\trich[field], _ = comment[field]\n\t\t}\n\t\tif ctx.Project != \"\" {\n\t\t\trich[\"project\"] = ctx.Project\n\t\t}\n\t\trich[\"type\"] = \"issue_comment\"\n\t\trich[\"item_type\"] = \"issue comment\"\n\t\trich[\"issue_comment\"] = true\n\t\trich[\"issue_created_at\"], _ = issue[\"created_at\"]\n\t\trich[\"issue_id\"] = issueID\n\t\trich[\"issue_number\"] = issueNumber\n\t\tiCID, _ := comment[\"id\"]\n\t\tcid := int64(iCID.(float64))\n\t\trich[\"id_in_repo\"] = cid\n\t\trich[\"issue_comment_id\"] = cid\n\t\trich[\"id\"] = id + \"/comment/\" + fmt.Sprintf(\"%d\", cid)\n\t\trich[\"url_id\"] = fmt.Sprintf(\"%s/issues/%d/comments/%d\", githubRepo, iNumber, cid)\n\t\treactions := 0\n\t\tiReactions, ok := Dig(comment, []string{\"reactions\", \"total_count\"}, false, true)\n\t\tif ok {\n\t\t\treactions = int(iReactions.(float64))\n\t\t}\n\t\trich[\"n_reactions\"] = reactions\n\t\trich[\"commenter_association\"], _ = comment[\"author_association\"]\n\t\trich[\"commenter_login\"], _ = Dig(comment, []string{\"user\", \"login\"}, false, true)\n\t\tiCommenterData, ok := comment[\"user_data\"]\n\t\tif ok && iCommenterData != nil {\n\t\t\tuser, _ := iCommenterData.(map[string]interface{})\n\t\t\trich[\"author_login\"], _ = user[\"login\"]\n\t\t\trich[\"author_name\"], _ = user[\"name\"]\n\t\t\trich[\"author_avatar_url\"], _ = user[\"avatar_url\"]\n\t\t\trich[\"commenter_avatar_url\"] = rich[\"author_avatar_url\"]\n\t\t\trich[\"commenter_name\"], _ = user[\"name\"]\n\t\t\trich[\"commenter_domain\"] = nil\n\t\t\tiEmail, ok := user[\"email\"]\n\t\t\tif ok {\n\t\t\t\temail, _ := iEmail.(string)\n\t\t\t\tary := strings.Split(email, \"@\")\n\t\t\t\tif len(ary) > 1 {\n\t\t\t\t\trich[\"commenter_domain\"] = strings.TrimSpace(ary[1])\n\t\t\t\t}\n\t\t\t}\n\t\t\trich[\"commenter_org\"], _ = user[\"company\"]\n\t\t\trich[\"commenter_location\"], _ = user[\"location\"]\n\t\t\trich[\"commenter_geolocation\"] = nil\n\t\t} else {\n\t\t\trich[\"author_login\"] = nil\n\t\t\trich[\"author_name\"] = nil\n\t\t\trich[\"author_avatar_url\"] = nil\n\t\t\trich[\"commenter_avatar_url\"] = nil\n\t\t\trich[\"commenter_name\"] = nil\n\t\t\trich[\"commenter_domain\"] = nil\n\t\t\trich[\"commenter_org\"] = nil\n\t\t\trich[\"commenter_location\"] = nil\n\t\t\trich[\"commenter_geolocation\"] = nil\n\t\t}\n\t\tiCreatedAt, _ := comment[\"created_at\"]\n\t\tcreatedAt, _ := TimeParseInterfaceString(iCreatedAt)\n\t\trich[j.DateField(ctx)] = createdAt\n\t\tif affs {\n\t\t\tauthorKey := \"user_data\"\n\t\t\tvar affsItems map[string]interface{}\n\t\t\taffsItems, err = j.AffsItems(ctx, comment, GitHubIssueCommentRoles, createdAt)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor prop, value := range affsItems {\n\t\t\t\trich[prop] = value\n\t\t\t}\n\t\t\tfor _, suff := range AffsFields {\n\t\t\t\trich[Author+suff] = rich[authorKey+suff]\n\t\t\t\trich[\"commenter\"+suff] = rich[authorKey+suff]\n\t\t\t}\n\t\t\torgsKey := authorKey + MultiOrgNames\n\t\t\t_, ok := Dig(rich, []string{orgsKey}, false, true)\n\t\t\tif !ok {\n\t\t\t\trich[orgsKey] = []interface{}{}\n\t\t\t}\n\t\t}\n\t\tfor prop, value := range CommonFields(j, createdAt, j.Category) {\n\t\t\trich[prop] = value\n\t\t}\n\t\tfor prop, value := range CommonFields(j, createdAt, j.Category+\"_comment\") {\n\t\t\trich[prop] = value\n\t\t}\n\t\trichItems = append(richItems, rich)\n\t}\n\treturn\n}", "func (s *Server) handleIssueCommentEvent(l *logrus.Entry, ic scm.IssueCommentHook) {\n\tl = l.WithFields(logrus.Fields{\n\t\tscmprovider.OrgLogField: ic.Repo.Namespace,\n\t\tscmprovider.RepoLogField: ic.Repo.Name,\n\t\tscmprovider.PrLogField: ic.Issue.Number,\n\t\t\"author\": ic.Comment.Author.Login,\n\t\t\"url\": ic.Comment.Link,\n\t})\n\tl.Infof(\"Issue comment %s.\", ic.Action)\n\tevent := &scmprovider.GenericCommentEvent{\n\t\tGUID: ic.GUID,\n\t\tIsPR: ic.Issue.PullRequest != nil,\n\t\tAction: ic.Action,\n\t\tBody: ic.Comment.Body,\n\t\tLink: ic.Comment.Link,\n\t\tNumber: ic.Issue.Number,\n\t\tRepo: ic.Repo,\n\t\tAuthor: ic.Comment.Author,\n\t\tIssueAuthor: ic.Issue.Author,\n\t\tAssignees: ic.Issue.Assignees,\n\t\tIssueState: ic.Issue.State,\n\t\tIssueBody: ic.Issue.Body,\n\t\tIssueLink: ic.Issue.Link,\n\t}\n\tif ic.Issue.PullRequest != nil {\n\t\tupdatedPR, _, err := s.ClientAgent.SCMProviderClient.PullRequests.Find(context.Background(), fmt.Sprintf(\"%s/%s\",\n\t\t\tic.Repo.Namespace, ic.Repo.Name), ic.Issue.Number)\n\t\tif err != nil {\n\t\t\tl.WithError(err).Error(\"Error fetching Pull Request details.\")\n\t\t} else {\n\t\t\tevent.HeadSha = updatedPR.Head.Sha\n\t\t}\n\t}\n\ts.handleGenericComment(\n\t\tl,\n\t\tevent,\n\t)\n}", "func (r *Repo) CheckIssueComments(issueNumber int, sender string) error {\n\turi := \"/repos/\" + r.Owner.Login + \"/issues/\" + strconv.Itoa(issueNumber) + \"/comments\"\n\trespBody, err := GithubAPICall(uri, \"GET\", nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar comments []Comment\n\tjson.Unmarshal(respBody, &comments)\n\n\tif r.CheckIssue(comments, sender) {\n\t\tfmt.Println(\"We are trying to merge PR the corresponding PR\\n\")\n\t\treturn MergePullRequest(r.Owner.Login, r.FullName, issueNumber)\n\t}\n\treturn nil\n}", "func (o *CaCertificateCreateReqWeb) SetComment(v string) {\n\to.Comment = v\n}", "func AddCommentHook(hookPoint boil.HookPoint, commentHook CommentHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\tcommentBeforeInsertHooks = append(commentBeforeInsertHooks, commentHook)\n\tcase boil.BeforeUpdateHook:\n\t\tcommentBeforeUpdateHooks = append(commentBeforeUpdateHooks, commentHook)\n\tcase boil.BeforeDeleteHook:\n\t\tcommentBeforeDeleteHooks = append(commentBeforeDeleteHooks, commentHook)\n\tcase boil.BeforeUpsertHook:\n\t\tcommentBeforeUpsertHooks = append(commentBeforeUpsertHooks, commentHook)\n\tcase boil.AfterInsertHook:\n\t\tcommentAfterInsertHooks = append(commentAfterInsertHooks, commentHook)\n\tcase boil.AfterSelectHook:\n\t\tcommentAfterSelectHooks = append(commentAfterSelectHooks, commentHook)\n\tcase boil.AfterUpdateHook:\n\t\tcommentAfterUpdateHooks = append(commentAfterUpdateHooks, commentHook)\n\tcase boil.AfterDeleteHook:\n\t\tcommentAfterDeleteHooks = append(commentAfterDeleteHooks, commentHook)\n\tcase boil.AfterUpsertHook:\n\t\tcommentAfterUpsertHooks = append(commentAfterUpsertHooks, commentHook)\n\t}\n}", "func (c *client) CreateComment(org, repo string, number int, comment string) error {\n\treturn c.CreateCommentWithContext(context.Background(), org, repo, number, comment)\n}", "func (o *DeployKey) SetComment(v string) {\n\to.Comment = &v\n}", "func (it IssueTracker) SubmitIssueChanges(issue *Issue, comment string) error {\n\terrFmt := \"Error updating issue \" + strconv.Itoa(issue.Id) + \": %s\"\n\tif !it.IsAuthenticated() {\n\t\treturn fmt.Errorf(errFmt, \"user is not authenticated!\")\n\t}\n\toldIssue, err := it.GetBug(issue.Project, issue.Id)\n\tif err != nil {\n\t\treturn fmt.Errorf(errFmt, err)\n\t}\n\tpostData := struct {\n\t\tContent string `json:\"content\"`\n\t\tUpdates struct {\n\t\t\tTitle *string `json:\"summary\"`\n\t\t\tLabels []string `json:\"labels\"`\n\t\t} `json:\"updates\"`\n\t}{\n\t\tContent: comment,\n\t}\n\tif issue.Title != oldIssue.Title {\n\t\tpostData.Updates.Title = &issue.Title\n\t}\n\t// TODO(borenet): Add other issue attributes, eg. Owner.\n\tlabels := make(map[string]int)\n\tfor _, label := range issue.Labels {\n\t\tlabels[label] = LABEL_ADDED\n\t}\n\tfor _, label := range oldIssue.Labels {\n\t\tif _, ok := labels[label]; ok {\n\t\t\tlabels[label] = LABEL_UNCHANGED\n\t\t} else {\n\t\t\tlabels[label] = LABEL_REMOVED\n\t\t}\n\t}\n\tlabelChanges := make([]string, 0)\n\tfor labelName, present := range labels {\n\t\tif present == LABEL_REMOVED {\n\t\t\tlabelChanges = append(labelChanges, \"-\"+labelName)\n\t\t} else if present == LABEL_ADDED {\n\t\t\tlabelChanges = append(labelChanges, labelName)\n\t\t}\n\t}\n\tif len(labelChanges) > 0 {\n\t\tpostData.Updates.Labels = labelChanges\n\t}\n\n\tpostBytes, err := json.Marshal(&postData)\n\tif err != nil {\n\t\treturn fmt.Errorf(errFmt, err)\n\t}\n\trequestURL := ISSUE_API_URL + issue.Project + \"/issues/\" +\n\t\tstrconv.Itoa(issue.Id) + \"/comments\"\n\tresp, err := it.OAuthTransport.Client().Post(\n\t\trequestURL, \"application/json\", bytes.NewReader(postBytes))\n\tif err != nil {\n\t\treturn fmt.Errorf(errFmt, err)\n\t}\n\tdefer resp.Body.Close()\n\tbody, _ := ioutil.ReadAll(resp.Body)\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn fmt.Errorf(errFmt, fmt.Sprintf(\n\t\t\t\"Issue tracker returned code %d:%v\", resp.StatusCode, string(body)))\n\t}\n\treturn nil\n}", "func (o *V0037JobProperties) SetComment(v string) {\n\to.Comment = &v\n}", "func (c *Client) DeleteIssueComment(issueIDOrKey string, commentID int) (*IssueComment, error) {\n\treturn c.DeleteIssueCommentContext(context.Background(), issueIDOrKey, commentID)\n}", "func (f *Fetcher) AddConflictCommentToPR(pr *github.PullRequest) error {\n\tif pr.User == nil || pr.User.Login == nil {\n\t\tlogrus.Infof(\"failed to get user from PR %d: empty User\", *(pr.Number))\n\t\treturn nil\n\t}\n\n\tcomments, err := f.client.ListComments(*(pr.Number))\n\tif err != nil {\n\t\treturn err\n\t}\n\t//logrus.Infof(\"PR %d: There are %d comments\", *(pr.Number), len(comments))\n\n\tbody := fmt.Sprintf(utils.PRConflictComment, *(pr.User.Login))\n\tnewComment := &github.IssueComment{\n\t\tBody: &body,\n\t}\n\n\tif len(comments) == 0 {\n\t\treturn f.client.AddCommentToIssue(*(pr.Number), newComment)\n\t}\n\n\tlatestComment := comments[len(comments)-1]\n\tif strings.Contains(*(latestComment.Body), utils.PRConflictSubStr) {\n\t\t//logrus.Infof(\"PR %d: latest comment %s \\nhas\\n %s\", *(pr.Number), *(latestComment.Body), utils.PRConflictSubStr)\n\t\t// remove all existing conflict comments\n\t\tfor _, comment := range comments[:(len(comments) - 1)] {\n\t\t\tif strings.Contains(*(comment.Body), utils.PRConflictSubStr) {\n\t\t\t\tif err := f.client.RemoveComment(*(comment.ID)); err != nil {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\t// remove all existing conflict comments\n\tfor _, comment := range comments {\n\t\tif strings.Contains(*(comment.Body), utils.PRConflictSubStr) {\n\t\t\tif err := f.client.RemoveComment(*(comment.ID)); err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t}\n\n\t// add a brand new conflict comment\n\treturn f.client.AddCommentToIssue(*(pr.Number), newComment)\n}", "func (c *client) CreatePullRequestReviewComment(org, repo string, number int, rc ReviewComment) error {\n\tc.log(\"CreatePullRequestReviewComment\", org, repo, number, rc)\n\n\t// TODO: remove custom Accept headers when their respective API fully launches.\n\tacceptHeaders := []string{\n\t\t// https://developer.github.com/changes/2016-05-12-reactions-api-preview/\n\t\t\"application/vnd.github.squirrel-girl-preview\",\n\t\t// https://developer.github.com/changes/2019-10-03-multi-line-comments/\n\t\t\"application/vnd.github.comfort-fade-preview+json\",\n\t}\n\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodPost,\n\t\taccept: strings.Join(acceptHeaders, \", \"),\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/pulls/%d/comments\", org, repo, number),\n\t\torg: org,\n\t\trequestBody: &rc,\n\t\texitCodes: []int{201},\n\t}, nil)\n\treturn err\n}", "func (mr *MockPostsRepoInterfaceMockRecorder) AddComment(arg0, arg1 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"AddComment\", reflect.TypeOf((*MockPostsRepoInterface)(nil).AddComment), arg0, arg1)\n}", "func (z *Client) CreateTicketComment(ctx context.Context, ticketID int64, ticketComment TicketComment) (TicketComment, error) {\n\ttype comment struct {\n\t\tTicket struct {\n\t\t\tTicketComment TicketComment `json:\"comment\"`\n\t\t} `json:\"ticket\"`\n\t}\n\n\tdata := &comment{}\n\tdata.Ticket.TicketComment = ticketComment\n\n\tbody, err := z.put(ctx, fmt.Sprintf(\"/tickets/%d.json\", ticketID), data)\n\tif err != nil {\n\t\treturn TicketComment{}, err\n\t}\n\n\tresult := TicketComment{}\n\terr = json.Unmarshal(body, &result)\n\tif err != nil {\n\t\treturn TicketComment{}, err\n\t}\n\n\treturn result, err\n}", "func (r *Repository) GetIssueCommentURL() string {\n\tif r == nil || r.IssueCommentURL == nil {\n\t\treturn \"\"\n\t}\n\treturn *r.IssueCommentURL\n}", "func (fc *fakeClient) CreateComment(owner, repo string, number int, comment string) error {\n\tfc.commentsAdded[number] = append(fc.commentsAdded[number], comment)\n\treturn nil\n}", "func (fc *fakeClient) CreateComment(owner, repo string, number int, comment string) error {\n\tfc.commentsAdded[number] = append(fc.commentsAdded[number], comment)\n\treturn nil\n}", "func (s *SubmitFeedbackInput) SetComment(v string) *SubmitFeedbackInput {\n\ts.Comment = &v\n\treturn s\n}", "func (c *Client) CreateIssueCommentsNotification(issueIDOrKey string, commentID int, input *CreateIssueCommentsNotificationInput) (*IssueComment, error) {\n\treturn c.CreateIssueCommentsNotificationContext(context.Background(), issueIDOrKey, commentID, input)\n}", "func (o *ReservationModel) SetComment(v string) {\n\to.Comment = &v\n}", "func (s *UpdateRevisionInput) SetComment(v string) *UpdateRevisionInput {\n\ts.Comment = &v\n\treturn s\n}", "func (o *WatchlistScreeningIndividualReviewCreateResponse) SetComment(v string) {\n\to.Comment.Set(&v)\n}", "func (c *client) CreateCommentReaction(org, repo string, id int, reaction string) error {\n\tc.log(\"CreateCommentReaction\", org, repo, id, reaction)\n\tr := Reaction{Content: reaction}\n\t_, err := c.request(&request{\n\t\tmethod: http.MethodPost,\n\t\tpath: fmt.Sprintf(\"/repos/%s/%s/issues/comments/%d/reactions\", org, repo, id),\n\t\taccept: \"application/vnd.github.squirrel-girl-preview\",\n\t\torg: org,\n\t\texitCodes: []int{201},\n\t\trequestBody: &r,\n\t}, nil)\n\treturn err\n}", "func (b *Client) ReactToComment(repo models.Repo, pullNum int, commentID int64, reaction string) error { // nolint revive\n\t// TODO: Bitbucket support for reactions\n\treturn nil\n}", "func (c *Client) GetIssueComment(issueIDOrKey string, commentID int) (*IssueComment, error) {\n\treturn c.GetIssueCommentContext(context.Background(), issueIDOrKey, commentID)\n}", "func (s *RevokeRevisionOutput) SetComment(v string) *RevokeRevisionOutput {\n\ts.Comment = &v\n\treturn s\n}", "func (m *sdt) SetComment(val string) {\n\tm.commentField = val\n}", "func (_options *ForceApproveOptions) SetComment(comment string) *ForceApproveOptions {\n\t_options.Comment = core.StringPtr(comment)\n\treturn _options\n}", "func (c *Client) UpdateIssueComment(issueIDOrKey string, commentID int, input *UpdateIssueCommentInput) (*IssueComment, error) {\n\treturn c.UpdateIssueCommentContext(context.Background(), issueIDOrKey, commentID, input)\n}", "func (s *RevisionEntry) SetComment(v string) *RevisionEntry {\n\ts.Comment = &v\n\treturn s\n}", "func (s *CreateRevisionInput) SetComment(v string) *CreateRevisionInput {\n\ts.Comment = &v\n\treturn s\n}", "func (s *DiscussionsService) AddSnippetDiscussionNote(pid interface{}, snippet int, discussion string, opt *AddSnippetDiscussionNoteOptions, options ...RequestOptionFunc) (*Note, *Response, error) {\n\tproject, err := parseID(pid)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tu := fmt.Sprintf(\"projects/%s/snippets/%d/discussions/%s/notes\",\n\t\tPathEscape(project),\n\t\tsnippet,\n\t\tdiscussion,\n\t)\n\n\treq, err := s.client.NewRequest(http.MethodPost, u, opt, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tn := new(Note)\n\tresp, err := s.client.Do(req, n)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn n, resp, nil\n}", "func AddComment(m *Comment) (id int64, err error) {\n\to := orm.NewOrm()\n\n\t//Query review authority\n\tvar b []*BlackMan\n\tqs := o.QueryTable(new(BlackMan))\n\tqs.Filter(\"UserId__exact\", m.NotifyUserId)\n\n\tif _, err = qs.All(&b); err == nil {\n\t\tfor v, item := range b {\n\t\t\tlog.Println(v, \"==\", item)\n\t\t\tif item.BlackMan == m.UserId {\n\t\t\t\treturn -1, errors.New(\"Users can not comment on the Blacklist\")\n\t\t\t}\n\t\t}\n\t\tid, err = o.Insert(m)\n\t\tif err == nil {\n\t\t\t//There are new comments to notify users\n\t\t\tu := User{Id: m.NotifyUserId}\n\t\t\to.Read(&u)\n\t\t\t//log.Println(u)\n\t\t\tu.CommentNotify++\n\t\t\to.Update(&u)\n\t\t}\n\t}\n\treturn\n}", "func (s *UpdateRevisionOutput) SetComment(v string) *UpdateRevisionOutput {\n\ts.Comment = &v\n\treturn s\n}", "func (s *GetRevisionOutput) SetComment(v string) *GetRevisionOutput {\n\ts.Comment = &v\n\treturn s\n}", "func (g Gitlab) PostResolutionComment(ctx context.Context, projectID, issueIID IntID, resolution Resolution) error {\n\tvar builder strings.Builder\n\terr := resolutionCommentTemplate.Execute(&builder, struct {\n\t\tResolution string\n\t\tResolveReason string\n\t}{\n\t\tstring(resolution.Tag),\n\t\tresolution.Reason,\n\t})\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tresp, err := g.NewRequest(ctx).\n\t\tSetBody(NoteParams{Body: builder.String()}).\n\t\tPost(g.APIV4URL(\"projects\", projectID, \"issues\", issueIID, \"notes\"))\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif resp.IsError() {\n\t\treturn trace.Wrap(responseError(resp))\n\t}\n\n\tlogger.Get(ctx).Debug(\"Successfully posted a resolution comment to the issue\")\n\treturn nil\n}", "func (o *RowCommentCreate) SetComment(v string) {\n\to.Comment = v\n}", "func (gh *GitHubTestEventsHandler) HandleIssueCommentEvent(logger log.Logger, comment *gogh.IssueCommentEvent) error {\n\tif !utils.Contains(handledCommentActions, *comment.Action) {\n\t\treturn nil\n\t}\n\n\tprLoader := ghservice.NewPullRequestLazyLoaderFromComment(gh.Client, comment)\n\tuserPerm := command.NewPermissionService(gh.Client, *comment.Sender.Login, prLoader)\n\n\tcmdHandler := command.CommentCmdHandler{Client: gh.Client}\n\n\tcmdHandler.Register(&command.RunCmd{\n\t\tPluginName: ProwPluginName,\n\t\tUserPermissionService: userPerm,\n\t\tWhenAddedOrEdited: func() error {\n\t\t\treturn gh.checkTestsAndSetStatus(logger, prLoader)\n\t\t}})\n\n\tcmdHandler.Register(&BypassCmd{\n\t\tuserPermissionService: userPerm,\n\t\twhenDeleted: func() error {\n\t\t\treturn gh.checkTestsAndSetStatus(logger, prLoader)\n\t\t},\n\t\twhenAddedOrEdited: func() error {\n\t\t\tpullRequest, err := prLoader.Load()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treportBypassCommand(pullRequest)\n\t\t\tstatusService := gh.newTestStatusService(logger, pullRequest)\n\t\t\treturn statusService.okWithoutTests(*comment.Sender.Login)\n\t\t}})\n\n\terr := cmdHandler.Handle(logger, comment)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t}\n\treturn err\n}", "func (issue *Issue) UpdateComment(id int, comment io.Reader) (*Comment, error) {\n\turl := fmt.Sprintf(\"%s/issue/%s/comment/%d\", BaseUrl, issue.Key, id)\n\tcode, body := execRequest(\"PUT\", url, comment)\n\tif code == http.StatusOK {\n\t\tvar jiraComment Comment\n\t\terr := json.Unmarshal(body, &jiraComment)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn &jiraComment, nil\n\t} else {\n\t\treturn nil, handleJiraError(body)\n\t}\n}", "func (t *T) AddComments(comments ...string) {\n\tt.commentExpressions = append(t.commentExpressions, comments...)\n}", "func (o *BucketsCollectionGetParams) SetComment(comment *string) {\n\to.Comment = comment\n}", "func (c *Client) CreateIssueCommentsNotificationContext(ctx context.Context, issueIDOrKey string, commentID int, input *CreateIssueCommentsNotificationInput) (*IssueComment, error) {\n\tu := fmt.Sprintf(\"/api/v2/issues/%v/comments/%v/notifications\", issueIDOrKey, commentID)\n\n\treq, err := c.NewRequest(\"POST\", u, input)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tissueComment := new(IssueComment)\n\tif err := c.Do(ctx, req, &issueComment); err != nil {\n\t\treturn nil, err\n\t}\n\treturn issueComment, nil\n}", "func (s *CreateRevisionOutput) SetComment(v string) *CreateRevisionOutput {\n\ts.Comment = &v\n\treturn s\n}", "func (k *Item) SetComment(s string) {\n\tk.SetString(CommentKey, s)\n}", "func (g Gitlab) PostReviewComment(ctx context.Context, projectID, issueIID IntID, review types.AccessReview) error {\n\tvar builder strings.Builder\n\terr := reviewCommentTemplate.Execute(&builder, struct {\n\t\ttypes.AccessReview\n\t\tProposedState string\n\t\tTimeFormat string\n\t}{\n\t\treview,\n\t\treview.ProposedState.String(),\n\t\ttime.RFC822,\n\t})\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tresp, err := g.NewRequest(ctx).\n\t\tSetBody(NoteParams{Body: builder.String()}).\n\t\tPost(g.APIV4URL(\"projects\", projectID, \"issues\", issueIID, \"notes\"))\n\tif err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\tif resp.IsError() {\n\t\treturn trace.Wrap(responseError(resp))\n\t}\n\n\tlogger.Get(ctx).Debug(\"Successfully posted a review comment to the issue\")\n\treturn nil\n}", "func (cfg *Config) AddComment(section, comment string) {\n\tsection_ := C.CString(section)\n\tcomment_ := C.CString(comment)\n\tdefer freeString(section_)\n\tdefer freeString(comment_)\n\tC.al_add_config_comment((*C.ALLEGRO_CONFIG)(cfg), section_, comment_)\n}", "func (ao *AggregateOptions) SetComment(s string) *AggregateOptions {\n\tao.Comment = &s\n\treturn ao\n}", "func (wouo *WorkOrderUpdateOne) AddComments(c ...*Comment) *WorkOrderUpdateOne {\n\tids := make([]string, len(c))\n\tfor i := range c {\n\t\tids[i] = c[i].ID\n\t}\n\treturn wouo.AddCommentIDs(ids...)\n}", "func AddCMFPaidprogramCommentHook(hookPoint boil.HookPoint, cmfPaidprogramCommentHook CMFPaidprogramCommentHook) {\n\tswitch hookPoint {\n\tcase boil.BeforeInsertHook:\n\t\tcmfPaidprogramCommentBeforeInsertHooks = append(cmfPaidprogramCommentBeforeInsertHooks, cmfPaidprogramCommentHook)\n\tcase boil.BeforeUpdateHook:\n\t\tcmfPaidprogramCommentBeforeUpdateHooks = append(cmfPaidprogramCommentBeforeUpdateHooks, cmfPaidprogramCommentHook)\n\tcase boil.BeforeDeleteHook:\n\t\tcmfPaidprogramCommentBeforeDeleteHooks = append(cmfPaidprogramCommentBeforeDeleteHooks, cmfPaidprogramCommentHook)\n\tcase boil.BeforeUpsertHook:\n\t\tcmfPaidprogramCommentBeforeUpsertHooks = append(cmfPaidprogramCommentBeforeUpsertHooks, cmfPaidprogramCommentHook)\n\tcase boil.AfterInsertHook:\n\t\tcmfPaidprogramCommentAfterInsertHooks = append(cmfPaidprogramCommentAfterInsertHooks, cmfPaidprogramCommentHook)\n\tcase boil.AfterSelectHook:\n\t\tcmfPaidprogramCommentAfterSelectHooks = append(cmfPaidprogramCommentAfterSelectHooks, cmfPaidprogramCommentHook)\n\tcase boil.AfterUpdateHook:\n\t\tcmfPaidprogramCommentAfterUpdateHooks = append(cmfPaidprogramCommentAfterUpdateHooks, cmfPaidprogramCommentHook)\n\tcase boil.AfterDeleteHook:\n\t\tcmfPaidprogramCommentAfterDeleteHooks = append(cmfPaidprogramCommentAfterDeleteHooks, cmfPaidprogramCommentHook)\n\tcase boil.AfterUpsertHook:\n\t\tcmfPaidprogramCommentAfterUpsertHooks = append(cmfPaidprogramCommentAfterUpsertHooks, cmfPaidprogramCommentHook)\n\t}\n}", "func NewCommentUsecase(p repository.Persistence, c repository.GithubCli) usecase.CommentUsecase {\n\treturn &commentUsecase{p, c}\n}", "func (issue *Issue) UpdateComment(\n\tid int, comment *Comment,\n) (*Comment, error) {\n\n\turl := fmt.Sprintf(\"%s/issue/%s/comment/%d\", BaseURL, issue.Key, id)\n\tencodedParams, err := json.Marshal(comment)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcode, body := execRequest(\"PUT\", url, bytes.NewBuffer(encodedParams))\n\tif code == http.StatusOK {\n\t\tvar jiraComment Comment\n\t\terr := json.Unmarshal(body, &jiraComment)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn &jiraComment, nil\n\t}\n\treturn nil, handleJiraError(body)\n}", "func (wou *WorkOrderUpdate) AddComments(c ...*Comment) *WorkOrderUpdate {\n\tids := make([]string, len(c))\n\tfor i := range c {\n\t\tids[i] = c[i].ID\n\t}\n\treturn wou.AddCommentIDs(ids...)\n}", "func (a *Client) AddCommentToTask(params *AddCommentToTaskParams, authInfo runtime.ClientAuthInfoWriter, opts ...ClientOption) (*AddCommentToTaskCreated, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewAddCommentToTaskParams()\n\t}\n\top := &runtime.ClientOperation{\n\t\tID: \"addCommentToTask\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/tasks/{id}/comments\",\n\t\tProducesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tConsumesMediaTypes: []string{\"application/vnd.goswagger.examples.task-tracker.v1+json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &AddCommentToTaskReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t}\n\tfor _, opt := range opts {\n\t\topt(op)\n\t}\n\n\tresult, err := a.transport.Submit(op)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*AddCommentToTaskCreated)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\tunexpectedSuccess := result.(*AddCommentToTaskDefault)\n\treturn nil, runtime.NewAPIError(\"unexpected success response: content available as default response in error\", unexpectedSuccess, unexpectedSuccess.Code())\n}", "func (o *AccountCollectionGetParams) SetComment(comment *string) {\n\to.Comment = comment\n}", "func (o *Project) AddComments(ctx context.Context, exec boil.ContextExecutor, insert bool, related ...*Comment) error {\n\tvar err error\n\tfor _, rel := range related {\n\t\tif insert {\n\t\t\trel.ProjectID = o.ID\n\t\t\tif err = rel.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t\t}\n\t\t} else {\n\t\t\tupdateQuery := fmt.Sprintf(\n\t\t\t\t\"UPDATE `comment` SET %s WHERE %s\",\n\t\t\t\tstrmangle.SetParamNames(\"`\", \"`\", 0, []string{\"project_id\"}),\n\t\t\t\tstrmangle.WhereClause(\"`\", \"`\", 0, commentPrimaryKeyColumns),\n\t\t\t)\n\t\t\tvalues := []interface{}{o.ID, rel.ID}\n\n\t\t\tif boil.IsDebug(ctx) {\n\t\t\t\twriter := boil.DebugWriterFrom(ctx)\n\t\t\t\tfmt.Fprintln(writer, updateQuery)\n\t\t\t\tfmt.Fprintln(writer, values)\n\t\t\t}\n\t\t\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to update foreign table\")\n\t\t\t}\n\n\t\t\trel.ProjectID = o.ID\n\t\t}\n\t}\n\n\tif o.R == nil {\n\t\to.R = &projectR{\n\t\t\tComments: related,\n\t\t}\n\t} else {\n\t\to.R.Comments = append(o.R.Comments, related...)\n\t}\n\n\tfor _, rel := range related {\n\t\tif rel.R == nil {\n\t\t\trel.R = &commentR{\n\t\t\t\tProject: o,\n\t\t\t}\n\t\t} else {\n\t\t\trel.R.Project = o\n\t\t}\n\t}\n\treturn nil\n}", "func resourceCommentCreate(d *schema.ResourceData, m interface{}) error {\n\tconfig := m.(*Config)\n\tbody := d.Get(\"body\").(string)\n\tissueKey := d.Get(\"issue_key\").(string)\n\n\tc := jira.Comment{Body: body}\n\n\tcomment, res, err := config.jiraClient.Issue.AddComment(issueKey, &c)\n\n\tif err != nil {\n\t\tbody, _ := ioutil.ReadAll(res.Body)\n\t\treturn errors.Wrapf(err, \"creating jira issue failed: %s\", body)\n\t}\n\n\td.SetId(comment.ID)\n\n\treturn resourceCommentRead(d, m)\n}", "func CreateIssueDependency(user *User, issue, dep *Issue) error {\n\tsess := x.NewSession()\n\tdefer sess.Close()\n\tif err := sess.Begin(); err != nil {\n\t\treturn err\n\t}\n\n\t// Check if it aleready exists\n\texists, err := issueDepExists(sess, issue.ID, dep.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif exists {\n\t\treturn ErrDependencyExists{issue.ID, dep.ID}\n\t}\n\t// And if it would be circular\n\tcircular, err := issueDepExists(sess, dep.ID, issue.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif circular {\n\t\treturn ErrCircularDependency{issue.ID, dep.ID}\n\t}\n\n\tif _, err := sess.Insert(&IssueDependency{\n\t\tUserID: user.ID,\n\t\tIssueID: issue.ID,\n\t\tDependencyID: dep.ID,\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\t// Add comment referencing the new dependency\n\tif err = createIssueDependencyComment(sess, user, issue, dep, true); err != nil {\n\t\treturn err\n\t}\n\n\treturn sess.Commit()\n}", "func (_options *ApproveOptions) SetComment(comment string) *ApproveOptions {\n\t_options.Comment = core.StringPtr(comment)\n\treturn _options\n}" ]
[ "0.75926155", "0.74070245", "0.692096", "0.6866009", "0.6845925", "0.68433267", "0.6652136", "0.65989584", "0.6485817", "0.6417826", "0.63988376", "0.63939494", "0.6198317", "0.61021245", "0.6083263", "0.60437244", "0.6038438", "0.6009101", "0.5939593", "0.59367543", "0.5919957", "0.5883192", "0.5874753", "0.5873795", "0.5673018", "0.5642936", "0.56388354", "0.5607376", "0.5553826", "0.544817", "0.5429191", "0.5400345", "0.53921455", "0.53831124", "0.5342129", "0.5314488", "0.52971566", "0.5252765", "0.52164507", "0.5202755", "0.5189898", "0.5176518", "0.51382613", "0.51270443", "0.51160043", "0.5109941", "0.508586", "0.50819814", "0.5064429", "0.5059989", "0.5050573", "0.5041562", "0.502666", "0.50105876", "0.49984175", "0.49910986", "0.49846232", "0.49811986", "0.49811986", "0.49771592", "0.49756348", "0.49750105", "0.4972048", "0.4949972", "0.49469683", "0.4935119", "0.49340585", "0.49338144", "0.49228796", "0.49219784", "0.49152613", "0.4912815", "0.4907291", "0.49045092", "0.4864912", "0.48518342", "0.48516235", "0.48470274", "0.48401794", "0.48341167", "0.48174265", "0.48054683", "0.47856963", "0.47814965", "0.47794876", "0.4772039", "0.47625208", "0.47549605", "0.47486928", "0.47207722", "0.47099477", "0.47087488", "0.47054893", "0.4694445", "0.46893677", "0.46825355", "0.46777147", "0.46702406", "0.4667181", "0.46670452" ]
0.7894321
0
setDryRun sets the CQ dry run bit on the issue.
func (r *AutoRoller) setDryRun(issue *autoroll.AutoRollIssue, dryRun bool) error { info, err := issue.ToGerritChangeInfo() if err != nil { return fmt.Errorf("Failed to convert issue to Gerrit ChangeInfo: %s", err) } if dryRun { if err := r.rm.SendToGerritDryRun(info, ""); err != nil { return err } } else { if err := r.rm.SendToGerritCQ(info, ""); err != nil { return err } } updated, err := r.retrieveRoll(issue.Issue) if err != nil { return err } return r.recent.Update(updated) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *CreateClientGatewayRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *UpdateVpnConnectionRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *CreateRouteRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *ReadConsumptionAccountRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *CreateSnapshotExportTaskRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *UpdateNetRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *DeregisterVmsInLoadBalancerRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *LinkPrivateIpsRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func setDryRun(ctx context.Context, val int64) context.Context {\n\tdval := &dryRunMarker{val}\n\treturn context.WithValue(ctx, dryRunMarker{}, dval)\n}", "func (o *UpdateServerCertificateRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *CreateUserRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *CreateLoadBalancerRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *LinkPublicIpRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *UpdateLoadBalancerRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (scl *SimpleConfigurationLayer) SetDryRun(dryRun *bool) {\n\tscl.DryRun = dryRun\n}", "func (c *Client) SetDryRun(enable bool) {\n\tc.transport.DryRun = enable\n}", "func (o *CreateLoadBalancerListenersRequest) SetDryRun(v bool) {\n\to.DryRun = &v\n}", "func (o *PostApplyManifestParams) SetDryRun(dryRun *bool) {\n\to.DryRun = dryRun\n}", "func (t *Txn) DryRun() {\n\tif t.dryRun {\n\t\tpanic(\"transaction is already in dry-run mode\")\n\t}\n\tt.dryRun = true\n}", "func (o ApplicationStatusOperationStateOperationSyncOutput) DryRun() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSync) *bool { return v.DryRun }).(pulumi.BoolPtrOutput)\n}", "func DryRun(g *types.Cmd) {\n\tg.AddOptions(\"--dry-run\")\n}", "func DryRun(g *types.Cmd) {\n\tg.AddOptions(\"--dry-run\")\n}", "func (o ApplicationOperationSyncOutput) DryRun() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperationSync) *bool { return v.DryRun }).(pulumi.BoolPtrOutput)\n}", "func WithDryRun(value bool) ClientOption {\n\treturn func(opts *ClientOptions) {\n\t\topts.DryRun = value\n\t}\n}", "func (n *Node) DryRun() {\n\tif n.commandMutators == nil {\n\t\tn.commandMutators = []commandMutator{}\n\t}\n\n\tn.commandMutators = append(n.commandMutators,\n\t\tfunc(c *exec.NodeCmd) *exec.NodeCmd {\n\t\t\treturn c.DryRun()\n\t\t},\n\t)\n}", "func DryRunOption(dryRun bool) Option {\n\treturn func(c *ExecutorConfig) { c.DryRun = dryRun }\n}", "func (d *DSP) SetWetDryMix(prewet, postwet, dry float64) error {\n\tres := C.FMOD_DSP_SetWetDryMix(d.cptr, C.float(prewet), C.float(postwet), C.float(dry))\n\treturn errs[res]\n}", "func (o ApplicationStatusOperationStateOperationSyncPtrOutput) DryRun() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperationSync) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.DryRun\n\t}).(pulumi.BoolPtrOutput)\n}", "func WithDryRun(dryRun bool) BuilderOptionFunc {\n\treturn func(b *Builder) error {\n\t\tb.dryRun = dryRun\n\t\treturn nil\n\t}\n}", "func (o ApplicationOperationSyncPtrOutput) DryRun() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationOperationSync) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.DryRun\n\t}).(pulumi.BoolPtrOutput)\n}", "func (drs DryRunStrategy) ClientDryRun() bool {\n\treturn drs == DryRunClient\n}", "func DryRun(dryRun bool) meterOption {\n\treturn func(m *ProgressMeter) {\n\t\tm.dryRun = dryRun\n\t}\n}", "func (o *CreateClientGatewayRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *UpdateVpnConnectionRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func IsDryRun(ctx context.Context) bool {\n\tr := getDryRun(ctx)\n\tif r == nil {\n\t\treturn false\n\t}\n\treturn true\n}", "func EnableDryRun(dryrun bool) func(*Context) {\n\treturn func(ctx *Context) {\n\t\tctx.options.DryRun = dryrun\n\t}\n}", "func (o *UpdateNetRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *ReadConsumptionAccountRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *UpdateServerCertificateRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (k *kubectlContext) DryRun(filename string) (string, error) {\n\tout, err := k.do(\"apply\", \"--server-dry-run\", \"--output\", \"yaml\", \"--filename\", filename)\n\treturn string(out), err\n}", "func DryRunCommand() *cobra.Command {\n\tvar table bool\n\n\tvar dryrun = &cobra.Command{\n\t\tUse: \"dryrun\",\n\t\tShort: \"dryrun simulates a payout\",\n\t\tLong: \"dryrun simulates a payout and prints the result in json or a table\",\n\t\tExample: `tzpay dryrun <cycle>`,\n\t\tRun: func(cmd *cobra.Command, args []string) {\n\t\t\tif len(args) == 0 {\n\t\t\t\tlog.Fatal(\"Missing cycle as argument.\")\n\t\t\t}\n\n\t\t\tdryrun := NewDryRun(args[0], table)\n\t\t\tdryrun.execute()\n\t\t},\n\t}\n\tdryrun.PersistentFlags().BoolVarP(&table, \"table\", \"t\", false, \"formats result into a table (Default: json)\")\n\n\treturn dryrun\n}", "func NewDryRunClient(getToken func() []byte, censor func([]byte) []byte, graphqlEndpoint string, bases ...string) (Client, error) {\n\treturn NewDryRunClientWithFields(logrus.Fields{}, getToken, censor, graphqlEndpoint, bases...)\n}", "func (o *CreateUserRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *ReadConsumptionAccountRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o ServicePerimetersServicePerimeterOutput) UseExplicitDryRunSpec() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v ServicePerimetersServicePerimeter) *bool { return v.UseExplicitDryRunSpec }).(pulumi.BoolPtrOutput)\n}", "func (o *UpdateVpnConnectionRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *LinkPrivateIpsRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *CreateSnapshotExportTaskRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *UpdateNetRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *CreateRouteRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (j *joinData) DryRun() bool {\n\treturn j.dryRun\n}", "func (o *DeregisterVmsInLoadBalancerRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (mr *MockAtomicLogicMockRecorder) DrySend(sender, value, fee, nonce, allowNonceGap, chainHeight interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DrySend\", reflect.TypeOf((*MockAtomicLogic)(nil).DrySend), sender, value, fee, nonce, allowNonceGap, chainHeight)\n}", "func NewDryRun(cycle string, table bool) DryRun {\n\tconfig, err := config.New()\n\tif err != nil {\n\t\tlog.WithField(\"error\", err.Error()).Fatal(\"Failed to load config.\")\n\t}\n\n\t// Clear sensitive data if loaded\n\tconfig.Key.Password = \"\"\n\tconfig.Key.Esk = \"\"\n\n\tc, err := strconv.Atoi(cycle)\n\tif err != nil {\n\t\tlog.WithField(\"error\", err.Error()).Fatal(\"Failed to parse cycle argument into integer.\")\n\t}\n\n\tpayout, err := payout.New(config, c, false, false)\n\tif err != nil {\n\t\tlog.WithField(\"error\", err.Error()).Fatal(\"Failed to intialize payout.\")\n\t}\n\n\treturn DryRun{\n\t\tpayout: payout,\n\t\tconfig: config,\n\t\tcycle: c,\n\t\ttable: table,\n\t\ttzkt: tzkt.NewTZKT(config.API.TZKT),\n\t}\n}", "func (o *UpdateServerCertificateRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func NewDryRunClient(token string) *Client {\n\treturn &Client{\n\t\tclient: &http.Client{},\n\t\ttoken: token,\n\t\tbase: githubBase,\n\t\tdry: true,\n\t}\n}", "func (o *UpdateLoadBalancerRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *LinkPublicIpRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *CreateClientGatewayRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func NewDryRunClient(c Client) Client {\n\treturn &dryRunClient{client: c}\n}", "func (mr *MockLogicMockRecorder) DrySend(sender, value, fee, nonce, allowNonceGap, chainHeight interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DrySend\", reflect.TypeOf((*MockLogic)(nil).DrySend), sender, value, fee, nonce, allowNonceGap, chainHeight)\n}", "func (in *DryRun) DeepCopy() *DryRun {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(DryRun)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func cmdRset(s *session, cmd *Command) {\n\ts.draft = nil\n\ts.Send(250, \"OK\")\n}", "func (o *CreateLoadBalancerRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (b *ManifestsApplyConfiguration) WithDryRunOption(value *DryRunOptionApplyConfiguration) *ManifestsApplyConfiguration {\n\tb.DryRunOption = value\n\treturn b\n}", "func (o *CreateUserRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func queryDryRun(w io.Writer, projectID string) error {\n\t// projectID := \"my-project-id\"\n\tctx := context.Background()\n\tclient, err := bigquery.NewClient(ctx, projectID)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"bigquery.NewClient: %w\", err)\n\t}\n\tdefer client.Close()\n\n\tq := client.Query(`\n\tSELECT\n\t\tname,\n\t\tCOUNT(*) as name_count\n\tFROM ` + \"`bigquery-public-data.usa_names.usa_1910_2013`\" + `\n\tWHERE state = 'WA'\n\tGROUP BY name`)\n\tq.DryRun = true\n\t// Location must match that of the dataset(s) referenced in the query.\n\tq.Location = \"US\"\n\n\tjob, err := q.Run(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Dry run is not asynchronous, so get the latest status and statistics.\n\tstatus := job.LastStatus()\n\tif err := status.Err(); err != nil {\n\t\treturn err\n\t}\n\tfmt.Fprintf(w, \"This query will process %d bytes\\n\", status.Statistics.TotalBytesProcessed)\n\treturn nil\n}", "func (o *DeregisterVmsInLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func TestDryRun(t *testing.T) {\n\n\t// start API server\n\ts, err := kubeapiservertesting.StartTestServer(t, kubeapiservertesting.NewDefaultTestServerOptions(), []string{\n\t\t\"--disable-admission-plugins=ServiceAccount,StorageObjectInUseProtection\",\n\t\t\"--runtime-config=api/all=true\",\n\t}, framework.SharedEtcd())\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer s.TearDownFn()\n\n\tclient, err := kubernetes.NewForConfig(s.ClientConfig)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdynamicClient, err := dynamic.NewForConfig(s.ClientConfig)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// create CRDs so we can make sure that custom resources do not get lost\n\tetcd.CreateTestCRDs(t, apiextensionsclientset.NewForConfigOrDie(s.ClientConfig), false, etcd.GetCustomResourceDefinitionData()...)\n\n\tif _, err := client.CoreV1().Namespaces().Create(context.TODO(), &v1.Namespace{ObjectMeta: metav1.ObjectMeta{Name: testNamespace}}, metav1.CreateOptions{}); err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tdryrunData := etcd.GetEtcdStorageData()\n\n\t// dry run specific stub overrides\n\tfor resource, stub := range map[schema.GroupVersionResource]string{\n\t\t// need to change event's namespace field to match dry run test\n\t\tgvr(\"\", \"v1\", \"events\"): `{\"involvedObject\": {\"namespace\": \"dryrunnamespace\"}, \"message\": \"some data here\", \"metadata\": {\"name\": \"event1\"}}`,\n\t} {\n\t\tdata := dryrunData[resource]\n\t\tdata.Stub = stub\n\t\tdryrunData[resource] = data\n\t}\n\n\t// gather resources to test\n\t_, resources, err := client.Discovery().ServerGroupsAndResources()\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to get ServerGroupsAndResources with error: %+v\", err)\n\t}\n\n\tfor _, resourceToTest := range etcd.GetResources(t, resources) {\n\t\tt.Run(resourceToTest.Mapping.Resource.String(), func(t *testing.T) {\n\t\t\tmapping := resourceToTest.Mapping\n\t\t\tgvk := resourceToTest.Mapping.GroupVersionKind\n\t\t\tgvResource := resourceToTest.Mapping.Resource\n\t\t\tkind := gvk.Kind\n\n\t\t\tif kindAllowList.Has(kind) {\n\t\t\t\tt.Skip(\"allowlisted\")\n\t\t\t}\n\n\t\t\ttestData, hasTest := dryrunData[gvResource]\n\n\t\t\tif !hasTest {\n\t\t\t\tt.Fatalf(\"no test data for %s. Please add a test for your new type to etcd.GetEtcdStorageData().\", gvResource)\n\t\t\t}\n\n\t\t\trsc, obj, err := etcd.JSONToUnstructured(testData.Stub, testNamespace, mapping, dynamicClient)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"failed to unmarshal stub (%v): %v\", testData.Stub, err)\n\t\t\t}\n\n\t\t\tname := obj.GetName()\n\n\t\t\tDryRunCreateTest(t, rsc, obj, gvResource)\n\n\t\t\tif _, err := rsc.Create(context.TODO(), obj, metav1.CreateOptions{}); err != nil {\n\t\t\t\tt.Fatalf(\"failed to create stub for %s: %#v\", gvResource, err)\n\t\t\t}\n\n\t\t\tDryRunUpdateTest(t, rsc, name)\n\t\t\tDryRunPatchTest(t, rsc, name)\n\t\t\tDryRunScalePatchTest(t, rsc, name)\n\t\t\tDryRunScaleUpdateTest(t, rsc, name)\n\t\t\tif resourceToTest.HasDeleteCollection {\n\t\t\t\tDryRunDeleteCollectionTest(t, rsc, name)\n\t\t\t}\n\t\t\tDryRunDeleteTest(t, rsc, name)\n\n\t\t\tif err = rsc.Delete(context.TODO(), obj.GetName(), *metav1.NewDeleteOptions(0)); err != nil {\n\t\t\t\tt.Fatalf(\"deleting final object failed: %v\", err)\n\t\t\t}\n\t\t})\n\t}\n}", "func (drs DryRunStrategy) ClientOrServerDryRun() bool {\n\treturn drs == DryRunClient || drs == DryRunServer\n}", "func getDryRun(ctx context.Context) *dryRunMarker {\n\tret := ctx.Value(dryRunMarker{})\n\tif ret == nil {\n\t\treturn nil\n\t}\n\tif dr, ok := ret.(*dryRunMarker); ok {\n\t\treturn dr\n\t}\n\treturn nil\n}", "func (o *CreateSnapshotExportTaskRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func newCmdDryRunDisabled(streams genericclioptions.IOStreams) *cobra.Command {\n\to := newDryrunOptions(streams)\n\to.enabledDryRun = false\n\n\tcmd := &cobra.Command{\n\t\tUse: \"disable [WatermarkPodAutoscaler name]\",\n\t\tShort: \"disable WPA(s) dry-run mode\",\n\t\tExample: fmt.Sprintf(dryrunExample, \"dry-run disable\"),\n\t\tSilenceUsage: true,\n\t\tRunE: func(c *cobra.Command, args []string) error {\n\t\t\tif err := o.complete(c, args); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := o.validate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn o.run(o.patch)\n\t\t},\n\t}\n\n\tcmd.Flags().StringVarP(&o.labelSelector, \"label-selector\", \"l\", \"\", \"Use to select WPA based in their labels\")\n\tcmd.Flags().BoolVarP(&o.allWPA, \"all\", \"\", false, \"Use select all existing WPA instances in a cluster\")\n\tcmd.Flags().BoolVarP(&o.allNamespaces, \"all-namespaces\", \"\", false, \"Use to search in all namespaces\")\n\to.configFlags.AddFlags(cmd.Flags())\n\n\treturn cmd\n}", "func (o *LinkPrivateIpsRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (mysqld *Mysqld) SetReadOnly(on bool) error {\n\t// temp logging, to be removed in v17\n\tvar newState string\n\tswitch on {\n\tcase false:\n\t\tnewState = \"ReadWrite\"\n\tcase true:\n\t\tnewState = \"ReadOnly\"\n\t}\n\tlog.Infof(\"SetReadOnly setting to : %s\", newState)\n\n\tquery := \"SET GLOBAL read_only = \"\n\tif on {\n\t\tquery += \"ON\"\n\t} else {\n\t\tquery += \"OFF\"\n\t}\n\treturn mysqld.ExecuteSuperQuery(context.TODO(), query)\n}", "func (o *LinkPublicIpRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *UpdateLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func NewDryRunClientWithFields(fields logrus.Fields, getToken func() []byte, censor func([]byte) []byte, graphqlEndpoint string, bases ...string) (Client, error) {\n\t_, _, client, err := NewClientFromOptions(fields, ClientOptions{\n\t\tCensor: censor,\n\t\tGetToken: getToken,\n\t\tGraphqlEndpoint: graphqlEndpoint,\n\t\tBases: bases,\n\t\tDryRun: true,\n\t}.Default())\n\treturn client, err\n}", "func newCmdDryRunEnabled(streams genericclioptions.IOStreams) *cobra.Command {\n\to := newDryrunOptions(streams)\n\to.enabledDryRun = true\n\n\tcmd := &cobra.Command{\n\t\tUse: \"enable [WatermarkPodAutoscaler name]\",\n\t\tShort: \"enable WPA(s) dry-run mode\",\n\t\tExample: fmt.Sprintf(dryrunExample, \"dry-run enable\"),\n\t\tSilenceUsage: true,\n\t\tRunE: func(c *cobra.Command, args []string) error {\n\t\t\tif err := o.complete(c, args); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := o.validate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treturn o.run(o.patch)\n\t\t},\n\t}\n\n\tcmd.Flags().StringVarP(&o.labelSelector, \"label-selector\", \"l\", \"\", \"Use to select WPA based in their labels\")\n\tcmd.Flags().BoolVarP(&o.allWPA, \"all\", \"\", false, \"Use select all existing WPA instances in a cluster\")\n\tcmd.Flags().BoolVarP(&o.allNamespaces, \"all-namespaces\", \"\", false, \"Use to search in all namespaces\")\n\to.configFlags.AddFlags(cmd.Flags())\n\n\treturn cmd\n}", "func NewCmdDryRun(streams genericclioptions.IOStreams) *cobra.Command {\n\to := newDryrunOptions(streams)\n\n\tcmd := &cobra.Command{\n\t\tUse: \"dry-run\",\n\t\tShort: \"configure WPA(s) dry-run\",\n\t\tSilenceUsage: true,\n\t}\n\n\tcmd.AddCommand(newCmdDryRunEnabled(streams))\n\tcmd.AddCommand(newCmdDryRunDisabled(streams))\n\tcmd.AddCommand(newCmdRevert(streams))\n\tcmd.AddCommand(newCmdDryRunList(streams))\n\n\to.configFlags.AddFlags(cmd.Flags())\n\n\treturn cmd\n}", "func (c MockClient) SetDebug(d bool) MockClient {\n\tc.restyClient.SetDebug(d)\n\treturn c\n}", "func (o LookupPolicyResultOutput) DryRunSpec() GoogleCloudOrgpolicyV2PolicySpecResponseOutput {\n\treturn o.ApplyT(func(v LookupPolicyResult) GoogleCloudOrgpolicyV2PolicySpecResponse { return v.DryRunSpec }).(GoogleCloudOrgpolicyV2PolicySpecResponseOutput)\n}", "func SetReplOrRun(rep bool) {\n\tIsRepl = rep\n\tif IsRepl {\n\t\tos.Setenv(\"COZY_RUNNING_IN_REPL\", \"true\")\n\t}\n}", "func NewDry(screen *ui.Screen, cfg Config) (*Dry, error) {\n\n\td, err := docker.ConnectToDaemon(cfg.dockerEnv())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdry, err := newDry(screen, d)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif cfg.MonitorMode {\n\t\tdry.changeView(Monitor)\n\t\twidgets.Monitor.RefreshRate(cfg.MonitorRefreshRate)\n\t}\n\treturn dry, nil\n}", "func (drs DryRunStrategy) ServerDryRun() bool {\n\treturn drs == DryRunServer\n}", "func NewDry(screen *ui.Screen, env *drydocker.DockerEnv) (*Dry, error) {\n\td, err := drydocker.ConnectToDaemon(env)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn newDry(screen, d)\n}", "func (o *CreateRouteRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (m *MockAtomicLogic) DrySend(sender interface{}, value, fee util.String, nonce uint64, allowNonceGap bool, chainHeight uint64) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DrySend\", sender, value, fee, nonce, allowNonceGap, chainHeight)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (fmd *FakeMysqlDaemon) SetReadOnly(on bool) error {\n\tfmd.ReadOnly = on\n\treturn nil\n}", "func (fmd *FakeMysqlDaemon) SetReadOnly(on bool) error {\n\tfmd.ReadOnly = on\n\treturn nil\n}", "func (o *CreateLoadBalancerListenersRequest) GetDryRun() bool {\n\tif o == nil || o.DryRun == nil {\n\t\tvar ret bool\n\t\treturn ret\n\t}\n\treturn *o.DryRun\n}", "func (o *CreateLoadBalancerRequest) HasDryRun() bool {\n\tif o != nil && o.DryRun != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (f *Session) SetBypassDocumentValidation(b bool) *Session {\n\tf.bulkWriteOptions = append(f.bulkWriteOptions, options.BulkWrite().SetBypassDocumentValidation(b))\n\tf.findOneAndReplaceOpts = append(f.findOneAndReplaceOpts,\n\t\toptions.FindOneAndReplace().SetBypassDocumentValidation(b))\n\tf.findOneAndUpdateOpts = append(f.findOneAndUpdateOpts, options.FindOneAndUpdate().SetBypassDocumentValidation(b))\n\tf.updateOpts = append(f.updateOpts, options.Update().SetBypassDocumentValidation(b))\n\n\treturn f\n}", "func (o *ReadConsumptionAccountRequest) GetDryRunOk() (*bool, bool) {\n\tif o == nil || o.DryRun == nil {\n\t\treturn nil, false\n\t}\n\treturn o.DryRun, true\n}", "func (st *fakeConn) SetReadOnly(readOnly bool) {\n\tst.readOnly = readOnly\n}", "func (m *MockLogic) DrySend(sender interface{}, value, fee util.String, nonce uint64, allowNonceGap bool, chainHeight uint64) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"DrySend\", sender, value, fee, nonce, allowNonceGap, chainHeight)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (p SetScriptSourceParams) WithDryRun(dryRun bool) *SetScriptSourceParams {\n\tp.DryRun = dryRun\n\treturn &p\n}", "func HasDryRunArg(r *http.Request) bool {\n\treturn urlQueryHasArg(r.URL, queryArgDryRun)\n}", "func (c *Client) Rset() (err error) {\n\tif _, err = c.Cmd(\"%s\\r\\n\", RSET); err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func ninjaDryRun(ctx context.Context, r ninjaRunner, targets []string) (string, string, error) {\n\t// -n means dry-run.\n\targs := []string{\"-d\", \"explain\", \"--verbose\", \"-n\"}\n\targs = append(args, targets...)\n\n\tvar stdout, stderr bytes.Buffer\n\terr := r.run(ctx, args, &stdout, &stderr)\n\tif err != nil {\n\t\t// stdout and stderr are normally not emitted because they're very\n\t\t// noisy, but if the dry run fails then they'll likely contain the\n\t\t// information necessary to understand the failure.\n\t\tstreams.Stdout(ctx).Write(stdout.Bytes())\n\t\tstreams.Stderr(ctx).Write(stderr.Bytes())\n\t}\n\treturn stdout.String(), stderr.String(), err\n}" ]
[ "0.73999643", "0.7393846", "0.7326219", "0.73214686", "0.73066396", "0.72535145", "0.7234877", "0.72340846", "0.72241324", "0.71935266", "0.71693945", "0.71318513", "0.7131546", "0.71252865", "0.7059958", "0.70524216", "0.68877614", "0.67712647", "0.6483322", "0.61109775", "0.60328364", "0.60328364", "0.6005266", "0.58921885", "0.58617246", "0.57790095", "0.57140744", "0.5674309", "0.56393015", "0.5616615", "0.5533708", "0.5530059", "0.550289", "0.5479813", "0.547306", "0.54580516", "0.5441079", "0.5431473", "0.5411837", "0.53797257", "0.5292658", "0.5277582", "0.5274284", "0.522878", "0.52211344", "0.5220839", "0.5190286", "0.5179864", "0.51717514", "0.5161537", "0.51499045", "0.51202625", "0.51178974", "0.5117111", "0.5107719", "0.5097578", "0.5096466", "0.5082272", "0.5058921", "0.50453824", "0.503661", "0.49868286", "0.49365044", "0.4936016", "0.49322164", "0.4926323", "0.49065825", "0.48835993", "0.48480627", "0.4809738", "0.48058158", "0.47826526", "0.47731766", "0.47665346", "0.47499466", "0.4727098", "0.47135752", "0.46971828", "0.4689146", "0.4682468", "0.46710232", "0.46671733", "0.46623453", "0.46528983", "0.46504387", "0.46253243", "0.46111053", "0.46038467", "0.45786977", "0.45786977", "0.4551199", "0.4484307", "0.4483834", "0.44740966", "0.44734606", "0.4464198", "0.4438088", "0.44265187", "0.4418551", "0.44172737" ]
0.75252444
0
updateCurrentRoll retrieves updated information about the current DEPS roll.
func (r *AutoRoller) updateCurrentRoll() error { currentRoll := r.recent.CurrentRoll() if currentRoll == nil { return nil } currentResult := currentRoll.Result updated, err := r.retrieveRoll(currentRoll.Issue) if err != nil { return err } // We have to rely on data we store for the dry run case. if !updated.Closed && util.In(currentResult, autoroll.DRY_RUN_RESULTS) { updated.Result = currentResult } // If the current roll succeeded, we need to make sure we update the // repo so that we see the roll commit. This can take some time, so // we have to repeatedly update until we see the commit. if updated.Committed { sklog.Infof("Roll succeeded (%d); syncing the repo until it lands.", currentRoll.Issue) for { sklog.Info("Syncing...") sklog.Infof("Looking for %s", currentRoll.RollingTo) if err := r.rm.ForceUpdate(); err != nil { return err } rolledPast, err := r.rm.RolledPast(currentRoll.RollingTo) if err != nil { return err } if rolledPast { break } time.Sleep(10 * time.Second) } r.liveness.Reset() } return r.recent.Update(updated) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (e *Election) updateCurrentVote(new *messages.LeaderLevelMessage) {\n\tnew.VolunteerPriority = e.GetVolunteerPriority(new.VolunteerMessage.Signer)\n\t// Add to display, add the previous vote\n\te.executeDisplay(new)\n\tif e.CurrentVote.Rank >= 0 {\n\t\tprev := e.CurrentVote.Copy()\n\t\tprev.Justification = []messages.LeaderLevelMessage{}\n\t\tprev.PreviousVote = nil\n\t\tnew.PreviousVote = prev\n\t}\n\n\t/**** Commitment checking ****/\n\t// Check if this is sequential\n\tif new.VolunteerPriority == e.CurrentVote.VolunteerPriority &&\n\t\t(e.CurrentVote.Rank+1 == new.Rank && e.CurrentVote.Level+1 == new.Level) {\n\t\te.CommitmentTally++\n\t} else {\n\t\t// Resetting the tally\n\t\te.CommitmentTally = 1\n\t}\n\n\tif new.Rank == 0 {\n\t\t// Rank 0 doesn't count towards the tally\n\t\te.CommitmentTally = 0\n\t}\n\te.CurrentVote = *new\n\n}", "func (s *server) updateCurrentTerm(term uint64, leaderName string) {\n\t//_assert(term > s.currentTerm,\n\t//\t\"upadteCurrentTerm: update is called when term is not larger than currentTerm\")\n\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\t// Store previous values temporarily.\n\t//prevTerm := s.currentTerm\n\t//prevLeader := s.leader\n\n\t// set currentTerm = T, convert to follower (§5.1)\n\t// stop heartbeats before step-down\n\tif s.state == Leader {\n\t\ts.mutex.Unlock()\n\t\tfor _, peer := range s.peers {\n\t\t\tpeer.stopHeartbeat(false)\n\t\t}\n\t\ts.mutex.Lock()\n\t}\n\t// update the term and clear vote for\n\tif s.state != Follower {\n\t\ts.mutex.Unlock()\n\t\ts.setState(Follower)\n\t\ts.mutex.Lock()\n\t}\n\ts.currentTerm = term\n\ts.leader = leaderName\n\ts.votedFor = \"\"\n}", "func (st *MemStorage) UpdateCurrent(gun string, update MetaUpdate) error {\n\tid := entryKey(gun, update.Role)\n\tst.lock.Lock()\n\tdefer st.lock.Unlock()\n\tif space, ok := st.tufMeta[id]; ok {\n\t\tfor _, v := range space {\n\t\t\tif v.version >= update.Version {\n\t\t\t\treturn &ErrOldVersion{}\n\t\t\t}\n\t\t}\n\t}\n\tversion := ver{version: update.Version, data: update.Data, createupdate: time.Now()}\n\tst.tufMeta[id] = append(st.tufMeta[id], &version)\n\tchecksumBytes := sha256.Sum256(update.Data)\n\tchecksum := hex.EncodeToString(checksumBytes[:])\n\n\t_, ok := st.checksums[gun]\n\tif !ok {\n\t\tst.checksums[gun] = make(map[string]ver)\n\t}\n\tst.checksums[gun][checksum] = version\n\treturn nil\n}", "func (d *dataUpdateTracker) current() uint64 {\n\td.mu.Lock()\n\tdefer d.mu.Unlock()\n\treturn d.Current.idx\n}", "func (auto *Autoscaler) updateScalerCurrentMetrics(scaler *commtypes.BcsAutoscaler) error {\n\trefKind := scaler.Spec.ScaleTargetRef.Kind\n\trefNs := scaler.Spec.ScaleTargetRef.Namespace\n\trefName := scaler.Spec.ScaleTargetRef.Name\n\n\tfor _, current := range scaler.Status.CurrentMetrics {\n\n\t\tswitch current.Type {\n\t\tcase commtypes.ResourceMetricSourceType:\n\t\t\tmetrics, err := auto.resourceMetrics.GetResourceMetric(current.Name, scaler.GetUuid())\n\t\t\tif err != nil {\n\t\t\t\tblog.Errorf(\"scaler %s ref(%s:%s:%s) get resources %s metrics error %s\", scaler.GetUuid(),\n\t\t\t\t\trefKind, refNs, refName, current.Name, err.Error())\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tvar totalValue float32\n\t\t\tvar num float32\n\t\t\tfor k, metric := range metrics {\n\t\t\t\tif metric.Value == 0 {\n\t\t\t\t\tblog.Warnf(\"scaler %s taskgroup %s metric value is zero\", scaler.GetUuid(), k)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tnum++\n\t\t\t\ttotalValue += metric.Value\n\t\t\t}\n\n\t\t\tif num > 0 {\n\t\t\t\tvalue, _ := strconv.ParseFloat(fmt.Sprintf(\"%.2f\", totalValue/num), 32)\n\t\t\t\tcurrent.Current.AverageUtilization = float32(value)\n\t\t\t\tcurrent.Timestamp = time.Now()\n\t\t\t}\n\n\t\tcase commtypes.ExternalMetricSourceType:\n\t\t\t//todo\n\n\t\tdefault:\n\t\t\tblog.Errorf(\"scaler %s metrics %s type %s is invalid\", scaler.GetUuid(), current.Name, current.Type)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *AutoRoller) GetActiveRoll() state_machine.RollCLImpl {\n\treturn r.currentRoll\n}", "func (i *ASN2ASDescClient) Current(ASN int) (string, error) {\n\tallDates, err := i.importedDates()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(allDates) < 0 {\n\t\treturn \"\", err\n\t}\n\tcurrent := allDates[len(allDates)-1]\n\tresult, err := redis.String(i.conn.Do(\"HGET\", fmt.Sprintf(\"asd:%d\", ASN), current))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn result, nil\n}", "func (game *Game) Roll(pinsDown int) {\n\tgame.currentFrame.Roll(pinsDown)\n\n\tif game.currentFrame.IsComplete() {\n\t\tgame.score += game.beforePreviousFrame.Bonus(\n\t\t\tgame.previousFrame,\n\t\t\tgame.currentFrame,\n\t\t)\n\t\tgame.score += game.currentFrame.Score()\n\t\tgame.beforePreviousFrame = game.previousFrame\n\t\tgame.previousFrame = game.currentFrame\n\t\tgame.currentFrame = *new(Frame)\n\t}\n}", "func (b *Bot) UpdateCurrentState(rs RobotState) error {\n\tb.mu.Lock()\n\tdefer b.mu.Unlock()\n\tif rs.X < 0 || rs.X > 9 {\n\t\treturn fmt.Errorf(\"Robot state X position (%d, y) exceeds warehouse dimensions\", rs.X)\n\t}\n\tif rs.Y < 0 || rs.Y > 9 {\n\t\treturn fmt.Errorf(\"Robot state Y position (x, %d) exceeds warehouse dimensions\", rs.Y)\n\t}\n\tb.state = rs\n\treturn nil\n}", "func UpdateCurrentChannel() error {\n\ttargetPath := \"/misc/ota/target_channel.json\"\n\tcontents, err := ioutil.ReadFile(targetPath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"no target channel recorded in %v: %w\", targetPath, err)\n\t}\n\tcurrentPath := \"/misc/ota/current_channel.json\"\n\tpartPath := currentPath + \".part\"\n\tf, err := os.Create(partPath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to write current channel to %v: %w\", partPath, err)\n\t}\n\tdefer f.Close()\n\tbuf := bytes.NewBuffer(contents)\n\t_, err = buf.WriteTo(f)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to write current channel to %v: %w\", currentPath, err)\n\t}\n\tf.Sync()\n\tf.Close()\n\tif err := os.Rename(partPath, currentPath); err != nil {\n\t\treturn fmt.Errorf(\"error moving %v to %v: %w\", partPath, currentPath, err)\n\t}\n\treturn nil\n}", "func (o StorageClusterSpecUpdateStrategyOutput) RollingUpdate() StorageClusterSpecUpdateStrategyRollingUpdatePtrOutput {\n\treturn o.ApplyT(func(v StorageClusterSpecUpdateStrategy) *StorageClusterSpecUpdateStrategyRollingUpdate {\n\t\treturn v.RollingUpdate\n\t}).(StorageClusterSpecUpdateStrategyRollingUpdatePtrOutput)\n}", "func (r *AutoRoller) rollFinished(ctx context.Context, justFinished codereview.RollImpl) error {\n\trecent := r.recent.GetRecentRolls()\n\t// Sanity check: pop any rolls which occurred after the one which just\n\t// finished.\n\tidx := -1\n\tvar currentRoll *autoroll.AutoRollIssue\n\tfor i, roll := range recent {\n\t\tissue := fmt.Sprintf(\"%d\", roll.Issue)\n\t\tif issue == justFinished.IssueID() {\n\t\t\tidx = i\n\t\t\tcurrentRoll = roll\n\t\t\tbreak\n\t\t}\n\t}\n\tif currentRoll == nil {\n\t\treturn skerr.Fmt(\"Unable to find just-finished roll %q in recent list!\", justFinished.IssueID())\n\t}\n\n\t// Feed AutoRoll stats into metrics.\n\tv := int64(0)\n\tif currentRoll.Closed && currentRoll.Committed {\n\t\tv = int64(1)\n\t}\n\tmetrics2.GetInt64Metric(\"autoroll_last_roll_result\", map[string]string{\"roller\": r.cfg.RollerName}).Update(v)\n\n\trecent = recent[idx:]\n\tvar lastRoll *autoroll.AutoRollIssue\n\tif len(recent) > 1 {\n\t\tlastRoll = recent[1]\n\t} else {\n\t\t// If there are no other rolls, then the below alerts do not apply.\n\t\treturn nil\n\t}\n\n\tissueURL := fmt.Sprintf(\"%s%d\", r.codereview.GetIssueUrlBase(), currentRoll.Issue)\n\n\t// Send notifications if this roll had a different result from the last\n\t// roll, ie. success -> failure or failure -> success.\n\tcurrentSuccess := util.In(currentRoll.Result, autoroll.SUCCESS_RESULTS)\n\tlastSuccess := util.In(lastRoll.Result, autoroll.SUCCESS_RESULTS)\n\tif lastRoll != nil {\n\t\tif currentSuccess && !lastSuccess {\n\t\t\tr.notifier.SendNewSuccess(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t} else if !currentSuccess && lastSuccess {\n\t\t\tr.notifier.SendNewFailure(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t}\n\t}\n\n\t// Send a notification if the last N rolls failed in a row.\n\tnFailed := 0\n\t// recent is in reverse chronological order.\n\tfor _, roll := range recent {\n\t\tif util.In(roll.Result, autoroll.SUCCESS_RESULTS) {\n\t\t\tbreak\n\t\t} else {\n\t\t\tnFailed++\n\t\t}\n\t}\n\tif nFailed == notifyIfLastNFailed {\n\t\tr.notifier.SendLastNFailed(ctx, notifyIfLastNFailed, issueURL)\n\t}\n\n\treturn nil\n}", "func (a *Accounting) callUpdateAccounting() (modules.AccountingInfo, error) {\n\tvar ai modules.AccountingInfo\n\n\t// Get Renter information\n\t//\n\t// NOTE: renter is optional so can be nil\n\tvar renterErr error\n\tif a.staticRenter != nil {\n\t\tvar spending modules.ContractorSpending\n\t\tspending, renterErr = a.staticRenter.PeriodSpending()\n\t\tif renterErr == nil {\n\t\t\t_, _, unspentUnallocated := spending.SpendingBreakdown()\n\t\t\tai.Renter.UnspentUnallocated = unspentUnallocated\n\t\t\tai.Renter.WithheldFunds = spending.WithheldFunds\n\t\t}\n\t}\n\n\t// Get Wallet information\n\tsc, sf, _, walletErr := a.staticWallet.ConfirmedBalance()\n\tif walletErr == nil {\n\t\tai.Wallet.ConfirmedSiacoinBalance = sc\n\t\tai.Wallet.ConfirmedSiafundBalance = sf\n\t}\n\n\t// Update the Accounting state\n\terr := errors.Compose(renterErr, walletErr)\n\tif err == nil {\n\t\ta.mu.Lock()\n\t\ta.persistence.Renter = ai.Renter\n\t\ta.persistence.Wallet = ai.Wallet\n\t\ta.persistence.Timestamp = time.Now().Unix()\n\t\ta.mu.Unlock()\n\t}\n\treturn ai, err\n}", "func (o StorageClusterSpecUpdateStrategyPtrOutput) RollingUpdate() StorageClusterSpecUpdateStrategyRollingUpdatePtrOutput {\n\treturn o.ApplyT(func(v *StorageClusterSpecUpdateStrategy) *StorageClusterSpecUpdateStrategyRollingUpdate {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.RollingUpdate\n\t}).(StorageClusterSpecUpdateStrategyRollingUpdatePtrOutput)\n}", "func (dsc *ReconcileDaemonSet) rollingUpdate(ds *appsv1alpha1.DaemonSet, hash string) (delay time.Duration, err error) {\n\n\tif ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.StandardRollingUpdateType {\n\t\treturn delay, dsc.standardRollingUpdate(ds, hash)\n\t} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.SurgingRollingUpdateType {\n\t\treturn dsc.surgingRollingUpdate(ds, hash)\n\t\t//} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.InplaceRollingUpdateType {\n\t\t//\treturn dsc.inplaceRollingUpdate(ds, hash)\n\t} else {\n\t\tklog.Errorf(\"no matched RollingUpdate type\")\n\t}\n\treturn\n}", "func SetRoll(iC *InterfaceConfig, degreeVal int) bool {\n dutyCycle := 0\n if degreeVal < 0 {\n dutyCycle = calcdutyCycleFromNeutralCenter(iC, Achannel, \"left\", (degreeVal*-1))\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n } else if degreeVal > 0{\n dutyCycle= calcdutyCycleFromNeutralCenter(iC, Achannel, \"right\", degreeVal)\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n } else if degreeVal == 0 {\n dutyCycle = calcdutyCycleFromNeutralCenter(iC, Achannel, \"left\", 0)\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n }\n return true\n}", "func (s *Transaction) MatchUpdateCurrentDealHolder(id, newHolderName string) (err error) {\n\tdefer func() {\n\t\tif err != nil {\n\t\t\ts.errors = append(s.errors, err)\n\t\t}\n\n\t\ts.matchNotifications = append(s.matchNotifications, store.MatchNotification{ID: id})\n\t}()\n\n\tm := s.inMemoryCopy.Match(id)\n\tif m == nil {\n\t\treturn store.DontExistError(fmt.Sprintf(`The match %s doesn't exist`, id))\n\t}\n\n\tm.Deals[len(m.Deals)-1].Holder.Name = newHolderName\n\n\treturn\n}", "func (m *PulseManager) Current() (*core.Pulse, error) {\n\tpulseNum := m.db.GetCurrentPulse()\n\tentropy, err := m.db.GetEntropy(pulseNum)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tpulse := core.Pulse{\n\t\tPulseNumber: pulseNum,\n\t\tEntropy: *entropy,\n\t}\n\treturn &pulse, nil\n}", "func (o *HistoricalBalance) SetCurrent(v float64) {\n\to.Current = v\n}", "func (k *Keeper) increaseCurrentRound(ctx sdk.Context) {\n\tcurrentRound := k.CurrentRound(ctx)\n\tcurrentRound = currentRound + 1\n\n\tstore := ctx.KVStore(k.storeKey)\n\n\troundBytes := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(roundBytes, currentRound)\n\tstore.Set([]byte(keyCurrentRound), roundBytes)\n}", "func (b *ballotMaster) getCurrentRound() uint64 {\n\tb.mutex.Lock()\n\tdefer b.mutex.Unlock()\n\n\treturn b.round\n}", "func GetCurrentAHRS() (float64, float64, float64, float64) {\r\n\treturn attitudeX, attitudeY, attitudeZ, heading\r\n}", "func (errors *Errors) Current() uint64 {\n\treturn errors.counter.Current()\n}", "func (gs *GameState) ApplyRoll(playerID xid.ID) error {\n\tif gs.Turn.Player == nil || *gs.Turn.Player != playerID {\n\t\treturn fmt.Errorf(\"You're not the rolling player\")\n\t}\n\troll := gs.NextRoll\n\tif roll == nil {\n\t\treturn fmt.Errorf(\"No roll ready\")\n\t}\n\n\tdefer gs.nextRoller()\n\n\tswitch roll.Action {\n\tcase ActionUp, ActionDown:\n\t\treturn gs.applyStockMove(roll)\n\tcase ActionDividend:\n\t\treturn gs.applyDividend(roll)\n\tdefault:\n\t\treturn fmt.Errorf(\"Unknown PlayerAction %v\", roll.Action)\n\t}\n}", "func (o PodsMetricStatusOutput) Current() MetricValueStatusOutput {\n\treturn o.ApplyT(func(v PodsMetricStatus) MetricValueStatus { return v.Current }).(MetricValueStatusOutput)\n}", "func (m *AgedAccountsPayable) SetCurrentAmount(value *float64)() {\n err := m.GetBackingStore().Set(\"currentAmount\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o PodsMetricStatusPatchOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v PodsMetricStatusPatch) *MetricValueStatusPatch { return v.Current }).(MetricValueStatusPatchPtrOutput)\n}", "func (a *Alert) Current() int {\n\treturn int(atomic.LoadInt32(&a.current))\n}", "func (m *Blueprint) Currents() (float64, float64, float64, error) {\n\treturn 0, 0, 0, api.ErrNotAvailable\n}", "func (d *Die) FullRoll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\n\t// Check if rolled too many times already\n\tif *CtxTotalRolls(ctx) >= CtxMaxRolls(ctx) {\n\t\treturn ErrMaxRolls\n\t}\n\n\tif err := d.Roll(ctx); err != nil {\n\t\treturn err\n\t}\n\n\t// Apply modifiers\n\tfor i := 0; i < len(d.Modifiers); i++ {\n\t\terr := d.Modifiers[i].Apply(ctx, d)\n\t\tswitch {\n\t\t// die rerolled, so restart validation checks with new modifiers\n\t\tcase err == ErrRerolled:\n\t\t\ti = -1\n\t\t\t// i++ => 0 to restart from first modifier\n\t\t\tbreak\n\t\tcase err != nil:\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func updateLastUpdate() error {\n\tnow := strconv.FormatInt(time.Now().Unix(), 10)\n\n\t_, err := dynamo.PutItem(&dynamodb.PutItemInput{\n\t\tTableName: aws.String(table),\n\t\tItem: map[string]*dynamodb.AttributeValue{\n\t\t\t\"LockName\": &dynamodb.AttributeValue{S: aws.String(\"Leader\")},\n\t\t\t\"LeaderName\": &dynamodb.AttributeValue{S: aws.String(name)},\n\t\t\t\"LastUpdate\": &dynamodb.AttributeValue{N: aws.String(now)},\n\t\t},\n\t\tConditionExpression: aws.String(\"LeaderName = :name\"),\n\t\tExpressionAttributeValues: map[string]*dynamodb.AttributeValue{\n\t\t\t\":name\": &dynamodb.AttributeValue{S: aws.String(name)},\n\t\t},\n\t})\n\tif err != nil {\n\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\tlog.Printf(\"Code=%s, Message=%s\", awsErr.Code(), awsErr.Message())\n\t\t}\n\t\t// TODO: If the condition expression fails, we've lost our leadership.\n\t\t// We'll have to convert this error and test for that failure.\n\t\tlog.Printf(\"updateLastUpdate(): %#v\", err)\n\t\tlog.Print(err.Error())\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *KebaUdp) currentPower() (float64, error) {\n\tvar kr keba.Report3\n\terr := c.roundtrip(\"report\", 3, &kr)\n\n\t// mW to W\n\treturn float64(kr.P) / 1e3, err\n}", "func (wb *Alfen) CurrentPower() (float64, error) {\n\tb, err := wb.conn.ReadHoldingRegisters(alfenRegPower, 2)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn rs485.RTUIeee754ToFloat64(b), err\n}", "func (cs *CommissionSchedule) CurrentRate(now epochtime.EpochTime) *quantity.Quantity {\n\tvar latestStartedStep *CommissionRateStep\n\tfor i := range cs.Rates {\n\t\tstep := &cs.Rates[i]\n\t\tif step.Start > now {\n\t\t\tbreak\n\t\t}\n\t\tlatestStartedStep = step\n\t}\n\tif latestStartedStep == nil {\n\t\treturn nil\n\t}\n\treturn &latestStartedStep.Rate\n}", "func (o *ExpenseApplicationsIndexResponseExpenseApplications) SetCurrentRound(v int32) {\n\to.CurrentRound = &v\n}", "func (pb *Bar) Current() int64 {\n\treturn atomic.LoadInt64(&pb.current)\n}", "func (o ResourceMetricStatusPatchOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v ResourceMetricStatusPatch) *MetricValueStatusPatch { return v.Current }).(MetricValueStatusPatchPtrOutput)\n}", "func (srv *Server) RenterCurrentPeriod() (types.BlockHeight, error) {\n\tif srv.node.Renter == nil {\n\t\treturn 0, errors.New(\"can't get renter settings for a non-renter node\")\n\t}\n\treturn srv.node.Renter.CurrentPeriod(), nil\n}", "func (c *KebaUdp) currents() (float64, float64, float64, error) {\n\tvar kr keba.Report3\n\terr := c.roundtrip(\"report\", 3, &kr)\n\n\t// 1mA to A\n\treturn float64(kr.I1) / 1e3, float64(kr.I2) / 1e3, float64(kr.I3) / 1e3, err\n}", "func (m *Maps) SetCurrent(lvl int) {\n\tlog.WithField(\"lvl\", lvl).Info(\"setting current level\")\n\tm.current = lvl\n\tm.active = m.mazes[m.current]\n}", "func (ema *Ema) Current() float64 {\n\treturn ema.points[len(ema.points)-1].Ema\n}", "func (o ElastigroupUpdatePolicyPtrOutput) RollConfig() ElastigroupUpdatePolicyRollConfigPtrOutput {\n\treturn o.ApplyT(func(v *ElastigroupUpdatePolicy) *ElastigroupUpdatePolicyRollConfig {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.RollConfig\n\t}).(ElastigroupUpdatePolicyRollConfigPtrOutput)\n}", "func (r *AutoRoller) retrieveRoll(ctx context.Context, roll *autoroll.AutoRollIssue, rollingTo *revision.Revision) (codereview.RollImpl, error) {\n\treturn r.codereview.RetrieveRoll(ctx, roll, r.recent, rollingTo, r.rollFinished)\n}", "func logUpdate(ch <-chan int, total int, msg string) {\n\tstart := time.Now()\n\ttimer := time.NewTimer(time.Minute)\n\tdefer timer.Stop()\n\tvar current int\n\tvar ok bool\n\tfor {\n\t\tselect {\n\t\tcase current, ok = <-ch:\n\t\t\tif !ok { // channel is closed\n\t\t\t\treturn\n\t\t\t}\n\t\tcase now := <-timer.C:\n\t\t\telapsed := now.Sub(start)\n\t\t\trate := elapsed / time.Duration(current)\n\t\t\teta := time.Duration(total-current) * rate\n\n\t\t\tlogrus.WithFields(logrus.Fields{\n\t\t\t\t\"current\": current,\n\t\t\t\t\"total\": total,\n\t\t\t\t\"percent\": (100 * current) / total,\n\t\t\t\t\"remain\": eta.Round(time.Minute),\n\t\t\t\t\"eta\": now.Add(eta).Round(time.Minute),\n\t\t\t}).Info(msg)\n\t\t\ttimer.Reset(time.Minute)\n\t\t}\n\t}\n}", "func (o ExternalMetricStatusPatchOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v ExternalMetricStatusPatch) *MetricValueStatusPatch { return v.Current }).(MetricValueStatusPatchPtrOutput)\n}", "func (_WyvernExchange *WyvernExchangeSession) CalculateCurrentPrice(addrs [7]common.Address, uints [9]*big.Int, feeMethod uint8, side uint8, saleKind uint8, howToCall uint8, calldata []byte, replacementPattern []byte, staticExtradata []byte) (*big.Int, error) {\n\treturn _WyvernExchange.Contract.CalculateCurrentPrice(&_WyvernExchange.CallOpts, addrs, uints, feeMethod, side, saleKind, howToCall, calldata, replacementPattern, staticExtradata)\n}", "func (s *skill) currentCoolDown() float64 {\n\ts.m.Lock()\n\tdefer s.m.Unlock()\n\treturn s.internalCD\n}", "func (o ResourceMetricStatusOutput) Current() MetricValueStatusOutput {\n\treturn o.ApplyT(func(v ResourceMetricStatus) MetricValueStatus { return v.Current }).(MetricValueStatusOutput)\n}", "func (c *Connection) Currents() (float64, float64, float64, error) {\n\tres, err := c.meterCache.Get()\n\treturn res.FloatValue(\"CURRENT\") / 1e3, 0, 0, err\n}", "func (o PodsMetricStatusPatchPtrOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v *PodsMetricStatusPatch) *MetricValueStatusPatch {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Current\n\t}).(MetricValueStatusPatchPtrOutput)\n}", "func DumpCurrentLedger(c echo.Context) (err error) {\n\tsnapshot := messagelayer.Tangle().LedgerState.SnapshotUTXO()\n\n\taMana, err := snapshotAccessMana()\n\tif err != nil {\n\t\treturn err\n\t}\n\tsnapshot.AccessManaByNode = aMana\n\n\tf, err := os.OpenFile(snapshotFileName, os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\tplugin.LogErrorf(\"unable to create snapshot file %s\", err)\n\t}\n\n\tn, err := snapshot.WriteTo(f)\n\tif err != nil {\n\t\tplugin.LogErrorf(\"unable to write snapshot content to file %s\", err)\n\t}\n\n\t// plugin.LogInfo(snapshot)\n\tplugin.LogInfo(\"Snapshot information: \")\n\tplugin.LogInfo(\" Number of snapshotted transactions: \", len(snapshot.Transactions))\n\tplugin.LogInfo(\" inputs, outputs, txID, unspentOutputs\")\n\tfor key, tx := range snapshot.Transactions {\n\t\tplugin.LogInfo(\" \", len(tx.Essence.Inputs()), len(tx.Essence.Outputs()), key)\n\t\tplugin.LogInfo(\" \", tx.UnspentOutputs)\n\t}\n\tplugin.LogInfo(\" Number of snapshotted accessManaEntries: \", len(snapshot.AccessManaByNode))\n\tplugin.LogInfo(\" nodeID, aMana, timestamp\")\n\tfor nodeID, accessMana := range snapshot.AccessManaByNode {\n\t\tplugin.LogInfo(\" \", nodeID, accessMana.Value, accessMana.Timestamp)\n\t}\n\n\tplugin.LogInfof(\"Bytes written %d\", n)\n\tf.Close()\n\n\treturn c.Attachment(snapshotFileName, snapshotFileName)\n}", "func (wb *Alfen) setCurrent(current float64) error {\n\tb := make([]byte, 4)\n\tbinary.BigEndian.PutUint32(b, math.Float32bits(float32(current)))\n\n\t_, err := wb.conn.WriteMultipleRegisters(alfenRegAmpsConfig, 2, b)\n\n\treturn err\n}", "func (b *Bar) Current() int {\n\tb.mtx.RLock()\n\tdefer b.mtx.RUnlock()\n\treturn b.current\n}", "func (o ContainerResourceMetricStatusPatchOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v ContainerResourceMetricStatusPatch) *MetricValueStatusPatch { return v.Current }).(MetricValueStatusPatchPtrOutput)\n}", "func (view *ViewPodDemo) DemoRollingUpdate() {\n\n\tpod := view.randomPod()\n\tlog.Printf(\"Random pod: %v\", pod)\n\n\tnpart := strings.Split(pod.Name, \"-\")\n\tnewPrefix := fmt.Sprintf(\"%v-%x\", npart[0], rand.Intn(1<<16))\n\n\tseq := 0\n\tfor {\n\t\tnn := fmt.Sprintf(\"%v-%v-%x\", npart[0], npart[1], seq)\n\t\toldPod, ok := view.Pods[nn]\n\t\tif !ok {\n\t\t\tbreak\n\t\t}\n\t\t// create new pod\n\t\tnewPod := view.RecreatePod(newPrefix, seq, oldPod)\n\t\tview.updateStatus(newPod, PodStatusSequenceStart)\n\n\t\t// delete old\n\t\tview.updateStatus(oldPod, PodStatusSequenceStop)\n\n\t\tseq++\n\t}\n\n}", "func (o ExternalMetricStatusOutput) Current() MetricValueStatusOutput {\n\treturn o.ApplyT(func(v ExternalMetricStatus) MetricValueStatus { return v.Current }).(MetricValueStatusOutput)\n}", "func (wb *Alfen) Currents() (float64, float64, float64, error) {\n\treturn wb.voltagesOrCurrents(alfenRegCurrents)\n}", "func UpdateExpiredCredit(c *gin.Context) {\n\tvar (\n\t\tdata interface{}\n\t\terr error\n\t)\n\tdata, err = util.PUpdateExpiredCredit()\n\tlib.GinResponse(c, data, err)\n}", "func (m *MockDeploymentClient) RollingUpdateDeployment(arg0 *v1.Deployment) (*v1.Deployment, bool, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"RollingUpdateDeployment\", arg0)\n\tret0, _ := ret[0].(*v1.Deployment)\n\tret1, _ := ret[1].(bool)\n\tret2, _ := ret[2].(error)\n\treturn ret0, ret1, ret2\n}", "func (o ResourceMetricStatusPatchPtrOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v *ResourceMetricStatusPatch) *MetricValueStatusPatch {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Current\n\t}).(MetricValueStatusPatchPtrOutput)\n}", "func (s DevilDice) UpdatePrevNewArrivals(arrivals []NewArrival) {\n\treturn\n}", "func rollingStateful(clientset *kubernetes.Clientset, namespace, name, image string) error {\n\n\tfmt.Printf(\"rolling upgrade : %s \\n\", name)\n\n\tupdateClient := clientset.AppsV1().StatefulSets(namespace)\n\n\tretryErr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t// Retrieve the latest version of Deployment before attempting update\n\t\t// RetryOnConflict uses exponential backoff to avoid exhausting the apiserver\n\t\tresult, getErr := updateClient.Get(name, metav1.GetOptions{})\n\t\tif getErr != nil {\n\t\t\tpanic(fmt.Errorf(\"Failed to get latest version of Deployment: %v\", getErr))\n\t\t}\n\n\t\tresult.Spec.Template.Spec.Containers[0].Image = image // change nginx version\n\n\t\t_, updateErr := updateClient.Update(result)\n\n\t\treturn updateErr\n\t})\n\n\tif retryErr != nil {\n\t\treturn retryErr\n\t}\n\n\treturn nil\n}", "func (o ObjectMetricStatusPatchOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v ObjectMetricStatusPatch) *MetricValueStatusPatch { return v.Current }).(MetricValueStatusPatchPtrOutput)\n}", "func (f *Frame) getPointValueFromLastRoll() int {\n\tlastRollString := f.rolls[f.currentBall-2]\n\tswitch lastRollString {\n\tcase \"X\":\n\t\treturn 10\n\tcase \"/\":\n\t\trollValue, _ := strconv.Atoi(f.rolls[f.currentBall-3])\n\t\treturn 10 - rollValue\n\tcase \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\":\n\t\trollValue, _ := strconv.Atoi(lastRollString)\n\t\treturn rollValue\n\t}\n\treturn 0\n}", "func (r *AutoRoller) updateStatus(ctx context.Context, replaceLastError bool, lastError string) error {\n\tr.statusMtx.Lock()\n\tdefer r.statusMtx.Unlock()\n\n\trecent := r.recent.GetRecentRolls()\n\n\tif !replaceLastError {\n\t\tlastError = r.status.Get().Error\n\t}\n\n\tfailureThrottledUntil := r.failureThrottle.ThrottledUntil().Unix()\n\tsafetyThrottledUntil := r.safetyThrottle.ThrottledUntil().Unix()\n\tsuccessThrottledUntil := r.successThrottle.ThrottledUntil().Unix()\n\tthrottledUntil := failureThrottledUntil\n\tif safetyThrottledUntil > throttledUntil {\n\t\tthrottledUntil = safetyThrottledUntil\n\t}\n\tif successThrottledUntil > throttledUntil {\n\t\tthrottledUntil = successThrottledUntil\n\t}\n\n\tnotRolledRevs := r.notRolledRevs\n\tnumNotRolled := len(notRolledRevs)\n\tsklog.Infof(\"Updating status (%d revisions behind)\", numNotRolled)\n\tif numNotRolled > maxNotRolledRevs {\n\t\tnotRolledRevs = notRolledRevs[:1]\n\t\tsklog.Warningf(\"Truncating NotRolledRevisions; %d is more than the maximum of %d\", numNotRolled, maxNotRolledRevs)\n\t}\n\tcurrentRollRev := \"\"\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tcurrentRollRev = currentRoll.RollingTo\n\t}\n\tif err := r.status.Set(ctx, r.roller, &status.AutoRollStatus{\n\t\tAutoRollMiniStatus: status.AutoRollMiniStatus{\n\t\t\tCurrentRollRev: currentRollRev,\n\t\t\tLastRollRev: r.lastRollRev.Id,\n\t\t\tMode: r.GetMode(),\n\t\t\tNumFailedRolls: r.recent.NumFailedRolls(),\n\t\t\tNumNotRolledCommits: numNotRolled,\n\t\t\tTimestamp: time.Now().UTC(),\n\t\t\tLastSuccessfulRollTimestamp: r.recent.LastSuccessfulRollTime(),\n\t\t},\n\t\tChildName: r.cfg.ChildDisplayName,\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastError,\n\t\tFullHistoryUrl: r.codereview.GetFullHistoryUrl(),\n\t\tIssueUrlBase: r.codereview.GetIssueUrlBase(),\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tNotRolledRevisions: notRolledRevs,\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t\tThrottledUntil: throttledUntil,\n\t\tValidModes: modes.ValidModes,\n\t\tValidStrategies: r.cfg.ValidStrategies(),\n\t}); err != nil {\n\t\treturn err\n\t}\n\t// Log the current reviewers(s).\n\tsklog.Infof(\"Current reviewers: %v\", r.GetEmails())\n\treturn r.status.Update(ctx)\n}", "func (d *Die) Reroll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\tif d.Result == nil {\n\t\treturn ErrUnrolled\n\t}\n\n\td.Result = nil\n\td.Rerolls++\n\t// reroll without reapplying all modifiers\n\treturn d.Roll(ctx)\n}", "func (o ContainerResourceMetricStatusOutput) Current() MetricValueStatusOutput {\n\treturn o.ApplyT(func(v ContainerResourceMetricStatus) MetricValueStatus { return v.Current }).(MetricValueStatusOutput)\n}", "func (m *Blueprint) CurrentPower() (float64, error) {\n\treturn 0, api.ErrNotAvailable\n}", "func (r *MockRepoManager) NextRollRev() string {\n\tr.mtx.RLock()\n\tdefer r.mtx.RUnlock()\n\treturn r.skiaHead\n}", "func (o *HistoricalBalance) GetCurrent() float64 {\n\tif o == nil {\n\t\tvar ret float64\n\t\treturn ret\n\t}\n\n\treturn o.Current\n}", "func (m *AgedAccountsPayable) GetCurrentAmount()(*float64) {\n val, err := m.GetBackingStore().Get(\"currentAmount\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*float64)\n }\n return nil\n}", "func (r *AutoRoller) GetCurrentRev() string {\n\treturn r.rm.LastRollRev()\n}", "func (bc *Catalog) CurrentEpoch() uint64 {\n\tbc.lock.RLock()\n\tdefer bc.lock.RUnlock()\n\treturn bc.epoch\n}", "func (o ElastigroupUpdatePolicyOutput) RollConfig() ElastigroupUpdatePolicyRollConfigPtrOutput {\n\treturn o.ApplyT(func(v ElastigroupUpdatePolicy) *ElastigroupUpdatePolicyRollConfig { return v.RollConfig }).(ElastigroupUpdatePolicyRollConfigPtrOutput)\n}", "func (o ExternalMetricStatusPatchPtrOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v *ExternalMetricStatusPatch) *MetricValueStatusPatch {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Current\n\t}).(MetricValueStatusPatchPtrOutput)\n}", "func (pb *Bar) SetCurrent(value int64) *Bar {\n\tatomic.StoreInt64(&pb.current, value)\n\tif pb.current >= pb.total && pb.total != -1 {\n\t\tpb.Finish()\n\t}\n\treturn pb\n}", "func (_Bindings *BindingsSession) ExchangeRateCurrent() (*types.Transaction, error) {\n\treturn _Bindings.Contract.ExchangeRateCurrent(&_Bindings.TransactOpts)\n}", "func (m *MockClientInterface) RollingUpdateDeployment(arg0 *v1.Deployment) (*v1.Deployment, bool, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"RollingUpdateDeployment\", arg0)\n\tret0, _ := ret[0].(*v1.Deployment)\n\tret1, _ := ret[1].(bool)\n\tret2, _ := ret[2].(error)\n\treturn ret0, ret1, ret2\n}", "func (db *DB) Roll() {\n\tr := rand.Intn(50)\n\n\tif db.ActiveReq.Requests[r].Shown == 0 {\n\t\treturn\n\t}\n\n\tdb.InactiveReq.Requests = append(db.InactiveReq.Requests, db.ActiveReq.Requests[r])\n\n\tdb.ActiveReq.Requests[r] = &Request{\n\t\tData: strRand.String(2),\n\t\tShown: 0,\n\t}\n}", "func (k *Keeper) CurrentRound(ctx sdk.Context) uint64 {\n\tstore := ctx.KVStore(k.storeKey)\n\tkeyBytes := []byte(keyCurrentRound)\n\tif !store.Has(keyBytes) {\n\t\troundBytes := make([]byte, 8)\n\t\tbinary.LittleEndian.PutUint64(roundBytes, 0)\n\t\tstore.Set(keyBytes, roundBytes)\n\t\treturn 0\n\t} else {\n\t\troundBytes := store.Get(keyBytes)\n\t\tround := binary.LittleEndian.Uint64(roundBytes)\n\t\treturn round\n\t}\n}", "func current(w http.ResponseWriter, r *http.Request) {\n\tu := Current(r.Context(), r)\n\tif u == nil {\n\t\tapi.RequestLogin(w, r)\n\t\treturn\n\t}\n\tapi.Encode(w, r, u, http.StatusOK)\n}", "func (o ContainerResourceMetricStatusPatchPtrOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v *ContainerResourceMetricStatusPatch) *MetricValueStatusPatch {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Current\n\t}).(MetricValueStatusPatchPtrOutput)\n}", "func (r *AutoRoller) Tick() error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence()\n\n\t// Update the status information.\n\tlastErrorStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrorStr = lastErr.Error()\n\t}\n\trecent := r.recent.GetRecentRolls()\n\tnumFailures := 0\n\tfor _, roll := range recent {\n\t\tif roll.Failed() {\n\t\t\tnumFailures++\n\t\t} else if roll.Succeeded() {\n\t\t\tbreak\n\t\t}\n\t}\n\tsklog.Infof(\"Updating status (%d)\", r.rm.CommitsNotRolled())\n\tif err := r.status.Set(&AutoRollStatus{\n\t\tAutoRollMiniStatus: AutoRollMiniStatus{\n\t\t\tNumFailedRolls: numFailures,\n\t\t\tNumNotRolledCommits: r.rm.CommitsNotRolled(),\n\t\t},\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tFullHistoryUrl: r.gerrit.Url(0) + \"/q/owner:\" + r.GetUser(),\n\t\tIssueUrlBase: r.gerrit.Url(0) + \"/c/\",\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t}); err != nil {\n\t\treturn err\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn lastErr\n}", "func Get(ctx context.Context, rollerName string) (*AutoRollStatus, error) {\n\tvar w DsStatusWrapper\n\tif err := ds.DS.Get(ctx, key(rollerName), &w); err != nil {\n\t\treturn nil, err\n\t}\n\trv := new(AutoRollStatus)\n\tif err := gob.NewDecoder(bytes.NewReader(w.Data)).Decode(rv); err != nil {\n\t\treturn nil, err\n\t}\n\treturn rv, nil\n}", "func (d *Download) Current() int {\r\n\treturn d.current\r\n}", "func (o ExternalMetricStatusPatchOutput) CurrentValue() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ExternalMetricStatusPatch) *string { return v.CurrentValue }).(pulumi.StringPtrOutput)\n}", "func (o PodsMetricStatusPtrOutput) Current() MetricValueStatusPtrOutput {\n\treturn o.ApplyT(func(v *PodsMetricStatus) *MetricValueStatus {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Current\n\t}).(MetricValueStatusPtrOutput)\n}", "func CalculateCurrent(voltage, resistance float64) float64 {\n\tif resistance <= 0 {\n\t\tpanic(\"Invalid resistance.\")\n\t}\n\treturn voltage / resistance\n}", "func (_WyvernExchange *WyvernExchangeCallerSession) CalculateCurrentPrice(addrs [7]common.Address, uints [9]*big.Int, feeMethod uint8, side uint8, saleKind uint8, howToCall uint8, calldata []byte, replacementPattern []byte, staticExtradata []byte) (*big.Int, error) {\n\treturn _WyvernExchange.Contract.CalculateCurrentPrice(&_WyvernExchange.CallOpts, addrs, uints, feeMethod, side, saleKind, howToCall, calldata, replacementPattern, staticExtradata)\n}", "func (a *Alert) SetCurrent(i int) {\n\tatomic.StoreInt32(&a.current, int32(i))\n}", "func (s *AutograderService) UpdateEnrollments(ctx context.Context, in *pb.CourseRequest) (*pb.Void, error) {\n\tusr, scm, err := s.getUserAndSCMForCourse(ctx, in.GetCourseID())\n\tif err != nil {\n\t\ts.logger.Errorf(\"UpdateEnrollments failed: authentication error: %w\", err)\n\t\treturn nil, ErrInvalidUserInfo\n\t}\n\tif !s.isTeacher(usr.GetID(), in.GetCourseID()) {\n\t\ts.logger.Error(\"UpdateEnrollments failed: user is not teacher\")\n\t\treturn nil, status.Errorf(codes.PermissionDenied, \"only teachers can update enrollment status\")\n\t}\n\terr = s.updateEnrollments(ctx, scm, in.GetCourseID())\n\tif err != nil {\n\t\ts.logger.Errorf(\"UpdateEnrollments failed: %w\", err)\n\t\tif contextCanceled(ctx) {\n\t\t\treturn nil, status.Error(codes.FailedPrecondition, ErrContextCanceled)\n\t\t}\n\t\tif ok, parsedErr := parseSCMError(err); ok {\n\t\t\treturn nil, parsedErr\n\t\t}\n\t\terr = status.Error(codes.InvalidArgument, \"failed to update pending enrollments\")\n\t}\n\treturn &pb.Void{}, err\n}", "func (s *AFDOStrategy) GetNextRollRev(ctx context.Context, _ []*vcsinfo.LongCommit) (string, error) {\n\t// Find the available AFDO versions, sorted newest to oldest, and store.\n\tavailable := []string{}\n\tif err := s.gcs.AllFilesInDirectory(ctx, AFDO_GS_PATH, func(item *storage.ObjectAttrs) {\n\t\tname := strings.TrimPrefix(item.Name, AFDO_GS_PATH)\n\t\tif _, err := parseAFDOVersion(name); err == nil {\n\t\t\tavailable = append(available, name)\n\t\t} else if err == errInvalidAFDOVersion {\n\t\t\t// There are files we don't care about in this bucket. Just ignore.\n\t\t} else {\n\t\t\tsklog.Error(err)\n\t\t}\n\t}); err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(available) == 0 {\n\t\treturn \"\", fmt.Errorf(\"No valid AFDO profile names found.\")\n\t}\n\tsort.Sort(afdoVersionSlice(available))\n\n\t// Store the available versions. Return the newest.\n\ts.mtx.Lock()\n\tdefer s.mtx.Unlock()\n\ts.versions = available\n\treturn s.versions[0], nil\n}", "func (s *Module) Current() string {\n\tvar rev string\n\tfor _, r := range s.Revision {\n\t\tif r.Name > rev {\n\t\t\trev = r.Name\n\t\t}\n\t}\n\treturn rev\n}", "func (s *Service) Current(w http.ResponseWriter, r *http.Request,\n\targs *Args, reply *Args) (err error) {\n\n\tc := context.NewContext(r)\n\tvar isSet bool\n\tuserID, _ := user.CurrentUserID(r)\n\t_, err = profile.Get(c, profile.GenAuthID(\"Password\", userID))\n\tif err == nil {\n\t\tisSet = true\n\t}\n\treply.Password = &Password{IsSet: isSet}\n\treturn nil\n}", "func (h *HeadResolver) Roll(ctx context.Context) (float64, error) {\n\treturn float64(h.Head.Roll), nil\n}", "func (r *AutoRoller) doAutoRollInner() (string, error) {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\t// Get updated info about the current roll.\n\tif err := r.updateCurrentRoll(); err != nil {\n\t\treturn STATUS_ERROR, err\n\t}\n\n\t// There's a currently-active roll. Determine whether or not it's still good.\n\t// If so, leave it open and exit. If not, close it so that we can open another.\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tsklog.Infof(\"Found current roll: %s\", r.issueUrl(currentRoll.Issue))\n\n\t\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\t\t// If we have a normal (non-dry-run) roll running,\n\t\t\t// switch it to a dry run.\n\t\t\tif currentRoll.CommitQueue {\n\t\t\t\tsklog.Infof(\"Setting dry-run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, true); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\n\t\t\t// If the CQ has finished, determine if it was a success\n\t\t\t// or failure.\n\t\t\trollDone, err := r.isDryRunDone(currentRoll)\n\t\t\tif err != nil {\n\t\t\t\treturn STATUS_ERROR, err\n\t\t\t}\n\t\t\tif rollDone {\n\t\t\t\tresult := autoroll.ROLL_RESULT_DRY_RUN_FAILURE\n\t\t\t\tstatus := STATUS_DRY_RUN_FAILURE\n\t\t\t\trollSuccessful, err := r.isDryRunSuccessful(currentRoll)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rollSuccessful {\n\t\t\t\t\tresult = autoroll.ROLL_RESULT_DRY_RUN_SUCCESS\n\t\t\t\t\tstatus = STATUS_DRY_RUN_SUCCESS\n\t\t\t\t}\n\t\t\t\tsklog.Infof(\"Dry run is finished: %v\", currentRoll)\n\t\t\t\tif currentRoll.RollingTo != r.rm.ChildHead() {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, result, fmt.Sprintf(\"Repo has passed %s; will open a new dry run.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else if currentRoll.Result != result {\n\t\t\t\t\t// The dry run just finished. Set its result.\n\t\t\t\t\tif result == autoroll.ROLL_RESULT_DRY_RUN_FAILURE {\n\t\t\t\t\t\tif err := r.closeIssue(currentRoll, result, \"Dry run failed. Closing, will open another.\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif err := r.addIssueComment(currentRoll, \"Dry run finished successfully; leaving open in case we want to land\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcurrentRoll.Result = result\n\t\t\t\t\t\tif err := r.recent.Update(currentRoll); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\treturn status, nil\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// The dry run is finished but still good. Leave it open.\n\t\t\t\t\tsklog.Infof(\"Dry run is finished and still good.\")\n\t\t\t\t\treturn status, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tsklog.Infof(\"Dry run still in progress.\")\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\t\t} else {\n\t\t\tif currentRoll.CommitQueueDryRun {\n\t\t\t\tsklog.Infof(\"Unsetting dry run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\t\t\t// If we're stopped, close the issue.\n\t\t\t\t// Respect the previous result of the roll.\n\t\t\t\tif err := r.closeIssue(currentRoll, r.makeRollResult(currentRoll), \"AutoRoller is stopped; closing the active roll.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else if !currentRoll.CommitQueue {\n\t\t\t\t// If the CQ failed, close the issue.\n\t\t\t\t// Special case: if the current roll was a dry run which succeeded, land it.\n\t\t\t\tif currentRoll.Result == autoroll.ROLL_RESULT_DRY_RUN_SUCCESS {\n\t\t\t\t\tsklog.Infof(\"Dry run succeeded. Attempting to land.\")\n\t\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t} else {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Commit queue failed; closing this roll.\"); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if time.Since(currentRoll.Modified) > 24*time.Hour {\n\t\t\t\t// If the roll has been open too long, close the issue.\n\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Roll has been open for over 24 hours; closing.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rolledPast {\n\t\t\t\t\t// If we've already rolled past the target revision, close the issue\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, fmt.Sprintf(\"Already rolled past %s; closing this roll.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// Current roll is still good.\n\t\t\t\t\tsklog.Infof(\"Roll is still active (%d): %s\", currentRoll.Issue, currentRoll.Subject)\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// If we're stopped, exit.\n\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\tsklog.Infof(\"Roller is stopped; not opening new rolls.\")\n\t\treturn STATUS_STOPPED, nil\n\t}\n\n\t// If we're up-to-date, exit.\n\tchildHead := r.rm.ChildHead()\n\tif r.rm.LastRollRev() == childHead {\n\t\tsklog.Infof(\"Repo is up-to-date.\")\n\t\treturn STATUS_UP_TO_DATE, nil\n\t}\n\n\t// Create a new roll.\n\tif r.attemptCounter.Get() >= ROLL_ATTEMPT_THROTTLE_NUM {\n\t\treturn STATUS_THROTTLED, nil\n\t}\n\tr.attemptCounter.Inc()\n\tdryRun := r.isMode(autoroll_modes.MODE_DRY_RUN)\n\tuploadedNum, err := r.rm.CreateNewRoll(r.strategy, r.GetEmails(), r.cqExtraTrybots, dryRun)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to upload a new roll: %s\", err)\n\t}\n\tuploaded, err := r.retrieveRoll(uploadedNum)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to retrieve uploaded roll: %s\", err)\n\t}\n\tif err := r.recent.Add(uploaded); err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to insert uploaded roll into database: %s\", err)\n\t}\n\n\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t}\n\treturn STATUS_IN_PROGRESS, nil\n}", "func (c *AutoRollStatusCache) Update(ctx context.Context) error {\n\tstatus, err := Get(ctx, c.roller)\n\tif err == datastore.ErrNoSuchEntity || status == nil {\n\t\t// This will occur the first time the roller starts,\n\t\t// before it sets the status for the first time. Ignore.\n\t\tsklog.Warningf(\"Unable to find AutoRollStatus for %s. Is this the first startup for this roller?\", c.roller)\n\t\tstatus = &AutoRollStatus{}\n\t} else if err != nil {\n\t\treturn err\n\t}\n\tc.mtx.Lock()\n\tdefer c.mtx.Unlock()\n\tc.status = status\n\treturn nil\n}", "func (o ObjectMetricStatusPatchPtrOutput) Current() MetricValueStatusPatchPtrOutput {\n\treturn o.ApplyT(func(v *ObjectMetricStatusPatch) *MetricValueStatusPatch {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Current\n\t}).(MetricValueStatusPatchPtrOutput)\n}", "func writeAutoRollDataPoint(dbClient *influxdb.Client, status *autoroll.AutoRollStatus) error {\n\tissue := -1\n\tif status.Status != autoroll.STATUS_IDLE {\n\t\tissue = status.CurrentRoll.Issue\n\t}\n\tpoint := []interface{}{\n\t\tinterface{}(status.LastRollRevision),\n\t\tinterface{}(status.CurrentRollRevision),\n\t\tinterface{}(issue),\n\t\tinterface{}(status.Head),\n\t\tinterface{}(status.Status),\n\t}\n\tseries := influxdb.Series{\n\t\tName: SERIES_AUTOROLL_CURRENTSTATUS,\n\t\tColumns: COLUMNS_AUTOROLL_CURRENTSTATUS,\n\t\tPoints: [][]interface{}{point},\n\t}\n\tseriesList := []*influxdb.Series{&series}\n\tglog.Infof(\"Pushing datapoint to %s: %v\", SERIES_AUTOROLL_CURRENTSTATUS, point)\n\treturn dbClient.WriteSeries(seriesList)\n}", "func (s *CheckpointManager) UpdateCurrent(taskID int, added int) error {\n\ts.mu.Lock()\n\tcp := s.checkpoints[taskID]\n\tcp.currentKeys += added\n\ts.mu.Unlock()\n\n\tflushed, imported, err := s.flushCtrl.Flush(s.indexID, FlushModeAuto)\n\tif !flushed || err != nil {\n\t\treturn err\n\t}\n\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\ts.progressLocalSyncMinKey()\n\tif imported && s.minKeySyncGlobal.Cmp(s.minKeySyncLocal) != 0 {\n\t\ts.minKeySyncGlobal = s.minKeySyncLocal\n\t\ts.globalCnt = s.localCnt\n\t\ts.dirty = true\n\n\t\ts.pidGlobal = s.pidLocal\n\t\ts.startGlobal = s.startLocal\n\t\ts.endGlobal = s.endLocal\n\t}\n\treturn nil\n}" ]
[ "0.50856096", "0.4899381", "0.48472196", "0.4755209", "0.4720856", "0.47145355", "0.47129878", "0.46703026", "0.46206605", "0.45635659", "0.45304877", "0.4475469", "0.44564143", "0.44422534", "0.4418267", "0.44162712", "0.44079155", "0.43984628", "0.4377662", "0.43716916", "0.4353333", "0.43436092", "0.43368152", "0.43323064", "0.432952", "0.4322028", "0.43187085", "0.43119928", "0.4306541", "0.43035153", "0.43032056", "0.42988175", "0.42907432", "0.42792508", "0.42780682", "0.42764887", "0.42534497", "0.4253019", "0.42489648", "0.42447984", "0.42043173", "0.42034373", "0.4200562", "0.41986117", "0.41931543", "0.4189546", "0.4187577", "0.41864845", "0.41864362", "0.41840005", "0.41789484", "0.4166808", "0.41649365", "0.4163005", "0.41621482", "0.41423538", "0.41313598", "0.41296354", "0.41292474", "0.41260004", "0.4117727", "0.41101056", "0.4109845", "0.41086206", "0.4104834", "0.41027647", "0.41000488", "0.4099563", "0.40992245", "0.40929395", "0.4092254", "0.40893203", "0.40881598", "0.40872136", "0.40871882", "0.40867773", "0.40866926", "0.4086437", "0.408275", "0.4076802", "0.40653905", "0.40608543", "0.4059604", "0.4058001", "0.4056244", "0.40516356", "0.405012", "0.4046186", "0.4044408", "0.40381372", "0.40344903", "0.4024217", "0.40221268", "0.40184182", "0.40039846", "0.40033993", "0.40006286", "0.3993669", "0.39913976", "0.39899114" ]
0.6866489
0
retrieveRoll obtains the given DEPS roll from the code review server.
func (r *AutoRoller) retrieveRoll(issueNum int64) (*autoroll.AutoRollIssue, error) { var a *autoroll.AutoRollIssue info, err := r.gerrit.GetIssueProperties(issueNum) if err != nil { return nil, fmt.Errorf("Failed to get issue properties: %s", err) } a, err = autoroll.FromGerritChangeInfo(info, r.rm.FullChildHash, r.rollIntoAndroid) if err != nil { return nil, fmt.Errorf("Failed to convert issue format: %s", err) } tryResults, err := autoroll.GetTryResultsFromGerrit(r.gerrit, a) if err != nil { return nil, fmt.Errorf("Failed to retrieve try results: %s", err) } a.TryResults = tryResults return a, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) retrieveRoll(ctx context.Context, roll *autoroll.AutoRollIssue, rollingTo *revision.Revision) (codereview.RollImpl, error) {\n\treturn r.codereview.RetrieveRoll(ctx, roll, r.recent, rollingTo, r.rollFinished)\n}", "func roll(r string) string {\n\tres, _, err := dice.Roll(r)\n\tif err != nil {\n\t\treturn DICE_USAGE\n\t}\n\treturn fmt.Sprintf(\"%v\", res.Int())\n}", "func RetrieveReplay(c *gin.Context) {\n\tid, err := strconv.ParseUint(c.Param(\"id\"), 10, 64)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, common.NewError(\"id\", errors.New(\"id must be valid number\")))\n\t\treturn\n\t}\n\n\treplay, err := GetReplay(uint(id))\n\tif err != nil {\n\t\tc.JSON(http.StatusNotFound, common.NewError(\"replay\", errors.New(\"No replay found for id\")))\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, replay)\n}", "func DRollN(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tsides := vars[\"sides\"]\n\tcount := vars[\"count\"]\n\n\ts := getSides(sides[1:])\n\tif s == 0 {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tc := getCount(count)\n\tif c == 0 {\n\t\tcountErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse(w, s, c)\n}", "func Roll(w http.ResponseWriter, r *http.Request) {\n\tsides := r.FormValue(\"sides\")\n\tcount := r.FormValue(\"count\")\n\n\ts := getSides(sides)\n\tif s == 0 {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tc := getCount(count)\n\tif c == 0 {\n\t\tcountErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse(w, s, c)\n}", "func NewRoll(s *S) *Roll {\n\tu := &Roll{\n\t\tS: s,\n\t\tC: NewCCap(s.Len() * 10),\n\t\tdmap: make([][]z.Lit, s.Len())}\n\treturn u\n}", "func doGetEnroll(id string){\n\tvar Url string\n\tif id==\"\" {\n\t\tUrl=BaseUrl+\"/studentlist/_design/getlistdata/_view/studentenrolled\"\n\t}\n\tif id!=\"\" {\n\tUrl=BaseUrl+\"/studentlist/_design/getlistdata/_view/studentenrolled?key=+\"+string(id)\n\t}\n\tresponse,_ := http.Get(Url)\n\tdefer response.Body.Close()\n\tdecoder := json.NewDecoder(response.Body)\n\terr := decoder.Decode(&studentEnrolledAPI)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func Get(ctx context.Context, rollerName string) (*AutoRollStatus, error) {\n\tvar w DsStatusWrapper\n\tif err := ds.DS.Get(ctx, key(rollerName), &w); err != nil {\n\t\treturn nil, err\n\t}\n\trv := new(AutoRollStatus)\n\tif err := gob.NewDecoder(bytes.NewReader(w.Data)).Decode(rv); err != nil {\n\t\treturn nil, err\n\t}\n\treturn rv, nil\n}", "func (d Dice) Roll() int {\n\td.m.Lock()\n\tdefer d.m.Unlock()\n\treturn util.RandRange(d.r, d.min, d.max)\n}", "func Get(repo repository.Repo, revision string) (*Review, error) {\n\tsummary, err := GetSummary(repo, revision)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif summary == nil {\n\t\treturn nil, nil\n\t}\n\treturn summary.Details()\n}", "func (r *MockRepoManager) NextRollRev() string {\n\tr.mtx.RLock()\n\tdefer r.mtx.RUnlock()\n\treturn r.skiaHead\n}", "func TestRollN(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Valid roll\", \"/d4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Valid roll\", \"/D4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Invalid variable\", \"/d5\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid variable\", \"/D5\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Valid query for count\", \"/d4?count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t\t{\"Valid query for count\", \"/D4?count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t\t{\"Invalid query for count\", \"/d4?count=0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid query for count\", \"/D4?count=0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t}\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func TestRoll(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Default roll\", \"\", response{http.StatusOK, ``}},\n\t\t{\"Valid query for sides\", \"?sides=4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Invalid query for sides\", \"?sides=5\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for count\", \"?count=2\", response{http.StatusOK, `\"count\":2`}},\n\t\t{\"Invalid query for count\", \"?count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for sides, invalid for count\", \"?sides=4&count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for count, invalid for sides\", \"?count=2&sides=1\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for sides and count\", \"?sides=4&count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t}\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func (server *PayrollServer) FindEmployeePayroll(ctx context.Context, req *hrm.FindEmployeePayrollRequest) (*hrm.FindEmployeePayrollResponse, error) {\n\n\tfilter := req.GetFilter()\n\tif filter == nil {\n\t\tserver.log.Error(\"filter is nil\")\n\t\treturn nil, status.Errorf(codes.InvalidArgument, \"filter is required\")\n\t}\n\n\t// handele context error\n\tif err := utils.ContextError(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar payroll db.Payroll\n\tvar err error\n\t// find record base on type of key\n\tswitch filter.GetKey().(type) {\n\tcase *hrm.PayrollFilter_EmployeeId:\n\t\t// check for valid uuid\n\t\tvar id uuid.UUID\n\t\tid, err = uuid.Parse(filter.GetEmployeeId())\n\t\tif err != nil {\n\t\t\tserver.log.Info(\"invalid uuid\", \"error\", err)\n\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"id is not a valid uuid: %v\", err)\n\t\t}\n\t\t// call helper func\n\t\tpayroll, err = findByEmployeeId(id, ctx, server)\n\tcase *hrm.PayrollFilter_EmployeeName:\n\t\tif len(req.Filter.GetEmployeeName()) == 0 {\n\t\t\tserver.log.Info(\"invalid employee name, empty\")\n\t\t\treturn nil, status.Errorf(codes.InvalidArgument, \"employee is required\")\n\t\t}\n\t\t// call helper func\n\t\tpayroll, err = findByEmployeeName(filter.GetEmployeeName(), ctx, server)\n\t}\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp := &hrm.Payroll{\n\t\tId: payroll.ID.String(),\n\t\tEmployeeId: payroll.Employee.ID.String(),\n\t\tCtc: payroll.Ctc,\n\t\tAllowance: payroll.Allowance,\n\t\tCreateBy: payroll.CreateBy.String(),\n\t\tCreatedAt: timestamppb.New(payroll.CreatedAt),\n\t\tUpdatedBy: payroll.UpdatedBy.String(),\n\t\tUpdatedAt: timestamppb.New(payroll.UpdatedAt),\n\t}\n\tres := &hrm.FindEmployeePayrollResponse{\n\t\tPayroll: p,\n\t}\n\n\treturn res, nil\n}", "func (d *Dice) Roll() int {\n\t// We named this call \"roll\" so we can easily mock the specific call out in our tests\n\treturn d.rand.Intn(\"roll\", d.sides)\n}", "func (d *Die) Reroll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\tif d.Result == nil {\n\t\treturn ErrUnrolled\n\t}\n\n\td.Result = nil\n\td.Rerolls++\n\t// reroll without reapplying all modifiers\n\treturn d.Roll(ctx)\n}", "func (d Die) Roll() int {\n\treturn rand.Intn(int(d)) + 1\n}", "func (c *Controller) getPract(id string, pract *store.Practitioner) error {\n\tp, err := c.Sc.GetPractitioner(id)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tpract.SCFHSRegistrationNumber = &p.Response.Info.Profile.RegistrationNumber\n\tpract.PractitionerID = pract.HealthID\n\tpract.IDNumber = &id\n\n\tpract.FirstNameAr = &p.Response.Info.Profile.Ar.FirstName\n\tpract.SecondNameAr = &p.Response.Info.Profile.Ar.SecondName\n\tpract.LastNameAr = &p.Response.Info.Profile.Ar.LastName\n\n\tpract.FirstNameEn = &p.Response.Info.Profile.En.FirstName\n\tpract.SecondNameEn = &p.Response.Info.Profile.En.SecondName\n\tpract.LastNameEn = &p.Response.Info.Profile.En.LastName\n\n\tpract.Gender_code = &p.Response.Info.Profile.Gender.Code\n\tpract.Gender_ar = &p.Response.Info.Profile.Gender.NameAr\n\tpract.Gender_en = &p.Response.Info.Profile.Gender.NameEn\n\n\tpract.SCFHSCategoryCode = &p.Response.Info.Professionality.Category.Code\n\tpract.SCFHSCategoryAr = &p.Response.Info.Professionality.Category.NameAr\n\tpract.SCFHSCategoryEn = &p.Response.Info.Professionality.Category.NameEn\n\n\tpract.SCFHSSpecialityCode = &p.Response.Info.Professionality.Specialty.Code\n\tpract.SCFHSSpecialityAr = &p.Response.Info.Professionality.Specialty.NameAr\n\tpract.SCFHSSpecialityEn = &p.Response.Info.Professionality.Specialty.NameEn\n\n\tpract.SCFHSPractitionerStatus = &p.Response.Info.Status.Code\n\tpract.SCFHSPractitionerStatusCode = &p.Response.Info.Status.DescAr\n\n\tpract.SCFHSRegistrationIssueDate = &p.Response.Info.Status.License.IssuedDate\n\tpract.SCFHSRegistrationExpiryDate = &p.Response.Info.Status.License.ExpiryDate\n\n\t// @TODO: change this\n\tpq := PatientQuery{ID: id}\n\tswitch pq.Kind() {\n\tcase KindCitizen:\n\t\tty := \"NationalId\"\n\t\tpract.IDType = &ty\n\tcase KindExpat:\n\t\tty := \"Iqama\"\n\t\tpract.IDType = &ty\n\t}\n\treturn nil\n}", "func GetPayrollEndpoint(ps pyr.PayrollService, userType string) http.HandlerFunc {\n\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tpayrollID := chi.URLParam(r, \"payrollID\")\n\t\tvar employeeID string\n\t\t// if userType == admin then get the employeeId from the request parameter\n\t\tif userType == \"admin\" {\n\t\t\temployeeID = chi.URLParam(r, \"employeeID\")\n\t\t} else {\n\t\t\t// Get Employee Token Data\n\t\t\ttokenData := r.Context().Value(\"tokenData\").(*md.EmployeeTokenData)\n\t\t\temployeeID = string(tokenData.EmployeeID)\n\t\t}\n\n\t\tvar payroll *md.Payroll\n\t\tpayroll = ps.GetPayroll(employeeID, payrollID)\n\t\tresp := ut.Message(true, \"\")\n\t\tresp[\"payroll\"] = payroll\n\t\tut.Respond(w, r, resp)\n\t}\n}", "func SimpleRoll(roll string) int {\n\tvar result, dieType int\n\tvar results []int\n\n\tif strings.Contains(roll, \"d\") {\n\t\tcomponents := strings.Split(roll, \"d\")\n\t\tnumDice, _ := strconv.Atoi(components[0])\n\t\tkeepHigh := 0\n\t\tkeepLow := 0\n\n\t\tif strings.Contains(components[1], \"h\") {\n\t\t\tcomps := strings.Split(components[1], \"h\")\n\t\t\tdieType, _ = strconv.Atoi(comps[0])\n\t\t\tkeepHigh, _ = strconv.Atoi(comps[1])\n\t\t} else if strings.Contains(components[1], \"l\") {\n\t\t\tcomps := strings.Split(components[1], \"l\")\n\t\t\tdieType, _ = strconv.Atoi(comps[0])\n\t\t\tkeepLow, _ = strconv.Atoi(comps[1])\n\t\t} else {\n\t\t\tdieType, _ = strconv.Atoi(components[1])\n\t\t}\n\n\t\tfor i := 0; i < numDice; i++ {\n\t\t\tresults = append(results, MakeRoll(dieType))\n\t\t\t//results = append(results, DummyRoll(dieType))\n\t\t}\n\t\tsort.Ints(results)\n\t\tif keepLow > 0 {\n\t\t\tfor i := 0; i < keepLow; i++ {\n\t\t\t\tresult += results[i]\n\t\t\t}\n\t\t} else if keepHigh > 0 {\n\t\t\tfor i := 0; i < keepHigh; i++ {\n\t\t\t\tresult += results[numDice-1-i]\n\t\t\t}\n\n\t\t} else {\n\t\t\tfor _, res := range results {\n\t\t\t\tresult += res\n\t\t\t}\n\t\t}\n\n\t} else {\n\t\tresult, _ = strconv.Atoi(roll)\n\t}\n\n\treturn result\n}", "func RollDie() (int, error) {\n\tr, err := rand.Int(rand.Reader, sides)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(r.Int64()) + 1, nil\n}", "func (c *client) Get(ctx context.Context, inputRa *security.RoleAssignment) (*[]security.RoleAssignment, error) {\n\trequest, err := c.getRoleAssignmentRequest(wssdcloudcommon.Operation_GET, inputRa)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresponse, err := c.RoleAssignmentAgentClient.Invoke(ctx, request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tras, err := c.getRoleAssignmentFromResponse(response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ras, nil\n}", "func (gs *GameState) Roll(player xid.ID) (*Roll, error) {\n\tstonk := gs.Stonks[rand.Intn(len(gs.Stonks))]\n\taction := PlayerAction(rand.Intn(3))\n\tmovement := allowedMovements[rand.Intn(len(allowedMovements))]\n\n\treturn &Roll{\n\t\tID: xid.New(),\n\t\tPlayer: player,\n\t\tStonk: stonk.ID,\n\t\tAction: action,\n\t\tValue: movement,\n\t}, nil\n}", "func GetPlane(name string, ds *mgo.Session, dbs string) (p model.Plane, err error) {\n\tsession := ds\n\tclone := session.Clone()\n\n\tdb := clone.DB(dbs)\n\tdal := NewMongoDBDAL(db)\n\tplanes := dal.C(\"planes\")\n\terr = planes.Find(bson.M{\"name\": name}).One(&p)\n\tif err != nil {\n\t\tlog.Println(\"Could not get a Record with that name!\")\n\t}\n\t// defer clone.Close()\n\treturn\n}", "func rollDie() int {\n\t//TODO test & fix range if needed (don't include 0)\n\t// hardcode for 6-sided die\n\treturn getRandNum(7)\n}", "func Roll(number int, sided int) Result {\n\treturn roll(number, sided)\n}", "func Roll(n, d, b int, s *discordgo.Session, m *discordgo.MessageCreate) {\n\tresult := \"Rolled: [\"\n\ttotal := b\n\tfor i := 0; i < n-1; i++ {\n\t\tval := rand.Intn(d) + 1\n\t\tresult += strconv.Itoa(val) + \", \"\n\t\ttotal += val\n\t}\n\tval := rand.Intn(d) + 1\n\tresult += strconv.Itoa(val)\n\ttotal += val\n\tif b > 0 {\n\t\tresult += \"] +\" + strconv.Itoa(b)\n\t} else if b == 0 {\n\t\tresult += \"]\"\n\t} else {\n\t\tresult += \"] \" + strconv.Itoa(b)\n\t}\n\n\tresult += \" = \" + strconv.Itoa(total)\n\ts.ChannelMessageSend(m.ChannelID, result)\n}", "func DummyRoll(dieType int) int {\n\treturn dieType / 2\n}", "func RollN(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tsides := vars[\"sides\"]\n\tcount := r.FormValue(\"count\")\n\n\ts := getSides(sides[1:])\n\tif s == 0 {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tc := getCount(count)\n\tif c == 0 {\n\t\tcountErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse(w, s, c)\n}", "func (db *DB) Roll() {\n\tr := rand.Intn(50)\n\n\tif db.ActiveReq.Requests[r].Shown == 0 {\n\t\treturn\n\t}\n\n\tdb.InactiveReq.Requests = append(db.InactiveReq.Requests, db.ActiveReq.Requests[r])\n\n\tdb.ActiveReq.Requests[r] = &Request{\n\t\tData: strRand.String(2),\n\t\tShown: 0,\n\t}\n}", "func RollDice(dice string) (string, error) {\n\t//This block of code turns a command line argument into two ints, one being the number of dice thrown, and the other being the type of dice\n\tclearFlags()\n\trollType, diceSlice, err := inputProofer(dice)\n\t//fmt.Println(diceSlice)\n\terr = assignMeaningToDiceSlice(diceSlice, rollType)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif numDice > 1000 {\n\t\treturn \"That's too many dice\", nil\n\t}\n\tif typeDice > 1000 {\n\t\treturn \"Golf balls are not dice\", nil\n\t}\n\tif numDice+typeDice > 1000 {\n\t\treturn \"Fuck you\", nil\n\t}\n\t//Create a source of random numbers seeded to the current time\n\n\tvar results []int\n\tresults, total, successes, err = rollDice(rollType, roundFlag)\n\tif err != nil {\n\t\treturn \"DAMNIT\", err\n\t}\n\treturn formatResults(results, rollType), nil\n\t//fmt.Println(results)\n\t//fmt.Println(total)\n\t//fmt.Println(successes)\n}", "func rollDice(sides, count int) (result int, err error) {\n\td := chooseDice(sides)\n\tif d == nil {\n\t\treturn 0, invalidDice{}\n\t}\n\n\tfor i := 0; i < count; i++ {\n\t\tresult += d()\n\t}\n\n\treturn result, nil\n}", "func (h *HeadResolver) Roll(ctx context.Context) (float64, error) {\n\treturn float64(h.Head.Roll), nil\n}", "func (*OSRS) Retrieve(r io.Reader) ([]SkillData, error) {\n\n\tdoc, err := html.Parse(r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttable, err := getTable(doc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tscores, err := getScoresOSRSList(table)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn scores, nil\n}", "func (s *Sailor) getAttempt(msg messages.Message, st *storage.State) error {\n\tgt := storage.GenerateTransaction(storage.GetOp, msg.Key, \"\")\n\titem, err := st.ApplyTransaction(gt)\n\n\treply := makeReply(s, &msg, \"getResponse\")\n\tif err != nil {\n\t\tif s.state == candidate {\n\t\t\treply.Error = \"|Src: \" + s.client.NodeName + \" | \" + err.Error() + \" : In Election Cycle\"\n\t\t} else {\n\t\t\treply.Error = \"|Src: \" + s.client.NodeName + \" | \" + err.Error() + \" : Current Leader is \" + s.leaderId\n\t\t}\n\t} else {\n\t\treply.Error = \"|Src: \" + s.client.NodeName + \" | \" + msg.Key + \"= \" + item + \" : Current Leader is \" + s.leaderId\n\t}\n\treturn s.client.SendToBroker(reply)\n}", "func ROLL(ci, mr operand.Op) { ctx.ROLL(ci, mr) }", "func (db *DB) GetReplicatedPulse(ctx context.Context, jet core.RecordID) (core.PulseNumber, error) {\n\tk := prefixkey(scopeIDSystem, jet[:], []byte{sysReplicatedPulse})\n\tbuf, err := db.get(ctx, k)\n\tif err != nil {\n\t\tif err == ErrNotFound {\n\t\t\terr = nil\n\t\t}\n\t\treturn 0, err\n\t}\n\treturn core.NewPulseNumber(buf), nil\n}", "func (r *MockRepoManager) LastRollRev() string {\n\tr.mtx.RLock()\n\tdefer r.mtx.RUnlock()\n\treturn r.lastRollRev\n}", "func rollDie() Die {\n\treturn Die{Pips: rand.Intn(6) + 1, Render: true}\n}", "func (a *V1EmployeesApiService) RetrieveCashDrawerShift(ctx context.Context, locationId string, shiftId string) (V1CashDrawerShift, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Get\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue V1CashDrawerShift\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/v1/{location_id}/cash-drawer-shifts/{shift_id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"location_id\"+\"}\", fmt.Sprintf(\"%v\", locationId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"shift_id\"+\"}\", fmt.Sprintf(\"%v\", shiftId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode < 300 {\n\t\t// If we succeed, return the data, otherwise pass on to decode error.\n\t\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"));\n\t\tif err == nil { \n\t\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t\t}\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v V1CashDrawerShift\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"));\n\t\t\t\tif err != nil {\n\t\t\t\t\tnewErr.error = err.Error()\n\t\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t\t}\n\t\t\t\tnewErr.model = v\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (c *Client) Rollback(p RollbackParameters) (*Release, error) {\n\tresponse, err := c.client.RollbackRelease(\n\t\tp.Release,\n\t\thelm.RollbackVersion(int32(p.Revision)))\n\tif err != nil {\n\t\treturn nil, trace.Wrap(err)\n\t}\n\treturn fromHelm(response.Release), nil\n}", "func roll(sides int) int {\n\tmin, max := 1, sides\n\troll := min + seededRand.Intn(max-min+1)\n\treturn roll\n}", "func (contract *ContractChaincode) GetSalary(stub shim.ChaincodeStubInterface, params []string) peer.Response {\n\t// This is returned\n\tresultString := \"{}\"\n\tEmpId := params[0]\n\n\t// Read the open data\n\tdataOpen, err1 := stub.GetPrivateData(\"HumanHitachiContract\", EmpId)\n\tif err1 != nil {\n\t\treturn shim.Error(\"Error1=\" + err1.Error())\n\t}\n\n\t// Read the acme private data\n\tdataSecret, err2 := stub.GetPrivateData(\"HumanPrivate\", EmpId)\n\n\taccessError := \"N.A.\"\n\tif err2 != nil {\n\t\t//return shim.Error(\"Error=\"+err1.Error())\n\t\tfmt.Println(\"Error2=\" + err2.Error())\n\t\taccessError = err2.Error()\n\t\tdataSecret = []byte(\"**** Not Allowed ***\")\n\t}\n\n\t// Returns the token value from 2 PDC + error\n\tresultString = \"{Human Hitachi Salary:\\\"\" + string(dataOpen) + \"\\\", HumanSecretSalary:\\\"\" + string(dataSecret) + \"\\\" , error:\\\"\" + accessError + \"\\\"}\"\n\n\treturn shim.Success([]byte(resultString))\n\n}", "func (c *MockRoleAssignmentsClient) Get(ctx context.Context, scope string, roleAssignmentName string) (result authorizationmgmt.RoleAssignment, err error) {\n\treturn c.MockGet(ctx, scope, roleAssignmentName)\n}", "func (r *repoManager) LastRollRev() string {\n\tr.infoMtx.RLock()\n\tdefer r.infoMtx.RUnlock()\n\treturn r.lastRollRev\n}", "func (t *SimpleChaincode) readDegree(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tvar enrolment, jsonResp string\n\tvar err error\n\n\tif len(args) != 1 {\n\t\treturn shim.Error(\"Incorrect number of arguments.Please give your enrolment number to fetch degree \")\n\t}\n\n\tenrolment = args[0]\n\tvalAsbytes, err := stub.GetState(enrolment) //get the degree from chaincode state\n\tif err != nil {\n\t\tjsonResp = \"{\\\"Error\\\":\\\"Failed to get state with this enrolment number \" + enrolment + \"\\\"}\"\n\t\treturn shim.Error(jsonResp)\n\t} else if valAsbytes == nil {\n\t\tjsonResp = \"{\\\"Error\\\":\\\"Degree does not exist: \" + enrolment + \"\\\"}\"\n\t\treturn shim.Error(jsonResp)\n\t}\n\n\treturn shim.Success(valAsbytes)\n}", "func (c *Contract) Repurchase(ctx TransactionContextInterface, jeweler string, paperNumber string, repurchaseDateTime string) (*InventoryFinancingPaper, error) {\r\n\tpaper, err := ctx.GetPaperList().GetPaper(jeweler, paperNumber)\r\n\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\tif paper.IsRepurchased() {\r\n\t\treturn nil, fmt.Errorf(\"paper %s:%s is already Repurchased\", jeweler, paperNumber)\r\n\t}\r\n\r\n\tpaper.SetRepurchased()\r\n\r\n\terr = ctx.GetPaperList().UpdatePaper(paper)\r\n\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\tfmt.Printf(\"inventory paper %q:%q is repurchased by %q,The repurchased date is %q. Current state = %q\\n\", jeweler, paperNumber, paper.GetRepurchaser(), repurchaseDateTime, paper.GetState())\r\n\treturn paper, nil\r\n}", "func GetStudentEnrolled(rw http.ResponseWriter, r *http.Request) {\n\tparams := r.URL.Query()\n\tid := params.Get(\":id\")\n\tdoGetEnroll(id)\n\trowReturnEnrolled.Value= studentEnrolledAPI.Rows[0].Value\n\trowReturnEnrolled.Key= studentEnrolledAPI.Rows[0].Key\n\trowReturnEnrolled.Id= studentEnrolledAPI.Rows[0].Id\n\ta, _ := json.Marshal(rowReturnEnrolled)\n\trw.WriteHeader(http.StatusOK)\n\trw.Write([]byte(a))\n}", "func GetGrades(config map[string]string) map[string]util.Module {\n\tclient := new(portalClient)\n\tclient.config = config\n\tclient.loginPortal()\n\tgrades := client.crawlPortal()\n\tclient.getPage(config[\"logout\"], false)\n\treturn grades\n}", "func (db *diceBae) LetsRoll() error {\n\tif err := db.session.Open(); err != nil {\n\t\treturn fmt.Errorf(\"failed to open Discord session: %v\", err)\n\t}\n\tdefer db.session.Close()\n\tdefer db.logFile.Close()\n\n\tdb.LogInfo(\"I have no dice, but I must roll. Press CTRL-C to exit.\")\n\tsc := make(chan os.Signal, 1)\n\t// Block on this channel until we get a termination signal.\n\tsignal.Notify(sc, syscall.SIGINT, syscall.SIGTERM, os.Interrupt, os.Kill)\n\t<-sc\n\tdb.LogInfo(\"Later dopes.\")\n\treturn nil\n}", "func TestDRollN(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Valid request\", \"/d4/1\", response{http.StatusOK, `\"count\":1,\"sides\":4`}},\n\t\t{\"Valid request\", \"/D4/1\", response{http.StatusOK, `\"count\":1,\"sides\":4`}},\n\t\t{\"Invalid dice variable\", \"/d5/1\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid dice variable\", \"/D5/1\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid count variable\", \"/d4/0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid count variable\", \"/D4/0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t// Not sure why count ocmes up first here, since the sides get parsed first in the code.\n\t\t// Moreover, not sure if I should be looking specifically for it or just for an error.\n\t\t{\"Invalid dice and count variable\", \"/d5/0\", response{http.StatusNotAcceptable, `\"error\":\"invalid count\"`}},\n\t\t{\"Invalid dice and count variable\", \"/D5/0\", response{http.StatusNotAcceptable, `\"error\":\"invalid count\"`}},\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func (r *AutoRoller) GetActiveRoll() state_machine.RollCLImpl {\n\treturn r.currentRoll\n}", "func (f *Frame) getPointValueFromLastRoll() int {\n\tlastRollString := f.rolls[f.currentBall-2]\n\tswitch lastRollString {\n\tcase \"X\":\n\t\treturn 10\n\tcase \"/\":\n\t\trollValue, _ := strconv.Atoi(f.rolls[f.currentBall-3])\n\t\treturn 10 - rollValue\n\tcase \"1\", \"2\", \"3\", \"4\", \"5\", \"6\", \"7\", \"8\", \"9\":\n\t\trollValue, _ := strconv.Atoi(lastRollString)\n\t\treturn rollValue\n\t}\n\treturn 0\n}", "func fetchLeaderboard(request request) (*Leaderboard, *Error) {\n\tresult := &leaderboardResponse{}\n\n\terr := httpClient.do(request, result)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &result.Data, nil\n}", "func (commit *Commit) GetRound() int32 {\n\tif commit == nil {\n\t\treturn -1\n\t}\n\treturn commit.Round\n}", "func (commit *Commit) GetRound() int32 {\n\treturn commit.Round\n}", "func getPunchline(seed string) (string, error) {\n\trapLyricsResp, err := http.PostForm(\n\t\trapLyricsApi,\n\t\turl.Values{\"input\": {seed}})\n\tif err != nil {\n\t\tlog.Printf(\"error while calling raplyrics %s\", err.Error())\n\t\treturn \"\", err\n\t}\n\tvar punchline Lyric\n\tif err := json.NewDecoder(rapLyricsResp.Body).Decode(&punchline); err != nil {\n\t\tlog.Printf(\"could not decode incoming punchline %s\", err.Error())\n\t\treturn \"\", err\n\t}\n\tdefer rapLyricsResp.Body.Close()\n\treturn punchline.Punch, nil\n}", "func (m *EducationAssignment) GetRubric()(EducationRubricable) {\n return m.rubric\n}", "func (r *AutoRoller) getRevision(ctx context.Context, id string) (*revision.Revision, error) {\n\tif id == r.lastRollRev.Id {\n\t\treturn r.lastRollRev, nil\n\t}\n\tif id == r.nextRollRev.Id {\n\t\treturn r.nextRollRev, nil\n\t}\n\tif id == r.tipRev.Id {\n\t\treturn r.tipRev, nil\n\t}\n\tfor _, rev := range r.notRolledRevs {\n\t\tif id == rev.Id {\n\t\t\treturn rev, nil\n\t\t}\n\t}\n\treturn r.rm.GetRevision(ctx, id)\n}", "func MakeRoll(dieType int) int {\n\treturn (rand.Intn(dieType) + 1)\n}", "func (game *Game) Roll(pinsDown int) {\n\tgame.currentFrame.Roll(pinsDown)\n\n\tif game.currentFrame.IsComplete() {\n\t\tgame.score += game.beforePreviousFrame.Bonus(\n\t\t\tgame.previousFrame,\n\t\t\tgame.currentFrame,\n\t\t)\n\t\tgame.score += game.currentFrame.Score()\n\t\tgame.beforePreviousFrame = game.previousFrame\n\t\tgame.previousFrame = game.currentFrame\n\t\tgame.currentFrame = *new(Frame)\n\t}\n}", "func getSale(userKey string, arg string) (Sale, error) {\n\tvar err error\n\tvar submission SaleSubmission\n\tvar sale Sale\n\n\terr = json.Unmarshal([]byte(arg), &submission)\n\tif err != nil {\n\t\treturn sale, fmt.Errorf(\"Cannot unmarshal sale submission.\")\n\t} else if submission.Quantity == 0 {\n\t\treturn sale, fmt.Errorf(\"Sale submission's quantity must be greater than 0.\")\n\t}\n\tsale.User = userKey\n\tsale.ItemId = submission.ItemId\n\tsale.Quantity = submission.Quantity\n\tsale.ShopId = submission.ShopId\n\tsale.DocType = \"Sale\"\n\treturn sale, nil\n}", "func (r *AutoRoller) rollFinished(ctx context.Context, justFinished codereview.RollImpl) error {\n\trecent := r.recent.GetRecentRolls()\n\t// Sanity check: pop any rolls which occurred after the one which just\n\t// finished.\n\tidx := -1\n\tvar currentRoll *autoroll.AutoRollIssue\n\tfor i, roll := range recent {\n\t\tissue := fmt.Sprintf(\"%d\", roll.Issue)\n\t\tif issue == justFinished.IssueID() {\n\t\t\tidx = i\n\t\t\tcurrentRoll = roll\n\t\t\tbreak\n\t\t}\n\t}\n\tif currentRoll == nil {\n\t\treturn skerr.Fmt(\"Unable to find just-finished roll %q in recent list!\", justFinished.IssueID())\n\t}\n\n\t// Feed AutoRoll stats into metrics.\n\tv := int64(0)\n\tif currentRoll.Closed && currentRoll.Committed {\n\t\tv = int64(1)\n\t}\n\tmetrics2.GetInt64Metric(\"autoroll_last_roll_result\", map[string]string{\"roller\": r.cfg.RollerName}).Update(v)\n\n\trecent = recent[idx:]\n\tvar lastRoll *autoroll.AutoRollIssue\n\tif len(recent) > 1 {\n\t\tlastRoll = recent[1]\n\t} else {\n\t\t// If there are no other rolls, then the below alerts do not apply.\n\t\treturn nil\n\t}\n\n\tissueURL := fmt.Sprintf(\"%s%d\", r.codereview.GetIssueUrlBase(), currentRoll.Issue)\n\n\t// Send notifications if this roll had a different result from the last\n\t// roll, ie. success -> failure or failure -> success.\n\tcurrentSuccess := util.In(currentRoll.Result, autoroll.SUCCESS_RESULTS)\n\tlastSuccess := util.In(lastRoll.Result, autoroll.SUCCESS_RESULTS)\n\tif lastRoll != nil {\n\t\tif currentSuccess && !lastSuccess {\n\t\t\tr.notifier.SendNewSuccess(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t} else if !currentSuccess && lastSuccess {\n\t\t\tr.notifier.SendNewFailure(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t}\n\t}\n\n\t// Send a notification if the last N rolls failed in a row.\n\tnFailed := 0\n\t// recent is in reverse chronological order.\n\tfor _, roll := range recent {\n\t\tif util.In(roll.Result, autoroll.SUCCESS_RESULTS) {\n\t\t\tbreak\n\t\t} else {\n\t\t\tnFailed++\n\t\t}\n\t}\n\tif nFailed == notifyIfLastNFailed {\n\t\tr.notifier.SendLastNFailed(ctx, notifyIfLastNFailed, issueURL)\n\t}\n\n\treturn nil\n}", "func (s *AFDOStrategy) GetNextRollRev(ctx context.Context, _ []*vcsinfo.LongCommit) (string, error) {\n\t// Find the available AFDO versions, sorted newest to oldest, and store.\n\tavailable := []string{}\n\tif err := s.gcs.AllFilesInDirectory(ctx, AFDO_GS_PATH, func(item *storage.ObjectAttrs) {\n\t\tname := strings.TrimPrefix(item.Name, AFDO_GS_PATH)\n\t\tif _, err := parseAFDOVersion(name); err == nil {\n\t\t\tavailable = append(available, name)\n\t\t} else if err == errInvalidAFDOVersion {\n\t\t\t// There are files we don't care about in this bucket. Just ignore.\n\t\t} else {\n\t\t\tsklog.Error(err)\n\t\t}\n\t}); err != nil {\n\t\treturn \"\", err\n\t}\n\tif len(available) == 0 {\n\t\treturn \"\", fmt.Errorf(\"No valid AFDO profile names found.\")\n\t}\n\tsort.Sort(afdoVersionSlice(available))\n\n\t// Store the available versions. Return the newest.\n\ts.mtx.Lock()\n\tdefer s.mtx.Unlock()\n\ts.versions = available\n\treturn s.versions[0], nil\n}", "func (d *DepartmentRepoImpl) Read(deptno int32) (models.Department, error) {\n\tdept := models.Department{}\n\tgetDepartment := d.DB.Table(\"departments\").Where(\"id = ?\", deptno).Find(&dept)\n\tif getDepartment.Error != nil {\n\t\tfmt.Println(\"[Repo Error] :\", getDepartment.Error)\n\t\treturn dept, getDepartment.Error\n\t}\n\n\treturn dept, nil\n}", "func (db *DB) GetReplicatedPulse(ctx context.Context) (core.PulseNumber, error) {\n\tbuf, err := db.get(ctx, prefixkey(scopeIDSystem, []byte{sysReplicatedPulse}))\n\tif err != nil {\n\t\tif err == ErrNotFound {\n\t\t\terr = nil\n\t\t}\n\t\treturn 0, err\n\t}\n\treturn core.NewPulseNumber(buf), nil\n}", "func (api *API) GetLeague(ctx context.Context, leagueID string, page int) (*models.Resource, error) {\n\tmethod := \"GET\"\n\tpageString := strconv.Itoa(page)\n\tpath := api.URI + \"/leagues-classic/\" + leagueID + \"/standings/?page_standings=\" + pageString\n\tlogData := log.Data{\"url\": path, \"method\": method}\n\n\tURL, err := url.Parse(path)\n\tif err != nil {\n\t\tlog.Event(ctx, \"failed to create url for api call\", log.ERROR, log.Error(err), logData)\n\t\treturn nil, err\n\t}\n\tpath = URL.String()\n\tlogData[\"url\"] = path\n\n\tbody, err := api.makeGetRequest(ctx, method, path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer body.Close()\n\n\tvar resource models.Resource\n\tif err := json.NewDecoder(body).Decode(&resource); err != nil {\n\t\tlog.Event(ctx, \"unable to unmarshal bytes into league resource\", log.ERROR, log.Error(err), logData)\n\t\treturn nil, err\n\t}\n\n\t// Check and retrieve other pages\n\tif resource.Standings != nil && resource.Standings.HasNext {\n\t\tpage++\n\t\tnext, err := api.GetLeague(ctx, leagueID, page)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tresource.Standings.Results = append(resource.Standings.Results, next.Standings.Results...)\n\t}\n\n\tlog.Event(ctx, \"successfully got league\", log.INFO, log.Data{\"resource\": resource})\n\n\treturn &resource, nil\n}", "func fetchDeal(id int) (Deal, error) {\n\tsession := connect()\n\tresult := Deal{}\n\tc := session.DB(\"test\").C(\"deals\")\n\terr := c.Find(bson.M{\"id\": id}).One(&result)\n\tif err != nil {\n\t\tfmt.Printf(\"Could not find deal %d in database. Err: %s\\n\", id, err.Error())\n\t\treturn Deal{}, errors.New(\"Invalid Id\")\n\t}\n\tfmt.Printf(\"Found deal %d in datastore.\\n\", id)\n\t// Set Expiry date to +1 week\n\tresult.Expires = time.Now().AddDate(0, 0, 7).Format(\"2006-01-02\")\n\tfmt.Println(\"Expiry date added.\")\n\n\treturn result, nil\n}", "func (r *reviewer) Review(group, resource, name string) (Review, error) {\n\taction := authorizer.AttributesRecord{\n\t\tVerb: \"get\",\n\t\tAPIGroup: group,\n\t\tResource: resource,\n\t\tName: name,\n\t\tResourceRequest: true,\n\t}\n\n\tsubjects, err := r.subjectAccessEvaluator.AllowedSubjects(action)\n\treview := &defaultReview{}\n\treview.users, review.groups = authorizationutil.RBACSubjectsToUsersAndGroups(subjects, action.GetNamespace())\n\tif err != nil {\n\t\treview.evaluationError = err.Error()\n\t}\n\treturn review, nil\n}", "func GetRoleAssignment(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *RoleAssignmentState, opts ...pulumi.ResourceOption) (*RoleAssignment, error) {\n\tvar resource RoleAssignment\n\terr := ctx.ReadResource(\"azure:marketplace/roleAssignment:RoleAssignment\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (cli *Service) GetRolePri(req *restful.Request, resp *restful.Response) {\n\n\tlanguage := util.GetActionLanguage(req)\n\townerID := util.GetOwnerID(req.Request.Header)\n\tdefErr := cli.Core.CCErr.CreateDefaultCCErrorIf(language)\n\tctx := util.GetDBContext(context.Background(), req.Request.Header)\n\tdb := cli.Instance.Clone()\n\n\tdefer req.Request.Body.Close()\n\tpathParams := req.PathParameters()\n\tobjID := pathParams[\"bk_obj_id\"]\n\tpropertyID := pathParams[\"bk_property_id\"]\n\tcond := make(map[string]interface{})\n\tcond[common.BKObjIDField] = objID\n\tcond[common.BKPropertyIDField] = propertyID\n\tvar result map[string]interface{}\n\tcond = util.SetModOwner(cond, ownerID)\n\n\tcnt, err := db.Table(common.BKTableNamePrivilege).Find(cond).Count(ctx)\n\tif nil != err {\n\t\tblog.Error(\"get user group privi error :%v\", err)\n\t\tresp.WriteError(http.StatusBadRequest, &meta.RespError{Msg: defErr.New(common.CCErrObjectDBOpErrno, err.Error())})\n\t\treturn\n\t}\n\tif 0 == cnt { // TODO:\n\t\tblog.V(3).Infof(\"failed to find the cnt\")\n\t\tinfo := make(map[string]interface{})\n\t\tresp.WriteEntity(meta.Response{BaseResp: meta.SuccessBaseResp, Data: info})\n\t\treturn\n\t}\n\n\terr = db.Table(common.BKTableNamePrivilege).Find(cond).One(ctx, &result)\n\tif nil != err {\n\t\tblog.Error(\"get role pri field error :%v\", err)\n\t\tresp.WriteError(http.StatusBadRequest, &meta.RespError{Msg: defErr.New(common.CCErrCommDBSelectFailed, err.Error())})\n\t\treturn\n\n\t}\n\tprivilege, ok := result[\"privilege\"]\n\tif !ok {\n\t\tblog.Errorf(\"not privilege, the origin data is %#v\", result)\n\t\tinfo := make(map[string]interface{})\n\t\tresp.WriteEntity(meta.Response{BaseResp: meta.SuccessBaseResp, Data: info})\n\t\treturn\n\n\t}\n\tresp.WriteEntity(meta.Response{BaseResp: meta.SuccessBaseResp, Data: privilege})\n}", "func (o *PostLeasesIDAuthReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostLeasesIDAuthCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewPostLeasesIDAuthUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewPostLeasesIDAuthForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewPostLeasesIDAuthInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (a *LaborApiService) GetEmployeeWage(ctx context.Context, id string) (GetEmployeeWageResponse, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Get\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue GetEmployeeWageResponse\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/v2/labor/employee-wages/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", fmt.Sprintf(\"%v\", id), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode < 300 {\n\t\t// If we succeed, return the data, otherwise pass on to decode error.\n\t\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"));\n\t\tif err == nil { \n\t\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t\t}\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v GetEmployeeWageResponse\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"));\n\t\t\t\tif err != nil {\n\t\t\t\t\tnewErr.error = err.Error()\n\t\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t\t}\n\t\t\t\tnewErr.model = v\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func (lps *LibraryPanelService) getLibraryPanel(c *models.ReqContext, uid string) (LibraryPanel, error) {\n\tvar libraryPanel LibraryPanel\n\terr := lps.SQLStore.WithDbSession(context.Background(), func(session *sqlstore.DBSession) error {\n\t\tvar err error\n\t\tlibraryPanel, err = getLibraryPanel(session, uid, c.SignedInUser.OrgId)\n\t\treturn err\n\t})\n\n\treturn libraryPanel, err\n}", "func (service *LoanSrvc) GetLoan(_ context.Context, p *loan.GetLoanPayload) (res *loan.Loan, view string, err error) {\n\tl := service.loan.GetLoan(p.ID)\n\n\tpaybackDate := l.PaybackDate.String()\n\tres = &loan.Loan{\n\t\tID: l.ID,\n\t\tAmount: l.Amount,\n\t\tLenderID: l.LenderID,\n\t\tBorrowerID: l.BorrowerID,\n\t\tPaybackDate: &paybackDate,\n\t}\n\n\treturn res, p.View, nil\n}", "func (r *MockRepoManager) CreateNewRoll(ctx context.Context, from, to string, emails []string, cqExtraTrybots string, dryRun bool) (int64, error) {\n\tr.mtx.RLock()\n\tdefer r.mtx.RUnlock()\n\treturn r.mockIssueNumber, nil\n}", "func HandleReview(res http.ResponseWriter, req *http.Request) {\n\tswitch req.Method {\n\tcase \"post\":\n\tcase \"POST\":\n\t\tmutex.Lock()\n\t\tfmt.Println(\"posting lunch review\")\n\t\tdate := req.FormValue(\"date\")\n\t\tmenuType := req.FormValue(\"menuType\")\n\t\tscore, scoreErr := strconv.ParseFloat(req.FormValue(\"score\"), 10)\n\t\tif menuType != menu.NEWMENUTYPE && menuType != menu.OLDMENUTYPE {\n\t\t\terrorString := \"incorrect menu type specified\"\n\t\t\tfmt.Println(errorString, \" \"+menuType)\n\t\t\thttp.Error(res, errorString, http.StatusBadRequest)\n\t\t\tmutex.Unlock()\n\t\t\treturn\n\t\t}\n\t\tif scoreErr != nil {\n\t\t\tfmt.Println(\"error in score parsing\")\n\t\t\thttp.Error(res, scoreErr.Error(), http.StatusBadRequest)\n\t\t\tmutex.Unlock()\n\t\t\treturn\n\t\t}\n\t\tif date == \"\" {\n\t\t\tfmt.Println(\"no date found\")\n\t\t\thttp.Error(res, \"no date found\", http.StatusBadRequest)\n\t\t\tmutex.Unlock()\n\t\t\treturn\n\t\t}\n\t\tspreadSheetClient, spreadSheetClientErr := authentication.GetClient()\n\t\tif spreadSheetClientErr != nil {\n\t\t\tfmt.Println(spreadSheetClientErr)\n\t\t\thttp.Error(res, spreadSheetClientErr.Error(), http.StatusInternalServerError)\n\t\t\tmutex.Unlock()\n\t\t\treturn\n\t\t}\n\t\tselectedMenu := menu.New(menuType, spreadSheetClient)\n\t\tcurrentScore, reviewErr := selectedMenu.PostReview(date, score)\n\t\tif reviewErr != nil {\n\t\t\tfmt.Println(reviewErr)\n\t\t\thttp.Error(res, reviewErr.Error(), http.StatusInternalServerError)\n\t\t\tmutex.Unlock()\n\t\t\treturn\n\t\t}\n\t\tjson.NewEncoder(res).Encode(currentScore)\n\t\tmutex.Unlock()\n\t\tbreak\n\tcase \"get\":\n\tcase \"GET\":\n\t\tdate := req.URL.Query().Get(\"date\")\n\t\tfmt.Println(\"retrieving score for date ->\", date)\n\t\tif date == \"\" {\n\t\t\tvar errString = \"incorrect date specified\"\n\t\t\tfmt.Println(errString)\n\t\t\thttp.Error(res, errString, http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tmenuType := req.URL.Query().Get(\"menuType\")\n\t\tfmt.Println(\"retrieving score for menu type -> \", menuType)\n\t\tif menuType != menu.NEWMENUTYPE && menuType != menu.OLDMENUTYPE {\n\t\t\tvar errString = \"menu type should either be new or old\"\n\t\t\tfmt.Println(errString)\n\t\t\thttp.Error(res, errString, http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tspreadSheetClient, spreadSheetClientErr := authentication.GetClient()\n\t\tif spreadSheetClientErr != nil {\n\t\t\tfmt.Println(spreadSheetClientErr)\n\t\t\thttp.Error(res, spreadSheetClientErr.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tselectedMenu := menu.New(menuType, spreadSheetClient)\n\t\tscore, err := selectedMenu.GetScore(date)\n\t\tif err != nil {\n\t\t\tfmt.Println(err.Error())\n\t\t\thttp.Error(res, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tjson.NewEncoder(res).Encode(score)\n\t\tbreak\n\t}\n}", "func (g *Game) getSpareBonus(rollIndex int) int {\n\treturn g.rolls[rollIndex+2]\n}", "func (m *SecureScoreControlProfile) GetTier()(*string) {\n return m.tier\n}", "func ExampleHCRPAssignmentReportsClient_Get() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armguestconfiguration.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewHCRPAssignmentReportsClient().Get(ctx, \"myResourceGroupName\", \"AuditSecureProtocol\", \"7367cbb8-ae99-47d0-a33b-a283564d2cb1\", \"myMachineName\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.AssignmentReport = armguestconfiguration.AssignmentReport{\n\t// \tName: to.Ptr(\"7367cbb8-ae99-47d0-a33b-a283564d2cb1\"),\n\t// \tID: to.Ptr(\"/subscriptions/mysubscriptionid/resourceGroups/myResourceGroupName/providers/Microsoft.HybridCompute/machines/myMachineName/providers/Microsoft.GuestConfiguration/guestConfigurationAssignments/AuditSecureProtocol/reports/7367cbb8-ae99-47d0-a33b-a283564d2cb1\"),\n\t// \tProperties: &armguestconfiguration.AssignmentReportProperties{\n\t// \t\tAssignment: &armguestconfiguration.AssignmentInfo{\n\t// \t\t\tName: to.Ptr(\"AuditSecureProtocol\"),\n\t// \t\t\tConfiguration: &armguestconfiguration.ConfigurationInfo{\n\t// \t\t\t\tName: to.Ptr(\"AuditSecureProtocol\"),\n\t// \t\t\t\tVersion: to.Ptr(\"1.0.0.0\"),\n\t// \t\t\t},\n\t// \t\t},\n\t// \t\tComplianceStatus: to.Ptr(armguestconfiguration.ComplianceStatusCompliant),\n\t// \t\tEndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2018-08-29T22:14:13Z\"); return t}()),\n\t// \t\tReportID: to.Ptr(\"7367cbb8-ae99-47d0-a33b-a283564d2cb1\"),\n\t// \t\tStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2018-08-29T22:13:53Z\"); return t}()),\n\t// \t\tVM: &armguestconfiguration.VMInfo{\n\t// \t\t\tID: to.Ptr(\"/subscriptions/mysubscriptionid/resourceGroups/myResourceGroupName/providers/Microsoft.HybridCompute/machines/myMachineName\"),\n\t// \t\t\tUUID: to.Ptr(\"vmuuid\"),\n\t// \t\t},\n\t// \t\tDetails: &armguestconfiguration.AssignmentReportDetails{\n\t// \t\t\tComplianceStatus: to.Ptr(armguestconfiguration.ComplianceStatusCompliant),\n\t// \t\t\tEndTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2018-08-29T22:14:13Z\"); return t}()),\n\t// \t\t\tJobID: to.Ptr(\"7367cbb8-ae99-47d0-a33b-a283564d2cb1\"),\n\t// \t\t\tOperationType: to.Ptr(armguestconfiguration.TypeConsistency),\n\t// \t\t\tResources: []*armguestconfiguration.AssignmentReportResource{\n\t// \t\t\t\t{\n\t// \t\t\t\t\tComplianceStatus: to.Ptr(armguestconfiguration.ComplianceStatusCompliant),\n\t// \t\t\t\t\tProperties: map[string]any{\n\t// \t\t\t\t\t\t\"ConfigurationName\": \"IsWebServerSecure\",\n\t// \t\t\t\t\t\t\"DependsOn\": nil,\n\t// \t\t\t\t\t\t\"IsSingleInstance\": \"Yes\",\n\t// \t\t\t\t\t\t\"ModuleName\": \"SecureProtocolWebServer\",\n\t// \t\t\t\t\t\t\"ModuleVersion\": \"1.0.0.3\",\n\t// \t\t\t\t\t\t\"Protocols\":[]any{\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"SSL 2.0\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"SSL 3.0\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"TLS 1.0\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"PCT 1.0\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"Multi-Protocol Unified Hello\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"TLS 1.1\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\tmap[string]any{\n\t// \t\t\t\t\t\t\t\t\"Ensure\": \"Absent\",\n\t// \t\t\t\t\t\t\t\t\"Protocol\": \"TLS 1.2\",\n\t// \t\t\t\t\t\t\t},\n\t// \t\t\t\t\t\t},\n\t// \t\t\t\t\t\t\"PsDscRunAsCredential\": nil,\n\t// \t\t\t\t\t\t\"Reasons\": nil,\n\t// \t\t\t\t\t\t\"ResourceId\": \"[SecureWebServer]s1\",\n\t// \t\t\t\t\t\t\"SourceInfo\": nil,\n\t// \t\t\t\t\t},\n\t// \t\t\t\t\tReasons: []*armguestconfiguration.AssignmentReportResourceComplianceReason{\n\t// \t\t\t\t\t\t{\n\t// \t\t\t\t\t\t\tCode: to.Ptr(\"DSC::RESOURCE::SUCCESS\"),\n\t// \t\t\t\t\t\t\tPhrase: to.Ptr(\"Operation successful.\"),\n\t// \t\t\t\t\t}},\n\t// \t\t\t}},\n\t// \t\t\tStartTime: to.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2018-08-29T22:13:53Z\"); return t}()),\n\t// \t\t},\n\t// \t},\n\t// }\n}", "func (rrl *RRL) debit(allowance int, t string) (int, error) {\n\tresult := rrl.table.UpdateAdd(t,\n\t\t// the 'update' function debits the account and returns the new balance\n\t\tfunc(el *interface{}) interface{} {\n\t\t\tra := (*el).(*ResponseAccount)\n\t\t\tif ra == nil {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tnow := time.Now()\n\t\t\tra.balance += allowance*int(now.Sub(ra.lastCheck).Seconds()) - 1\n\t\t\tif ra.balance >= rrl.window {\n\t\t\t\t// balance can't exceed window\n\t\t\t\tra.balance = rrl.window - 1\n\t\t\t} else if min := -1 * rrl.window * allowance; ra.balance < min {\n\t\t\t\t// balance can't be more negative than window * allowance\n\t\t\t\tra.balance = min\n\t\t\t}\n\t\t\tra.lastCheck = now\n\t\t\treturn ra.balance\n\t\t},\n\t\t// the 'add' function returns a new ResponseAccount for the response type\n\t\tfunc() interface{} {\n\t\t\tra := &ResponseAccount{\n\t\t\t\tallowance: allowance,\n\t\t\t\tlastCheck: time.Now(),\n\t\t\t\tbalance: rrl.window - 1,\n\t\t\t}\n\t\t\treturn ra\n\t\t})\n\n\tif result == nil {\n\t\treturn 0, nil\n\t}\n\tif err, ok := result.(error); ok {\n\t\treturn 0, err\n\t}\n\tif balance, ok := result.(int); ok {\n\t\treturn balance, nil\n\t}\n\treturn 0, errors.New(\"unexpected result type\")\n}", "func ViewRedeem(rw http.ResponseWriter, r *http.Request) {\n\t// get the token\n\treqToken := r.Header.Get(\"Authorization\")\n\n\t// get the claims\n\tclaims, isNotValid := GetClaims(reqToken, rw)\n\tif isNotValid {\n\t\treturn\n\t}\n\n\tdt, err := db.GetUserRedeem(claims.Roll)\n\tif err != nil {\n\t\trw.WriteHeader(http.StatusInternalServerError)\n\t\trw.Write(Rsp(err.Error(), \"Server Error\"))\n\t\treturn\n\t}\n\trw.WriteHeader(http.StatusOK)\n\tres := c.RespData{\n\t\tMessage: \"All data\",\n\t\tData: dt,\n\t}\n\tjson.NewEncoder(rw).Encode(res)\n}", "func (o *DlcOracle) FetchRPoint(datafeedId, timestamp uint64) ([33]byte, error) {\n\tvar rPoint [33]byte\n\tif len(o.Url) == 0 {\n\t\treturn rPoint, fmt.Errorf(\"Oracle was not imported from the web -\" +\n\t\t\t\" cannot fetch R point. Enter manually using the\" +\n\t\t\t\" [dlc contract setrpoint] command\")\n\t}\n\n\turl := fmt.Sprintf(\"%s/api/rpoint/%d/%d\", o.Url, datafeedId, timestamp)\n\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn rPoint, err\n\t}\n\tclient := &http.Client{}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn rPoint, err\n\t}\n\tdefer resp.Body.Close()\n\n\tvar response DlcOracleRPointResponse\n\n\tif err := json.NewDecoder(resp.Body).Decode(&response); err != nil {\n\t\treturn rPoint, err\n\t}\n\n\tR, err := hex.DecodeString(response.RHex)\n\tif err != nil {\n\t\treturn rPoint, err\n\t}\n\n\tcopy(rPoint[:], R[:])\n\treturn rPoint, nil\n\n}", "func (s *authenticationReviewLister) Get(name string) (*v3.AuthenticationReview, error) {\n\tobj, exists, err := s.indexer.GetByKey(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, errors.NewNotFound(v3.Resource(\"authenticationreview\"), name)\n\t}\n\treturn obj.(*v3.AuthenticationReview), nil\n}", "func (client *Client) GetRetcodeLogstore(request *GetRetcodeLogstoreRequest) (response *GetRetcodeLogstoreResponse, err error) {\n\tresponse = CreateGetRetcodeLogstoreResponse()\n\terr = client.DoAction(request, response)\n\treturn\n}", "func (haa *HistoryArchiveAdapter) GetLedger(sequence uint32) (io.ArchiveLedgerReader, error) {\n\treturn nil, fmt.Errorf(\"not implemented yet\")\n}", "func journalAssessment(ctx context.Context, xbiz *XBusiness, d time.Time, a *Assessment, d1, d2 *time.Time) (Journal, error) {\n\tfuncname := \"journalAssessment\"\n\t// Console(\"*** Entered %s\\n\", funcname)\n\t// Console(\"%s: d = %s, d1 = %s, d2 = %s\\n\", funcname, d.Format(RRDATEREPORTFMT), d1.Format(RRDATEREPORTFMT), d2.Format(RRDATEREPORTFMT))\n\t// Console(\"%s: Assessment: PASMID = %d, RentCycle = %d, ProrationCycle = %d, Start = %s, Stop = %s\\n\", funcname, a.PASMID, a.RentCycle, a.ProrationCycle, a.Start.Format(RRDATETIMEW2UIFMT), a.Stop.Format(RRDATETIMEW2UIFMT))\n\tvar j Journal\n\n\t// pf, num, den, start, stop, err := ProrateAssessment(ctx, xbiz, a, &d, d1, d2)\n\tpf, _, _, _, _, err := ProrateAssessment(ctx, xbiz, a, &d, d1, d2)\n\tif err != nil {\n\t\t// Console(\"%s: exiting. err = %s\\n\", funcname, err.Error())\n\t\treturn j, err\n\t}\n\n\t// Console(\"%s: A:: **** AFTER PRORATION CHECK **** pf = %6.4f, num = %d, den = %d, start = %s, stop = %s\\n\", funcname, pf, num, den, start.Format(RRDATEFMT3), stop.Format(RRDATEFMT3))\n\t// Console(\"%s: B:: After ProrateAssessment: start = %s, stop = %s\\n\", funcname, start.Format(RRDATETIMEW2UIFMT), stop.Format(RRDATETIMEW2UIFMT))\n\n\t//--------------------------------------------------------------------------------------\n\t// This is a safeguard against issues encountered in Feb 2018 where rent assessments\n\t// continued after the RentalAgreement RentStop date.\n\t//--------------------------------------------------------------------------------------\n\tif pf < float64(0) {\n\t\tpf = float64(0)\n\t}\n\n\t// Console(\"%s: a.ASMTID = %d, d = %s, d1 = %s, d2 = %s\\n\", funcname, a.ASMID, d.Format(RRDATEFMT4), d1.Format(RRDATEFMT4), d2.Format(RRDATEFMT4))\n\t// Console(\"%s: pf = %f, num = %d, den = %d, start = %s, stop = %s\\n\", funcname, pf, num, den, start.Format(RRDATEFMT4), stop.Format(RRDATEFMT4))\n\n\tj = Journal{BID: a.BID, Dt: d, Type: JNLTYPEASMT, ID: a.ASMID}\n\n\tasmRules, err := GetAssessmentAccountRule(ctx, a)\n\tif err != nil {\n\t\t// Console(\"%s: exiting. err = %s\\n\", funcname, err.Error())\n\t\treturn j, err\n\t}\n\n\t// Console(\"%s: C:: Parsing account rule: %s Amount = %8.2f\\n\", funcname, asmRules, a.Amount)\n\tm, err := ParseAcctRule(ctx, xbiz, a.RID, d1, d2, asmRules, a.Amount, pf) // a rule such as \"d 11001 1000.0, c 40001 1100.0, d 41004 100.00\"\n\tif err != nil {\n\t\t// Console(\"%s: C1:: exiting. err = %s\\n\", funcname, err.Error())\n\t\treturn j, err\n\t}\n\n\t// Console(\"%s: m = %#v\\n\", funcname, m)\n\t// for i := 0; i < len(m); i++ {\n\t// \tConsole(\"D:: m[%d].Amount = %f, .Action = %s .Expr = %s\\n\", i, m[i].Amount, m[i].Action, m[i].Expr)\n\t// }\n\n\t_, j.Amount = sumAllocations(&m)\n\tj.Amount = RoundToCent(j.Amount)\n\n\t// Console(\"%s: E:: j.Amount = %8.2f, pf = %8.5f\\n\", funcname, j.Amount, pf)\n\n\t//------------------------------------------------------------------------------------------------------\n\t// THIS BLOCK OF CODE SHOULD BE DELETED. IT SHOULD BE HANDLED IN ASSESSMENT CODE, NOT JOURNAL CODE.\n\t//=====================================================================================================\n\t// the assessment amount may have\n\t// been prorated as it was a newly created recurring assessment for a RentalAgreement that was either\n\t// just beginning or just ending. If so, we'll update the assessment amount here the calculated\n\t// j.Amount != a.Amount\n\t//------------------------------------------------------------------------------------------------------\n\t// if pf < 1.0 {\n\t// \tConsole(\"%s: F:: will update assessment\\n\", funcname)\n\t// \ta.Amount = j.Amount // update to the prorated amount\n\t// \ta.Start = start // adjust to the dates used in the proration\n\t// \ta.Stop = stop // adjust to the dates used in the proration\n\t// \ta.Comment = fmt.Sprintf(\"Prorated for %d of %d %s\", num, den, ProrationUnits(a.ProrationCycle))\n\t// \tConsole(\"%s: G:: a.Amount = %8.2f\\n\", funcname, a.Amount)\n\t// \tif err := UpdateAssessment(ctx, a); err != nil {\n\t// \t\terr = fmt.Errorf(\"Error updating prorated assessment amount: %s\", err.Error())\n\t// \t\tConsole(\"%s: H:: exiting. err = %s\\n\", funcname, err.Error())\n\t// \t\treturn j, err\n\t// \t}\n\t// \tConsole(\"%s: I:: Updating ASMID = %d, Amount = %8.2f\\n\", funcname, a.ASMID, a.Amount)\n\t// }\n\t// Console(\"%s: J:: ASMID = %d, Amount = %8.2f\\n\", funcname, a.ASMID, a.Amount)\n\n\t//-------------------------------------------------------------------------------------------\n\t// In the event that we need to prorate, pull together the pieces and determine the\n\t// fractional amounts so that all the entries can net to 0.00. Essentially, this means\n\t// handling the $0.01 off problem when dealing with fractional numbers. The way we'll\n\t// handle this is to apply the extra cent to the largest number\n\t//-------------------------------------------------------------------------------------------\n\tif pf < 1.0 {\n\t\t// new method using ProcessSum\n\t\tvar asum []SumFloat\n\t\tfor i := 0; i < len(m); i++ {\n\t\t\tvar b SumFloat\n\t\t\tif m[i].Action == \"c\" {\n\t\t\t\tb.Val = -m[i].Amount\n\t\t\t} else {\n\t\t\t\tb.Val = m[i].Amount\n\t\t\t}\n\t\t\tb.Amount = RoundToCent(b.Val)\n\t\t\tb.Remainder = b.Amount - b.Val\n\t\t\tasum = append(asum, b)\n\t\t}\n\t\tProcessSumFloats(asum)\n\t\tfor i := 0; i < len(asum); i++ {\n\t\t\tif m[i].Action == \"c\" {\n\t\t\t\tm[i].Amount = -asum[i].Amount // the adjusted value after ProcessSumFloats\n\t\t\t} else {\n\t\t\t\tm[i].Amount = asum[i].Amount // the adjusted value after ProcessSumFloats\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// Console(\"INSERTING JOURNAL: Date = %s, Type = %d, amount = %f\\n\", j.Dt, j.Type, j.Amount)\n\n\tjid, err := InsertJournal(ctx, &j)\n\tif err != nil {\n\t\tLogAndPrintError(funcname, err)\n\t\t// Console(\"%s: exiting. err = %s\\n\", funcname, err.Error())\n\t\treturn j, err\n\t}\n\n\ts := \"\"\n\tfor i := 0; i < len(m); i++ {\n\t\ts += fmt.Sprintf(\"%s %s %.2f\", m[i].Action, m[i].AcctExpr, RoundToCent(m[i].Amount))\n\t\tif i+1 < len(m) {\n\t\t\ts += \", \"\n\t\t}\n\t}\n\tif jid > 0 {\n\t\tvar ja JournalAllocation\n\t\tja.JID = jid\n\t\tja.RID = a.RID\n\t\tja.ASMID = a.ASMID\n\t\tja.Amount = RoundToCent(j.Amount)\n\t\tja.AcctRule = s\n\t\tja.BID = a.BID\n\t\tja.RAID = a.RAID\n\n\t\t// Console(\"INSERTING JOURNAL-ALLOCATION: ja.JID = %d, ja.ASMID = %d, ja.RAID = %d\\n\", ja.JID, ja.ASMID, ja.RAID)\n\t\tif _, err = InsertJournalAllocationEntry(ctx, &ja); err != nil {\n\t\t\tLogAndPrintError(funcname, err)\n\t\t\t// Console(\"%s: exiting. err = %s\\n\", funcname, err.Error())\n\t\t\treturn j, err\n\t\t}\n\t\tj.JA = append(j.JA, ja)\n\t}\n\n\t// Console(\"%s: exiting\\n\", funcname)\n\treturn j, err\n}", "func (o *PostRollbackReleaseNameReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostRollbackReleaseNameOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 401:\n\t\tresult := NewPostRollbackReleaseNameUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 403:\n\t\tresult := NewPostRollbackReleaseNameForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 500:\n\t\tresult := NewPostRollbackReleaseNameInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func GetLender(c *gin.Context) {\n\tvar pathVar schemas.LenderPath\n\tvar lender lender.Lender\n\tif err := c.BindUri(&pathVar); err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tif err := lender.Select(pathVar.LID); err != nil {\n\t\tc.JSON(http.StatusNotFound, gin.H{\"message\": \"not found\"})\n\t\treturn\n\t}\n\tdata := schemas.LenderSerializer(lender)\n\tc.JSON(http.StatusOK, data)\n\n}", "func OverReadingsCreateGet(c buffalo.Context) error {\n\ttx := c.Value(\"tx\").(*pop.Connection)\n\tparticipant := &models.Participant{}\n\tif err := tx.Eager(\"Screenings\", \"OverReadings\").Find(participant, c.Param(\"pid\")); err != nil {\n\t\treturn c.Error(404, err)\n\t}\n\tscreening := participant.Screenings[0]\n\tif !(c.Param(\"pid\") == participant.ID.String() && c.Param(\"sid\") == screening.ID.String()) {\n\t\tc.Flash().Add(\"danger\", \"Not Found\")\n\t\treturn c.Redirect(302, \"/cases/index\")\n\t}\n\tif len(participant.OverReadings) > 0 {\n\t\tovr := participant.OverReadings[0]\n\t\tred := \"/participants/\" + c.Param(\"pid\") + \"/screenings/\" + c.Param(\"sid\") + \"/overreadings/edit/\" + ovr.ID.String()\n\t\treturn c.Redirect(302, red)\n\t}\n\tc.Set(\"participant\", participant)\n\tc.Set(\"screening\", screening)\n\tc.Set(\"overReading\", &models.OverReading{})\n\n\tright, left, err := getImage(participant.ParticipantID)\n\tif err != nil {\n\t\tleft = \"\"\n\t\tright = \"\"\n\t}\n\n\tc.Set(\"leftEyeLink\", left)\n\tc.Set(\"rightEyeLink\", right)\n\n\tb := c.Value(\"breadcrumb\").(helpers.Breadcrumbs)\n\tb = append(b, helpers.Breadcrumb{Title: \"Cases\", Path: \"/cases/index\"})\n\tb = append(b, helpers.Breadcrumb{Title: \"New Over Reading\", Path: \"#\"})\n\tc.Set(\"breadcrumb\", b)\n\n\treturn c.Render(200, r.HTML(\"over_readings/create.html\"))\n}", "func chooseDice(sides int) dice.Dice {\n\tswitch sides {\n\tcase 4:\n\t\treturn dice.D4\n\tcase 6:\n\t\treturn dice.D6\n\tcase 8:\n\t\treturn dice.D8\n\tcase 10:\n\t\treturn dice.D10\n\tcase 12:\n\t\treturn dice.D12\n\tcase 20:\n\t\treturn dice.D20\n\tcase 100:\n\t\treturn dice.D100\n\tdefault:\n\t\treturn nil\n\t}\n}", "func (db *DB) GetReview(bookid string) (*Review, error) {\n\n\t// Query statement\n\tstmt := `SELECT bookid, username, rating, review, created FROM reviews WHERE bookid = $1`\n\n\t// Execute query\n\trow := db.QueryRow(stmt, bookid)\n\tr := &Review{}\n\n\t// Pull data into request*DB\n\terr := row.Scan(&r.BookID, &r.Username, &r.Rating, &r.Review, &r.Created)\n\tif err == sql.ErrNoRows {\n\t\treturn nil, nil\n\t} else if err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn r, nil\n}", "func (m Message) GetRepurchaseRate(f *field.RepurchaseRateField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetRepurchaseRate(f *field.RepurchaseRateField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetRepurchaseRate(f *field.RepurchaseRateField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func (m Message) GetRepurchaseRate(f *field.RepurchaseRateField) quickfix.MessageRejectError {\n\treturn m.Body.Get(f)\n}", "func Roll(value float64) *SimpleElement { return newSEFloat(\"roll\", value) }", "func (r *AutoRoller) doAutoRollInner() (string, error) {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\t// Get updated info about the current roll.\n\tif err := r.updateCurrentRoll(); err != nil {\n\t\treturn STATUS_ERROR, err\n\t}\n\n\t// There's a currently-active roll. Determine whether or not it's still good.\n\t// If so, leave it open and exit. If not, close it so that we can open another.\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tsklog.Infof(\"Found current roll: %s\", r.issueUrl(currentRoll.Issue))\n\n\t\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\t\t// If we have a normal (non-dry-run) roll running,\n\t\t\t// switch it to a dry run.\n\t\t\tif currentRoll.CommitQueue {\n\t\t\t\tsklog.Infof(\"Setting dry-run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, true); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\n\t\t\t// If the CQ has finished, determine if it was a success\n\t\t\t// or failure.\n\t\t\trollDone, err := r.isDryRunDone(currentRoll)\n\t\t\tif err != nil {\n\t\t\t\treturn STATUS_ERROR, err\n\t\t\t}\n\t\t\tif rollDone {\n\t\t\t\tresult := autoroll.ROLL_RESULT_DRY_RUN_FAILURE\n\t\t\t\tstatus := STATUS_DRY_RUN_FAILURE\n\t\t\t\trollSuccessful, err := r.isDryRunSuccessful(currentRoll)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rollSuccessful {\n\t\t\t\t\tresult = autoroll.ROLL_RESULT_DRY_RUN_SUCCESS\n\t\t\t\t\tstatus = STATUS_DRY_RUN_SUCCESS\n\t\t\t\t}\n\t\t\t\tsklog.Infof(\"Dry run is finished: %v\", currentRoll)\n\t\t\t\tif currentRoll.RollingTo != r.rm.ChildHead() {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, result, fmt.Sprintf(\"Repo has passed %s; will open a new dry run.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else if currentRoll.Result != result {\n\t\t\t\t\t// The dry run just finished. Set its result.\n\t\t\t\t\tif result == autoroll.ROLL_RESULT_DRY_RUN_FAILURE {\n\t\t\t\t\t\tif err := r.closeIssue(currentRoll, result, \"Dry run failed. Closing, will open another.\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif err := r.addIssueComment(currentRoll, \"Dry run finished successfully; leaving open in case we want to land\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcurrentRoll.Result = result\n\t\t\t\t\t\tif err := r.recent.Update(currentRoll); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\treturn status, nil\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// The dry run is finished but still good. Leave it open.\n\t\t\t\t\tsklog.Infof(\"Dry run is finished and still good.\")\n\t\t\t\t\treturn status, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tsklog.Infof(\"Dry run still in progress.\")\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\t\t} else {\n\t\t\tif currentRoll.CommitQueueDryRun {\n\t\t\t\tsklog.Infof(\"Unsetting dry run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\t\t\t// If we're stopped, close the issue.\n\t\t\t\t// Respect the previous result of the roll.\n\t\t\t\tif err := r.closeIssue(currentRoll, r.makeRollResult(currentRoll), \"AutoRoller is stopped; closing the active roll.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else if !currentRoll.CommitQueue {\n\t\t\t\t// If the CQ failed, close the issue.\n\t\t\t\t// Special case: if the current roll was a dry run which succeeded, land it.\n\t\t\t\tif currentRoll.Result == autoroll.ROLL_RESULT_DRY_RUN_SUCCESS {\n\t\t\t\t\tsklog.Infof(\"Dry run succeeded. Attempting to land.\")\n\t\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t} else {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Commit queue failed; closing this roll.\"); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if time.Since(currentRoll.Modified) > 24*time.Hour {\n\t\t\t\t// If the roll has been open too long, close the issue.\n\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Roll has been open for over 24 hours; closing.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rolledPast {\n\t\t\t\t\t// If we've already rolled past the target revision, close the issue\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, fmt.Sprintf(\"Already rolled past %s; closing this roll.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// Current roll is still good.\n\t\t\t\t\tsklog.Infof(\"Roll is still active (%d): %s\", currentRoll.Issue, currentRoll.Subject)\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// If we're stopped, exit.\n\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\tsklog.Infof(\"Roller is stopped; not opening new rolls.\")\n\t\treturn STATUS_STOPPED, nil\n\t}\n\n\t// If we're up-to-date, exit.\n\tchildHead := r.rm.ChildHead()\n\tif r.rm.LastRollRev() == childHead {\n\t\tsklog.Infof(\"Repo is up-to-date.\")\n\t\treturn STATUS_UP_TO_DATE, nil\n\t}\n\n\t// Create a new roll.\n\tif r.attemptCounter.Get() >= ROLL_ATTEMPT_THROTTLE_NUM {\n\t\treturn STATUS_THROTTLED, nil\n\t}\n\tr.attemptCounter.Inc()\n\tdryRun := r.isMode(autoroll_modes.MODE_DRY_RUN)\n\tuploadedNum, err := r.rm.CreateNewRoll(r.strategy, r.GetEmails(), r.cqExtraTrybots, dryRun)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to upload a new roll: %s\", err)\n\t}\n\tuploaded, err := r.retrieveRoll(uploadedNum)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to retrieve uploaded roll: %s\", err)\n\t}\n\tif err := r.recent.Add(uploaded); err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to insert uploaded roll into database: %s\", err)\n\t}\n\n\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t}\n\treturn STATUS_IN_PROGRESS, nil\n}", "func (c *Client) GetBalance() (*BalanceSheet, error) {\n\turl := fmt.Sprintf(\"%v%v\", c.Host, totalOwedURL())\n\tledgerRequest, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not make ledger request: %v\", err)\n\t}\n\tledgerResponse, err := c.Do(ledgerRequest)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error making request: %v\", err)\n\t}\n\tif ledgerResponse.StatusCode >= 400 {\n\t\treturn nil, fmt.Errorf(\"bad response code from ledger request: %v\", ledgerResponse.StatusCode)\n\t}\n\tdefer ledgerResponse.Body.Close()\n\tledgerBody, err := ioutil.ReadAll(ledgerResponse.Body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to read ledger response body: %v\", err)\n\t}\n\tbalance, err := balanceFromHTML(ledgerBody)\n\tif err != nil {\n\t\tfmt.Println(\" == == == Ledger Body == == ==\")\n\t\tfmt.Println(string(ledgerBody))\n\t\tfmt.Println(\" == == == == == == == == == ==\")\n\t\treturn nil, err\n\t}\n\treturn balance, nil\n}", "func RollDiceRedo(dice string) (string, int, int, error) {\n\terr := clearFlags()\n\tif err != nil {\n\t\tfmt.Println(\"error clearing flags\")\n\t}\n\tcommand, flags, err := sFlags.CreateFlags(dice)\n\tparseFlags(flags)\n\trollType, diceSlice, err := inputProofer(command)\n\terr = assignMeaningToDiceSlice(diceSlice, rollType)\n\tif err != nil {\n\t\treturn \"\", 0, 0, err\n\t}\n\tif numDice > 1000 {\n\t\treturn \"That's too many dice\", 0, 0, nil\n\t}\n\tif typeDice > 1000 {\n\t\treturn \"Golf balls are not dice\", 0, 0, nil\n\t}\n\tif numDice+typeDice > 1000 {\n\t\treturn \"Fuck you\", 0, 0, nil\n\t}\n\tvar results []int\n\tresults, total, successes, err = rollDice(rollType, roundFlag)\n\tif err != nil {\n\t\treturn \"DAMNIT\", 0, 0, err\n\t}\n\treturn formatResults(results, rollType), successes, total, err\n}" ]
[ "0.72072464", "0.5197747", "0.4978265", "0.48919395", "0.47729144", "0.47611463", "0.4735145", "0.47127548", "0.46137926", "0.46028164", "0.45983827", "0.45616758", "0.45554212", "0.45538542", "0.4522338", "0.45083523", "0.44950864", "0.4485159", "0.44710016", "0.44660395", "0.44630724", "0.4438446", "0.44329363", "0.4396209", "0.43895534", "0.43634123", "0.43606344", "0.43497103", "0.43257916", "0.4322977", "0.43145168", "0.43144128", "0.42939237", "0.42881536", "0.42855638", "0.4280072", "0.424713", "0.4224333", "0.4212577", "0.42041552", "0.4191452", "0.41857147", "0.41818854", "0.41784057", "0.4178005", "0.41770002", "0.41599956", "0.41595504", "0.41571224", "0.41568923", "0.41493592", "0.41488227", "0.41385663", "0.40938216", "0.4093553", "0.4079964", "0.40788153", "0.40771744", "0.40745473", "0.4068037", "0.40679762", "0.40677834", "0.40535995", "0.4049765", "0.40457392", "0.40174863", "0.40134308", "0.40051085", "0.4003455", "0.40033218", "0.400008", "0.39987698", "0.39904746", "0.39875075", "0.39857557", "0.39845523", "0.39806423", "0.39656338", "0.3963572", "0.39627045", "0.3960003", "0.39551234", "0.39517114", "0.39479136", "0.3946876", "0.39434576", "0.39286244", "0.3928495", "0.39267468", "0.392483", "0.391293", "0.39120075", "0.39115724", "0.39115724", "0.39115724", "0.39115724", "0.39083514", "0.39082462", "0.3905997", "0.39044416" ]
0.6336865
1
doAutoRoll is the primary method of the AutoRoll Bot. It runs on a timer, updates checkouts, manages active roll CLs, and uploads new rolls. It sets the status of the bot which may be read by users.
func (r *AutoRoller) doAutoRoll() error { status, lastError := r.doAutoRollInner() lastErrorStr := "" if lastError != nil { lastErrorStr = lastError.Error() } gerritUrl := r.gerrit.Url(0) // Update status information. if err := r.status.set(&AutoRollStatus{ CurrentRoll: r.recent.CurrentRoll(), Error: lastErrorStr, GerritUrl: gerritUrl, LastRoll: r.recent.LastRoll(), LastRollRev: r.rm.LastRollRev(), Mode: r.modeHistory.CurrentMode(), Recent: r.recent.GetRecentRolls(), Status: status, }); err != nil { return err } return lastError }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) doAutoRollInner() (string, error) {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\t// Get updated info about the current roll.\n\tif err := r.updateCurrentRoll(); err != nil {\n\t\treturn STATUS_ERROR, err\n\t}\n\n\t// There's a currently-active roll. Determine whether or not it's still good.\n\t// If so, leave it open and exit. If not, close it so that we can open another.\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tsklog.Infof(\"Found current roll: %s\", r.issueUrl(currentRoll.Issue))\n\n\t\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\t\t// If we have a normal (non-dry-run) roll running,\n\t\t\t// switch it to a dry run.\n\t\t\tif currentRoll.CommitQueue {\n\t\t\t\tsklog.Infof(\"Setting dry-run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, true); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\n\t\t\t// If the CQ has finished, determine if it was a success\n\t\t\t// or failure.\n\t\t\trollDone, err := r.isDryRunDone(currentRoll)\n\t\t\tif err != nil {\n\t\t\t\treturn STATUS_ERROR, err\n\t\t\t}\n\t\t\tif rollDone {\n\t\t\t\tresult := autoroll.ROLL_RESULT_DRY_RUN_FAILURE\n\t\t\t\tstatus := STATUS_DRY_RUN_FAILURE\n\t\t\t\trollSuccessful, err := r.isDryRunSuccessful(currentRoll)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rollSuccessful {\n\t\t\t\t\tresult = autoroll.ROLL_RESULT_DRY_RUN_SUCCESS\n\t\t\t\t\tstatus = STATUS_DRY_RUN_SUCCESS\n\t\t\t\t}\n\t\t\t\tsklog.Infof(\"Dry run is finished: %v\", currentRoll)\n\t\t\t\tif currentRoll.RollingTo != r.rm.ChildHead() {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, result, fmt.Sprintf(\"Repo has passed %s; will open a new dry run.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else if currentRoll.Result != result {\n\t\t\t\t\t// The dry run just finished. Set its result.\n\t\t\t\t\tif result == autoroll.ROLL_RESULT_DRY_RUN_FAILURE {\n\t\t\t\t\t\tif err := r.closeIssue(currentRoll, result, \"Dry run failed. Closing, will open another.\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif err := r.addIssueComment(currentRoll, \"Dry run finished successfully; leaving open in case we want to land\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcurrentRoll.Result = result\n\t\t\t\t\t\tif err := r.recent.Update(currentRoll); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\treturn status, nil\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// The dry run is finished but still good. Leave it open.\n\t\t\t\t\tsklog.Infof(\"Dry run is finished and still good.\")\n\t\t\t\t\treturn status, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tsklog.Infof(\"Dry run still in progress.\")\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\t\t} else {\n\t\t\tif currentRoll.CommitQueueDryRun {\n\t\t\t\tsklog.Infof(\"Unsetting dry run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\t\t\t// If we're stopped, close the issue.\n\t\t\t\t// Respect the previous result of the roll.\n\t\t\t\tif err := r.closeIssue(currentRoll, r.makeRollResult(currentRoll), \"AutoRoller is stopped; closing the active roll.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else if !currentRoll.CommitQueue {\n\t\t\t\t// If the CQ failed, close the issue.\n\t\t\t\t// Special case: if the current roll was a dry run which succeeded, land it.\n\t\t\t\tif currentRoll.Result == autoroll.ROLL_RESULT_DRY_RUN_SUCCESS {\n\t\t\t\t\tsklog.Infof(\"Dry run succeeded. Attempting to land.\")\n\t\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t} else {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Commit queue failed; closing this roll.\"); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if time.Since(currentRoll.Modified) > 24*time.Hour {\n\t\t\t\t// If the roll has been open too long, close the issue.\n\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Roll has been open for over 24 hours; closing.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rolledPast {\n\t\t\t\t\t// If we've already rolled past the target revision, close the issue\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, fmt.Sprintf(\"Already rolled past %s; closing this roll.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// Current roll is still good.\n\t\t\t\t\tsklog.Infof(\"Roll is still active (%d): %s\", currentRoll.Issue, currentRoll.Subject)\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// If we're stopped, exit.\n\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\tsklog.Infof(\"Roller is stopped; not opening new rolls.\")\n\t\treturn STATUS_STOPPED, nil\n\t}\n\n\t// If we're up-to-date, exit.\n\tchildHead := r.rm.ChildHead()\n\tif r.rm.LastRollRev() == childHead {\n\t\tsklog.Infof(\"Repo is up-to-date.\")\n\t\treturn STATUS_UP_TO_DATE, nil\n\t}\n\n\t// Create a new roll.\n\tif r.attemptCounter.Get() >= ROLL_ATTEMPT_THROTTLE_NUM {\n\t\treturn STATUS_THROTTLED, nil\n\t}\n\tr.attemptCounter.Inc()\n\tdryRun := r.isMode(autoroll_modes.MODE_DRY_RUN)\n\tuploadedNum, err := r.rm.CreateNewRoll(r.strategy, r.GetEmails(), r.cqExtraTrybots, dryRun)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to upload a new roll: %s\", err)\n\t}\n\tuploaded, err := r.retrieveRoll(uploadedNum)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to retrieve uploaded roll: %s\", err)\n\t}\n\tif err := r.recent.Add(uploaded); err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to insert uploaded roll into database: %s\", err)\n\t}\n\n\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t}\n\treturn STATUS_IN_PROGRESS, nil\n}", "func (r *AutoRoller) Tick() error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence()\n\n\t// Update the status information.\n\tlastErrorStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrorStr = lastErr.Error()\n\t}\n\trecent := r.recent.GetRecentRolls()\n\tnumFailures := 0\n\tfor _, roll := range recent {\n\t\tif roll.Failed() {\n\t\t\tnumFailures++\n\t\t} else if roll.Succeeded() {\n\t\t\tbreak\n\t\t}\n\t}\n\tsklog.Infof(\"Updating status (%d)\", r.rm.CommitsNotRolled())\n\tif err := r.status.Set(&AutoRollStatus{\n\t\tAutoRollMiniStatus: AutoRollMiniStatus{\n\t\t\tNumFailedRolls: numFailures,\n\t\t\tNumNotRolledCommits: r.rm.CommitsNotRolled(),\n\t\t},\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tFullHistoryUrl: r.gerrit.Url(0) + \"/q/owner:\" + r.GetUser(),\n\t\tIssueUrlBase: r.gerrit.Url(0) + \"/c/\",\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t}); err != nil {\n\t\treturn err\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn lastErr\n}", "func NewAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, tickFrequency, repoFrequency time.Duration, depot_tools string, rollIntoAndroid bool, strategy string) (*AutoRoller, error) {\n\tvar err error\n\tvar rm repo_manager.RepoManager\n\tif rollIntoAndroid {\n\t\trm, err = repo_manager.NewAndroidRepoManager(workdir, parentBranch, childPath, childBranch, repoFrequency, gerrit)\n\t} else {\n\t\trm, err = repo_manager.NewDEPSRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, repoFrequency, depot_tools, gerrit)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := autoroll_modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tattemptCounter: util.NewAutoDecrementCounter(ROLL_ATTEMPT_THROTTLE_TIME),\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tincludeCommitLog: true,\n\t\tliveness: metrics2.NewLiveness(\"last-autoroll-landed\", map[string]string{\"child-path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\trm: rm,\n\t\tstatus: &autoRollStatusCache{},\n\t\tstrategy: strategy,\n\t\trollIntoAndroid: rollIntoAndroid,\n\t}\n\n\t// Cycle once to fill out the current status.\n\tif err := arb.doAutoRoll(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tgo func() {\n\t\tfor range time.Tick(tickFrequency) {\n\t\t\tutil.LogErr(arb.doAutoRoll())\n\t\t}\n\t}()\n\n\treturn arb, nil\n}", "func (r *AutoRoller) Tick(ctx context.Context) error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\n\t// Update the config vars.\n\tif err := r.reg.Update(ctx); err != nil {\n\t\tsklog.Errorf(\"Failed to update config registry; continuing, but config may be out of date: %s\", err)\n\t}\n\n\t// Determine if we should unthrottle.\n\tshouldUnthrottle, err := r.throttle.Get(ctx, r.roller)\n\tif err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to determine whether we should unthrottle\")\n\t}\n\tif shouldUnthrottle {\n\t\tif err := r.unthrottle(ctx); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to unthrottle\")\n\t\t}\n\t\tif err := r.throttle.Reset(ctx, r.roller); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to reset unthrottle counter\")\n\t\t}\n\t}\n\n\t// Update modes and strategies.\n\tif err := r.modeHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update mode history\")\n\t}\n\toldStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif err := r.strategyHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update strategy history\")\n\t}\n\tnewStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif oldStrategy != newStrategy {\n\t\tstrat, err := strategy.GetNextRollStrategy(newStrategy)\n\t\tif err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to get next roll strategy\")\n\t\t}\n\t\tr.strategyMtx.Lock()\n\t\tr.strategy = strat\n\t\tr.strategyMtx.Unlock()\n\t}\n\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence(ctx)\n\tlastErrStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrStr = lastErr.Error()\n\t}\n\n\t// Update the status information.\n\tif err := r.updateStatus(ctx, true, lastErrStr); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update status\")\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn skerr.Wrapf(lastErr, \"Failed state transition sequence\")\n}", "func (r *AutoRoller) Start(ctx context.Context, tickFrequency time.Duration) {\n\tsklog.Infof(\"Starting autoroller.\")\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(tickFrequency, func(_ context.Context) {\n\t\t// Explicitly ignore the passed-in context; this allows us to\n\t\t// continue running even if the context is canceled, which helps\n\t\t// to prevent errors due to interrupted syncs, etc.\n\t\tctx := context.Background()\n\t\tif err := r.Tick(ctx); err != nil {\n\t\t\t// Hack: we frequently get failures from GoB which trigger error-rate alerts.\n\t\t\t// These alerts are noise and sometimes hide real failures. If the error is\n\t\t\t// due to a sync failure, log it as a warning instead of an error. We'll rely\n\t\t\t// on the liveness alert in the case where we see persistent sync failures.\n\t\t\tif isSyncError(err) {\n\t\t\t\tsklog.Warningf(\"Failed to run autoroll: %s\", err)\n\t\t\t} else {\n\t\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t}, nil)\n\n\t// Update the current reviewers in a loop.\n\tlvReviewers := metrics2.NewLiveness(\"last_successful_reviewers_retrieval\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(30*time.Minute, func(ctx context.Context) {\n\t\temails := GetReviewers(r.client, r.cfg.RollerName, r.cfg.Reviewer, r.cfg.ReviewerBackup)\n\t\tr.emailsMtx.Lock()\n\t\tdefer r.emailsMtx.Unlock()\n\t\tr.emails = emails\n\n\t\tconfigCopies := replaceReviewersPlaceholder(r.cfg.Notifiers, emails)\n\t\tif err := r.notifier.ReloadConfigs(ctx, configCopies); err != nil {\n\t\t\tsklog.Errorf(\"Failed to reload configs: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tlvReviewers.Reset()\n\t}, nil)\n\n\t// Handle requests for manual rolls.\n\tif r.cfg.SupportsManualRolls {\n\t\tlvManualRolls := metrics2.NewLiveness(\"last_successful_manual_roll_check\", map[string]string{\"roller\": r.roller})\n\t\tcleanup.Repeat(time.Minute, func(_ context.Context) {\n\t\t\t// Explicitly ignore the passed-in context; this allows\n\t\t\t// us to continue handling manual rolls even if the\n\t\t\t// context is canceled, which helps to prevent errors\n\t\t\t// due to interrupted syncs, etc.\n\t\t\tctx := context.Background()\n\t\t\tif err := r.handleManualRolls(ctx); err != nil {\n\t\t\t\tsklog.Error(err)\n\t\t\t} else {\n\t\t\t\tlvManualRolls.Reset()\n\t\t\t}\n\t\t}, nil)\n\t}\n}", "func (r *AutoRoller) handleManualRolls(ctx context.Context) error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tif r.GetMode() == modes.ModeOffline {\n\t\treturn nil\n\t}\n\n\tsklog.Infof(\"Searching manual roll requests for %s\", r.cfg.RollerName)\n\treqs, err := r.manualRollDB.GetIncomplete(r.cfg.RollerName)\n\tif err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to get incomplete rolls\")\n\t}\n\tsklog.Infof(\"Found %d requests.\", len(reqs))\n\tfor _, req := range reqs {\n\t\tvar issue *autoroll.AutoRollIssue\n\t\tvar to *revision.Revision\n\t\tif req.NoResolveRevision {\n\t\t\tto = &revision.Revision{Id: req.Revision}\n\t\t} else {\n\t\t\tto, err = r.getRevision(ctx, req.Revision)\n\t\t\tif err != nil {\n\t\t\t\terr := skerr.Wrapf(err, \"failed to resolve revision %q\", req.Revision)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif req.ExternalChangeId != \"\" {\n\t\t\tto.ExternalChangeId = req.ExternalChangeId\n\t\t}\n\t\tif req.Status == manual.STATUS_PENDING {\n\t\t\t// Avoid creating rolls to the current revision.\n\t\t\tfrom := r.GetCurrentRev()\n\t\t\tif to.Id == from.Id {\n\t\t\t\terr := skerr.Fmt(\"Already at revision %q\", from.Id)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\temails := []string{}\n\t\t\tif !req.NoEmail {\n\t\t\t\temails = r.GetEmails()\n\t\t\t\tif !util.In(req.Requester, emails) {\n\t\t\t\t\temails = append(emails, req.Requester)\n\t\t\t\t}\n\t\t\t}\n\t\t\tvar err error\n\t\t\tsklog.Infof(\"Creating manual roll to %s as requested by %s...\", req.Revision, req.Requester)\n\n\t\t\tissue, err = r.createNewRoll(ctx, from, to, emails, req.DryRun, req.Canary, req.Requester)\n\t\t\tif err != nil {\n\t\t\t\terr := skerr.Wrapf(err, \"failed to create manual roll for %s\", req.Id)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t} else if req.Status == manual.STATUS_STARTED {\n\t\t\tsplit := strings.Split(req.Url, \"/\")\n\t\t\ti, err := strconv.Atoi(split[len(split)-1])\n\t\t\tif err != nil {\n\t\t\t\treturn skerr.Wrapf(err, \"Failed to parse issue number from %s for %s: %s\", req.Url, req.Id, err)\n\t\t\t}\n\t\t\tissue = &autoroll.AutoRollIssue{\n\t\t\t\tRollingTo: req.Revision,\n\t\t\t\tIsDryRun: req.DryRun,\n\t\t\t\tIssue: int64(i),\n\t\t\t}\n\t\t} else {\n\t\t\tsklog.Errorf(\"Found manual roll request %s in unknown status %q\", req.Id, req.Status)\n\t\t\tcontinue\n\t\t}\n\t\tsklog.Infof(\"Getting status for manual roll # %d\", issue.Issue)\n\t\troll, err := r.retrieveRoll(ctx, issue, to)\n\t\tif err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to retrieve manual roll %s: %s\", req.Id, err)\n\t\t}\n\t\treq.Status = manual.STATUS_STARTED\n\t\treq.Url = roll.IssueURL()\n\n\t\tif req.DryRun {\n\t\t\tif roll.IsDryRunFinished() {\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\tif roll.IsDryRunSuccess() {\n\t\t\t\t\treq.Result = manual.RESULT_SUCCESS\n\t\t\t\t} else {\n\t\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\t}\n\t\t\t}\n\t\t} else if roll.IsFinished() {\n\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\tif roll.IsSuccess() {\n\t\t\t\treq.Result = manual.RESULT_SUCCESS\n\t\t\t} else {\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t}\n\t\t}\n\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t}\n\t}\n\treturn nil\n}", "func NewAutoRoller(ctx context.Context, c *config.Config, emailer emailclient.Client, chatBotConfigReader chatbot.ConfigReader, g *gerrit.Gerrit, githubClient *github.GitHub, workdir, recipesCfgFile, serverURL string, gcsClient gcs.GCSClient, client *http.Client, rollerName string, local bool, statusDB status.DB, manualRollDB manual.DB) (*AutoRoller, error) {\n\t// Validation and setup.\n\tif err := c.Validate(); err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to validate config\")\n\t}\n\tvar cr codereview.CodeReview\n\tvar err error\n\tif c.GetGerrit() != nil {\n\t\tcr, err = codereview.NewGerrit(c.GetGerrit(), g, client)\n\t} else if c.GetGithub() != nil {\n\t\tcr, err = codereview.NewGitHub(c.GetGithub(), githubClient)\n\t} else {\n\t\treturn nil, skerr.Fmt(\"Either GitHub or Gerrit is required.\")\n\t}\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to initialize code review\")\n\t}\n\treg, err := config_vars.NewRegistry(ctx, chrome_branch.NewClient(client))\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create config var registry\")\n\t}\n\n\t// Create the RepoManager.\n\trm, err := repo_manager.New(ctx, c.GetRepoManagerConfig(), reg, workdir, rollerName, recipesCfgFile, serverURL, c.ServiceAccount, client, cr, c.IsInternal, local)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\n\tsklog.Info(\"Creating strategy history.\")\n\tsh, err := strategy.NewDatastoreStrategyHistory(ctx, rollerName, c.ValidStrategies())\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create strategy history\")\n\t}\n\tcurrentStrategy := sh.CurrentStrategy()\n\tif currentStrategy == nil {\n\t\t// If there's no history, set the initial strategy.\n\t\tsklog.Infof(\"Setting initial strategy for %s to %q\", rollerName, c.DefaultStrategy())\n\t\tif err := sh.Add(ctx, c.DefaultStrategy(), \"AutoRoll Bot\", \"Setting initial strategy.\"); err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to set initial strategy\")\n\t\t}\n\t\tcurrentStrategy = sh.CurrentStrategy()\n\t}\n\tsklog.Info(\"Setting strategy.\")\n\tstrat, err := strategy.GetNextRollStrategy(currentStrategy.Strategy)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to get next roll strategy\")\n\t}\n\n\tsklog.Info(\"Running repo_manager.Update()\")\n\tlastRollRev, tipRev, notRolledRevs, err := rm.Update(ctx)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed initial repo manager update\")\n\t}\n\tnextRollRev := strat.GetNextRollRev(notRolledRevs)\n\tif nextRollRev == nil {\n\t\tnextRollRev = lastRollRev\n\t}\n\n\tsklog.Info(\"Creating roll history\")\n\trecent, err := recent_rolls.NewRecentRolls(ctx, recent_rolls.NewDatastoreRollsDB(ctx), rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create recent rolls DB\")\n\t}\n\tsklog.Info(\"Creating mode history\")\n\tmh, err := modes.NewDatastoreModeHistory(ctx, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create mode history\")\n\t}\n\tif mh.CurrentMode() == nil {\n\t\tsklog.Info(\"Setting initial mode.\")\n\t\tif err := mh.Add(ctx, modes.ModeRunning, \"AutoRoll Bot\", \"Setting initial mode.\"); err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to set initial mode\")\n\t\t}\n\t}\n\n\t// Throttling counters.\n\tsklog.Info(\"Creating throttlers\")\n\tsafetyThrottleCfg := config.DefaultSafetyThrottleConfig\n\tif c.SafetyThrottle != nil {\n\t\tsafetyThrottleCfg = c.SafetyThrottle\n\t}\n\tsafetyThrottleDuration, err := human.ParseDuration(safetyThrottleCfg.TimeWindow)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to parse safety throttle time window\")\n\t}\n\tsafetyThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/attempt_counter\", safetyThrottleDuration, int64(safetyThrottleCfg.AttemptCount))\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create safety throttler\")\n\t}\n\n\tfailureThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/fail_counter\", time.Hour, 1)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create failure throttler\")\n\t}\n\n\tvar rollCooldown time.Duration\n\tif c.RollCooldown != \"\" {\n\t\trollCooldown, err = human.ParseDuration(c.RollCooldown)\n\t}\n\tsuccessThrottle, err := state_machine.NewThrottler(ctx, gcsClient, rollerName+\"/success_counter\", rollCooldown, 1)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create success throttler\")\n\t}\n\tsklog.Info(\"Getting reviewers\")\n\temails := GetReviewers(client, c.RollerName, c.Reviewer, c.ReviewerBackup)\n\tsklog.Info(\"Creating notifier\")\n\tconfigCopies := replaceReviewersPlaceholder(c.Notifiers, emails)\n\tn, err := arb_notifier.New(ctx, c.ChildDisplayName, c.ParentDisplayName, serverURL, client, emailer, chatBotConfigReader, configCopies)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create notifier\")\n\t}\n\tsklog.Info(\"Creating status cache.\")\n\tstatusCache, err := status.NewCache(ctx, statusDB, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create status cache\")\n\t}\n\tsklog.Info(\"Creating TimeWindow.\")\n\tvar tw *time_window.TimeWindow\n\tif c.TimeWindow != \"\" {\n\t\ttw, err = time_window.Parse(c.TimeWindow)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to create TimeWindow\")\n\t\t}\n\t}\n\tcommitMsgBuilder, err := commit_msg.NewBuilder(c.CommitMsg, reg, c.ChildDisplayName, c.ParentDisplayName, serverURL, c.ChildBugLink, c.ParentBugLink, c.TransitiveDeps)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tarb := &AutoRoller{\n\t\tcfg: c,\n\t\tclient: client,\n\t\tcodereview: cr,\n\t\tcommitMsgBuilder: commitMsgBuilder,\n\t\temails: emails,\n\t\tfailureThrottle: failureThrottle,\n\t\tlastRollRev: lastRollRev,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"roller\": c.RollerName}),\n\t\tmanualRollDB: manualRollDB,\n\t\tmodeHistory: mh,\n\t\tnextRollRev: nextRollRev,\n\t\tnotifier: n,\n\t\tnotRolledRevs: notRolledRevs,\n\t\trecent: recent,\n\t\treg: reg,\n\t\trm: rm,\n\t\troller: rollerName,\n\t\trollUploadAttempts: metrics2.GetCounter(\"autoroll_cl_upload_attempts\", map[string]string{\"roller\": c.RollerName}),\n\t\trollUploadFailures: metrics2.GetCounter(\"autoroll_cl_upload_failures\", map[string]string{\"roller\": c.RollerName}),\n\t\tsafetyThrottle: safetyThrottle,\n\t\tserverURL: serverURL,\n\t\treviewers: c.Reviewer,\n\t\treviewersBackup: c.ReviewerBackup,\n\t\tstatus: statusCache,\n\t\tstrategy: strat,\n\t\tstrategyHistory: sh,\n\t\tsuccessThrottle: successThrottle,\n\t\tthrottle: unthrottle.NewDatastore(ctx),\n\t\ttimeWindow: tw,\n\t\ttipRev: tipRev,\n\t}\n\tsklog.Info(\"Creating state machine\")\n\tsm, err := state_machine.New(ctx, arb, n, gcsClient, rollerName)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to create state machine\")\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\trollingTo, err := arb.getRevision(ctx, current.RollingTo)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrap(err)\n\t\t}\n\t\troll, err := arb.retrieveRoll(ctx, current, rollingTo)\n\t\tif err != nil {\n\t\t\treturn nil, skerr.Wrapf(err, \"Failed to retrieve current roll\")\n\t\t}\n\t\tif err := roll.InsertIntoDB(ctx); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\tsklog.Info(\"Done creating autoroller\")\n\treturn arb, nil\n}", "func (r *AutoRoller) Start(tickFrequency, repoFrequency time.Duration, ctx context.Context) {\n\tsklog.Infof(\"Starting autoroller.\")\n\trepo_manager.Start(r.rm, repoFrequency, ctx)\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\")\n\tgo util.RepeatCtx(tickFrequency, ctx, func() {\n\t\tif err := r.Tick(); err != nil {\n\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\tutil.LogErr(r.recent.Close())\n\t\t\t\tutil.LogErr(r.modeHistory.Close())\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\t}()\n}", "func (r *AutoRoller) rollFinished(ctx context.Context, justFinished codereview.RollImpl) error {\n\trecent := r.recent.GetRecentRolls()\n\t// Sanity check: pop any rolls which occurred after the one which just\n\t// finished.\n\tidx := -1\n\tvar currentRoll *autoroll.AutoRollIssue\n\tfor i, roll := range recent {\n\t\tissue := fmt.Sprintf(\"%d\", roll.Issue)\n\t\tif issue == justFinished.IssueID() {\n\t\t\tidx = i\n\t\t\tcurrentRoll = roll\n\t\t\tbreak\n\t\t}\n\t}\n\tif currentRoll == nil {\n\t\treturn skerr.Fmt(\"Unable to find just-finished roll %q in recent list!\", justFinished.IssueID())\n\t}\n\n\t// Feed AutoRoll stats into metrics.\n\tv := int64(0)\n\tif currentRoll.Closed && currentRoll.Committed {\n\t\tv = int64(1)\n\t}\n\tmetrics2.GetInt64Metric(\"autoroll_last_roll_result\", map[string]string{\"roller\": r.cfg.RollerName}).Update(v)\n\n\trecent = recent[idx:]\n\tvar lastRoll *autoroll.AutoRollIssue\n\tif len(recent) > 1 {\n\t\tlastRoll = recent[1]\n\t} else {\n\t\t// If there are no other rolls, then the below alerts do not apply.\n\t\treturn nil\n\t}\n\n\tissueURL := fmt.Sprintf(\"%s%d\", r.codereview.GetIssueUrlBase(), currentRoll.Issue)\n\n\t// Send notifications if this roll had a different result from the last\n\t// roll, ie. success -> failure or failure -> success.\n\tcurrentSuccess := util.In(currentRoll.Result, autoroll.SUCCESS_RESULTS)\n\tlastSuccess := util.In(lastRoll.Result, autoroll.SUCCESS_RESULTS)\n\tif lastRoll != nil {\n\t\tif currentSuccess && !lastSuccess {\n\t\t\tr.notifier.SendNewSuccess(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t} else if !currentSuccess && lastSuccess {\n\t\t\tr.notifier.SendNewFailure(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t}\n\t}\n\n\t// Send a notification if the last N rolls failed in a row.\n\tnFailed := 0\n\t// recent is in reverse chronological order.\n\tfor _, roll := range recent {\n\t\tif util.In(roll.Result, autoroll.SUCCESS_RESULTS) {\n\t\t\tbreak\n\t\t} else {\n\t\t\tnFailed++\n\t\t}\n\t}\n\tif nFailed == notifyIfLastNFailed {\n\t\tr.notifier.SendLastNFailed(ctx, notifyIfLastNFailed, issueURL)\n\t}\n\n\treturn nil\n}", "func LoadAutoRollData(dbClient *influxdb.Client, workdir string) {\n\trollCheckoutsDir := path.Join(workdir, \"autoroll_git\")\n\tskiaRepo, err := gitinfo.CloneOrUpdate(SKIA_REPO, path.Join(rollCheckoutsDir, \"skia\"), false)\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to check out skia: %s\", err)\n\t\treturn\n\t}\n\tchromiumRepo, err := gitinfo.CloneOrUpdate(CHROMIUM_REPO, path.Join(rollCheckoutsDir, \"chromium\"), false)\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to check out chromium: %s\", err)\n\t\treturn\n\t}\n\n\tfor _ = range time.Tick(time.Minute) {\n\t\ts, err := autoroll.CurrentStatus(skiaRepo, chromiumRepo)\n\t\tif err != nil {\n\t\t\tglog.Error(err)\n\t\t} else {\n\t\t\terr := writeAutoRollDataPoint(dbClient, s)\n\t\t\tif err != nil {\n\t\t\t\tglog.Error(err)\n\t\t\t}\n\t\t}\n\t\tskiaRepo.Update(true, false)\n\t\tchromiumRepo.Update(true, false)\n\t}\n}", "func (r *AutoRoller) updateCurrentRoll() error {\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll == nil {\n\t\treturn nil\n\t}\n\tcurrentResult := currentRoll.Result\n\n\tupdated, err := r.retrieveRoll(currentRoll.Issue)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// We have to rely on data we store for the dry run case.\n\tif !updated.Closed && util.In(currentResult, autoroll.DRY_RUN_RESULTS) {\n\t\tupdated.Result = currentResult\n\t}\n\n\t// If the current roll succeeded, we need to make sure we update the\n\t// repo so that we see the roll commit. This can take some time, so\n\t// we have to repeatedly update until we see the commit.\n\tif updated.Committed {\n\t\tsklog.Infof(\"Roll succeeded (%d); syncing the repo until it lands.\", currentRoll.Issue)\n\t\tfor {\n\t\t\tsklog.Info(\"Syncing...\")\n\t\t\tsklog.Infof(\"Looking for %s\", currentRoll.RollingTo)\n\t\t\tif err := r.rm.ForceUpdate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif rolledPast {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttime.Sleep(10 * time.Second)\n\t\t}\n\t\tr.liveness.Reset()\n\t}\n\treturn r.recent.Update(updated)\n}", "func (r *EndpointsRolloutV1RestClient) AutoUpdateRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (e EndpointsRolloutV1Client) AutoUpdateRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoUpdateRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func newAutoRoller(workdir, childPath, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, rm repo_manager.RepoManager, retrieveRoll func(*AutoRoller, int64) (RollImpl, error)) (*AutoRoller, error) {\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"child_path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\tretrieveRoll: retrieveRoll,\n\t\trm: rm,\n\t\tstatus: &AutoRollStatusCache{},\n\t}\n\tsm, err := state_machine.New(arb, workdir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\troll, err := arb.retrieveRoll(arb, current.Issue)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\treturn arb, nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoWatchRollout(ctx context.Context, options *api.ListWatchOptions) (kvstore.Watcher, error) {\n\tpath := r.instance + makeURIRolloutV1AutoWatchRolloutWatchOper(options)\n\tpath = strings.Replace(path, \"http://\", \"ws://\", 1)\n\tpath = strings.Replace(path, \"https://\", \"wss://\", 1)\n\tparams := apiutils.GetQueryStringFromListWatchOptions(options)\n\tif params != \"\" {\n\t\tpath = path + \"?\" + params\n\t}\n\theader := http.Header{}\n\tr.updateHTTPHeader(ctx, &header)\n\tdialer := websocket.DefaultDialer\n\tdialer.TLSClientConfig = &tls.Config{InsecureSkipVerify: true}\n\tconn, hresp, err := dialer.Dial(path, header)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to connect web socket to [%s](%s)[%+v]\", path, err, hresp)\n\t}\n\tbridgefn := func(lw *listerwatcher.WatcherClient) {\n\t\tfor {\n\t\t\tin := &AutoMsgRolloutWatchHelper{}\n\t\t\terr := conn.ReadJSON(in)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor _, e := range in.Events {\n\t\t\t\tev := kvstore.WatchEvent{\n\t\t\t\t\tType: kvstore.WatchEventType(e.Type),\n\t\t\t\t\tObject: e.Object,\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase lw.OutCh <- &ev:\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tclose(lw.OutCh)\n\t\t\t\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tlw := listerwatcher.NewWatcherClient(nil, bridgefn)\n\tlw.Run()\n\tgo func() {\n\t\t<-ctx.Done()\n\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t}()\n\treturn lw, nil\n}", "func (e EndpointsRolloutV1Server) AutoUpdateRollout(ctx context.Context, in Rollout) (Rollout, error) {\n\tresp, err := e.AutoUpdateRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn Rollout{}, err\n\t}\n\treturn *resp.(*Rollout), nil\n}", "func (r *MockRepoManager) RollerWillUpload(issueNum int64, from, to string, dryRun bool) *gerrit.ChangeInfo {\n\t// Gerrit API only has millisecond precision.\n\tnow := time.Now().UTC().Round(time.Millisecond)\n\tdescription := fmt.Sprintf(`Roll src/third_party/skia/ %s..%s (42 commits).\n\nblah blah\nTBR=some-sheriff\n`, from[:12], to[:12])\n\tr.mockIssueNumber = issueNum\n\trev := &gerrit.Revision{\n\t\tID: \"1\",\n\t\tNumber: 1,\n\t\tCreatedString: now.Format(gerrit.TIME_FORMAT),\n\t\tCreated: now,\n\t}\n\tcqLabel := gerrit.COMMITQUEUE_LABEL_SUBMIT\n\tif dryRun {\n\t\tif r.rollIntoAndroid {\n\t\t\tcqLabel = gerrit.AUTOSUBMIT_LABEL_NONE\n\t\t} else {\n\t\t\tcqLabel = gerrit.COMMITQUEUE_LABEL_DRY_RUN\n\t\t}\n\t}\n\troll := &gerrit.ChangeInfo{\n\t\tCreated: now,\n\t\tCreatedString: now.Format(gerrit.TIME_FORMAT),\n\t\tSubject: description,\n\t\tChangeId: fmt.Sprintf(\"%d\", r.mockIssueNumber),\n\t\tIssue: r.mockIssueNumber,\n\t\tOwner: &gerrit.Owner{\n\t\t\tEmail: \"[email protected]\",\n\t\t},\n\t\tProject: \"skia\",\n\t\tRevisions: map[string]*gerrit.Revision{\n\t\t\t\"1\": rev,\n\t\t},\n\t\tPatchsets: []*gerrit.Revision{rev},\n\t\tUpdated: now,\n\t\tUpdatedString: now.Format(gerrit.TIME_FORMAT),\n\t}\n\tif r.rollIntoAndroid {\n\t\troll.Labels = map[string]*gerrit.LabelEntry{\n\t\t\tgerrit.PRESUBMIT_VERIFIED_LABEL: {\n\t\t\t\tAll: []*gerrit.LabelDetail{},\n\t\t\t},\n\t\t\tgerrit.AUTOSUBMIT_LABEL: {\n\t\t\t\tAll: []*gerrit.LabelDetail{\n\t\t\t\t\t{\n\t\t\t\t\t\tValue: gerrit.AUTOSUBMIT_LABEL_SUBMIT,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t} else {\n\t\troll.Labels = map[string]*gerrit.LabelEntry{\n\t\t\tgerrit.CODEREVIEW_LABEL: {\n\t\t\t\tAll: []*gerrit.LabelDetail{\n\t\t\t\t\t{\n\t\t\t\t\t\tValue: gerrit.CODEREVIEW_LABEL_APPROVE,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tgerrit.COMMITQUEUE_LABEL: {\n\t\t\t\tAll: []*gerrit.LabelDetail{\n\t\t\t\t\t{\n\t\t\t\t\t\tValue: cqLabel,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\treturn roll\n}", "func (e EndpointsRolloutV1Client) AutoUpdateRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\tresp, err := e.AutoUpdateRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutAction{}, err\n\t}\n\treturn resp.(*RolloutAction), nil\n}", "func (e EndpointsRolloutV1Client) AutoWatchRollout(ctx context.Context, in *api.ListWatchOptions) (RolloutV1_AutoWatchRolloutClient, error) {\n\treturn e.Client.AutoWatchRollout(ctx, in)\n}", "func (e EndpointsRolloutV1Server) AutoUpdateRolloutAction(ctx context.Context, in RolloutAction) (RolloutAction, error) {\n\tresp, err := e.AutoUpdateRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutAction{}, err\n\t}\n\treturn *resp.(*RolloutAction), nil\n}", "func (r *AutoRoller) updateStatus(ctx context.Context, replaceLastError bool, lastError string) error {\n\tr.statusMtx.Lock()\n\tdefer r.statusMtx.Unlock()\n\n\trecent := r.recent.GetRecentRolls()\n\n\tif !replaceLastError {\n\t\tlastError = r.status.Get().Error\n\t}\n\n\tfailureThrottledUntil := r.failureThrottle.ThrottledUntil().Unix()\n\tsafetyThrottledUntil := r.safetyThrottle.ThrottledUntil().Unix()\n\tsuccessThrottledUntil := r.successThrottle.ThrottledUntil().Unix()\n\tthrottledUntil := failureThrottledUntil\n\tif safetyThrottledUntil > throttledUntil {\n\t\tthrottledUntil = safetyThrottledUntil\n\t}\n\tif successThrottledUntil > throttledUntil {\n\t\tthrottledUntil = successThrottledUntil\n\t}\n\n\tnotRolledRevs := r.notRolledRevs\n\tnumNotRolled := len(notRolledRevs)\n\tsklog.Infof(\"Updating status (%d revisions behind)\", numNotRolled)\n\tif numNotRolled > maxNotRolledRevs {\n\t\tnotRolledRevs = notRolledRevs[:1]\n\t\tsklog.Warningf(\"Truncating NotRolledRevisions; %d is more than the maximum of %d\", numNotRolled, maxNotRolledRevs)\n\t}\n\tcurrentRollRev := \"\"\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tcurrentRollRev = currentRoll.RollingTo\n\t}\n\tif err := r.status.Set(ctx, r.roller, &status.AutoRollStatus{\n\t\tAutoRollMiniStatus: status.AutoRollMiniStatus{\n\t\t\tCurrentRollRev: currentRollRev,\n\t\t\tLastRollRev: r.lastRollRev.Id,\n\t\t\tMode: r.GetMode(),\n\t\t\tNumFailedRolls: r.recent.NumFailedRolls(),\n\t\t\tNumNotRolledCommits: numNotRolled,\n\t\t\tTimestamp: time.Now().UTC(),\n\t\t\tLastSuccessfulRollTimestamp: r.recent.LastSuccessfulRollTime(),\n\t\t},\n\t\tChildName: r.cfg.ChildDisplayName,\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastError,\n\t\tFullHistoryUrl: r.codereview.GetFullHistoryUrl(),\n\t\tIssueUrlBase: r.codereview.GetIssueUrlBase(),\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tNotRolledRevisions: notRolledRevs,\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t\tThrottledUntil: throttledUntil,\n\t\tValidModes: modes.ValidModes,\n\t\tValidStrategies: r.cfg.ValidStrategies(),\n\t}); err != nil {\n\t\treturn err\n\t}\n\t// Log the current reviewers(s).\n\tsklog.Infof(\"Current reviewers: %v\", r.GetEmails())\n\treturn r.status.Update(ctx)\n}", "func (r *EndpointsRolloutV1RestClient) AutoAddRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (r *EndpointsRolloutV1RestClient) AutoUpdateRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (e EndpointsRolloutV1Server) AutoWatchRollout(in *api.ListWatchOptions, stream RolloutV1_AutoWatchRolloutServer) error {\n\treturn e.watchHandlerRollout(in, stream)\n}", "func (db *diceBae) LetsRoll() error {\n\tif err := db.session.Open(); err != nil {\n\t\treturn fmt.Errorf(\"failed to open Discord session: %v\", err)\n\t}\n\tdefer db.session.Close()\n\tdefer db.logFile.Close()\n\n\tdb.LogInfo(\"I have no dice, but I must roll. Press CTRL-C to exit.\")\n\tsc := make(chan os.Signal, 1)\n\t// Block on this channel until we get a termination signal.\n\tsignal.Notify(sc, syscall.SIGINT, syscall.SIGTERM, os.Interrupt, os.Kill)\n\t<-sc\n\tdb.LogInfo(\"Later dopes.\")\n\treturn nil\n}", "func Set(ctx context.Context, rollerName string, st *AutoRollStatus) error {\n\tbuf := bytes.NewBuffer(nil)\n\tif err := gob.NewEncoder(buf).Encode(st); err != nil {\n\t\treturn err\n\t}\n\tw := &DsStatusWrapper{\n\t\tData: buf.Bytes(),\n\t\tRoller: rollerName,\n\t}\n\t_, err := ds.DS.RunInTransaction(ctx, func(tx *datastore.Transaction) error {\n\t\t_, err := tx.Put(key(rollerName), w)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Optionally export the mini version of the internal roller's status\n\t// to the external datastore.\n\tif util.In(rollerName, EXPORT_WHITELIST) {\n\t\texportStatus := &AutoRollStatus{\n\t\t\tAutoRollMiniStatus: st.AutoRollMiniStatus,\n\t\t}\n\t\tbuf := bytes.NewBuffer(nil)\n\t\tif err := gob.NewEncoder(buf).Encode(exportStatus); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tw := &DsStatusWrapper{\n\t\t\tData: buf.Bytes(),\n\t\t\tRoller: rollerName,\n\t\t}\n\t\t_, err := ds.DS.RunInTransaction(ctx, func(tx *datastore.Transaction) error {\n\t\t\tk := key(rollerName)\n\t\t\tk.Namespace = ds.AUTOROLL_NS\n\t\t\tk.Parent.Namespace = ds.AUTOROLL_NS\n\t\t\t_, err := tx.Put(k, w)\n\t\t\treturn err\n\t\t})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoWatchRolloutAction(ctx context.Context, options *api.ListWatchOptions) (kvstore.Watcher, error) {\n\tpath := r.instance + makeURIRolloutV1AutoWatchRolloutActionWatchOper(options)\n\tpath = strings.Replace(path, \"http://\", \"ws://\", 1)\n\tpath = strings.Replace(path, \"https://\", \"wss://\", 1)\n\tparams := apiutils.GetQueryStringFromListWatchOptions(options)\n\tif params != \"\" {\n\t\tpath = path + \"?\" + params\n\t}\n\theader := http.Header{}\n\tr.updateHTTPHeader(ctx, &header)\n\tdialer := websocket.DefaultDialer\n\tdialer.TLSClientConfig = &tls.Config{InsecureSkipVerify: true}\n\tconn, hresp, err := dialer.Dial(path, header)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to connect web socket to [%s](%s)[%+v]\", path, err, hresp)\n\t}\n\tbridgefn := func(lw *listerwatcher.WatcherClient) {\n\t\tfor {\n\t\t\tin := &AutoMsgRolloutActionWatchHelper{}\n\t\t\terr := conn.ReadJSON(in)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor _, e := range in.Events {\n\t\t\t\tev := kvstore.WatchEvent{\n\t\t\t\t\tType: kvstore.WatchEventType(e.Type),\n\t\t\t\t\tObject: e.Object,\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase lw.OutCh <- &ev:\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tclose(lw.OutCh)\n\t\t\t\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tlw := listerwatcher.NewWatcherClient(nil, bridgefn)\n\tlw.Run()\n\tgo func() {\n\t\t<-ctx.Done()\n\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t}()\n\treturn lw, nil\n}", "func Do(retryFunc func() error, config *Config) error {\n\tvar err error\n\tconfig.Logger.Infof(\"Job attempt: %d\", 1)\n\terr = retryFunc()\n\tif err == nil {\n\t\treturn nil\n\t}\n\tfor n := uint(1); n < config.Attempts; n++ {\n\t\tdelayTime := config.Delay * (1 << (n - 1))\n\t\ttime.Sleep((time.Duration)(delayTime) * config.Units)\n\t\tconfig.Logger.Infof(\"Job attempt: %d\", n+1)\n\t\terr = retryFunc()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn err\n}", "func (e EndpointsRolloutV1Client) AutoListRollout(ctx context.Context, in *api.ListWatchOptions) (*RolloutList, error) {\n\tresp, err := e.AutoListRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutList{}, err\n\t}\n\treturn resp.(*RolloutList), nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoListRollout(ctx context.Context, options *api.ListWatchOptions) (*RolloutList, error) {\n\tpath := makeURIRolloutV1AutoListRolloutListOper(options)\n\tif r.bufferId != \"\" {\n\t\tpath = strings.Replace(path, \"/configs\", \"/staging/\"+r.bufferId, 1)\n\t}\n\treq, err := r.getHTTPRequest(ctx, options, \"GET\", path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := r.client.Do(req.WithContext(ctx))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"request failed (%s)\", err)\n\t}\n\tdefer resp.Body.Close()\n\tret, err := decodeHTTPrespRolloutV1AutoListRollout(ctx, resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ret.(*RolloutList), err\n}", "func (r *AutoRoller) createNewRoll(ctx context.Context, from, to *revision.Revision, emails []string, dryRun, canary bool, manualRollRequester string) (rv *autoroll.AutoRollIssue, rvErr error) {\n\t// Track roll CL upload attempts vs failures.\n\tdefer func() {\n\t\tr.rollUploadAttempts.Inc(1)\n\t\tif rvErr == nil {\n\t\t\tr.rollUploadFailures.Reset()\n\t\t} else {\n\t\t\tr.rollUploadFailures.Inc(1)\n\t\t}\n\t}()\n\tr.statusMtx.RLock()\n\tvar revs []*revision.Revision\n\tfound := false\n\tfor _, rev := range r.notRolledRevs {\n\t\tif rev.Id == to.Id {\n\t\t\tfound = true\n\t\t}\n\t\tif found {\n\t\t\trevs = append(revs, rev)\n\t\t}\n\t}\n\tr.statusMtx.RUnlock()\n\n\tcommitMsg, err := r.commitMsgBuilder.Build(from, to, revs, emails, r.cfg.Contacts, canary, manualRollRequester)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tsklog.Infof(\"Creating new roll with commit message: \\n%s\", commitMsg)\n\tissueNum, err := r.rm.CreateNewRoll(ctx, from, to, revs, emails, dryRun, commitMsg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tissue := &autoroll.AutoRollIssue{\n\t\tIsDryRun: dryRun,\n\t\tIssue: issueNum,\n\t\tRollingFrom: from.Id,\n\t\tRollingTo: to.Id,\n\t}\n\treturn issue, nil\n}", "func (e EndpointsRolloutV1Client) AutoAddRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoAddRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (e EndpointsRolloutV1Client) AutoAddRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\tresp, err := e.AutoAddRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutAction{}, err\n\t}\n\treturn resp.(*RolloutAction), nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoAddRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (e EndpointsRolloutV1Server) AutoAddRolloutAction(ctx context.Context, in RolloutAction) (RolloutAction, error) {\n\tresp, err := e.AutoAddRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutAction{}, err\n\t}\n\treturn *resp.(*RolloutAction), nil\n}", "func (r *AutoRoller) setDryRun(issue *autoroll.AutoRollIssue, dryRun bool) error {\n\tinfo, err := issue.ToGerritChangeInfo()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to convert issue to Gerrit ChangeInfo: %s\", err)\n\t}\n\tif dryRun {\n\t\tif err := r.rm.SendToGerritDryRun(info, \"\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tif err := r.rm.SendToGerritCQ(info, \"\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tupdated, err := r.retrieveRoll(issue.Issue)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn r.recent.Update(updated)\n}", "func (r *Roller) onRoller(i interface{}, next func(g interface{})) {\n\tr.Munch(i)\n\tnext(nil)\n}", "func (gs *GameState) ApplyRoll(playerID xid.ID) error {\n\tif gs.Turn.Player == nil || *gs.Turn.Player != playerID {\n\t\treturn fmt.Errorf(\"You're not the rolling player\")\n\t}\n\troll := gs.NextRoll\n\tif roll == nil {\n\t\treturn fmt.Errorf(\"No roll ready\")\n\t}\n\n\tdefer gs.nextRoller()\n\n\tswitch roll.Action {\n\tcase ActionUp, ActionDown:\n\t\treturn gs.applyStockMove(roll)\n\tcase ActionDividend:\n\t\treturn gs.applyDividend(roll)\n\tdefault:\n\t\treturn fmt.Errorf(\"Unknown PlayerAction %v\", roll.Action)\n\t}\n}", "func (e EndpointsRolloutV1Client) AutoWatchRolloutAction(ctx context.Context, in *api.ListWatchOptions) (RolloutV1_AutoWatchRolloutActionClient, error) {\n\treturn e.Client.AutoWatchRolloutAction(ctx, in)\n}", "func (e EndpointsRolloutV1Server) AutoListRollout(ctx context.Context, in api.ListWatchOptions) (RolloutList, error) {\n\tresp, err := e.AutoListRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutList{}, err\n\t}\n\treturn *resp.(*RolloutList), nil\n}", "func (e EndpointsRolloutV1Server) AutoWatchRolloutAction(in *api.ListWatchOptions, stream RolloutV1_AutoWatchRolloutActionServer) error {\n\treturn e.watchHandlerRolloutAction(in, stream)\n}", "func (e EndpointsRolloutV1Server) AutoAddRollout(ctx context.Context, in Rollout) (Rollout, error) {\n\tresp, err := e.AutoAddRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn Rollout{}, err\n\t}\n\treturn *resp.(*Rollout), nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoLabelRollout(ctx context.Context, in *api.Label) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (e EndpointsRolloutV1Client) AutoLabelRollout(ctx context.Context, in *api.Label) (*Rollout, error) {\n\tresp, err := e.AutoLabelRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (r *EndpointsRolloutV1RestClient) AutoListRolloutAction(ctx context.Context, options *api.ListWatchOptions) (*RolloutActionList, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func Roll(n, d, b int, s *discordgo.Session, m *discordgo.MessageCreate) {\n\tresult := \"Rolled: [\"\n\ttotal := b\n\tfor i := 0; i < n-1; i++ {\n\t\tval := rand.Intn(d) + 1\n\t\tresult += strconv.Itoa(val) + \", \"\n\t\ttotal += val\n\t}\n\tval := rand.Intn(d) + 1\n\tresult += strconv.Itoa(val)\n\ttotal += val\n\tif b > 0 {\n\t\tresult += \"] +\" + strconv.Itoa(b)\n\t} else if b == 0 {\n\t\tresult += \"]\"\n\t} else {\n\t\tresult += \"] \" + strconv.Itoa(b)\n\t}\n\n\tresult += \" = \" + strconv.Itoa(total)\n\ts.ChannelMessageSend(m.ChannelID, result)\n}", "func (s *Scaler) do() {\n\tvar (\n\t\twg sync.WaitGroup\n\t\ttr *http.Transport\n\t\tcli *http.Client\n\t)\n\ttr = &http.Transport{}\n\tcli = &http.Client{\n\t\tTransport: tr,\n\t\tTimeout: httpTimeout,\n\t}\n\n\tfor _, a := range s.Actions {\n\t\tgo func(a *model.ActionHTTP) {\n\t\t\twg.Add(1)\n\t\t\tdelay, _ := time.ParseDuration(a.Delay)\n\t\t\turl := a.URL.String()\n\t\t\terr := retry.Do(\n\t\t\t\tfunc() error {\n\t\t\t\t\t// TODO(kiennt): Check kind of action url -> Authen or not?\n\t\t\t\t\treq, err := http.NewRequest(a.Method, url, nil)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tresp, err := cli.Do(req)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tdefer resp.Body.Close()\n\t\t\t\t\treturn nil\n\t\t\t\t},\n\t\t\t\tretry.DelayType(func(n uint, config *retry.Config) time.Duration {\n\t\t\t\t\tvar f retry.DelayTypeFunc\n\t\t\t\t\tswitch a.DelayType {\n\t\t\t\t\tcase \"fixed\":\n\t\t\t\t\t\tf = retry.FixedDelay\n\t\t\t\t\tcase \"backoff\":\n\t\t\t\t\t\tf = retry.BackOffDelay\n\t\t\t\t\t}\n\t\t\t\t\treturn f(n, config)\n\t\t\t\t}),\n\t\t\t\tretry.Attempts(a.Attempts),\n\t\t\t\tretry.Delay(delay),\n\t\t\t\tretry.RetryIf(func(err error) bool {\n\t\t\t\t\treturn common.RetryableError(err)\n\t\t\t\t}),\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\tlevel.Error(s.logger).Log(\"msg\", \"Error doing scale action\", \"url\", url, \"err\", err)\n\t\t\t\texporter.ReportFailureScalerActionCounter(cluster.ClusterID, \"http\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\texporter.ReportSuccessScalerActionCounter(cluster.ClusterID, \"http\")\n\t\t\tlevel.Info(s.logger).Log(\"msg\", \"Sending request\",\n\t\t\t\t\"url\", url, \"method\", a.Method)\n\t\t\ts.alert.Fire(time.Now())\n\t\t\tdefer wg.Done()\n\t\t}(a)\n\t}\n\t// Wait until all actions were performed\n\twg.Wait()\n}", "func (e EndpointsRolloutV1Client) AutoDeleteRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoDeleteRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (a *Auto) Run(args []string) int {\n\tlog.Println(\"Initializing Vault with auto-seal\")\n\ta.Vault.EncryptKeyFun = vaultautoinit.EncryptKeyFun\n\ta.Vault.ProcessKeyFun = vaultautoinit.ProcessKeyFun\n\ta.options = append(a.options, vaultinterface.EnableGPGEncryption())\n\ta.options = append(a.options, vaultinterface.SetGPGCryptoConfig(a.config.GetGPGCryptoConfig()))\n\ta.options = append(a.options, vaultinterface.EnableAutoInitialization())\n\ta.options = append(a.options, vaultinterface.SetVaultConfig(a.config.GetVaultConfig()))\n\ta.options = append(a.options, vaultinterface.SetCASConfig(a.config.GetCASConfig()))\n\terr := a.Vault.Finalize(a.options...)\n\tif err != nil {\n\t\tlog.Printf(\"Error finalizing vaultinterface %v\", err)\n\t\treturn 1\n\t}\n\ta.Vault.Run()\n\treturn 0\n}", "func (r *EndpointsRolloutV1RestClient) AutoDeleteRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (cli *CLI) AutoAuthenticate() {\n\tcli.api.SetToken(os.Getenv(\"SLACK_TOKEN\"))\n\tcli.api.SetCookie(os.Getenv(\"SLACK_COOKIE\"))\n}", "func (view *ViewPodDemo) DemoRollingUpdate() {\n\n\tpod := view.randomPod()\n\tlog.Printf(\"Random pod: %v\", pod)\n\n\tnpart := strings.Split(pod.Name, \"-\")\n\tnewPrefix := fmt.Sprintf(\"%v-%x\", npart[0], rand.Intn(1<<16))\n\n\tseq := 0\n\tfor {\n\t\tnn := fmt.Sprintf(\"%v-%v-%x\", npart[0], npart[1], seq)\n\t\toldPod, ok := view.Pods[nn]\n\t\tif !ok {\n\t\t\tbreak\n\t\t}\n\t\t// create new pod\n\t\tnewPod := view.RecreatePod(newPrefix, seq, oldPod)\n\t\tview.updateStatus(newPod, PodStatusSequenceStart)\n\n\t\t// delete old\n\t\tview.updateStatus(oldPod, PodStatusSequenceStop)\n\n\t\tseq++\n\t}\n\n}", "func (r *EndpointsRolloutV1RestClient) AutoGetRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tpath := makeURIRolloutV1AutoGetRolloutGetOper(in)\n\tif r.bufferId != \"\" {\n\t\tpath = strings.Replace(path, \"/configs\", \"/staging/\"+r.bufferId, 1)\n\t}\n\treq, err := r.getHTTPRequest(ctx, in, \"GET\", path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := r.client.Do(req.WithContext(ctx))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"request failed (%s)\", err)\n\t}\n\tdefer resp.Body.Close()\n\tret, err := decodeHTTPrespRolloutV1AutoGetRollout(ctx, resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ret.(*Rollout), err\n}", "func (dsc *ReconcileDaemonSet) rollingUpdate(ds *appsv1alpha1.DaemonSet, hash string) (delay time.Duration, err error) {\n\n\tif ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.StandardRollingUpdateType {\n\t\treturn delay, dsc.standardRollingUpdate(ds, hash)\n\t} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.SurgingRollingUpdateType {\n\t\treturn dsc.surgingRollingUpdate(ds, hash)\n\t\t//} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.InplaceRollingUpdateType {\n\t\t//\treturn dsc.inplaceRollingUpdate(ds, hash)\n\t} else {\n\t\tklog.Errorf(\"no matched RollingUpdate type\")\n\t}\n\treturn\n}", "func (e EndpointsRolloutV1Client) AutoListRolloutAction(ctx context.Context, in *api.ListWatchOptions) (*RolloutActionList, error) {\n\tresp, err := e.AutoListRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutActionList{}, err\n\t}\n\treturn resp.(*RolloutActionList), nil\n}", "func NewAndroidAutoRoller(workdir, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, strategy repo_manager.NextRollStrategy, preUploadSteps []string, serverURL string) (*AutoRoller, error) {\n\trm, err := repo_manager.NewAndroidRepoManager(workdir, parentBranch, childPath, childBranch, gerrit, strategy, preUploadSteps, serverURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tretrieveRoll := func(arb *AutoRoller, issue int64) (RollImpl, error) {\n\t\treturn newGerritAndroidRoll(arb.gerrit, arb.rm, arb.recent, issue)\n\t}\n\treturn newAutoRoller(workdir, childPath, cqExtraTrybots, emails, gerrit, rm, retrieveRoll)\n}", "func Get(ctx context.Context, rollerName string) (*AutoRollStatus, error) {\n\tvar w DsStatusWrapper\n\tif err := ds.DS.Get(ctx, key(rollerName), &w); err != nil {\n\t\treturn nil, err\n\t}\n\trv := new(AutoRollStatus)\n\tif err := gob.NewDecoder(bytes.NewReader(w.Data)).Decode(rv); err != nil {\n\t\treturn nil, err\n\t}\n\treturn rv, nil\n}", "func (e EndpointsRolloutV1Client) AutoGetRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoGetRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (a *AlertRules) run() {\n\tc := time.Tick(time.Second * time.Duration(a.AlertsInterval))\n\n\tfor range c {\n\t\ta.checkAlerts()\n\t}\n}", "func (o WorkloadStatusConfigAutomaticRuleOutput) Rollup() WorkloadStatusConfigAutomaticRuleRollupOutput {\n\treturn o.ApplyT(func(v WorkloadStatusConfigAutomaticRule) WorkloadStatusConfigAutomaticRuleRollup { return v.Rollup }).(WorkloadStatusConfigAutomaticRuleRollupOutput)\n}", "func (r *EndpointsRolloutV1RestClient) AutoGetRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (e EndpointsRolloutV1Client) AutoGetRolloutAction(ctx context.Context, in *RolloutAction) (*RolloutAction, error) {\n\tresp, err := e.AutoGetRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutAction{}, err\n\t}\n\treturn resp.(*RolloutAction), nil\n}", "func (a *Client) AutoDnssecKeyRollover(params *AutoDnssecKeyRolloverParams) (*AutoDnssecKeyRolloverOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewAutoDnssecKeyRolloverParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"autoDnssecKeyRollover\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/domain/{name}/_autoDnssecKeyRollover\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &AutoDnssecKeyRolloverReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*AutoDnssecKeyRolloverOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for autoDnssecKeyRollover: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (e EndpointsRolloutV1Server) AutoListRolloutAction(ctx context.Context, in api.ListWatchOptions) (RolloutActionList, error) {\n\tresp, err := e.AutoListRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutActionList{}, err\n\t}\n\treturn *resp.(*RolloutActionList), nil\n}", "func (e EndpointsRolloutV1Server) AutoGetRolloutAction(ctx context.Context, in RolloutAction) (RolloutAction, error) {\n\tresp, err := e.AutoGetRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutAction{}, err\n\t}\n\treturn *resp.(*RolloutAction), nil\n}", "func (c *ConfigMapScaler) DoWebhook(context *contextinternal.Context,\n\tclusterStateRegistry *clusterstate.ClusterStateRegistry, sd *ScaleDown,\n\tnodes []*corev1.Node, pods []*corev1.Pod) errors.AutoscalerError {\n\tnodeNameToNodeInfo := scheduler_util.CreateNodeNameToInfoMap(pods, nodes)\n\n\toptions, candidates, err := c.GetResponses(context, clusterStateRegistry,\n\t\tnodeNameToNodeInfo, nodes, sd)\n\tif err != nil {\n\t\treturn errors.NewAutoscalerError(errors.InternalError,\n\t\t\t\"failed to get response from configmap: %v\", err)\n\t}\n\n\tcheckErr := checkResourcesLimits(context, nodes, options, candidates)\n\tif checkErr != nil {\n\t\treturn checkErr\n\t}\n\n\terr = c.ExecuteScale(context, clusterStateRegistry, sd, nodes, options, candidates, nodeNameToNodeInfo)\n\tif err != nil {\n\t\treturn errors.NewAutoscalerError(errors.CloudProviderError,\n\t\t\t\"failed to execute scale from configmap: %v\", err)\n\t}\n\treturn nil\n}", "func (o BeanstalkDeploymentPreferencesOutput) AutomaticRoll() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v BeanstalkDeploymentPreferences) *bool { return v.AutomaticRoll }).(pulumi.BoolPtrOutput)\n}", "func MakeRolloutV1AutoUpdateRolloutEndpoint(s ServiceRolloutV1Server, logger log.Logger) endpoint.Endpoint {\n\tf := func(ctx context.Context, request interface{}) (response interface{}, err error) {\n\t\treq := request.(*Rollout)\n\t\tv, err := s.AutoUpdateRollout(ctx, *req)\n\t\treturn respRolloutV1AutoUpdateRollout{\n\t\t\tV: v,\n\t\t\tErr: err,\n\t\t}, nil\n\t}\n\treturn trace.ServerEndpoint(\"RolloutV1:AutoUpdateRollout\")(f)\n}", "func PerformRollingUpgrade(client kubernetes.Interface, config util.Config, envarPostfix string, upgradeFuncs callbacks.RollingUpgradeFuncs) error {\n\titems := upgradeFuncs.ItemsFunc(client, config.Namespace)\n\tvar err error\n\tfor _, i := range items {\n\t\tcontainers := upgradeFuncs.ContainersFunc(i)\n\t\tresourceName := util.ToObjectMeta(i).Name\n\t\t// find correct annotation and update the resource\n\t\tannotationValue := util.ToObjectMeta(i).Annotations[config.Annotation]\n\t\tif annotationValue != \"\" {\n\t\t\tvalues := strings.Split(annotationValue, \",\")\n\t\t\tfor _, value := range values {\n\t\t\t\tif value == config.ResourceName {\n\t\t\t\t\tupdated := updateContainers(containers, value, config.SHAValue, envarPostfix)\n\t\t\t\t\tif !updated {\n\t\t\t\t\t\tlogrus.Warnf(\"Rolling upgrade failed because no container found to add environment variable in %s of type %s in namespace: %s\", resourceName, upgradeFuncs.ResourceType, config.Namespace)\n\t\t\t\t\t} else {\n\t\t\t\t\t\terr = upgradeFuncs.UpdateFunc(client, config.Namespace, i)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tlogrus.Errorf(\"Update for %s of type %s in namespace %s failed with error %v\", resourceName, upgradeFuncs.ResourceType, config.Namespace, err)\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tlogrus.Infof(\"Updated %s of type %s in namespace: %s \", resourceName, upgradeFuncs.ResourceType, config.Namespace)\n\t\t\t\t\t\t}\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn err\n}", "func main() {\n\tvar sides int = 10000\n\tlog.Printf(\"rolling a %v sided dice\\n\", sides)\n\tunique_rolls := map[int]bool{}\n\ti := 1\n\tfor len(unique_rolls) < sides {\n\t\tvar roll int = roll(sides)\n\t\tfmt.Println(\"rolled a\", roll)\n\t\tunique_rolls = updateUniqueRolls(roll, unique_rolls)\n\t\ti += 1\n\t}\n\tlog.Printf(\"got %v unique rolls in %v attempts \\n\", len(unique_rolls), i)\n}", "func main() {\n\trand.Seed(time.Now().Unix())\n\tdice := NewDice(6, &randomock.Random{})\n\n\tfor roll := 0; roll < 10; roll++ {\n\t\tresult := dice.Roll()\n\t\tfmt.Printf(\"Rolling dice attempt %d.... got %d\\n\", roll, result)\n\t}\n}", "func (e EndpointsRolloutV1Server) AutoLabelRollout(ctx context.Context, in api.Label) (Rollout, error) {\n\tresp, err := e.AutoLabelRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn Rollout{}, err\n\t}\n\treturn *resp.(*Rollout), nil\n}", "func (e EndpointsRolloutV1Client) AutoLabelRolloutAction(ctx context.Context, in *api.Label) (*RolloutAction, error) {\n\tresp, err := e.AutoLabelRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &RolloutAction{}, err\n\t}\n\treturn resp.(*RolloutAction), nil\n}", "func (r *checker) DoChecks() {\n\tfor _, check := range r.checks {\n\t\tcheck.result.LastCheck = time.Now()\n\t\tcheck.result.Count++\n\t\tcheck.result.LastStatus, check.result.LastMessage = check.Exec()\n\n\t\tlog.Debugf(\"%s health check: status=%s, message=%s\", check.Name, check.result.LastStatus, check.result.LastMessage)\n\t}\n}", "func NewDEPSAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, depot_tools string, strategy repo_manager.NextRollStrategy, preUploadSteps []string, includeLog bool, depsCustomVars []string, serverURL string) (*AutoRoller, error) {\n\trm, err := repo_manager.NewDEPSRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, depot_tools, gerrit, strategy, preUploadSteps, includeLog, depsCustomVars, serverURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tretrieveRoll := func(arb *AutoRoller, issue int64) (RollImpl, error) {\n\t\treturn newGerritRoll(arb.gerrit, arb.rm, arb.recent, issue)\n\t}\n\treturn newAutoRoller(workdir, childPath, cqExtraTrybots, emails, gerrit, rm, retrieveRoll)\n}", "func (o *forwardIndexDice) roll(timestamp xtime.UnixNano) bool {\n\tif !o.enabled {\n\t\treturn false\n\t}\n\n\tthreshold := timestamp.Truncate(o.blockSize).Add(o.forwardIndexThreshold)\n\tif !timestamp.Before(threshold) {\n\t\treturn o.forwardIndexDice.Roll()\n\t}\n\n\treturn false\n}", "func (e EndpointsRolloutV1Server) AutoDeleteRollout(ctx context.Context, in Rollout) (Rollout, error) {\n\tresp, err := e.AutoDeleteRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn Rollout{}, err\n\t}\n\treturn *resp.(*Rollout), nil\n}", "func writeAutoRollDataPoint(dbClient *influxdb.Client, status *autoroll.AutoRollStatus) error {\n\tissue := -1\n\tif status.Status != autoroll.STATUS_IDLE {\n\t\tissue = status.CurrentRoll.Issue\n\t}\n\tpoint := []interface{}{\n\t\tinterface{}(status.LastRollRevision),\n\t\tinterface{}(status.CurrentRollRevision),\n\t\tinterface{}(issue),\n\t\tinterface{}(status.Head),\n\t\tinterface{}(status.Status),\n\t}\n\tseries := influxdb.Series{\n\t\tName: SERIES_AUTOROLL_CURRENTSTATUS,\n\t\tColumns: COLUMNS_AUTOROLL_CURRENTSTATUS,\n\t\tPoints: [][]interface{}{point},\n\t}\n\tseriesList := []*influxdb.Series{&series}\n\tglog.Infof(\"Pushing datapoint to %s: %v\", SERIES_AUTOROLL_CURRENTSTATUS, point)\n\treturn dbClient.WriteSeries(seriesList)\n}", "func (e EndpointsRolloutV1Server) AutoGetRollout(ctx context.Context, in Rollout) (Rollout, error) {\n\tresp, err := e.AutoGetRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn Rollout{}, err\n\t}\n\treturn *resp.(*Rollout), nil\n}", "func (c *Controller) manageRollingUpdate(admin submarine.AdminInterface, cluster *rapi.SubmarineCluster, rCluster *submarine.Cluster, nodes submarine.Nodes) (bool, error) {\n\treturn true, nil\n}", "func MakeRolloutV1AutoUpdateRolloutActionEndpoint(s ServiceRolloutV1Server, logger log.Logger) endpoint.Endpoint {\n\tf := func(ctx context.Context, request interface{}) (response interface{}, err error) {\n\t\treq := request.(*RolloutAction)\n\t\tv, err := s.AutoUpdateRolloutAction(ctx, *req)\n\t\treturn respRolloutV1AutoUpdateRolloutAction{\n\t\t\tV: v,\n\t\t\tErr: err,\n\t\t}, nil\n\t}\n\treturn trace.ServerEndpoint(\"RolloutV1:AutoUpdateRolloutAction\")(f)\n}", "func (r *EndpointsRolloutV1RestClient) AutoLabelRolloutAction(ctx context.Context, in *api.Label) (*RolloutAction, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (self *AltaActor) runLoop() {\n\tfor {\n\t\tselect {\n\t\tcase event := <-self.EventChan:\n\t\t\tself.Model.Fsm.FsmEvent(event)\n\n\t\t\t// Save state after each transition\n\t\t\tself.saveModel()\n\t\tcase <-self.ticker.C:\n\t\t\t// FIXME: Use this timer to perform retries when things fail\n\t\t\tlog.Debugf(\"Alta: %s, FSM state: %s, state: %#v\", self.Model.Spec.AltaName,\n\t\t\t\tself.Model.Fsm.FsmState, self)\n\n\t\t\t// If we are stuck in created state, retry it periodically\n\t\t\tif self.Model.Fsm.FsmState == \"created\" {\n\t\t\t\tself.AltaEvent(\"schedule\")\n\t\t\t}\n\t\t}\n\t}\n}", "func MakeRoll(dieType int) int {\n\treturn (rand.Intn(dieType) + 1)\n}", "func Do(retries int, trier Trier, sleeps ...Sleep) error {\n\tif retries <= 0 {\n\t\treturn nil\n\t}\n\n\ttotal := retries\n\nLBBEGIN:\n\n\tif err := trier.Try(); err != nil {\n\t\tretries--\n\t\tif retries == 0 {\n\t\t\treturn ErrMaxRetries\n\t\t}\n\n\t\tif len(sleeps) > 0 {\n\t\t\tif d := sleeps[0](total - retries); d > 0 {\n\t\t\t\ttime.Sleep(d)\n\t\t\t}\n\t\t}\n\n\t\tgoto LBBEGIN\n\t}\n\n\treturn nil\n}", "func (game *Game) Roll(pinsDown int) {\n\tgame.currentFrame.Roll(pinsDown)\n\n\tif game.currentFrame.IsComplete() {\n\t\tgame.score += game.beforePreviousFrame.Bonus(\n\t\t\tgame.previousFrame,\n\t\t\tgame.currentFrame,\n\t\t)\n\t\tgame.score += game.currentFrame.Score()\n\t\tgame.beforePreviousFrame = game.previousFrame\n\t\tgame.previousFrame = game.currentFrame\n\t\tgame.currentFrame = *new(Frame)\n\t}\n}", "func (e EndpointsRolloutV1Server) AutoLabelRolloutAction(ctx context.Context, in api.Label) (RolloutAction, error) {\n\tresp, err := e.AutoLabelRolloutActionEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn RolloutAction{}, err\n\t}\n\treturn *resp.(*RolloutAction), nil\n}", "func Loop(){\n\tfor {\n\t\t\t <-ElectionTimer.C\n\t\t\tif r.Id == r.LeaderId { \n\t\t\t\t\t//r.ResetTimer()\n\t\t\t\t}else{\n\t\t\t\t\tr.CallElection()\t\t\t\t\t\t\t\n\t\t\t\t}\n\t\t}//end of for\t\n}", "func (p *player) rolled_dice(num int) bool {\r\n\tfmt.Println(\"Rolled: \", num)\r\n\tvar total_pos int\r\n\tif p.position.x != initial_player_position_x && p.position.y != initial_player_position_y {\r\n\t\ttotal_pos = p.position.x*grid_size + p.position.y + num\r\n\t} else {\r\n\t\ttotal_pos = num - 1\r\n\t}\r\n\tif total_pos >= grid_size*grid_size {\r\n\t\treturn false\r\n\t}\r\n\tif status, returned_func := hit_by_snake(total_pos); status {\r\n\t\ttotal_pos = returned_func()\r\n\t}\r\n\tif status, returned_func := got_elevated(total_pos); status {\r\n\t\ttotal_pos = returned_func()\r\n\t}\r\n\tp.position.total_pos = total_pos\r\n\tp.position.x = total_pos / grid_size\r\n\tp.position.y = total_pos % grid_size\r\n\treturn true\r\n}", "func Setup(conf *config.RedisConf, logger *logrus.Logger) error {\n\tcli := helper.NewRedisClient(conf, nil)\n\tsha1, err := cli.ScriptLoad(dummyCtx, throttleIncrLuaScript).Result()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"load the throttle incr script: %s\", err.Error())\n\t}\n\tsha2, err := cli.ScriptLoad(dummyCtx, throttleDecrLuaScript).Result()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"load the throttle decr script: %s\", err.Error())\n\t}\n\t_throttler = &Throttler{\n\t\tredisCli: cli,\n\t\tlogger: logger,\n\t\tincrSHA: sha1,\n\t\tdecrSHA: sha2,\n\t\tstop: make(chan bool),\n\t\tcache: make(map[string]*Limiter, 0),\n\t}\n\tgo _throttler.asyncLoop()\n\treturn nil\n}", "func (c *Checkout) CreateNewRoll(ctx context.Context, from, to *revision.Revision, rolling []*revision.Revision, emails []string, dryRun bool, commitMsg string, createRoll CreateRollFunc, uploadRoll UploadRollFunc) (int64, error) {\n\t// Create the roll branch.\n\t_, upstreamBranch, err := c.Update(ctx)\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\t_, _ = c.Git(ctx, \"branch\", \"-D\", RollBranch) // Fails if the branch does not exist.\n\tif _, err := c.Git(ctx, \"checkout\", \"-b\", RollBranch, \"-t\", fmt.Sprintf(\"origin/%s\", upstreamBranch)); err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif _, err := c.Git(ctx, \"reset\", \"--hard\", upstreamBranch); err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\n\t// Run the provided function to create the changes for the roll.\n\thash, err := createRoll(ctx, c.Checkout, from, to, rolling, commitMsg)\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\n\t// Ensure that createRoll generated at least one commit downstream of\n\t// p.baseCommit, and that it did not leave uncommitted changes.\n\tcommits, err := c.RevList(ctx, \"--ancestry-path\", \"--first-parent\", fmt.Sprintf(\"%s..%s\", upstreamBranch, hash))\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif len(commits) == 0 {\n\t\treturn 0, skerr.Fmt(\"createRoll generated no commits!\")\n\t}\n\tif _, err := c.Git(ctx, \"diff\", \"--quiet\"); err != nil {\n\t\treturn 0, skerr.Wrapf(err, \"createRoll left uncommitted changes\")\n\t}\n\tout, err := c.Git(ctx, \"ls-files\", \"--others\", \"--exclude-standard\")\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif len(strings.Fields(out)) > 0 {\n\t\treturn 0, skerr.Fmt(\"createRoll left untracked files:\\n%s\", out)\n\t}\n\n\t// Upload the CL.\n\treturn uploadRoll(ctx, c.Checkout, upstreamBranch, hash, emails, dryRun, commitMsg)\n}", "func (j *Job) doCheckAndRunByTicks(currentTimerTicks int64) {\n\t// Ticks check.\n\tif currentTimerTicks < j.nextTicks.Val() {\n\t\treturn\n\t}\n\tj.nextTicks.Set(currentTimerTicks + j.ticks)\n\t// Perform job checking.\n\tswitch j.status.Val() {\n\tcase StatusRunning:\n\t\tif j.IsSingleton() {\n\t\t\treturn\n\t\t}\n\tcase StatusReady:\n\t\tif !j.status.Cas(StatusReady, StatusRunning) {\n\t\t\treturn\n\t\t}\n\tcase StatusStopped:\n\t\treturn\n\tcase StatusClosed:\n\t\treturn\n\t}\n\t// Perform job running.\n\tj.Run()\n}", "func (db *DB) Roll() {\n\tr := rand.Intn(50)\n\n\tif db.ActiveReq.Requests[r].Shown == 0 {\n\t\treturn\n\t}\n\n\tdb.InactiveReq.Requests = append(db.InactiveReq.Requests, db.ActiveReq.Requests[r])\n\n\tdb.ActiveReq.Requests[r] = &Request{\n\t\tData: strRand.String(2),\n\t\tShown: 0,\n\t}\n}", "func (o *RentalRower) doAfterUpsertHooks(exec boil.Executor) (err error) {\n\tfor _, hook := range rentalRowerAfterUpsertHooks {\n\t\tif err := hook(exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (h *StandHystrix) doCheck() {\n\tif h.checkAliveFunc == nil || h.checkHystrixFunc == nil {\n\t\treturn\n\t}\n\tif h.IsHystrix() {\n\t\tisAlive := h.checkAliveFunc()\n\t\tif isAlive {\n\t\t\th.TriggerAlive()\n\t\t\th.GetCounter().Clear()\n\t\t\ttime.AfterFunc(time.Duration(h.checkHystrixInterval)*time.Second, h.doCheck)\n\t\t} else {\n\t\t\ttime.AfterFunc(time.Duration(h.checkAliveInterval)*time.Second, h.doCheck)\n\t\t}\n\t} else {\n\t\tisHystrix := h.checkHystrixFunc()\n\t\tif isHystrix {\n\t\t\th.TriggerHystrix()\n\t\t\ttime.AfterFunc(time.Duration(h.checkAliveInterval)*time.Second, h.doCheck)\n\t\t} else {\n\t\t\ttime.AfterFunc(time.Duration(h.checkHystrixInterval)*time.Second, h.doCheck)\n\t\t}\n\t}\n}", "func (d *Scheduler) Run() {\n\tgo d.webhookSched.Run()\n\tlog.Println(\"Starting scheduler...\")\n\tif err := d.updateChecks(); err != nil {\n\t\tpanic(err)\n\t}\n\tnow := time.Now().UTC()\n\td.running = true\n\tvar checkTime time.Time\n\tfor {\n\t\tsort.Sort(byTime(d.checks))\n\t\tif d.checks == nil {\n\t\t\td.updateChecks()\n\t\t}\n\t\tif d.checks != nil && len(d.checks) == 0 {\n\t\t\t// Sleep for 5 years until the config change\n\t\t\tcheckTime = now.AddDate(5, 0, 0)\n\t\t} else {\n\t\t\tcheckTime = d.checks[0].Next\n\t\t}\n\t\tselect {\n\t\tcase now = <-time.After(checkTime.Sub(now)):\n\t\t\tfor _, check := range d.checks {\n\t\t\t\tif now.Sub(check.Next) < 0 {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tif !check.Next.IsZero() {\n\t\t\t\t\tcheck.Prev = check.Next\n\t\t\t\t}\n\t\t\t\tgo func(check *Check) {\n\t\t\t\t\toldStatus := check.Up\n\t\t\t\t\tLeaderCheck(d.raft, check)\n\t\t\t\t\tif !check.Next.IsZero() {\n\t\t\t\t\t\tcheck.LastCheck = check.Next.Unix()\n\t\t\t\t\t}\n\t\t\t\t\t// Re-compute the uptime percentage\n\t\t\t\t\tif check.TimeDown > 0 {\n\t\t\t\t\t\ttotal := check.Interval * check.Pings\n\t\t\t\t\t\tcheck.Uptime = float32(int64(total)-check.TimeDown) / float32(total)\n\t\t\t\t\t\tlog.Printf(\"uptime:%+v\", check)\n\t\t\t\t\t}\n\t\t\t\t\tif check.Up != oldStatus {\n\t\t\t\t\t\tlog.Printf(\"Check %v status changed from %v to %v\", check.ID, oldStatus, check.Up)\n\t\t\t\t\t\tvar wg sync.WaitGroup\n\t\t\t\t\t\twg.Add(3)\n\t\t\t\t\t\tgo func(check *Check) {\n\t\t\t\t\t\t\tdefer wg.Done()\n\t\t\t\t\t\t\tif d.raft.Producer == nil {\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tjs, err := json.Marshal(check)\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif err := d.raft.Producer.Publish(\"neverdown\", js); err != nil {\n\t\t\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}(check)\n\t\t\t\t\t\tgo func(check *Check) {\n\t\t\t\t\t\t\tdefer wg.Done()\n\t\t\t\t\t\t\tif err := NotifyEmails(check); err != nil {\n\t\t\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}(check)\n\t\t\t\t\t\tgo func(check *Check) {\n\t\t\t\t\t\t\tdefer wg.Done()\n\t\t\t\t\t\t\tif err := ExecuteWebhooks(d.raft, d.webhookSched, check); err != nil {\n\t\t\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}(check)\n\t\t\t\t\t\twg.Wait()\n\t\t\t\t\t}\n\t\t\t\t\tif err := d.raft.ExecCommand(check.ToPostCmd()); err != nil {\n\t\t\t\t\t\tpanic(err)\n\t\t\t\t\t}\n\t\t\t\t}(check)\n\t\t\t\tcheck.ComputeNext(now)\n\t\t\t\tcontinue\n\t\t\t}\n\t\tcase <-d.stop:\n\t\t\td.running = false\n\t\t\treturn\n\t\tcase <-d.Reloadch:\n\t\t\td.updateChecks()\n\t\t}\n\t}\n}", "func (c *AutoRollStatusCache) Update(ctx context.Context) error {\n\tstatus, err := Get(ctx, c.roller)\n\tif err == datastore.ErrNoSuchEntity || status == nil {\n\t\t// This will occur the first time the roller starts,\n\t\t// before it sets the status for the first time. Ignore.\n\t\tsklog.Warningf(\"Unable to find AutoRollStatus for %s. Is this the first startup for this roller?\", c.roller)\n\t\tstatus = &AutoRollStatus{}\n\t} else if err != nil {\n\t\treturn err\n\t}\n\tc.mtx.Lock()\n\tdefer c.mtx.Unlock()\n\tc.status = status\n\treturn nil\n}", "func NewManifestAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, depot_tools string, strategy repo_manager.NextRollStrategy, preUploadSteps []string, serverURL string) (*AutoRoller, error) {\n\trm, err := repo_manager.NewManifestRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, depot_tools, gerrit, strategy, preUploadSteps, serverURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tretrieveRoll := func(arb *AutoRoller, issue int64) (RollImpl, error) {\n\t\treturn newGerritRoll(arb.gerrit, arb.rm, arb.recent, issue)\n\t}\n\treturn newAutoRoller(workdir, childPath, cqExtraTrybots, emails, gerrit, rm, retrieveRoll)\n}", "func (o *Onboarder) DoOnboard(ce cloudevents.Event, loggingDone chan bool) error {\n\n\tdefer func() { loggingDone <- true }()\n\n\tevent := &keptnevents.ServiceCreateEventData{}\n\tif err := ce.DataAs(event); err != nil {\n\t\to.logger.Error(fmt.Sprintf(\"Got Data Error: %s\", err.Error()))\n\t\treturn err\n\t}\n\n\tif _, ok := event.DeploymentStrategies[\"*\"]; ok {\n\t\tdeplStrategies, err := FixDeploymentStrategies(event.Project, event.DeploymentStrategies[\"*\"])\n\t\tif err != nil {\n\t\t\to.logger.Error(fmt.Sprintf(\"Error when getting deployment strategies: %s\" + err.Error()))\n\t\t\treturn err\n\t\t}\n\t\tevent.DeploymentStrategies = deplStrategies\n\t} else if os.Getenv(\"PRE_WORKFLOW_ENGINE\") == \"true\" && (event.DeploymentStrategies == nil || len(event.DeploymentStrategies) == 0) {\n\t\tdeplStrategies, err := GetDeploymentStrategies(event.Project)\n\t\tif err != nil {\n\t\t\to.logger.Error(fmt.Sprintf(\"Error when getting deployment strategies: %s\" + err.Error()))\n\t\t\treturn err\n\t\t}\n\t\tevent.DeploymentStrategies = deplStrategies\n\t}\n\n\to.logger.Info(fmt.Sprintf(\"Start creating service %s in project %s\", event.Service, event.Project))\n\n\turl, err := serviceutils.GetConfigServiceURL()\n\tif err != nil {\n\t\to.logger.Error(fmt.Sprintf(\"Error when getting config service url: %s\", err.Error()))\n\t\treturn err\n\t}\n\n\tstageHandler := keptnutils.NewStageHandler(url.String())\n\tstages, err := stageHandler.GetAllStages(event.Project)\n\tif err != nil {\n\t\to.logger.Error(\"Error when getting all stages: \" + err.Error())\n\t\treturn err\n\t}\n\n\tfirstService, err := o.isFirstServiceOfProject(event, stages)\n\tif err != nil {\n\t\to.logger.Error(\"Error when checking whether any service was created before: \" + err.Error())\n\t\treturn err\n\t}\n\tif firstService {\n\t\to.logger.Info(\"Create Helm umbrella charts\")\n\t\tumbrellaChartHandler := helm.NewUmbrellaChartHandler(o.mesh)\n\t\tif err := o.initAndApplyUmbrellaChart(event, umbrellaChartHandler, stages); err != nil {\n\t\t\to.logger.Error(fmt.Sprintf(\"Error when initalizing and applying umbrella charts for project %s: %s\", event.Project, err.Error()))\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, stage := range stages {\n\t\tif err := o.onboardService(stage.StageName, event, url.String()); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\to.logger.Info(fmt.Sprintf(\"Finished creating service %s in project %s\", event.Service, event.Project))\n\treturn nil\n}", "func TestRolling_deployInitialHooks(t *testing.T) {\n\tvar hookError error\n\n\tstrategy := &RollingDeploymentStrategy{\n\t\trcClient: fake.NewSimpleClientset().CoreV1(),\n\t\teventClient: fake.NewSimpleClientset().CoreV1(),\n\t\tinitialStrategy: &testStrategy{\n\t\t\tdeployFn: func(from *corev1.ReplicationController, to *corev1.ReplicationController, desiredReplicas int,\n\t\t\t\tupdateAcceptor strat.UpdateAcceptor) error {\n\t\t\t\treturn nil\n\t\t\t},\n\t\t},\n\t\trollingUpdate: func(config *RollingUpdaterConfig) error {\n\t\t\treturn nil\n\t\t},\n\t\thookExecutor: &hookExecutorImpl{\n\t\t\texecuteFunc: func(hook *appsv1.LifecycleHook, deployment *corev1.ReplicationController, suffix, label string) error {\n\t\t\t\treturn hookError\n\t\t\t},\n\t\t},\n\t\tgetUpdateAcceptor: getUpdateAcceptor,\n\t\tapiRetryPeriod: 1 * time.Millisecond,\n\t\tapiRetryTimeout: 10 * time.Millisecond,\n\t}\n\n\tcases := []struct {\n\t\tparams *appsv1.RollingDeploymentStrategyParams\n\t\thookShouldFail bool\n\t\tdeploymentShouldFail bool\n\t}{\n\t\t{rollingParams(appsv1.LifecycleHookFailurePolicyAbort, \"\"), true, true},\n\t\t{rollingParams(appsv1.LifecycleHookFailurePolicyAbort, \"\"), false, false},\n\t\t{rollingParams(\"\", appsv1.LifecycleHookFailurePolicyAbort), true, true},\n\t\t{rollingParams(\"\", appsv1.LifecycleHookFailurePolicyAbort), false, false},\n\t}\n\n\tfor i, tc := range cases {\n\t\tconfig := appstest.OkDeploymentConfig(2)\n\t\tconfig.Spec.Strategy.RollingParams = tc.params\n\t\tdeployment, _ := appsutil.MakeDeployment(config)\n\t\thookError = nil\n\t\tif tc.hookShouldFail {\n\t\t\thookError = fmt.Errorf(\"hook failure\")\n\t\t}\n\t\tstrategy.out, strategy.errOut = &bytes.Buffer{}, &bytes.Buffer{}\n\t\terr := strategy.Deploy(nil, deployment, 2)\n\t\tif err != nil && tc.deploymentShouldFail {\n\t\t\tt.Logf(\"got expected error: %v\", err)\n\t\t}\n\t\tif err == nil && tc.deploymentShouldFail {\n\t\t\tt.Errorf(\"%d: expected an error for case: %v\", i, tc)\n\t\t}\n\t\tif err != nil && !tc.deploymentShouldFail {\n\t\t\tt.Errorf(\"%d: unexpected error for case: %v: %v\", i, tc, err)\n\t\t}\n\t}\n}" ]
[ "0.6744203", "0.6604339", "0.63867587", "0.6273548", "0.6151653", "0.6139569", "0.5802719", "0.5772128", "0.573523", "0.57012045", "0.54317874", "0.53480554", "0.5339104", "0.5238146", "0.5197582", "0.5124661", "0.51030296", "0.5057785", "0.5050795", "0.50184435", "0.49690998", "0.49456725", "0.49433577", "0.4920629", "0.49079248", "0.49060917", "0.48738182", "0.485386", "0.48389992", "0.4816229", "0.47779426", "0.4762787", "0.47199073", "0.46897376", "0.46796927", "0.46773404", "0.46099186", "0.4587671", "0.4578478", "0.4564735", "0.45541403", "0.45439932", "0.452798", "0.45231965", "0.45133528", "0.4491133", "0.4490782", "0.4452945", "0.44426554", "0.44362608", "0.44334805", "0.4430441", "0.44271478", "0.44117177", "0.4409867", "0.43893242", "0.43524083", "0.4325162", "0.43242612", "0.43238905", "0.43055224", "0.4302925", "0.4302299", "0.4288972", "0.42837563", "0.4254833", "0.42494595", "0.4236013", "0.42339677", "0.42271447", "0.4223106", "0.42162874", "0.42142838", "0.42053497", "0.41891688", "0.41820946", "0.41779464", "0.4167891", "0.41668805", "0.4164404", "0.41635603", "0.41576716", "0.41481787", "0.4142108", "0.4109242", "0.41083968", "0.4100347", "0.40906665", "0.40867123", "0.40797204", "0.40723842", "0.40599927", "0.40333453", "0.40300116", "0.4022065", "0.40218574", "0.40183243", "0.40056643", "0.40056017", "0.39986017" ]
0.7747144
0
makeRollResult determines what the result of a roll should be, given that it is going to be closed.
func (r *AutoRoller) makeRollResult(roll *autoroll.AutoRollIssue) string { if util.In(roll.Result, autoroll.DRY_RUN_RESULTS) { if roll.Result == autoroll.ROLL_RESULT_DRY_RUN_IN_PROGRESS { return autoroll.ROLL_RESULT_DRY_RUN_FAILURE } else { return roll.Result } } return autoroll.ROLL_RESULT_FAILURE }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func MakeRoll(dieType int) int {\n\treturn (rand.Intn(dieType) + 1)\n}", "func Roll(number int, sided int) Result {\n\treturn roll(number, sided)\n}", "func (r *AutoRoller) doAutoRollInner() (string, error) {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\t// Get updated info about the current roll.\n\tif err := r.updateCurrentRoll(); err != nil {\n\t\treturn STATUS_ERROR, err\n\t}\n\n\t// There's a currently-active roll. Determine whether or not it's still good.\n\t// If so, leave it open and exit. If not, close it so that we can open another.\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll != nil {\n\t\tsklog.Infof(\"Found current roll: %s\", r.issueUrl(currentRoll.Issue))\n\n\t\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\t\t// If we have a normal (non-dry-run) roll running,\n\t\t\t// switch it to a dry run.\n\t\t\tif currentRoll.CommitQueue {\n\t\t\t\tsklog.Infof(\"Setting dry-run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, true); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\n\t\t\t// If the CQ has finished, determine if it was a success\n\t\t\t// or failure.\n\t\t\trollDone, err := r.isDryRunDone(currentRoll)\n\t\t\tif err != nil {\n\t\t\t\treturn STATUS_ERROR, err\n\t\t\t}\n\t\t\tif rollDone {\n\t\t\t\tresult := autoroll.ROLL_RESULT_DRY_RUN_FAILURE\n\t\t\t\tstatus := STATUS_DRY_RUN_FAILURE\n\t\t\t\trollSuccessful, err := r.isDryRunSuccessful(currentRoll)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rollSuccessful {\n\t\t\t\t\tresult = autoroll.ROLL_RESULT_DRY_RUN_SUCCESS\n\t\t\t\t\tstatus = STATUS_DRY_RUN_SUCCESS\n\t\t\t\t}\n\t\t\t\tsklog.Infof(\"Dry run is finished: %v\", currentRoll)\n\t\t\t\tif currentRoll.RollingTo != r.rm.ChildHead() {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, result, fmt.Sprintf(\"Repo has passed %s; will open a new dry run.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else if currentRoll.Result != result {\n\t\t\t\t\t// The dry run just finished. Set its result.\n\t\t\t\t\tif result == autoroll.ROLL_RESULT_DRY_RUN_FAILURE {\n\t\t\t\t\t\tif err := r.closeIssue(currentRoll, result, \"Dry run failed. Closing, will open another.\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\tif err := r.addIssueComment(currentRoll, \"Dry run finished successfully; leaving open in case we want to land\"); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\tcurrentRoll.Result = result\n\t\t\t\t\t\tif err := r.recent.Update(currentRoll); err != nil {\n\t\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t\t}\n\t\t\t\t\t\treturn status, nil\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// The dry run is finished but still good. Leave it open.\n\t\t\t\t\tsklog.Infof(\"Dry run is finished and still good.\")\n\t\t\t\t\treturn status, nil\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tsklog.Infof(\"Dry run still in progress.\")\n\t\t\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t\t\t}\n\t\t} else {\n\t\t\tif currentRoll.CommitQueueDryRun {\n\t\t\t\tsklog.Infof(\"Unsetting dry run bit on %s\", r.gerrit.Url(currentRoll.Issue))\n\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\t\t\t// If we're stopped, close the issue.\n\t\t\t\t// Respect the previous result of the roll.\n\t\t\t\tif err := r.closeIssue(currentRoll, r.makeRollResult(currentRoll), \"AutoRoller is stopped; closing the active roll.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else if !currentRoll.CommitQueue {\n\t\t\t\t// If the CQ failed, close the issue.\n\t\t\t\t// Special case: if the current roll was a dry run which succeeded, land it.\n\t\t\t\tif currentRoll.Result == autoroll.ROLL_RESULT_DRY_RUN_SUCCESS {\n\t\t\t\t\tsklog.Infof(\"Dry run succeeded. Attempting to land.\")\n\t\t\t\t\tif err := r.setDryRun(currentRoll, false); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t} else {\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Commit queue failed; closing this roll.\"); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if time.Since(currentRoll.Modified) > 24*time.Hour {\n\t\t\t\t// If the roll has been open too long, close the issue.\n\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, \"Roll has been open for over 24 hours; closing.\"); err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t}\n\t\t\t\tif rolledPast {\n\t\t\t\t\t// If we've already rolled past the target revision, close the issue\n\t\t\t\t\tif err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, fmt.Sprintf(\"Already rolled past %s; closing this roll.\", currentRoll.RollingTo)); err != nil {\n\t\t\t\t\t\treturn STATUS_ERROR, err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\t// Current roll is still good.\n\t\t\t\t\tsklog.Infof(\"Roll is still active (%d): %s\", currentRoll.Issue, currentRoll.Subject)\n\t\t\t\t\treturn STATUS_IN_PROGRESS, nil\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// If we're stopped, exit.\n\tif r.isMode(autoroll_modes.MODE_STOPPED) {\n\t\tsklog.Infof(\"Roller is stopped; not opening new rolls.\")\n\t\treturn STATUS_STOPPED, nil\n\t}\n\n\t// If we're up-to-date, exit.\n\tchildHead := r.rm.ChildHead()\n\tif r.rm.LastRollRev() == childHead {\n\t\tsklog.Infof(\"Repo is up-to-date.\")\n\t\treturn STATUS_UP_TO_DATE, nil\n\t}\n\n\t// Create a new roll.\n\tif r.attemptCounter.Get() >= ROLL_ATTEMPT_THROTTLE_NUM {\n\t\treturn STATUS_THROTTLED, nil\n\t}\n\tr.attemptCounter.Inc()\n\tdryRun := r.isMode(autoroll_modes.MODE_DRY_RUN)\n\tuploadedNum, err := r.rm.CreateNewRoll(r.strategy, r.GetEmails(), r.cqExtraTrybots, dryRun)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to upload a new roll: %s\", err)\n\t}\n\tuploaded, err := r.retrieveRoll(uploadedNum)\n\tif err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to retrieve uploaded roll: %s\", err)\n\t}\n\tif err := r.recent.Add(uploaded); err != nil {\n\t\treturn STATUS_ERROR, fmt.Errorf(\"Failed to insert uploaded roll into database: %s\", err)\n\t}\n\n\tif r.isMode(autoroll_modes.MODE_DRY_RUN) {\n\t\treturn STATUS_DRY_RUN_IN_PROGRESS, nil\n\t}\n\treturn STATUS_IN_PROGRESS, nil\n}", "func (r *AutoRoller) createNewRoll(ctx context.Context, from, to *revision.Revision, emails []string, dryRun, canary bool, manualRollRequester string) (rv *autoroll.AutoRollIssue, rvErr error) {\n\t// Track roll CL upload attempts vs failures.\n\tdefer func() {\n\t\tr.rollUploadAttempts.Inc(1)\n\t\tif rvErr == nil {\n\t\t\tr.rollUploadFailures.Reset()\n\t\t} else {\n\t\t\tr.rollUploadFailures.Inc(1)\n\t\t}\n\t}()\n\tr.statusMtx.RLock()\n\tvar revs []*revision.Revision\n\tfound := false\n\tfor _, rev := range r.notRolledRevs {\n\t\tif rev.Id == to.Id {\n\t\t\tfound = true\n\t\t}\n\t\tif found {\n\t\t\trevs = append(revs, rev)\n\t\t}\n\t}\n\tr.statusMtx.RUnlock()\n\n\tcommitMsg, err := r.commitMsgBuilder.Build(from, to, revs, emails, r.cfg.Contacts, canary, manualRollRequester)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tsklog.Infof(\"Creating new roll with commit message: \\n%s\", commitMsg)\n\tissueNum, err := r.rm.CreateNewRoll(ctx, from, to, revs, emails, dryRun, commitMsg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tissue := &autoroll.AutoRollIssue{\n\t\tIsDryRun: dryRun,\n\t\tIssue: issueNum,\n\t\tRollingFrom: from.Id,\n\t\tRollingTo: to.Id,\n\t}\n\treturn issue, nil\n}", "func DummyRoll(dieType int) int {\n\treturn dieType / 2\n}", "func (r *AutoRoller) rollFinished(ctx context.Context, justFinished codereview.RollImpl) error {\n\trecent := r.recent.GetRecentRolls()\n\t// Sanity check: pop any rolls which occurred after the one which just\n\t// finished.\n\tidx := -1\n\tvar currentRoll *autoroll.AutoRollIssue\n\tfor i, roll := range recent {\n\t\tissue := fmt.Sprintf(\"%d\", roll.Issue)\n\t\tif issue == justFinished.IssueID() {\n\t\t\tidx = i\n\t\t\tcurrentRoll = roll\n\t\t\tbreak\n\t\t}\n\t}\n\tif currentRoll == nil {\n\t\treturn skerr.Fmt(\"Unable to find just-finished roll %q in recent list!\", justFinished.IssueID())\n\t}\n\n\t// Feed AutoRoll stats into metrics.\n\tv := int64(0)\n\tif currentRoll.Closed && currentRoll.Committed {\n\t\tv = int64(1)\n\t}\n\tmetrics2.GetInt64Metric(\"autoroll_last_roll_result\", map[string]string{\"roller\": r.cfg.RollerName}).Update(v)\n\n\trecent = recent[idx:]\n\tvar lastRoll *autoroll.AutoRollIssue\n\tif len(recent) > 1 {\n\t\tlastRoll = recent[1]\n\t} else {\n\t\t// If there are no other rolls, then the below alerts do not apply.\n\t\treturn nil\n\t}\n\n\tissueURL := fmt.Sprintf(\"%s%d\", r.codereview.GetIssueUrlBase(), currentRoll.Issue)\n\n\t// Send notifications if this roll had a different result from the last\n\t// roll, ie. success -> failure or failure -> success.\n\tcurrentSuccess := util.In(currentRoll.Result, autoroll.SUCCESS_RESULTS)\n\tlastSuccess := util.In(lastRoll.Result, autoroll.SUCCESS_RESULTS)\n\tif lastRoll != nil {\n\t\tif currentSuccess && !lastSuccess {\n\t\t\tr.notifier.SendNewSuccess(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t} else if !currentSuccess && lastSuccess {\n\t\t\tr.notifier.SendNewFailure(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t}\n\t}\n\n\t// Send a notification if the last N rolls failed in a row.\n\tnFailed := 0\n\t// recent is in reverse chronological order.\n\tfor _, roll := range recent {\n\t\tif util.In(roll.Result, autoroll.SUCCESS_RESULTS) {\n\t\t\tbreak\n\t\t} else {\n\t\t\tnFailed++\n\t\t}\n\t}\n\tif nFailed == notifyIfLastNFailed {\n\t\tr.notifier.SendLastNFailed(ctx, notifyIfLastNFailed, issueURL)\n\t}\n\n\treturn nil\n}", "func (r Result) Reroll() Result {\n\treturn Roll(len(r.Ints()), r.die)\n}", "func ROLL(ci, mr operand.Op) { ctx.ROLL(ci, mr) }", "func (r *MockRepoManager) CreateNewRoll(ctx context.Context, from, to string, emails []string, cqExtraTrybots string, dryRun bool) (int64, error) {\n\tr.mtx.RLock()\n\tdefer r.mtx.RUnlock()\n\treturn r.mockIssueNumber, nil\n}", "func (_m *RollerMetricsRecorder) MeasureDieRollResult(ctx context.Context, rollerType string, dieRoll *model.DieRoll) {\n\t_m.Called(ctx, rollerType, dieRoll)\n}", "func (r *AutoRoller) retrieveRoll(ctx context.Context, roll *autoroll.AutoRollIssue, rollingTo *revision.Revision) (codereview.RollImpl, error) {\n\treturn r.codereview.RetrieveRoll(ctx, roll, r.recent, rollingTo, r.rollFinished)\n}", "func SimpleRoll(roll string) int {\n\tvar result, dieType int\n\tvar results []int\n\n\tif strings.Contains(roll, \"d\") {\n\t\tcomponents := strings.Split(roll, \"d\")\n\t\tnumDice, _ := strconv.Atoi(components[0])\n\t\tkeepHigh := 0\n\t\tkeepLow := 0\n\n\t\tif strings.Contains(components[1], \"h\") {\n\t\t\tcomps := strings.Split(components[1], \"h\")\n\t\t\tdieType, _ = strconv.Atoi(comps[0])\n\t\t\tkeepHigh, _ = strconv.Atoi(comps[1])\n\t\t} else if strings.Contains(components[1], \"l\") {\n\t\t\tcomps := strings.Split(components[1], \"l\")\n\t\t\tdieType, _ = strconv.Atoi(comps[0])\n\t\t\tkeepLow, _ = strconv.Atoi(comps[1])\n\t\t} else {\n\t\t\tdieType, _ = strconv.Atoi(components[1])\n\t\t}\n\n\t\tfor i := 0; i < numDice; i++ {\n\t\t\tresults = append(results, MakeRoll(dieType))\n\t\t\t//results = append(results, DummyRoll(dieType))\n\t\t}\n\t\tsort.Ints(results)\n\t\tif keepLow > 0 {\n\t\t\tfor i := 0; i < keepLow; i++ {\n\t\t\t\tresult += results[i]\n\t\t\t}\n\t\t} else if keepHigh > 0 {\n\t\t\tfor i := 0; i < keepHigh; i++ {\n\t\t\t\tresult += results[numDice-1-i]\n\t\t\t}\n\n\t\t} else {\n\t\t\tfor _, res := range results {\n\t\t\t\tresult += res\n\t\t\t}\n\t\t}\n\n\t} else {\n\t\tresult, _ = strconv.Atoi(roll)\n\t}\n\n\treturn result\n}", "func TestRoll(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Default roll\", \"\", response{http.StatusOK, ``}},\n\t\t{\"Valid query for sides\", \"?sides=4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Invalid query for sides\", \"?sides=5\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for count\", \"?count=2\", response{http.StatusOK, `\"count\":2`}},\n\t\t{\"Invalid query for count\", \"?count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for sides, invalid for count\", \"?sides=4&count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for count, invalid for sides\", \"?count=2&sides=1\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for sides and count\", \"?sides=4&count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t}\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func (c *Checkout) CreateNewRoll(ctx context.Context, from, to *revision.Revision, rolling []*revision.Revision, emails []string, dryRun bool, commitMsg string, createRoll CreateRollFunc, uploadRoll UploadRollFunc) (int64, error) {\n\t// Create the roll branch.\n\t_, upstreamBranch, err := c.Update(ctx)\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\t_, _ = c.Git(ctx, \"branch\", \"-D\", RollBranch) // Fails if the branch does not exist.\n\tif _, err := c.Git(ctx, \"checkout\", \"-b\", RollBranch, \"-t\", fmt.Sprintf(\"origin/%s\", upstreamBranch)); err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif _, err := c.Git(ctx, \"reset\", \"--hard\", upstreamBranch); err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\n\t// Run the provided function to create the changes for the roll.\n\thash, err := createRoll(ctx, c.Checkout, from, to, rolling, commitMsg)\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\n\t// Ensure that createRoll generated at least one commit downstream of\n\t// p.baseCommit, and that it did not leave uncommitted changes.\n\tcommits, err := c.RevList(ctx, \"--ancestry-path\", \"--first-parent\", fmt.Sprintf(\"%s..%s\", upstreamBranch, hash))\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif len(commits) == 0 {\n\t\treturn 0, skerr.Fmt(\"createRoll generated no commits!\")\n\t}\n\tif _, err := c.Git(ctx, \"diff\", \"--quiet\"); err != nil {\n\t\treturn 0, skerr.Wrapf(err, \"createRoll left uncommitted changes\")\n\t}\n\tout, err := c.Git(ctx, \"ls-files\", \"--others\", \"--exclude-standard\")\n\tif err != nil {\n\t\treturn 0, skerr.Wrap(err)\n\t}\n\tif len(strings.Fields(out)) > 0 {\n\t\treturn 0, skerr.Fmt(\"createRoll left untracked files:\\n%s\", out)\n\t}\n\n\t// Upload the CL.\n\treturn uploadRoll(ctx, c.Checkout, upstreamBranch, hash, emails, dryRun, commitMsg)\n}", "func (r *repoManager) CreateNewRoll(emails []string, cqExtraTrybots string, dryRun bool) (int64, error) {\n\tr.repoMtx.Lock()\n\tdefer r.repoMtx.Unlock()\n\n\t// Clean the checkout, get onto a fresh branch.\n\tif err := r.cleanChromium(); err != nil {\n\t\treturn 0, err\n\t}\n\tif _, err := exec.RunCwd(r.chromiumDir, \"git\", \"checkout\", \"-b\", DEPS_ROLL_BRANCH, \"-t\", \"origin/master\", \"-f\"); err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Defer some more cleanup.\n\tdefer func() {\n\t\tutil.LogErr(r.cleanChromium())\n\t}()\n\n\t// Create the roll CL.\n\tif _, err := exec.RunCwd(r.chromiumDir, \"git\", \"config\", \"user.name\", autoroll.ROLL_AUTHOR); err != nil {\n\t\treturn 0, err\n\t}\n\tif _, err := exec.RunCwd(r.chromiumDir, \"git\", \"config\", \"user.email\", autoroll.ROLL_AUTHOR); err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Find Chromium bugs.\n\tbugs := []string{}\n\tcr := r.childRepo\n\tcommits, err := cr.RevList(fmt.Sprintf(\"%s..%s\", r.lastRollRev, r.childHead))\n\tif err != nil {\n\t\treturn 0, fmt.Errorf(\"Failed to list revisions: %s\", err)\n\t}\n\tfor _, c := range commits {\n\t\td, err := cr.Details(c, false)\n\t\tif err != nil {\n\t\t\treturn 0, fmt.Errorf(\"Failed to obtain commit details: %s\", err)\n\t\t}\n\t\tb := util.BugsFromCommitMsg(d.Body)\n\t\tfor _, bug := range b[util.PROJECT_CHROMIUM] {\n\t\t\tbugs = append(bugs, bug)\n\t\t}\n\t}\n\n\targs := []string{r.childPath, r.childHead}\n\tif len(bugs) > 0 {\n\t\targs = append(args, \"--bug\", strings.Join(bugs, \",\"))\n\t}\n\tglog.Infof(\"Running command: roll-dep %s\", strings.Join(args, \" \"))\n\tif _, err := exec.RunCommand(&exec.Command{\n\t\tDir: r.chromiumDir,\n\t\tEnv: getEnv(r.depot_tools),\n\t\tName: r.rollDep,\n\t\tArgs: args,\n\t}); err != nil {\n\t\treturn 0, err\n\t}\n\t// Build the commit message, starting with the message provided by roll-dep.\n\tcommitMsg, err := exec.RunCwd(r.chromiumDir, \"git\", \"log\", \"-n1\", \"--format=%B\", \"HEAD\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tcommitMsg += `\nDocumentation for the AutoRoller is here:\nhttps://skia.googlesource.com/buildbot/+/master/autoroll/README.md\n\nIf the roll is causing failures, see:\nhttp://www.chromium.org/developers/tree-sheriffs/sheriff-details-chromium#TOC-Failures-due-to-DEPS-rolls\n\n`\n\tif cqExtraTrybots != \"\" {\n\t\tcommitMsg += \"\\n\" + fmt.Sprintf(TMPL_CQ_INCLUDE_TRYBOTS, cqExtraTrybots)\n\t}\n\tuploadCmd := &exec.Command{\n\t\tDir: r.chromiumDir,\n\t\tEnv: getEnv(r.depot_tools),\n\t\tName: \"git\",\n\t\tArgs: []string{\"cl\", \"upload\", \"--bypass-hooks\", \"-f\"},\n\t}\n\tif dryRun {\n\t\tuploadCmd.Args = append(uploadCmd.Args, \"--cq-dry-run\")\n\t} else {\n\t\tuploadCmd.Args = append(uploadCmd.Args, \"--use-commit-queue\")\n\t}\n\ttbr := \"\\nTBR=\"\n\tif emails != nil && len(emails) > 0 {\n\t\temailStr := strings.Join(emails, \",\")\n\t\ttbr += emailStr\n\t\tuploadCmd.Args = append(uploadCmd.Args, \"--send-mail\", \"--cc\", emailStr)\n\t}\n\tcommitMsg += tbr\n\tuploadCmd.Args = append(uploadCmd.Args, \"-m\", commitMsg)\n\n\t// Upload the CL.\n\tif _, err := exec.RunCommand(uploadCmd); err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Obtain the issue number.\n\ttmp, err := ioutil.TempDir(\"\", \"\")\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer util.RemoveAll(tmp)\n\tjsonFile := path.Join(tmp, \"issue.json\")\n\tif _, err := exec.RunCommand(&exec.Command{\n\t\tDir: r.chromiumDir,\n\t\tEnv: getEnv(r.depot_tools),\n\t\tName: \"git\",\n\t\tArgs: []string{\"cl\", \"issue\", fmt.Sprintf(\"--json=%s\", jsonFile)},\n\t}); err != nil {\n\t\treturn 0, err\n\t}\n\tf, err := os.Open(jsonFile)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tvar issue issueJson\n\tif err := json.NewDecoder(f).Decode(&issue); err != nil {\n\t\treturn 0, err\n\t}\n\treturn issue.Issue, nil\n}", "func roll(r string) string {\n\tres, _, err := dice.Roll(r)\n\tif err != nil {\n\t\treturn DICE_USAGE\n\t}\n\treturn fmt.Sprintf(\"%v\", res.Int())\n}", "func (future *MonitorsCreateFuture) result(client MonitorsClient) (mr MonitorResource, err error) {\n\tvar done bool\n\tdone, err = future.DoneWithContext(context.Background(), client)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"logz.MonitorsCreateFuture\", \"Result\", future.Response(), \"Polling failure\")\n\t\treturn\n\t}\n\tif !done {\n\t\tmr.Response.Response = future.Response()\n\t\terr = azure.NewAsyncOpIncompleteError(\"logz.MonitorsCreateFuture\")\n\t\treturn\n\t}\n\tsender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n\tif mr.Response.Response, err = future.GetResult(sender); err == nil && mr.Response.Response.StatusCode != http.StatusNoContent {\n\t\tmr, err = client.CreateResponder(mr.Response.Response)\n\t\tif err != nil {\n\t\t\terr = autorest.NewErrorWithError(err, \"logz.MonitorsCreateFuture\", \"Result\", mr.Response.Response, \"Failure responding to request\")\n\t\t}\n\t}\n\treturn\n}", "func (r *AutoRoller) retrieveRoll(issueNum int64) (*autoroll.AutoRollIssue, error) {\n\tvar a *autoroll.AutoRollIssue\n\tinfo, err := r.gerrit.GetIssueProperties(issueNum)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get issue properties: %s\", err)\n\t}\n\ta, err = autoroll.FromGerritChangeInfo(info, r.rm.FullChildHash, r.rollIntoAndroid)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to convert issue format: %s\", err)\n\t}\n\ttryResults, err := autoroll.GetTryResultsFromGerrit(r.gerrit, a)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to retrieve try results: %s\", err)\n\t}\n\ta.TryResults = tryResults\n\treturn a, nil\n}", "func (c *Context) ROLL(ci, mr operand.Op) {\n\tc.addinstruction(x86.ROLL(ci, mr))\n}", "func (t *ParseTransaction) MakeResult(value interface{}) ParseResult {\n\tt.Commit()\n\treturn ParseResult{value, t.Range()}\n}", "func (d *Die) Reroll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\tif d.Result == nil {\n\t\treturn ErrUnrolled\n\t}\n\n\td.Result = nil\n\td.Rerolls++\n\t// reroll without reapplying all modifiers\n\treturn d.Roll(ctx)\n}", "func newAutoRoller(workdir, childPath, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, rm repo_manager.RepoManager, retrieveRoll func(*AutoRoller, int64) (RollImpl, error)) (*AutoRoller, error) {\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"child_path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\tretrieveRoll: retrieveRoll,\n\t\trm: rm,\n\t\tstatus: &AutoRollStatusCache{},\n\t}\n\tsm, err := state_machine.New(arb, workdir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\troll, err := arb.retrieveRoll(arb, current.Issue)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\treturn arb, nil\n}", "func Roll(n, d, b int, s *discordgo.Session, m *discordgo.MessageCreate) {\n\tresult := \"Rolled: [\"\n\ttotal := b\n\tfor i := 0; i < n-1; i++ {\n\t\tval := rand.Intn(d) + 1\n\t\tresult += strconv.Itoa(val) + \", \"\n\t\ttotal += val\n\t}\n\tval := rand.Intn(d) + 1\n\tresult += strconv.Itoa(val)\n\ttotal += val\n\tif b > 0 {\n\t\tresult += \"] +\" + strconv.Itoa(b)\n\t} else if b == 0 {\n\t\tresult += \"]\"\n\t} else {\n\t\tresult += \"] \" + strconv.Itoa(b)\n\t}\n\n\tresult += \" = \" + strconv.Itoa(total)\n\ts.ChannelMessageSend(m.ChannelID, result)\n}", "func TestRollN(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Valid roll\", \"/d4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Valid roll\", \"/D4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Invalid variable\", \"/d5\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid variable\", \"/D5\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Valid query for count\", \"/d4?count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t\t{\"Valid query for count\", \"/D4?count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t\t{\"Invalid query for count\", \"/d4?count=0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid query for count\", \"/D4?count=0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t}\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func NewRoll(s *S) *Roll {\n\tu := &Roll{\n\t\tS: s,\n\t\tC: NewCCap(s.Len() * 10),\n\t\tdmap: make([][]z.Lit, s.Len())}\n\treturn u\n}", "func NewResultWin(c Color) Result {\n\tif c == White {\n\t\treturn WhiteWin\n\t} else if c == Black {\n\t\treturn BlackWin\n\t}\n\treturn Draw\n}", "func MakeResult(rowType reflect.Type) Result {\n\treturn Result{\n\t\trowType: rowType,\n\t}\n}", "func (r *AutoRoller) updateCurrentRoll() error {\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll == nil {\n\t\treturn nil\n\t}\n\tcurrentResult := currentRoll.Result\n\n\tupdated, err := r.retrieveRoll(currentRoll.Issue)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// We have to rely on data we store for the dry run case.\n\tif !updated.Closed && util.In(currentResult, autoroll.DRY_RUN_RESULTS) {\n\t\tupdated.Result = currentResult\n\t}\n\n\t// If the current roll succeeded, we need to make sure we update the\n\t// repo so that we see the roll commit. This can take some time, so\n\t// we have to repeatedly update until we see the commit.\n\tif updated.Committed {\n\t\tsklog.Infof(\"Roll succeeded (%d); syncing the repo until it lands.\", currentRoll.Issue)\n\t\tfor {\n\t\t\tsklog.Info(\"Syncing...\")\n\t\t\tsklog.Infof(\"Looking for %s\", currentRoll.RollingTo)\n\t\t\tif err := r.rm.ForceUpdate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif rolledPast {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttime.Sleep(10 * time.Second)\n\t\t}\n\t\tr.liveness.Reset()\n\t}\n\treturn r.recent.Update(updated)\n}", "func (future *SubAccountCreateFuture) result(client SubAccountClient) (mr MonitorResource, err error) {\n\tvar done bool\n\tdone, err = future.DoneWithContext(context.Background(), client)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"logz.SubAccountCreateFuture\", \"Result\", future.Response(), \"Polling failure\")\n\t\treturn\n\t}\n\tif !done {\n\t\tmr.Response.Response = future.Response()\n\t\terr = azure.NewAsyncOpIncompleteError(\"logz.SubAccountCreateFuture\")\n\t\treturn\n\t}\n\tsender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n\tif mr.Response.Response, err = future.GetResult(sender); err == nil && mr.Response.Response.StatusCode != http.StatusNoContent {\n\t\tmr, err = client.CreateResponder(mr.Response.Response)\n\t\tif err != nil {\n\t\t\terr = autorest.NewErrorWithError(err, \"logz.SubAccountCreateFuture\", \"Result\", mr.Response.Response, \"Failure responding to request\")\n\t\t}\n\t}\n\treturn\n}", "func (d Die) Roll() int {\n\treturn rand.Intn(int(d)) + 1\n}", "func TestRollDice(t *testing.T) {\n\tresult := Roll()\n\tfor i := 1; i == 20; i++ {\n\t\tif result > 6 || result < 0 {\n\t\t\tt.Errorf(\"the rolled result is %v\", result)\n\t\t}\n\t}\n}", "func NewRetry(\n\tconf Config,\n\tmgr types.Manager,\n\tlog log.Modular,\n\tstats metrics.Type,\n) (Type, error) {\n\tif conf.Retry.Output == nil {\n\t\treturn nil, errors.New(\"cannot create retry output without a child\")\n\t}\n\n\twrapped, err := New(*conf.Retry.Output, mgr, log, stats)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to create output '%v': %v\", conf.Retry.Output.Type, err)\n\t}\n\n\tvar boffCtor func() backoff.BackOff\n\tif boffCtor, err = conf.Retry.GetCtor(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Retry{\n\t\trunning: 1,\n\t\tconf: conf.Retry,\n\n\t\tlog: log,\n\t\tstats: stats,\n\t\twrapped: wrapped,\n\t\tbackoffCtor: boffCtor,\n\t\ttransactionsOut: make(chan types.Transaction),\n\n\t\tcloseChan: make(chan struct{}),\n\t\tclosedChan: make(chan struct{}),\n\t}, nil\n}", "func (fr *FakeResult) Close() {\n}", "func NewResultHandle(ctx context.Context, cc *client.Client, opt client.SolveOpt, product string, buildFunc gateway.BuildFunc, ch chan *client.SolveStatus) (*ResultHandle, *client.SolveResponse, error) {\n\t// Create a new context to wrap the original, and cancel it when the\n\t// caller-provided context is cancelled.\n\t//\n\t// We derive the context from the background context so that we can forbid\n\t// cancellation of the build request after <-done is closed (which we do\n\t// before returning the ResultHandle).\n\tbaseCtx := ctx\n\tctx, cancel := context.WithCancelCause(context.Background())\n\tdone := make(chan struct{})\n\tgo func() {\n\t\tselect {\n\t\tcase <-baseCtx.Done():\n\t\t\tcancel(baseCtx.Err())\n\t\tcase <-done:\n\t\t\t// Once done is closed, we've recorded a ResultHandle, so we\n\t\t\t// shouldn't allow cancelling the underlying build request anymore.\n\t\t}\n\t}()\n\n\t// Create a new channel to forward status messages to the original.\n\t//\n\t// We do this so that we can discard status messages after the main portion\n\t// of the build is complete. This is necessary for the solve error case,\n\t// where the original gateway is kept open until the ResultHandle is\n\t// closed - we don't want progress messages from operations in that\n\t// ResultHandle to display after this function exits.\n\t//\n\t// Additionally, callers should wait for the progress channel to be closed.\n\t// If we keep the session open and never close the progress channel, the\n\t// caller will likely hang.\n\tbaseCh := ch\n\tch = make(chan *client.SolveStatus)\n\tgo func() {\n\t\tfor {\n\t\t\ts, ok := <-ch\n\t\t\tif !ok {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tselect {\n\t\t\tcase <-baseCh:\n\t\t\t\t// base channel is closed, discard status messages\n\t\t\tdefault:\n\t\t\t\tbaseCh <- s\n\t\t\t}\n\t\t}\n\t}()\n\tdefer close(baseCh)\n\n\tvar resp *client.SolveResponse\n\tvar respErr error\n\tvar respHandle *ResultHandle\n\n\tgo func() {\n\t\tdefer cancel(context.Canceled) // ensure no dangling processes\n\n\t\tvar res *gateway.Result\n\t\tvar err error\n\t\tresp, err = cc.Build(ctx, opt, product, func(ctx context.Context, c gateway.Client) (*gateway.Result, error) {\n\t\t\tvar err error\n\t\t\tres, err = buildFunc(ctx, c)\n\n\t\t\tif res != nil && err == nil {\n\t\t\t\t// Force evaluation of the build result (otherwise, we likely\n\t\t\t\t// won't get a solve error)\n\t\t\t\tdef, err2 := getDefinition(ctx, res)\n\t\t\t\tif err2 != nil {\n\t\t\t\t\treturn nil, err2\n\t\t\t\t}\n\t\t\t\tres, err = evalDefinition(ctx, c, def)\n\t\t\t}\n\n\t\t\tif err != nil {\n\t\t\t\t// Scenario 1: we failed to evaluate a node somewhere in the\n\t\t\t\t// build graph.\n\t\t\t\t//\n\t\t\t\t// In this case, we construct a ResultHandle from this\n\t\t\t\t// original Build session, and return it alongside the original\n\t\t\t\t// build error. We then need to keep the gateway session open\n\t\t\t\t// until the caller explicitly closes the ResultHandle.\n\n\t\t\t\tvar se *errdefs.SolveError\n\t\t\t\tif errors.As(err, &se) {\n\t\t\t\t\trespHandle = &ResultHandle{\n\t\t\t\t\t\tdone: make(chan struct{}),\n\t\t\t\t\t\tsolveErr: se,\n\t\t\t\t\t\tgwClient: c,\n\t\t\t\t\t\tgwCtx: ctx,\n\t\t\t\t\t}\n\t\t\t\t\trespErr = se\n\t\t\t\t\tclose(done)\n\n\t\t\t\t\t// Block until the caller closes the ResultHandle.\n\t\t\t\t\tselect {\n\t\t\t\t\tcase <-respHandle.done:\n\t\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn res, err\n\t\t}, ch)\n\t\tif respHandle != nil {\n\t\t\treturn\n\t\t}\n\t\tif err != nil {\n\t\t\t// Something unexpected failed during the build, we didn't succeed,\n\t\t\t// but we also didn't make it far enough to create a ResultHandle.\n\t\t\trespErr = err\n\t\t\tclose(done)\n\t\t\treturn\n\t\t}\n\n\t\t// Scenario 2: we successfully built the image with no errors.\n\t\t//\n\t\t// In this case, the original gateway session has now been closed\n\t\t// since the Build has been completed. So, we need to create a new\n\t\t// gateway session to populate the ResultHandle. To do this, we\n\t\t// need to re-evaluate the target result, in this new session. This\n\t\t// should be instantaneous since the result should be cached.\n\n\t\tdef, err := getDefinition(ctx, res)\n\t\tif err != nil {\n\t\t\trespErr = err\n\t\t\tclose(done)\n\t\t\treturn\n\t\t}\n\n\t\t// NOTE: ideally this second connection should be lazily opened\n\t\topt := opt\n\t\topt.Ref = \"\"\n\t\topt.Exports = nil\n\t\topt.CacheExports = nil\n\t\topt.Internal = true\n\t\t_, respErr = cc.Build(ctx, opt, \"buildx\", func(ctx context.Context, c gateway.Client) (*gateway.Result, error) {\n\t\t\tres, err := evalDefinition(ctx, c, def)\n\t\t\tif err != nil {\n\t\t\t\t// This should probably not happen, since we've previously\n\t\t\t\t// successfully evaluated the same result with no issues.\n\t\t\t\treturn nil, errors.Wrap(err, \"inconsistent solve result\")\n\t\t\t}\n\t\t\trespHandle = &ResultHandle{\n\t\t\t\tdone: make(chan struct{}),\n\t\t\t\tres: res,\n\t\t\t\tgwClient: c,\n\t\t\t\tgwCtx: ctx,\n\t\t\t}\n\t\t\tclose(done)\n\n\t\t\t// Block until the caller closes the ResultHandle.\n\t\t\tselect {\n\t\t\tcase <-respHandle.done:\n\t\t\tcase <-ctx.Done():\n\t\t\t}\n\t\t\treturn nil, ctx.Err()\n\t\t}, nil)\n\t\tif respHandle != nil {\n\t\t\treturn\n\t\t}\n\t\tclose(done)\n\t}()\n\n\t// Block until the other thread signals that it's completed the build.\n\tselect {\n\tcase <-done:\n\tcase <-baseCtx.Done():\n\t\tif respErr == nil {\n\t\t\trespErr = baseCtx.Err()\n\t\t}\n\t}\n\treturn respHandle, resp, respErr\n}", "func response(w http.ResponseWriter, s, c int) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\tresult, err := rollDice(s, c)\n\tif err != nil {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse := rollResponse{c, s, result}\n\tw.WriteHeader(http.StatusOK)\n\tenc := json.NewEncoder(w)\n\tjsonEncode(w, enc, response)\n}", "func (future *SingleSignOnCreateOrUpdateFuture) result(client SingleSignOnClient) (ssor SingleSignOnResource, err error) {\n\tvar done bool\n\tdone, err = future.DoneWithContext(context.Background(), client)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"logz.SingleSignOnCreateOrUpdateFuture\", \"Result\", future.Response(), \"Polling failure\")\n\t\treturn\n\t}\n\tif !done {\n\t\tssor.Response.Response = future.Response()\n\t\terr = azure.NewAsyncOpIncompleteError(\"logz.SingleSignOnCreateOrUpdateFuture\")\n\t\treturn\n\t}\n\tsender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n\tif ssor.Response.Response, err = future.GetResult(sender); err == nil && ssor.Response.Response.StatusCode != http.StatusNoContent {\n\t\tssor, err = client.CreateOrUpdateResponder(ssor.Response.Response)\n\t\tif err != nil {\n\t\t\terr = autorest.NewErrorWithError(err, \"logz.SingleSignOnCreateOrUpdateFuture\", \"Result\", ssor.Response.Response, \"Failure responding to request\")\n\t\t}\n\t}\n\treturn\n}", "func (sdeep *SSDEEP) rollHash(c byte) uint32 {\n\trs := &sdeep.rollingState\n\trs.h2 -= rs.h1\n\trs.h2 += rollingWindow * uint32(c)\n\trs.h1 += uint32(c)\n\trs.h1 -= uint32(rs.window[rs.n])\n\trs.window[rs.n] = c\n\trs.n++\n\tif rs.n == rollingWindow {\n\t\trs.n = 0\n\t}\n\trs.h3 = rs.h3 << 5\n\trs.h3 ^= uint32(c)\n\treturn rs.h1 + rs.h2 + rs.h3\n}", "func (o ApplicationStatusOperationStateOperationOutput) Retry() ApplicationStatusOperationStateOperationRetryPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperation) *ApplicationStatusOperationStateOperationRetry {\n\t\treturn v.Retry\n\t}).(ApplicationStatusOperationStateOperationRetryPtrOutput)\n}", "func newResultState(sharedConfig jsonio.GoldResults, config *GoldClientConfig) *resultState {\n\tgoldURL := config.OverrideGoldURL\n\tif goldURL == \"\" {\n\t\tgoldURL = getGoldInstanceURL(config.InstanceID)\n\t}\n\tbucket := config.OverrideBucket\n\tif bucket == \"\" {\n\t\tbucket = getBucket(config.InstanceID)\n\t}\n\n\tret := &resultState{\n\t\tSharedConfig: sharedConfig,\n\t\tPerTestPassFail: config.PassFailStep,\n\t\tFailureFile: config.FailureFile,\n\t\tInstanceID: config.InstanceID,\n\t\tUploadOnly: config.UploadOnly,\n\t\tGoldURL: goldURL,\n\t\tBucket: bucket,\n\t}\n\n\treturn ret\n}", "func MakeResTransaction(creator utils.Addr,key utils.Key,\n inps []TrInput,out TrOutput,\n jobblock,jobtrans,hashsol string,\n evaluation float64,isMin bool)*ResTransaction{\n tr := new(ResTransaction)\n tr.Timestamp = time.Now()\n tr.Output = out\n tr.Inputs = inps\n tr.Creator = creator\n tr.JobBlock = jobblock\n tr.JobTrans = jobtrans\n tr.HashSol = hashsol\n tr.Evaluation = evaluation\n tr.IsMin = isMin\n tr.Hash = tr.GetHash()\n tr.Signature = fmt.Sprintf(\"%x\",utils.GetSignatureFromHash(tr.Hash,key))\n return tr\n}", "func (r *AutoRoller) doAutoRoll() error {\n\tstatus, lastError := r.doAutoRollInner()\n\n\tlastErrorStr := \"\"\n\tif lastError != nil {\n\t\tlastErrorStr = lastError.Error()\n\t}\n\n\tgerritUrl := r.gerrit.Url(0)\n\n\t// Update status information.\n\tif err := r.status.set(&AutoRollStatus{\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tGerritUrl: gerritUrl,\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: r.recent.GetRecentRolls(),\n\t\tStatus: status,\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\treturn lastError\n}", "func MakeResult(host, ip string, ports ...int) Result {\n\tp := make([]Port, len(ports))\n\tfor i := 0; i < len(ports); i++ {\n\t\tport := Port{ports[i], portdes.GetPortDescription(ports[i])}\n\t\tp[i] = port\n\t}\n\n\treturn Result{\n\t\tHost: host,\n\t\tIP: ip,\n\t\tPorts: p,\n\t\tFinished: true,\n\t}\n}", "func (fl *FileLogger) Roll() error {\n\tif e := fl.File.Close(); e != nil {\n\t\tfmt.Printf(\"fileLog: error closing file %v\\n\", e)\n\t}\n\n\tif e := os.Rename(fmt.Sprintf(fl.string, 1), fmt.Sprintf(fl.string, 2)); e != nil {\n\t\t// we can't rename, there is little point in try truncating\n\t\tfmt.Printf(\"filelog: rename failed, attempting to truncate current %v\\n\", e)\n\t}\n\n\tf, e := os.Create(fmt.Sprintf(fl.string, 1))\n\tif e != nil {\n\t\tfl.File = nil\n\t\tfmt.Printf(\"filelog: opening new log file failed %v\\n\", e)\n\t\treturn e\n\t}\n\n\tfl.File = f\n\treturn nil\n}", "func Roll(w http.ResponseWriter, r *http.Request) {\n\tsides := r.FormValue(\"sides\")\n\tcount := r.FormValue(\"count\")\n\n\ts := getSides(sides)\n\tif s == 0 {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tc := getCount(count)\n\tif c == 0 {\n\t\tcountErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse(w, s, c)\n}", "func (d *Die) Roll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\n\t// Check if rolled too many times already\n\tvar maxRolls = int64(MaxRolls)\n\tif ctxMaxRolls, ok := ctx.Value(CtxKeyMaxRolls).(int64); ok {\n\t\tmaxRolls = ctxMaxRolls\n\t}\n\n\tif *CtxTotalRolls(ctx) >= uint64(maxRolls) {\n\t\treturn ErrMaxRolls\n\t}\n\n\t// bump context roll count\n\tatomic.AddUint64(CtxTotalRolls(ctx), 1)\n\n\tif d.Size == 0 {\n\t\td.Result = NewResult(0)\n\t\treturn nil\n\t}\n\n\tswitch d.Type {\n\tcase TypeFudge:\n\t\td.Result = NewResult(float64(Source.Intn(int(d.Size*2+1)) - int(d.Size)))\n\t\tif d.Result.Value == -float64(d.Size) {\n\t\t\td.CritFailure = true\n\t\t}\n\tdefault:\n\t\td.Result = NewResult(float64(1 + Source.Intn(int(d.Size))))\n\t\tif d.Result.Value == 1 {\n\t\t\td.CritFailure = true\n\t\t}\n\t}\n\t// default critical success on max roll; override via modifiers\n\tif d.Result.Value == float64(d.Size) {\n\t\td.CritSuccess = true\n\t}\n\n\treturn nil\n}", "func newRunner(output string, err error) *MockRunner {\n\tm := &MockRunner{}\n\tm.On(\"Run\", mock.Anything).Return([]byte(output), err)\n\treturn m\n}", "func (d Dice) Roll() int {\n\td.m.Lock()\n\tdefer d.m.Unlock()\n\treturn util.RandRange(d.r, d.min, d.max)\n}", "func (d *RabinKarp64) Roll(c byte) {\n\t// This check costs 10-15% performance. If we disable it, we crash\n\t// when the window is empty. If we enable it, we are always correct\n\t// (an empty window never changes no matter how much you roll it).\n\t//if len(d.window) == 0 {\n\t//\treturn\n\t//}\n\t// extract the entering/leaving bytes and update the circular buffer.\n\tenter := c\n\tleave := uint64(d.window[d.oldest])\n\td.window[d.oldest] = c\n\td.oldest += 1\n\tif d.oldest >= len(d.window) {\n\t\td.oldest = 0\n\t}\n\n\td.value ^= d.tables.out[leave]\n\tindex := byte(d.value >> d.polShift)\n\td.value <<= 8\n\td.value |= Pol(enter)\n\td.value ^= d.tables.mod[index]\n}", "func (as *activeRuleSet) toRollupResults(\n\tid []byte,\n\tcutoverNanos int64,\n\ttargets []rollupTarget,\n\tkeepOriginal bool,\n\ttags [][]models.Tag,\n\tmatchOpts MatchOptions,\n) (rollupResults, error) {\n\tif len(targets) == 0 {\n\t\treturn rollupResults{}, nil\n\t}\n\n\t// If we cannot extract tags from the id, this is likely an invalid\n\t// metric and we bail early.\n\t_, sortedTagPairBytes, err := matchOpts.NameAndTagsFn(id)\n\tif err != nil {\n\t\treturn rollupResults{}, err\n\t}\n\n\tvar (\n\t\tmultiErr = xerrors.NewMultiError()\n\t\tpipelines = make([]metadata.PipelineMetadata, 0, len(targets))\n\t\tnewRollupIDResults = make([]idWithMatchResults, 0, len(targets))\n\t\ttagPairs []metricid.TagPair\n\t)\n\n\tfor idx, target := range targets {\n\t\tpipeline := target.Pipeline\n\t\t// A rollup target should always have a non-empty pipeline but\n\t\t// just being defensive here.\n\t\tif pipeline.IsEmpty() {\n\t\t\terr = fmt.Errorf(\"target %v has empty pipeline\", target)\n\t\t\tmultiErr = multiErr.Add(err)\n\t\t\tcontinue\n\t\t}\n\t\tvar (\n\t\t\taggregationID aggregation.ID\n\t\t\trollupID []byte\n\t\t\tnumSteps = pipeline.Len()\n\t\t\tfirstOp = pipeline.At(0)\n\t\t\ttoApply mpipeline.Pipeline\n\t\t)\n\t\tswitch firstOp.Type {\n\t\tcase mpipeline.AggregationOpType:\n\t\t\taggregationID, err = aggregation.CompressTypes(firstOp.Aggregation.Type)\n\t\t\tif err != nil {\n\t\t\t\terr = fmt.Errorf(\"target %v operation 0 aggregation type compression error: %v\", target, err)\n\t\t\t\tmultiErr = multiErr.Add(err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttoApply = pipeline.SubPipeline(1, numSteps)\n\t\tcase mpipeline.TransformationOpType:\n\t\t\taggregationID = aggregation.DefaultID\n\t\t\ttoApply = pipeline\n\t\tcase mpipeline.RollupOpType:\n\t\t\ttagPairs = tagPairs[:0]\n\t\t\tvar matched bool\n\t\t\trollupID, matched, err = as.matchRollupTarget(\n\t\t\t\tsortedTagPairBytes,\n\t\t\t\tfirstOp.Rollup,\n\t\t\t\ttagPairs,\n\t\t\t\ttags[idx],\n\t\t\t\tmatchRollupTargetOptions{generateRollupID: true},\n\t\t\t\tmatchOpts)\n\t\t\tif err != nil {\n\t\t\t\tmultiErr = multiErr.Add(err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif !matched {\n\t\t\t\t// The incoming metric ID did not match the rollup target.\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\taggregationID = firstOp.Rollup.AggregationID\n\t\t\ttoApply = pipeline.SubPipeline(1, numSteps)\n\t\tdefault:\n\t\t\terr = fmt.Errorf(\"target %v operation 0 has unknown type: %v\", target, firstOp.Type)\n\t\t\tmultiErr = multiErr.Add(err)\n\t\t\tcontinue\n\t\t}\n\t\ttagPairs = tagPairs[:0]\n\t\tapplied, err := as.applyIDToPipeline(sortedTagPairBytes, toApply, tagPairs, tags[idx], matchOpts)\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"failed to apply id %s to pipeline %v: %v\", id, toApply, err)\n\t\t\tmultiErr = multiErr.Add(err)\n\t\t\tcontinue\n\t\t}\n\t\tnewPipeline := metadata.PipelineMetadata{\n\t\t\tAggregationID: aggregationID,\n\t\t\tStoragePolicies: target.StoragePolicies,\n\t\t\tPipeline: applied,\n\t\t\tResendEnabled: target.ResendEnabled,\n\t\t}\n\t\tif rollupID == nil {\n\t\t\t// The applied pipeline applies to the incoming ID.\n\t\t\tpipelines = append(pipelines, newPipeline)\n\t\t} else {\n\t\t\tif len(tags[idx]) > 0 {\n\t\t\t\tnewPipeline.Tags = tags[idx]\n\t\t\t}\n\t\t\t// The applied pipeline applies to a new rollup ID.\n\t\t\tmatchResults := ruleMatchResults{\n\t\t\t\tcutoverNanos: cutoverNanos,\n\t\t\t\tpipelines: []metadata.PipelineMetadata{newPipeline},\n\t\t\t}\n\t\t\tnewRollupIDResult := idWithMatchResults{id: rollupID, matchResults: matchResults}\n\t\t\tnewRollupIDResults = append(newRollupIDResults, newRollupIDResult)\n\t\t}\n\t}\n\n\treturn rollupResults{\n\t\tforExistingID: ruleMatchResults{cutoverNanos: cutoverNanos, pipelines: pipelines},\n\t\tforNewRollupIDs: newRollupIDResults,\n\t\tkeepOriginal: keepOriginal,\n\t}, multiErr.FinalError()\n}", "func (e *engineImpl) rollSM(c context.Context, job *CronJob, cb func(*StateMachine) error) error {\n\tsched, err := job.parseSchedule()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"bad schedule %q - %s\", job.effectiveSchedule(), err)\n\t}\n\tnow := clock.Now(c).UTC()\n\trnd := mathrand.Get(c)\n\tsm := StateMachine{\n\t\tState: job.State,\n\t\tNow: now,\n\t\tSchedule: sched,\n\t\tNonce: func() int64 { return rnd.Int63() + 1 },\n\t\tContext: c,\n\t}\n\t// All errors returned by state machine transition changes are transient.\n\t// Fatal errors (when we have them) should be reflected as a state changing\n\t// into \"BROKEN\" state.\n\tif err := cb(&sm); err != nil {\n\t\treturn errors.WrapTransient(err)\n\t}\n\tif len(sm.Actions) != 0 {\n\t\tif err := e.enqueueActions(c, job.JobID, sm.Actions); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif sm.State.State != job.State.State {\n\t\tlogging.Infof(c, \"%s -> %s\", job.State.State, sm.State.State)\n\t}\n\tjob.State = sm.State\n\treturn nil\n}", "func handleResults(result hooks.InternalMessage) {\n\tres, ok := result.Results.(TableRow)\n\tmanager.Status.AddCurrentScans(-1)\n\n\tif !ok {\n\t\tmanager.Logger.Errorf(\"Couldn't assert type of result for %v\", result.Domain.DomainName)\n\t\tres = TableRow{}\n\t\tresult.StatusCode = hooks.InternalFatalError\n\t}\n\n\tswitch result.StatusCode {\n\tcase hooks.InternalFatalError:\n\t\tres.ScanStatus = hooks.StatusError\n\t\tmanager.Status.AddFatalErrorScans(1)\n\t\tmanager.Logger.Infof(\"Assessment of %v failed ultimately\", result.Domain.DomainName)\n\tcase hooks.InternalSuccess:\n\t\tres.ScanStatus = hooks.StatusDone\n\t\tmanager.Logger.Debugf(\"Assessment of %v was successful\", result.Domain.DomainName)\n\t\tmanager.Status.AddFinishedScans(1)\n\t}\n\twhere := hooks.ScanWhereCond{\n\t\tDomainID: result.Domain.DomainID,\n\t\tScanID: manager.ScanID,\n\t\tTestWithSSL: result.Domain.TestWithSSL}\n\terr := backend.SaveResults(manager.GetTableName(), structs.New(where), structs.New(res))\n\tif err != nil {\n\t\tmanager.Logger.Errorf(\"Couldn't save results for %v: %v\", result.Domain.DomainName, err)\n\t\treturn\n\t}\n\tmanager.Logger.Debugf(\"Results for %v saved\", result.Domain.DomainName)\n\n}", "func RollDice(dice string) (string, error) {\n\t//This block of code turns a command line argument into two ints, one being the number of dice thrown, and the other being the type of dice\n\tclearFlags()\n\trollType, diceSlice, err := inputProofer(dice)\n\t//fmt.Println(diceSlice)\n\terr = assignMeaningToDiceSlice(diceSlice, rollType)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif numDice > 1000 {\n\t\treturn \"That's too many dice\", nil\n\t}\n\tif typeDice > 1000 {\n\t\treturn \"Golf balls are not dice\", nil\n\t}\n\tif numDice+typeDice > 1000 {\n\t\treturn \"Fuck you\", nil\n\t}\n\t//Create a source of random numbers seeded to the current time\n\n\tvar results []int\n\tresults, total, successes, err = rollDice(rollType, roundFlag)\n\tif err != nil {\n\t\treturn \"DAMNIT\", err\n\t}\n\treturn formatResults(results, rollType), nil\n\t//fmt.Println(results)\n\t//fmt.Println(total)\n\t//fmt.Println(successes)\n}", "func (gs *GameState) Roll(player xid.ID) (*Roll, error) {\n\tstonk := gs.Stonks[rand.Intn(len(gs.Stonks))]\n\taction := PlayerAction(rand.Intn(3))\n\tmovement := allowedMovements[rand.Intn(len(allowedMovements))]\n\n\treturn &Roll{\n\t\tID: xid.New(),\n\t\tPlayer: player,\n\t\tStonk: stonk.ID,\n\t\tAction: action,\n\t\tValue: movement,\n\t}, nil\n}", "func (dwr *DifferentialWheeledRobot) RollPosition(distLeft, distRight float64, prev Position) Position {\n\n\t// Straight line\n\tif distLeft == distRight {\n\t\treturn Position{\n\t\t\tprev.X + distLeft*math.Cos(prev.Theta),\n\t\t\tprev.Y + distLeft*math.Sin(prev.Theta),\n\t\t\tprev.Theta,\n\t\t}\n\t}\n\n\t// Turning\n\tturnRadius := dwr.BaseWidth * (distRight + distLeft) / (2 * (distRight - distLeft))\n\tangle := (distRight-distLeft)/dwr.BaseWidth + prev.Theta\n\treturn Position{\n\t\tprev.X + turnRadius*(math.Sin(angle)-math.Sin(prev.Theta)),\n\t\tprev.Y - turnRadius*(math.Cos(angle)-math.Cos(prev.Theta)),\n\t\tangle,\n\t}\n\n\t// s := (distLeft + distRight) / 2.0\n\t// theta := (distRight-distLeft)/dwr.BaseWidth + prev.Theta\n\t// x := s*math.Cos(theta) + prev.X\n\t// y := s*math.Sin(theta) + prev.Y\n\n\t// return Position{x, y, theta}\n\n}", "func rollDice(sides, count int) (result int, err error) {\n\td := chooseDice(sides)\n\tif d == nil {\n\t\treturn 0, invalidDice{}\n\t}\n\n\tfor i := 0; i < count; i++ {\n\t\tresult += d()\n\t}\n\n\treturn result, nil\n}", "func (o ApplicationOperationOutput) Retry() ApplicationOperationRetryPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperation) *ApplicationOperationRetry { return v.Retry }).(ApplicationOperationRetryPtrOutput)\n}", "func makeRole(t *testing.T, setup RoleSetup, numStages int) (r role) {\n\tt.Helper()\n\tr = role{\n\t\tchans: &channelMap{entries: make(map[channel.ID]*paymentChannel)},\n\t\tsetup: setup,\n\t\ttimeout: setup.Timeout,\n\t\terrs: setup.Errors,\n\t\tnumStages: numStages,\n\t\tchallengeDuration: setup.ChallengeDuration,\n\t}\n\tcl, err := client.New(r.setup.Identity.Address(),\n\t\tr.setup.Bus, r.setup.Funder, r.setup.Adjudicator, r.setup.Wallet, r.setup.Watcher)\n\tif err != nil {\n\t\tt.Fatal(\"Error creating client: \", err)\n\t}\n\tr.setClient(cl) // init client\n\treturn r\n}", "func createprogressReader(reader *io.ReadCloser, len int64) *progressReader {\n\tret := &progressReader{reader, len, 0, time.Now()}\n\treturn ret\n}", "func (r Result) Drop(n int, hl MatchType) Result {\n\tout := Result{die: r.die}\n\n\t// And here.\n\tif n < 1 || n > len(r.rolls) {\n\t\tout.rolls = r.rolls\n\t\treturn out\n\t}\n\n\tsort.Sort(r)\n\tswitch hl {\n\tcase HIGH:\n\t\tout.rolls = r.rolls[:len(r.rolls)-n]\n\tcase LOW:\n\t\tout.rolls = r.rolls[n:]\n\t}\n\n\treturn out\n}", "func (h *HeadResolver) Roll(ctx context.Context) (float64, error) {\n\treturn float64(h.Head.Roll), nil\n}", "func newRollText(ctx context.Context) (*text.Text, error) {\n\tt, err := text.New(text.RollContent())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn t, nil\n}", "func WithRollType(typ RollType) Option {\n\treturn func(opts *options) {\n\t\topts.rollType = typ\n\t}\n}", "func RollDiceRedo(dice string) (string, int, int, error) {\n\terr := clearFlags()\n\tif err != nil {\n\t\tfmt.Println(\"error clearing flags\")\n\t}\n\tcommand, flags, err := sFlags.CreateFlags(dice)\n\tparseFlags(flags)\n\trollType, diceSlice, err := inputProofer(command)\n\terr = assignMeaningToDiceSlice(diceSlice, rollType)\n\tif err != nil {\n\t\treturn \"\", 0, 0, err\n\t}\n\tif numDice > 1000 {\n\t\treturn \"That's too many dice\", 0, 0, nil\n\t}\n\tif typeDice > 1000 {\n\t\treturn \"Golf balls are not dice\", 0, 0, nil\n\t}\n\tif numDice+typeDice > 1000 {\n\t\treturn \"Fuck you\", 0, 0, nil\n\t}\n\tvar results []int\n\tresults, total, successes, err = rollDice(rollType, roundFlag)\n\tif err != nil {\n\t\treturn \"DAMNIT\", 0, 0, err\n\t}\n\treturn formatResults(results, rollType), successes, total, err\n}", "func (r Result) Len() int { return len(r.rolls) }", "func OpenShardedRDB(dirs []string, lldirs []string,\n\tbatched bool, check bool, kvf kvFactory) (*ShardedRDB, error) {\n\tshards := make([]*rdb, 0)\n\tif batched {\n\t\tplog.Infof(\"using batched ShardedRDB\")\n\t} else {\n\t\tplog.Infof(\"using plain ShardedRDB\")\n\t}\n\tif check && batched {\n\t\tpanic(\"check && batched both set to true\")\n\t}\n\tvar err error\n\tif check {\n\t\tplog.Infof(\"checking all LogDB shards...\")\n\t\tbatched, err = checkAllShards(dirs, lldirs, kvf)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tplog.Infof(\"all shards checked, batched: %t\", batched)\n\t}\n\tfor i := uint64(0); i < numOfRocksDBInstance; i++ {\n\t\tdir := filepath.Join(dirs[i], fmt.Sprintf(\"logdb-%d\", i))\n\t\tlldir := \"\"\n\t\tif len(lldirs) > 0 {\n\t\t\tlldir = filepath.Join(lldirs[i], fmt.Sprintf(\"logdb-%d\", i))\n\t\t}\n\t\tdb, err := openRDB(dir, lldir, batched, kvf)\n\t\tif err != nil {\n\t\t\tfor _, s := range shards {\n\t\t\t\ts.close()\n\t\t\t}\n\t\t\treturn nil, err\n\t\t}\n\t\tshards = append(shards, db)\n\t}\n\tpartitioner := server.NewDoubleFixedPartitioner(numOfRocksDBInstance,\n\t\tnumOfStepEngineWorker)\n\tmw := &ShardedRDB{\n\t\tshards: shards,\n\t\tpartitioner: partitioner,\n\t\tcompactions: newCompactions(),\n\t\tcompactionCh: make(chan struct{}, 1),\n\t\tstopper: syncutil.NewStopper(),\n\t}\n\tmw.stopper.RunWorker(func() {\n\t\tmw.compactionWorkerMain()\n\t})\n\treturn mw, nil\n}", "func (o *CreateRuleWaiverReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewCreateRuleWaiverCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewCreateRuleWaiverBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 401:\n\t\tresult := NewCreateRuleWaiverUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewCreateRuleWaiverForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewCreateRuleWaiverInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func RollDie() (int, error) {\n\tr, err := rand.Int(rand.Reader, sides)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(r.Int64()) + 1, nil\n}", "func (i *indexer) makeHoverResult(pkgs []*packages.Package, p *packages.Package, f *ast.File, obj types.Object) (string, error) {\n\tcontents, err := findContents(pkgs, p, f, obj)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"find contents: %v\", err)\n\t}\n\n\thoverResultID, err := i.w.EmitHoverResult(contents)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(`emit \"hoverResult\": %v`, err)\n\t}\n\n\treturn hoverResultID, nil\n}", "func NewRoller() *Roller {\n\troll := &Roller{\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t\t[]Callable{},\n\t}\n\treturn roll\n}", "func dispose(res *http.Response, err error) (*http.Response, error) {\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\tlog.Printf(\"[INFO] response: %d (%s)\", res.StatusCode, res.Status)\n\tvar buf bytes.Buffer\n\tif _, err := io.Copy(&buf, res.Body); err != nil {\n\t\tlog.Printf(\"[ERR] response: error copying response body\")\n\t} else {\n\t\tlog.Printf(\"[DEBUG] response: %s\", buf.String())\n\t\tres.Body.Close()\n\t\tres.Body = &bytesReadCloser{&buf}\n\t}\n\n\tswitch res.StatusCode {\n\tcase 200:\n\t\treturn res, nil\n\tcase 201:\n\t\treturn res, nil\n\tcase 202:\n\t\treturn res, nil\n\tcase 204:\n\t\treturn res, nil\n\tcase 400:\n\t\treturn nil, parseErr(res)\n\tcase 401:\n\t\treturn nil, fmt.Errorf(\"authentication failed\")\n\tcase 404:\n\t\treturn nil, fmt.Errorf(\"resource not found\")\n\tcase 422:\n\t\treturn nil, parseErr(res)\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"client: %s\", res.Status)\n\t}\n}", "func (mct *MinorCalcTask) publishResult(result float64) {\n\t*mct.ResultsHolder = result\n}", "func (c *DataChannel) makeReader(ctx context.Context, id clientID) *dataReader {\n\tvar m map[string]*dataReader\n\tvar ok bool\n\tif m, ok = c.readers[id.instID]; !ok {\n\t\tm = make(map[string]*dataReader)\n\t\tc.readers[id.instID] = m\n\t}\n\n\tif r, ok := m[id.ptransformID]; ok {\n\t\treturn r\n\t}\n\n\tr := &dataReader{id: id, buf: make(chan []byte, bufElements), done: make(chan bool, 1), channel: c}\n\n\t// Just in case initial data for an instruction arrives *after* an instructon has ended.\n\t// eg. it was blocked by another reader being slow, or the other instruction failed.\n\t// So we provide a pre-completed reader, and do not cache it, as there's no further cleanup for it.\n\tif _, ok := c.endedInstructions[id.instID]; ok {\n\t\tr.completed = true\n\t\tclose(r.buf)\n\t\tr.err = io.EOF // In case of any actual data readers, so they terminate without error.\n\t\treturn r\n\t}\n\n\tm[id.ptransformID] = r\n\treturn r\n}", "func (me XsdGoPkgHasElem_Roll) RollDefault() Tangle180Type {\r\n\tvar x = new(Tangle180Type)\r\n\tx.Set(\"0.0\")\r\n\treturn *x\r\n}", "func TestDRollN(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Valid request\", \"/d4/1\", response{http.StatusOK, `\"count\":1,\"sides\":4`}},\n\t\t{\"Valid request\", \"/D4/1\", response{http.StatusOK, `\"count\":1,\"sides\":4`}},\n\t\t{\"Invalid dice variable\", \"/d5/1\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid dice variable\", \"/D5/1\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid count variable\", \"/d4/0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t{\"Invalid count variable\", \"/D4/0\", response{http.StatusNotAcceptable, `\"error\"`}},\n\t\t// Not sure why count ocmes up first here, since the sides get parsed first in the code.\n\t\t// Moreover, not sure if I should be looking specifically for it or just for an error.\n\t\t{\"Invalid dice and count variable\", \"/d5/0\", response{http.StatusNotAcceptable, `\"error\":\"invalid count\"`}},\n\t\t{\"Invalid dice and count variable\", \"/D5/0\", response{http.StatusNotAcceptable, `\"error\":\"invalid count\"`}},\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func (future *MaintenanceWindowsCreateOrUpdateFuture) Result(client MaintenanceWindowsClient) (mw MaintenanceWindow, err error) {\n var done bool\n done, err = future.DoneWithContext(context.Background(), client)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"mysqlflexibleservers.MaintenanceWindowsCreateOrUpdateFuture\", \"Result\", future.Response(), \"Polling failure\")\n return\n }\n if !done {\n err = azure.NewAsyncOpIncompleteError(\"mysqlflexibleservers.MaintenanceWindowsCreateOrUpdateFuture\")\n return\n }\n sender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n if mw.Response.Response, err = future.GetResult(sender); err == nil && mw.Response.Response.StatusCode != http.StatusNoContent {\n mw, err = client.CreateOrUpdateResponder(mw.Response.Response)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"mysqlflexibleservers.MaintenanceWindowsCreateOrUpdateFuture\", \"Result\", mw.Response.Response, \"Failure responding to request\")\n }\n }\n return\n }", "func (d *Dice) Roll() int {\n\t// We named this call \"roll\" so we can easily mock the specific call out in our tests\n\treturn d.rand.Intn(\"roll\", d.sides)\n}", "func makeInvocation(j *schedulerJob, i *engine.Invocation) *invocation {\n\t// Invocations with Multistage == false trait are never in \"RUNNING\" state,\n\t// they perform all their work in 'LaunchTask' while technically being in\n\t// \"STARTING\" state. We display them as \"RUNNING\" instead. See comment for\n\t// task.Traits.Multistage for more info.\n\tstatus := i.Status\n\tif !j.traits.Multistage && status == task.StatusStarting {\n\t\tstatus = task.StatusRunning\n\t}\n\n\ttriggeredBy := \"-\"\n\tif i.TriggeredBy != \"\" {\n\t\ttriggeredBy = string(i.TriggeredBy)\n\t\tif i.TriggeredBy.Email() != \"\" {\n\t\t\ttriggeredBy = i.TriggeredBy.Email() // triggered by a user (not a service)\n\t\t}\n\t}\n\n\tfinished := i.Finished\n\tif finished.IsZero() {\n\t\tfinished = j.now\n\t}\n\tduration := humanize.RelTime(i.Started, finished, \"\", \"\")\n\tif duration == \"now\" {\n\t\tduration = \"1 second\" // \"now\" looks weird for durations\n\t}\n\n\treturn &invocation{\n\t\tProjectID: j.ProjectID,\n\t\tJobName: j.JobName,\n\t\tInvID: i.ID,\n\t\tAttempt: i.RetryCount + 1,\n\t\tRevision: i.Revision,\n\t\tRevisionURL: i.RevisionURL,\n\t\tDefinition: taskToText(i.Task),\n\t\tTriggeredBy: triggeredBy,\n\t\tStarted: humanize.RelTime(i.Started, j.now, \"ago\", \"from now\"),\n\t\tDuration: duration,\n\t\tStatus: string(status),\n\t\tDebugLog: i.DebugLog,\n\t\tRowClass: statusToRowClass[status],\n\t\tLabelClass: statusToLabelClass[status],\n\t\tViewURL: i.ViewURL,\n\t}\n}", "func rollDie() int {\n\t//TODO test & fix range if needed (don't include 0)\n\t// hardcode for 6-sided die\n\treturn getRandNum(7)\n}", "func (r *RbacV1RoleRule) createRule(role *rbacV1.Role, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(role)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(role)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(role)\n\t\t},\n\t}\n\treturn rule\n}", "func HandleResultsCreate(a API, s checks.API, w http.ResponseWriter, r *http.Request) {\n\tlogger := shared.GetLogger(a.Context())\n\n\tif AuthenticateUploader(a, r) != InternalUsername {\n\t\thttp.Error(w, \"This is a private API.\", http.StatusUnauthorized)\n\n\t\treturn\n\t}\n\tbody, err := io.ReadAll(r.Body)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\n\t\treturn\n\t}\n\n\tvar testRun shared.TestRun\n\tif err := json.Unmarshal(body, &testRun); err != nil {\n\t\thttp.Error(w, \"Failed to parse JSON: \"+err.Error(), http.StatusBadRequest)\n\n\t\treturn\n\t}\n\n\tif testRun.TimeStart.IsZero() {\n\t\ttestRun.TimeStart = time.Now()\n\t}\n\tif testRun.TimeEnd.IsZero() {\n\t\ttestRun.TimeEnd = testRun.TimeStart\n\t}\n\ttestRun.CreatedAt = time.Now()\n\n\t// nolint:staticcheck // TODO: Fix staticcheck lint error (SA1019).\n\tif len(testRun.FullRevisionHash) != 40 {\n\t\thttp.Error(w, \"full_revision_hash must be the full SHA (40 chars)\", http.StatusBadRequest)\n\n\t\treturn\n\t} else if testRun.Revision != \"\" && strings.Index(testRun.FullRevisionHash, testRun.Revision) != 0 {\n\t\thttp.Error(w,\n\t\t\tfmt.Sprintf(\n\t\t\t\t\"Mismatch of full_revision_hash and revision fields: %s vs %s\",\n\t\t\t\ttestRun.FullRevisionHash,\n\t\t\t\ttestRun.Revision,\n\t\t\t),\n\t\t\thttp.StatusBadRequest)\n\n\t\treturn\n\t}\n\t// nolint:staticcheck // TODO: Fix staticcheck lint error (SA1019).\n\ttestRun.Revision = testRun.FullRevisionHash[:10]\n\n\tkey, err := a.AddTestRun(&testRun)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\n\t\treturn\n\t}\n\t// Copy int64 representation of key into TestRun.ID so that clients can\n\t// inspect/use key value.\n\ttestRun.ID = key.IntID()\n\n\t// Do not schedule on pr_base to avoid redundancy with pr_head.\n\tif !testRun.LabelsSet().Contains(shared.PRBaseLabel) {\n\t\tspec := shared.ProductSpec{} // nolint:exhaustruct // TODO: Fix exhaustruct lint error\n\t\tspec.BrowserName = testRun.BrowserName\n\t\tspec.Labels = mapset.NewSet(testRun.Channel())\n\t\terr = s.ScheduleResultsProcessing(testRun.FullRevisionHash, spec)\n\t\tif err != nil {\n\t\t\tlogger.Warningf(\"Failed to schedule results: %s\", err.Error())\n\t\t}\n\t}\n\n\t// nolint:exhaustruct // TODO: Fix exhaustruct lint error.\n\tpendingRun := shared.PendingTestRun{\n\t\tID: testRun.ID,\n\t\tStage: shared.StageValid,\n\t\tProductAtRevision: testRun.ProductAtRevision,\n\t}\n\tif err := a.UpdatePendingTestRun(pendingRun); err != nil {\n\t\t// This is a non-fatal error; don't return.\n\t\tlogger.Errorf(\"Failed to update pending test run: %s\", err.Error())\n\t}\n\n\tjsonOutput, err := json.Marshal(testRun)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\n\t\treturn\n\t}\n\tlogger.Infof(\"Successfully created run %v (%s)\", testRun.ID, testRun.String())\n\tw.WriteHeader(http.StatusCreated)\n\t_, err = w.Write(jsonOutput)\n\tif err != nil {\n\t\tlogger.Warningf(\"Failed to write data in api/results/create handler: %s\", err.Error())\n\t}\n}", "func (r *Response) toResult(useSummaryResult bool) (result common.KYCResult, err error) {\n\tdetailsCreateIfNil := func(details **common.KYCDetails) {\n\t\tif *details == nil {\n\t\t\t*details = &common.KYCDetails{}\n\t\t}\n\t}\n\n\tswitch useSummaryResult {\n\tcase true:\n\t\tswitch r.SummaryResult.Key {\n\t\tcase Success:\n\t\t\tresult.Status = common.Approved\n\t\tcase Failure:\n\t\t\tresult.Status = common.Denied\n\t\tcase Partial:\n\t\t\tresult.Status = common.Unclear\n\t\t}\n\tcase false:\n\t\tswitch r.Results.Key {\n\t\tcase Match:\n\t\t\tresult.Status = common.Approved\n\t\tcase NoMatch, MatchRestricted:\n\t\t\tresult.Status = common.Denied\n\t\t}\n\t}\n\n\tif r.Restriction != nil {\n\t\tdetailsCreateIfNil(&result.Details)\n\t\tresult.Details.Reasons = []string{\n\t\t\tr.Restriction.Message,\n\t\t\tr.Restriction.PatriotAct.List,\n\t\t\tfmt.Sprintf(\"Patriot Act score: %d\", r.Restriction.PatriotAct.Score),\n\t\t}\n\t}\n\n\tif r.Qualifiers != nil {\n\t\tdetailsCreateIfNil(&result.Details)\n\t\tfor _, q := range r.Qualifiers.Qualifiers {\n\t\t\tresult.Details.Reasons = append(result.Details.Reasons, q.Message)\n\t\t}\n\t}\n\n\treturn\n}", "func (me XsdGoPkgHasElems_Roll) RollDefault() Tangle180Type {\r\n\tvar x = new(Tangle180Type)\r\n\tx.Set(\"0.0\")\r\n\treturn *x\r\n}", "func (gs *GameState) ApplyRoll(playerID xid.ID) error {\n\tif gs.Turn.Player == nil || *gs.Turn.Player != playerID {\n\t\treturn fmt.Errorf(\"You're not the rolling player\")\n\t}\n\troll := gs.NextRoll\n\tif roll == nil {\n\t\treturn fmt.Errorf(\"No roll ready\")\n\t}\n\n\tdefer gs.nextRoller()\n\n\tswitch roll.Action {\n\tcase ActionUp, ActionDown:\n\t\treturn gs.applyStockMove(roll)\n\tcase ActionDividend:\n\t\treturn gs.applyDividend(roll)\n\tdefault:\n\t\treturn fmt.Errorf(\"Unknown PlayerAction %v\", roll.Action)\n\t}\n}", "func (r Results) Reroll() Results {\n\tvar results Results\n\n\tfor _, result := range r {\n\t\tresults = append(results, result.Reroll())\n\t}\n\n\treturn results\n}", "func makeLogsOutput(ch chan plog.Logs) *otelcol.ConsumerArguments {\n\tlogsConsumer := fakeconsumer.Consumer{\n\t\tConsumeLogsFunc: func(ctx context.Context, l plog.Logs) error {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\treturn ctx.Err()\n\t\t\tcase ch <- l:\n\t\t\t\treturn nil\n\t\t\t}\n\t\t},\n\t}\n\n\treturn &otelcol.ConsumerArguments{\n\t\tLogs: []otelcol.Consumer{&logsConsumer},\n\t}\n}", "func RollN(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\tsides := vars[\"sides\"]\n\tcount := r.FormValue(\"count\")\n\n\ts := getSides(sides[1:])\n\tif s == 0 {\n\t\tsidesErrResponse(w)\n\t\treturn\n\t}\n\n\tc := getCount(count)\n\tif c == 0 {\n\t\tcountErrResponse(w)\n\t\treturn\n\t}\n\n\tresponse(w, s, c)\n}", "func (db *diceBae) LetsRoll() error {\n\tif err := db.session.Open(); err != nil {\n\t\treturn fmt.Errorf(\"failed to open Discord session: %v\", err)\n\t}\n\tdefer db.session.Close()\n\tdefer db.logFile.Close()\n\n\tdb.LogInfo(\"I have no dice, but I must roll. Press CTRL-C to exit.\")\n\tsc := make(chan os.Signal, 1)\n\t// Block on this channel until we get a termination signal.\n\tsignal.Notify(sc, syscall.SIGINT, syscall.SIGTERM, os.Interrupt, os.Kill)\n\t<-sc\n\tdb.LogInfo(\"Later dopes.\")\n\treturn nil\n}", "func SetRoll(iC *InterfaceConfig, degreeVal int) bool {\n dutyCycle := 0\n if degreeVal < 0 {\n dutyCycle = calcdutyCycleFromNeutralCenter(iC, Achannel, \"left\", (degreeVal*-1))\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n } else if degreeVal > 0{\n dutyCycle= calcdutyCycleFromNeutralCenter(iC, Achannel, \"right\", degreeVal)\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n } else if degreeVal == 0 {\n dutyCycle = calcdutyCycleFromNeutralCenter(iC, Achannel, \"left\", 0)\n iC.pca.SetChannel(Achannel, 0, dutyCycle)\n }\n return true\n}", "func (o *CreateImmuRuleReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewCreateImmuRuleCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewCreateImmuRuleBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 401:\n\t\tresult := NewCreateImmuRuleUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewCreateImmuRuleForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewCreateImmuRuleNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewCreateImmuRuleInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (f *FakeLoggerEventStreamingClient) StreamResult(addr string, args []string) (logutil.EventStream, error) {\n\tf.mu.Lock()\n\tdefer f.mu.Unlock()\n\n\tk := generateKey(args)\n\tresult, ok := f.results[k]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"no response was registered for args: %v\", args)\n\t}\n\tif result.addr != \"\" && addr != result.addr {\n\t\treturn nil, fmt.Errorf(\"client sent request to wrong server address. got: %v want: %v\", addr, result.addr)\n\t}\n\tresult.count--\n\tif result.count == 0 {\n\t\tdelete(f.results, k)\n\t}\n\n\treturn &streamResultAdapter{\n\t\tlines: strings.Split(result.output, \"\\n\"),\n\t\tindex: 0,\n\t\terr: result.err,\n\t}, nil\n}", "func (future *JobsCreateFuture) Result(client JobsClient) (jr JobResource, err error) {\n\tvar done bool\n\tdone, err = future.Done(client)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"databox.JobsCreateFuture\", \"Result\", future.Response(), \"Polling failure\")\n\t\treturn\n\t}\n\tif !done {\n\t\terr = azure.NewAsyncOpIncompleteError(\"databox.JobsCreateFuture\")\n\t\treturn\n\t}\n\tsender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n\tif jr.Response.Response, err = future.GetResult(sender); err == nil && jr.Response.Response.StatusCode != http.StatusNoContent {\n\t\tjr, err = client.CreateResponder(jr.Response.Response)\n\t\tif err != nil {\n\t\t\terr = autorest.NewErrorWithError(err, \"databox.JobsCreateFuture\", \"Result\", jr.Response.Response, \"Failure responding to request\")\n\t\t}\n\t}\n\treturn\n}", "func (_m *Service) CreateDiceRoll(ctx context.Context, r dice.CreateDiceRollRequest) (*dice.CreateDiceRollResponse, error) {\n\tret := _m.Called(ctx, r)\n\n\tvar r0 *dice.CreateDiceRollResponse\n\tif rf, ok := ret.Get(0).(func(context.Context, dice.CreateDiceRollRequest) *dice.CreateDiceRollResponse); ok {\n\t\tr0 = rf(ctx, r)\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).(*dice.CreateDiceRollResponse)\n\t\t}\n\t}\n\n\tvar r1 error\n\tif rf, ok := ret.Get(1).(func(context.Context, dice.CreateDiceRollRequest) error); ok {\n\t\tr1 = rf(ctx, r)\n\t} else {\n\t\tr1 = ret.Error(1)\n\t}\n\n\treturn r0, r1\n}", "func (obj *FileRoller) Roll(bytesToAdd int) (rolled bool, err error) {\n\n\trolled = false\n\terr = nil\n\n\tvar f afero.File\n\tvar fi os.FileInfo\n\n\tvar exists bool\n\n\texists, err = afero.Exists(DepFS, obj.FileName)\n\n\tif false == exists {\n\t\treturn\n\t}\n\n\tf, err = DepFS.OpenFile(obj.FileName, os.O_WRONLY, 0666)\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tdefer f.Close()\n\n\tfi, err = f.Stat()\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tif (fi.Size() + int64(bytesToAdd)) > obj.MaxSize {\n\t\tf.Close()\n\n\t\tvar logFrom string\n\t\tvar logTo string\n\t\tvar exists bool\n\n\t\tif obj.LogsToKeep <= 1 {\n\t\t\terr = DepFS.Remove(obj.FileName)\n\t\t} else {\n\t\t\tfor i := 0; i < obj.LogsToKeep-1; i++ {\n\t\t\t\tif i >= (obj.LogsToKeep - 2) {\n\t\t\t\t\tlogFrom = fmt.Sprintf(\"%s\", obj.FileName)\n\t\t\t\t\tlogTo = fmt.Sprintf(\"%s%d\", obj.FileName, obj.LogsToKeep-i-2)\n\t\t\t\t} else {\n\t\t\t\t\tlogFrom = fmt.Sprintf(\"%s%d\", obj.FileName, obj.LogsToKeep-i-3)\n\t\t\t\t\tlogTo = fmt.Sprintf(\"%s%d\", obj.FileName, obj.LogsToKeep-i-2)\n\t\t\t\t}\n\n\t\t\t\texists, err = afero.Exists(DepFS, logFrom)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\tif true == exists {\n\t\t\t\t\terr = DepFS.Rename(logFrom, logTo)\n\t\t\t\t}\n\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\trolled = true\n\n\t\t/* Create our file after we moved the old one */\n\t\tf, err = DepFS.OpenFile(obj.FileName, os.O_CREATE, 0666)\n\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tdefer f.Close()\n\t}\n\n\treturn\n}", "func (c *Conn) newReturn() (_ rpccp.Return, sendMsg func(*lockedConn), _ *rc.Releaser, _ error) {\n\toutMsg, err := c.transport.NewMessage()\n\tif err != nil {\n\t\treturn rpccp.Return{}, nil, nil, rpcerr.WrapFailed(\"create return\", err)\n\t}\n\tret, err := outMsg.Message.NewReturn()\n\tif err != nil {\n\t\toutMsg.Release()\n\t\treturn rpccp.Return{}, nil, nil, rpcerr.WrapFailed(\"create return\", err)\n\t}\n\n\t// Before releasing the message, we need to wait both until it is sent and\n\t// until the local vat is done with it. We therefore implement a simple\n\t// ref-counting mechanism such that 'release' must be called twice before\n\t// 'releaseMsg' is called.\n\treleaser := rc.NewReleaser(2, outMsg.Release)\n\n\tunlockedConn := c\n\treturn ret, func(c *lockedConn) {\n\t\tc.assertIs(unlockedConn)\n\t\tc.lk.sendTx.Send(asyncSend{\n\t\t\tsend: outMsg.Send,\n\t\t\trelease: releaser.Decr,\n\t\t\tonSent: func(err error) {\n\t\t\t\tif err != nil {\n\t\t\t\t\tc.er.ReportError(exc.WrapError(\"send return\", err))\n\t\t\t\t}\n\t\t\t},\n\t\t})\n\t}, releaser, nil\n}", "func (o ApplicationStatusOperationStateOperationPtrOutput) Retry() ApplicationStatusOperationStateOperationRetryPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperation) *ApplicationStatusOperationStateOperationRetry {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Retry\n\t}).(ApplicationStatusOperationStateOperationRetryPtrOutput)\n}", "func (future *FirewallRulesCreateOrUpdateFuture) Result(client FirewallRulesClient) (fr FirewallRule, err error) {\n var done bool\n done, err = future.DoneWithContext(context.Background(), client)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"mysqlflexibleservers.FirewallRulesCreateOrUpdateFuture\", \"Result\", future.Response(), \"Polling failure\")\n return\n }\n if !done {\n err = azure.NewAsyncOpIncompleteError(\"mysqlflexibleservers.FirewallRulesCreateOrUpdateFuture\")\n return\n }\n sender := autorest.DecorateSender(client, autorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n if fr.Response.Response, err = future.GetResult(sender); err == nil && fr.Response.Response.StatusCode != http.StatusNoContent {\n fr, err = client.CreateOrUpdateResponder(fr.Response.Response)\n if err != nil {\n err = autorest.NewErrorWithError(err, \"mysqlflexibleservers.FirewallRulesCreateOrUpdateFuture\", \"Result\", fr.Response.Response, \"Failure responding to request\")\n }\n }\n return\n }", "func MakeReweigh(db *pop.Connection, assertions Assertions) models.Reweigh {\n\tshipment := assertions.MTOShipment\n\tif isZeroUUID(shipment.ID) {\n\t\tassertions.MTOShipment.Status = models.MTOShipmentStatusApproved\n\t\tshipment = MakeMTOShipment(db, assertions)\n\t}\n\n\treweigh := models.Reweigh{\n\t\tRequestedAt: time.Now(),\n\t\tRequestedBy: models.ReweighRequesterTOO,\n\t\tShipment: shipment,\n\t\tShipmentID: shipment.ID,\n\t}\n\n\tmergeModels(&reweigh, assertions.Reweigh)\n\n\tmustCreate(db, &reweigh, assertions.Stub)\n\n\treturn reweigh\n}", "func (r *RbacV1Beta1RoleBindingRule) createRule(rolebinding *rbacV1beta1.RoleBinding, ydr *YamlDerivedResource) *rule {\n\trule := &rule{\n\t\tID: r.ID,\n\t\tPrereqs: r.Prereqs,\n\t\tCondition: func() bool {\n\t\t\tif r.Condition == nil {\n\t\t\t\treturn true\n\t\t\t}\n\t\t\treturn r.Condition(rolebinding)\n\t\t},\n\t\tMessage: r.Message,\n\t\tLevel: r.Level,\n\t\tResources: []*YamlDerivedResource{ydr},\n\t\tFix: func() bool {\n\t\t\tif r.Fix == nil {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\treturn r.Fix(rolebinding)\n\t\t},\n\t\tFixDescription: func() string {\n\t\t\tif r.FixDescription == nil {\n\t\t\t\treturn \"\"\n\t\t\t}\n\t\t\treturn r.FixDescription(rolebinding)\n\t\t},\n\t}\n\treturn rule\n}", "func (l *Logger) rotate() (err error) {\n\t_ = l.fileHandle.Close()\n\t_ = os.Rename(l.filename, fmt.Sprintf(\"%s.%d\", l.filename, l.splitRotateIndex))\n\tl.fileHandle, err = os.OpenFile(l.filename, os.O_RDWR|os.O_CREATE|os.O_APPEND, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tl.out = l.fileHandle\n\tl.splitRotateIndex++\n\tif l.splitRotateIndex > l.totalRotateSplit {\n\t\tl.splitRotateIndex = 0\n\t}\n\treturn err\n}", "func roll(sides int) int {\n\tmin, max := 1, sides\n\troll := min + seededRand.Intn(max-min+1)\n\treturn roll\n}" ]
[ "0.5401736", "0.5311997", "0.52291656", "0.50614893", "0.5017364", "0.49559066", "0.48634666", "0.4850141", "0.4791151", "0.47884125", "0.47248992", "0.46033987", "0.46009803", "0.4534906", "0.45265788", "0.44927794", "0.44854492", "0.4425973", "0.43776146", "0.4364654", "0.4363637", "0.43547466", "0.43235046", "0.43065244", "0.42919815", "0.4228411", "0.41874373", "0.4184938", "0.41831028", "0.41730922", "0.41569018", "0.41042522", "0.40979227", "0.40961832", "0.40936068", "0.40919295", "0.40688953", "0.40640354", "0.4063995", "0.40610918", "0.40254608", "0.4012322", "0.40110272", "0.40107954", "0.39797214", "0.3977783", "0.3975616", "0.39652327", "0.39638934", "0.3958185", "0.39557517", "0.39453542", "0.39166722", "0.39073586", "0.39072943", "0.38947982", "0.3885407", "0.3881819", "0.38089415", "0.38052547", "0.3791487", "0.37908226", "0.37905672", "0.37760398", "0.37734503", "0.3768859", "0.37576574", "0.3752755", "0.37481532", "0.3733758", "0.37324026", "0.37239584", "0.37187743", "0.3697261", "0.36936003", "0.367326", "0.36683136", "0.36639786", "0.36529317", "0.36480197", "0.36457962", "0.3637295", "0.3634881", "0.36320424", "0.36262524", "0.3626126", "0.36259335", "0.3625899", "0.36251315", "0.36173224", "0.36148804", "0.36131865", "0.36093423", "0.36086193", "0.3600993", "0.359773", "0.3595106", "0.35899505", "0.35881752", "0.35866633" ]
0.6764335
0
doAutoRollInner does the actual work of the AutoRoll.
func (r *AutoRoller) doAutoRollInner() (string, error) { r.runningMtx.Lock() defer r.runningMtx.Unlock() // Get updated info about the current roll. if err := r.updateCurrentRoll(); err != nil { return STATUS_ERROR, err } // There's a currently-active roll. Determine whether or not it's still good. // If so, leave it open and exit. If not, close it so that we can open another. currentRoll := r.recent.CurrentRoll() if currentRoll != nil { sklog.Infof("Found current roll: %s", r.issueUrl(currentRoll.Issue)) if r.isMode(autoroll_modes.MODE_DRY_RUN) { // If we have a normal (non-dry-run) roll running, // switch it to a dry run. if currentRoll.CommitQueue { sklog.Infof("Setting dry-run bit on %s", r.gerrit.Url(currentRoll.Issue)) if err := r.setDryRun(currentRoll, true); err != nil { return STATUS_ERROR, err } return STATUS_DRY_RUN_IN_PROGRESS, nil } // If the CQ has finished, determine if it was a success // or failure. rollDone, err := r.isDryRunDone(currentRoll) if err != nil { return STATUS_ERROR, err } if rollDone { result := autoroll.ROLL_RESULT_DRY_RUN_FAILURE status := STATUS_DRY_RUN_FAILURE rollSuccessful, err := r.isDryRunSuccessful(currentRoll) if err != nil { return STATUS_ERROR, err } if rollSuccessful { result = autoroll.ROLL_RESULT_DRY_RUN_SUCCESS status = STATUS_DRY_RUN_SUCCESS } sklog.Infof("Dry run is finished: %v", currentRoll) if currentRoll.RollingTo != r.rm.ChildHead() { if err := r.closeIssue(currentRoll, result, fmt.Sprintf("Repo has passed %s; will open a new dry run.", currentRoll.RollingTo)); err != nil { return STATUS_ERROR, err } } else if currentRoll.Result != result { // The dry run just finished. Set its result. if result == autoroll.ROLL_RESULT_DRY_RUN_FAILURE { if err := r.closeIssue(currentRoll, result, "Dry run failed. Closing, will open another."); err != nil { return STATUS_ERROR, err } } else { if err := r.addIssueComment(currentRoll, "Dry run finished successfully; leaving open in case we want to land"); err != nil { return STATUS_ERROR, err } currentRoll.Result = result if err := r.recent.Update(currentRoll); err != nil { return STATUS_ERROR, err } return status, nil } } else { // The dry run is finished but still good. Leave it open. sklog.Infof("Dry run is finished and still good.") return status, nil } } else { sklog.Infof("Dry run still in progress.") return STATUS_DRY_RUN_IN_PROGRESS, nil } } else { if currentRoll.CommitQueueDryRun { sklog.Infof("Unsetting dry run bit on %s", r.gerrit.Url(currentRoll.Issue)) if err := r.setDryRun(currentRoll, false); err != nil { return STATUS_ERROR, err } } if r.isMode(autoroll_modes.MODE_STOPPED) { // If we're stopped, close the issue. // Respect the previous result of the roll. if err := r.closeIssue(currentRoll, r.makeRollResult(currentRoll), "AutoRoller is stopped; closing the active roll."); err != nil { return STATUS_ERROR, err } } else if !currentRoll.CommitQueue { // If the CQ failed, close the issue. // Special case: if the current roll was a dry run which succeeded, land it. if currentRoll.Result == autoroll.ROLL_RESULT_DRY_RUN_SUCCESS { sklog.Infof("Dry run succeeded. Attempting to land.") if err := r.setDryRun(currentRoll, false); err != nil { return STATUS_ERROR, nil } return STATUS_IN_PROGRESS, nil } else { if err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, "Commit queue failed; closing this roll."); err != nil { return STATUS_ERROR, err } } } else if time.Since(currentRoll.Modified) > 24*time.Hour { // If the roll has been open too long, close the issue. if err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, "Roll has been open for over 24 hours; closing."); err != nil { return STATUS_ERROR, err } } else { rolledPast, err := r.rm.RolledPast(currentRoll.RollingTo) if err != nil { return STATUS_ERROR, err } if rolledPast { // If we've already rolled past the target revision, close the issue if err := r.closeIssue(currentRoll, autoroll.ROLL_RESULT_FAILURE, fmt.Sprintf("Already rolled past %s; closing this roll.", currentRoll.RollingTo)); err != nil { return STATUS_ERROR, err } } else { // Current roll is still good. sklog.Infof("Roll is still active (%d): %s", currentRoll.Issue, currentRoll.Subject) return STATUS_IN_PROGRESS, nil } } } } // If we're stopped, exit. if r.isMode(autoroll_modes.MODE_STOPPED) { sklog.Infof("Roller is stopped; not opening new rolls.") return STATUS_STOPPED, nil } // If we're up-to-date, exit. childHead := r.rm.ChildHead() if r.rm.LastRollRev() == childHead { sklog.Infof("Repo is up-to-date.") return STATUS_UP_TO_DATE, nil } // Create a new roll. if r.attemptCounter.Get() >= ROLL_ATTEMPT_THROTTLE_NUM { return STATUS_THROTTLED, nil } r.attemptCounter.Inc() dryRun := r.isMode(autoroll_modes.MODE_DRY_RUN) uploadedNum, err := r.rm.CreateNewRoll(r.strategy, r.GetEmails(), r.cqExtraTrybots, dryRun) if err != nil { return STATUS_ERROR, fmt.Errorf("Failed to upload a new roll: %s", err) } uploaded, err := r.retrieveRoll(uploadedNum) if err != nil { return STATUS_ERROR, fmt.Errorf("Failed to retrieve uploaded roll: %s", err) } if err := r.recent.Add(uploaded); err != nil { return STATUS_ERROR, fmt.Errorf("Failed to insert uploaded roll into database: %s", err) } if r.isMode(autoroll_modes.MODE_DRY_RUN) { return STATUS_DRY_RUN_IN_PROGRESS, nil } return STATUS_IN_PROGRESS, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *AutoRoller) doAutoRoll() error {\n\tstatus, lastError := r.doAutoRollInner()\n\n\tlastErrorStr := \"\"\n\tif lastError != nil {\n\t\tlastErrorStr = lastError.Error()\n\t}\n\n\tgerritUrl := r.gerrit.Url(0)\n\n\t// Update status information.\n\tif err := r.status.set(&AutoRollStatus{\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tGerritUrl: gerritUrl,\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: r.recent.GetRecentRolls(),\n\t\tStatus: status,\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\treturn lastError\n}", "func (r *AutoRoller) rollFinished(ctx context.Context, justFinished codereview.RollImpl) error {\n\trecent := r.recent.GetRecentRolls()\n\t// Sanity check: pop any rolls which occurred after the one which just\n\t// finished.\n\tidx := -1\n\tvar currentRoll *autoroll.AutoRollIssue\n\tfor i, roll := range recent {\n\t\tissue := fmt.Sprintf(\"%d\", roll.Issue)\n\t\tif issue == justFinished.IssueID() {\n\t\t\tidx = i\n\t\t\tcurrentRoll = roll\n\t\t\tbreak\n\t\t}\n\t}\n\tif currentRoll == nil {\n\t\treturn skerr.Fmt(\"Unable to find just-finished roll %q in recent list!\", justFinished.IssueID())\n\t}\n\n\t// Feed AutoRoll stats into metrics.\n\tv := int64(0)\n\tif currentRoll.Closed && currentRoll.Committed {\n\t\tv = int64(1)\n\t}\n\tmetrics2.GetInt64Metric(\"autoroll_last_roll_result\", map[string]string{\"roller\": r.cfg.RollerName}).Update(v)\n\n\trecent = recent[idx:]\n\tvar lastRoll *autoroll.AutoRollIssue\n\tif len(recent) > 1 {\n\t\tlastRoll = recent[1]\n\t} else {\n\t\t// If there are no other rolls, then the below alerts do not apply.\n\t\treturn nil\n\t}\n\n\tissueURL := fmt.Sprintf(\"%s%d\", r.codereview.GetIssueUrlBase(), currentRoll.Issue)\n\n\t// Send notifications if this roll had a different result from the last\n\t// roll, ie. success -> failure or failure -> success.\n\tcurrentSuccess := util.In(currentRoll.Result, autoroll.SUCCESS_RESULTS)\n\tlastSuccess := util.In(lastRoll.Result, autoroll.SUCCESS_RESULTS)\n\tif lastRoll != nil {\n\t\tif currentSuccess && !lastSuccess {\n\t\t\tr.notifier.SendNewSuccess(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t} else if !currentSuccess && lastSuccess {\n\t\t\tr.notifier.SendNewFailure(ctx, fmt.Sprintf(\"%d\", currentRoll.Issue), issueURL)\n\t\t}\n\t}\n\n\t// Send a notification if the last N rolls failed in a row.\n\tnFailed := 0\n\t// recent is in reverse chronological order.\n\tfor _, roll := range recent {\n\t\tif util.In(roll.Result, autoroll.SUCCESS_RESULTS) {\n\t\t\tbreak\n\t\t} else {\n\t\t\tnFailed++\n\t\t}\n\t}\n\tif nFailed == notifyIfLastNFailed {\n\t\tr.notifier.SendLastNFailed(ctx, notifyIfLastNFailed, issueURL)\n\t}\n\n\treturn nil\n}", "func (r *Roller) onRoller(i interface{}, next func(g interface{})) {\n\tr.Munch(i)\n\tnext(nil)\n}", "func (r *AutoRoller) Tick() error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence()\n\n\t// Update the status information.\n\tlastErrorStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrorStr = lastErr.Error()\n\t}\n\trecent := r.recent.GetRecentRolls()\n\tnumFailures := 0\n\tfor _, roll := range recent {\n\t\tif roll.Failed() {\n\t\t\tnumFailures++\n\t\t} else if roll.Succeeded() {\n\t\t\tbreak\n\t\t}\n\t}\n\tsklog.Infof(\"Updating status (%d)\", r.rm.CommitsNotRolled())\n\tif err := r.status.Set(&AutoRollStatus{\n\t\tAutoRollMiniStatus: AutoRollMiniStatus{\n\t\t\tNumFailedRolls: numFailures,\n\t\t\tNumNotRolledCommits: r.rm.CommitsNotRolled(),\n\t\t},\n\t\tCurrentRoll: r.recent.CurrentRoll(),\n\t\tError: lastErrorStr,\n\t\tFullHistoryUrl: r.gerrit.Url(0) + \"/q/owner:\" + r.GetUser(),\n\t\tIssueUrlBase: r.gerrit.Url(0) + \"/c/\",\n\t\tLastRoll: r.recent.LastRoll(),\n\t\tLastRollRev: r.rm.LastRollRev(),\n\t\tMode: r.modeHistory.CurrentMode(),\n\t\tRecent: recent,\n\t\tStatus: string(r.sm.Current()),\n\t}); err != nil {\n\t\treturn err\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn lastErr\n}", "func (w *WaitTask) startInner(taskContext *TaskContext) {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\tklog.V(3).Infof(\"wait task progress: %d/%d\", 0, len(w.Ids))\n\n\tpending := object.ObjMetadataSet{}\n\tfor _, id := range w.Ids {\n\t\tswitch {\n\t\tcase w.skipped(taskContext, id):\n\t\t\terr := taskContext.InventoryManager().SetSkippedReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as skipped reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSkipped)\n\t\tcase w.changedUID(taskContext, id):\n\t\t\t// replaced\n\t\t\tw.handleChangedUID(taskContext, id)\n\t\tcase w.reconciledByID(taskContext, id):\n\t\t\terr := taskContext.InventoryManager().SetSuccessfulReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as successful reconcile: %v\", err)\n\t\t\t}\n\t\t\tw.sendEvent(taskContext, id, event.ReconcileSuccessful)\n\t\tdefault:\n\t\t\terr := taskContext.InventoryManager().SetPendingReconcile(id)\n\t\t\tif err != nil {\n\t\t\t\t// Object never applied or deleted!\n\t\t\t\tklog.Errorf(\"Failed to mark object as pending reconcile: %v\", err)\n\t\t\t}\n\t\t\tpending = append(pending, id)\n\t\t\tw.sendEvent(taskContext, id, event.ReconcilePending)\n\t\t}\n\t}\n\tw.pending = pending\n\n\tklog.V(3).Infof(\"wait task progress: %d/%d\", len(w.Ids)-len(w.pending), len(w.Ids))\n\n\tif len(pending) == 0 {\n\t\t// all reconciled - clear pending and exit\n\t\tklog.V(3).Infof(\"all objects reconciled or skipped (name: %q)\", w.TaskName)\n\t\tw.cancelFunc()\n\t}\n}", "func NewAutoRoller(workdir, parentRepo, parentBranch, childPath, childBranch, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, tickFrequency, repoFrequency time.Duration, depot_tools string, rollIntoAndroid bool, strategy string) (*AutoRoller, error) {\n\tvar err error\n\tvar rm repo_manager.RepoManager\n\tif rollIntoAndroid {\n\t\trm, err = repo_manager.NewAndroidRepoManager(workdir, parentBranch, childPath, childBranch, repoFrequency, gerrit)\n\t} else {\n\t\trm, err = repo_manager.NewDEPSRepoManager(workdir, parentRepo, parentBranch, childPath, childBranch, repoFrequency, depot_tools, gerrit)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := autoroll_modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tattemptCounter: util.NewAutoDecrementCounter(ROLL_ATTEMPT_THROTTLE_TIME),\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tincludeCommitLog: true,\n\t\tliveness: metrics2.NewLiveness(\"last-autoroll-landed\", map[string]string{\"child-path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\trm: rm,\n\t\tstatus: &autoRollStatusCache{},\n\t\tstrategy: strategy,\n\t\trollIntoAndroid: rollIntoAndroid,\n\t}\n\n\t// Cycle once to fill out the current status.\n\tif err := arb.doAutoRoll(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tgo func() {\n\t\tfor range time.Tick(tickFrequency) {\n\t\t\tutil.LogErr(arb.doAutoRoll())\n\t\t}\n\t}()\n\n\treturn arb, nil\n}", "func (c Client) doRetry(ctx context.Context, method, uri string, body []byte, result any) error {\n\toperation := func() error {\n\t\treturn c.do(ctx, method, uri, body, result)\n\t}\n\n\tnotify := func(err error, duration time.Duration) {\n\t\tlog.Printf(\"client retries because of %v\", err)\n\t}\n\n\tbo := backoff.NewExponentialBackOff()\n\tbo.InitialInterval = 1 * time.Second\n\n\terr := backoff.RetryNotify(operation, bo, notify)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *AutoRoller) handleManualRolls(ctx context.Context) error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tif r.GetMode() == modes.ModeOffline {\n\t\treturn nil\n\t}\n\n\tsklog.Infof(\"Searching manual roll requests for %s\", r.cfg.RollerName)\n\treqs, err := r.manualRollDB.GetIncomplete(r.cfg.RollerName)\n\tif err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to get incomplete rolls\")\n\t}\n\tsklog.Infof(\"Found %d requests.\", len(reqs))\n\tfor _, req := range reqs {\n\t\tvar issue *autoroll.AutoRollIssue\n\t\tvar to *revision.Revision\n\t\tif req.NoResolveRevision {\n\t\t\tto = &revision.Revision{Id: req.Revision}\n\t\t} else {\n\t\t\tto, err = r.getRevision(ctx, req.Revision)\n\t\t\tif err != nil {\n\t\t\t\terr := skerr.Wrapf(err, \"failed to resolve revision %q\", req.Revision)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif req.ExternalChangeId != \"\" {\n\t\t\tto.ExternalChangeId = req.ExternalChangeId\n\t\t}\n\t\tif req.Status == manual.STATUS_PENDING {\n\t\t\t// Avoid creating rolls to the current revision.\n\t\t\tfrom := r.GetCurrentRev()\n\t\t\tif to.Id == from.Id {\n\t\t\t\terr := skerr.Fmt(\"Already at revision %q\", from.Id)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\temails := []string{}\n\t\t\tif !req.NoEmail {\n\t\t\t\temails = r.GetEmails()\n\t\t\t\tif !util.In(req.Requester, emails) {\n\t\t\t\t\temails = append(emails, req.Requester)\n\t\t\t\t}\n\t\t\t}\n\t\t\tvar err error\n\t\t\tsklog.Infof(\"Creating manual roll to %s as requested by %s...\", req.Revision, req.Requester)\n\n\t\t\tissue, err = r.createNewRoll(ctx, from, to, emails, req.DryRun, req.Canary, req.Requester)\n\t\t\tif err != nil {\n\t\t\t\terr := skerr.Wrapf(err, \"failed to create manual roll for %s\", req.Id)\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\treq.ResultDetails = err.Error()\n\t\t\t\tsklog.Errorf(\"Failed to create manual roll: %s\", req.ResultDetails)\n\t\t\t\tr.notifier.SendManualRollCreationFailed(ctx, req.Requester, req.Revision, err)\n\t\t\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t} else if req.Status == manual.STATUS_STARTED {\n\t\t\tsplit := strings.Split(req.Url, \"/\")\n\t\t\ti, err := strconv.Atoi(split[len(split)-1])\n\t\t\tif err != nil {\n\t\t\t\treturn skerr.Wrapf(err, \"Failed to parse issue number from %s for %s: %s\", req.Url, req.Id, err)\n\t\t\t}\n\t\t\tissue = &autoroll.AutoRollIssue{\n\t\t\t\tRollingTo: req.Revision,\n\t\t\t\tIsDryRun: req.DryRun,\n\t\t\t\tIssue: int64(i),\n\t\t\t}\n\t\t} else {\n\t\t\tsklog.Errorf(\"Found manual roll request %s in unknown status %q\", req.Id, req.Status)\n\t\t\tcontinue\n\t\t}\n\t\tsklog.Infof(\"Getting status for manual roll # %d\", issue.Issue)\n\t\troll, err := r.retrieveRoll(ctx, issue, to)\n\t\tif err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to retrieve manual roll %s: %s\", req.Id, err)\n\t\t}\n\t\treq.Status = manual.STATUS_STARTED\n\t\treq.Url = roll.IssueURL()\n\n\t\tif req.DryRun {\n\t\t\tif roll.IsDryRunFinished() {\n\t\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\t\tif roll.IsDryRunSuccess() {\n\t\t\t\t\treq.Result = manual.RESULT_SUCCESS\n\t\t\t\t} else {\n\t\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t\t}\n\t\t\t}\n\t\t} else if roll.IsFinished() {\n\t\t\treq.Status = manual.STATUS_COMPLETE\n\t\t\tif roll.IsSuccess() {\n\t\t\t\treq.Result = manual.RESULT_SUCCESS\n\t\t\t} else {\n\t\t\t\treq.Result = manual.RESULT_FAILURE\n\t\t\t}\n\t\t}\n\t\tif err := r.manualRollDB.Put(req); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to update manual roll request\")\n\t\t}\n\t}\n\treturn nil\n}", "func (r *AutoRoller) updateCurrentRoll() error {\n\tcurrentRoll := r.recent.CurrentRoll()\n\tif currentRoll == nil {\n\t\treturn nil\n\t}\n\tcurrentResult := currentRoll.Result\n\n\tupdated, err := r.retrieveRoll(currentRoll.Issue)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// We have to rely on data we store for the dry run case.\n\tif !updated.Closed && util.In(currentResult, autoroll.DRY_RUN_RESULTS) {\n\t\tupdated.Result = currentResult\n\t}\n\n\t// If the current roll succeeded, we need to make sure we update the\n\t// repo so that we see the roll commit. This can take some time, so\n\t// we have to repeatedly update until we see the commit.\n\tif updated.Committed {\n\t\tsklog.Infof(\"Roll succeeded (%d); syncing the repo until it lands.\", currentRoll.Issue)\n\t\tfor {\n\t\t\tsklog.Info(\"Syncing...\")\n\t\t\tsklog.Infof(\"Looking for %s\", currentRoll.RollingTo)\n\t\t\tif err := r.rm.ForceUpdate(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\trolledPast, err := r.rm.RolledPast(currentRoll.RollingTo)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif rolledPast {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttime.Sleep(10 * time.Second)\n\t\t}\n\t\tr.liveness.Reset()\n\t}\n\treturn r.recent.Update(updated)\n}", "func LoadAutoRollData(dbClient *influxdb.Client, workdir string) {\n\trollCheckoutsDir := path.Join(workdir, \"autoroll_git\")\n\tskiaRepo, err := gitinfo.CloneOrUpdate(SKIA_REPO, path.Join(rollCheckoutsDir, \"skia\"), false)\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to check out skia: %s\", err)\n\t\treturn\n\t}\n\tchromiumRepo, err := gitinfo.CloneOrUpdate(CHROMIUM_REPO, path.Join(rollCheckoutsDir, \"chromium\"), false)\n\tif err != nil {\n\t\tglog.Errorf(\"Failed to check out chromium: %s\", err)\n\t\treturn\n\t}\n\n\tfor _ = range time.Tick(time.Minute) {\n\t\ts, err := autoroll.CurrentStatus(skiaRepo, chromiumRepo)\n\t\tif err != nil {\n\t\t\tglog.Error(err)\n\t\t} else {\n\t\t\terr := writeAutoRollDataPoint(dbClient, s)\n\t\t\tif err != nil {\n\t\t\t\tglog.Error(err)\n\t\t\t}\n\t\t}\n\t\tskiaRepo.Update(true, false)\n\t\tchromiumRepo.Update(true, false)\n\t}\n}", "func main() {\n\tvar sides int = 10000\n\tlog.Printf(\"rolling a %v sided dice\\n\", sides)\n\tunique_rolls := map[int]bool{}\n\ti := 1\n\tfor len(unique_rolls) < sides {\n\t\tvar roll int = roll(sides)\n\t\tfmt.Println(\"rolled a\", roll)\n\t\tunique_rolls = updateUniqueRolls(roll, unique_rolls)\n\t\ti += 1\n\t}\n\tlog.Printf(\"got %v unique rolls in %v attempts \\n\", len(unique_rolls), i)\n}", "func (o WorkloadStatusConfigAutomaticRuleOutput) Rollup() WorkloadStatusConfigAutomaticRuleRollupOutput {\n\treturn o.ApplyT(func(v WorkloadStatusConfigAutomaticRule) WorkloadStatusConfigAutomaticRuleRollup { return v.Rollup }).(WorkloadStatusConfigAutomaticRuleRollupOutput)\n}", "func ROLL(ci, mr operand.Op) { ctx.ROLL(ci, mr) }", "func Do(retryFunc func() error, config *Config) error {\n\tvar err error\n\tconfig.Logger.Infof(\"Job attempt: %d\", 1)\n\terr = retryFunc()\n\tif err == nil {\n\t\treturn nil\n\t}\n\tfor n := uint(1); n < config.Attempts; n++ {\n\t\tdelayTime := config.Delay * (1 << (n - 1))\n\t\ttime.Sleep((time.Duration)(delayTime) * config.Units)\n\t\tconfig.Logger.Infof(\"Job attempt: %d\", n+1)\n\t\terr = retryFunc()\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn err\n}", "func (r *AutoRoller) Tick(ctx context.Context) error {\n\tr.runningMtx.Lock()\n\tdefer r.runningMtx.Unlock()\n\n\tsklog.Infof(\"Running autoroller.\")\n\n\t// Update the config vars.\n\tif err := r.reg.Update(ctx); err != nil {\n\t\tsklog.Errorf(\"Failed to update config registry; continuing, but config may be out of date: %s\", err)\n\t}\n\n\t// Determine if we should unthrottle.\n\tshouldUnthrottle, err := r.throttle.Get(ctx, r.roller)\n\tif err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to determine whether we should unthrottle\")\n\t}\n\tif shouldUnthrottle {\n\t\tif err := r.unthrottle(ctx); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to unthrottle\")\n\t\t}\n\t\tif err := r.throttle.Reset(ctx, r.roller); err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to reset unthrottle counter\")\n\t\t}\n\t}\n\n\t// Update modes and strategies.\n\tif err := r.modeHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update mode history\")\n\t}\n\toldStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif err := r.strategyHistory.Update(ctx); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update strategy history\")\n\t}\n\tnewStrategy := r.strategyHistory.CurrentStrategy().Strategy\n\tif oldStrategy != newStrategy {\n\t\tstrat, err := strategy.GetNextRollStrategy(newStrategy)\n\t\tif err != nil {\n\t\t\treturn skerr.Wrapf(err, \"Failed to get next roll strategy\")\n\t\t}\n\t\tr.strategyMtx.Lock()\n\t\tr.strategy = strat\n\t\tr.strategyMtx.Unlock()\n\t}\n\n\t// Run the state machine.\n\tlastErr := r.sm.NextTransitionSequence(ctx)\n\tlastErrStr := \"\"\n\tif lastErr != nil {\n\t\tlastErrStr = lastErr.Error()\n\t}\n\n\t// Update the status information.\n\tif err := r.updateStatus(ctx, true, lastErrStr); err != nil {\n\t\treturn skerr.Wrapf(err, \"Failed to update status\")\n\t}\n\tsklog.Infof(\"Autoroller state %s\", r.sm.Current())\n\tif lastRoll := r.recent.LastRoll(); lastRoll != nil && util.In(lastRoll.Result, []string{autoroll.ROLL_RESULT_DRY_RUN_SUCCESS, autoroll.ROLL_RESULT_SUCCESS}) {\n\t\tr.liveness.ManualReset(lastRoll.Modified)\n\t}\n\treturn skerr.Wrapf(lastErr, \"Failed state transition sequence\")\n}", "func Do(retries int, trier Trier, sleeps ...Sleep) error {\n\tif retries <= 0 {\n\t\treturn nil\n\t}\n\n\ttotal := retries\n\nLBBEGIN:\n\n\tif err := trier.Try(); err != nil {\n\t\tretries--\n\t\tif retries == 0 {\n\t\t\treturn ErrMaxRetries\n\t\t}\n\n\t\tif len(sleeps) > 0 {\n\t\t\tif d := sleeps[0](total - retries); d > 0 {\n\t\t\t\ttime.Sleep(d)\n\t\t\t}\n\t\t}\n\n\t\tgoto LBBEGIN\n\t}\n\n\treturn nil\n}", "func newAutoRoller(workdir, childPath, cqExtraTrybots string, emails []string, gerrit *gerrit.Gerrit, rm repo_manager.RepoManager, retrieveRoll func(*AutoRoller, int64) (RollImpl, error)) (*AutoRoller, error) {\n\trecent, err := recent_rolls.NewRecentRolls(path.Join(workdir, \"recent_rolls.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmh, err := modes.NewModeHistory(path.Join(workdir, \"autoroll_modes.db\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tarb := &AutoRoller{\n\t\tcqExtraTrybots: cqExtraTrybots,\n\t\temails: emails,\n\t\tgerrit: gerrit,\n\t\tliveness: metrics2.NewLiveness(\"last_autoroll_landed\", map[string]string{\"child_path\": childPath}),\n\t\tmodeHistory: mh,\n\t\trecent: recent,\n\t\tretrieveRoll: retrieveRoll,\n\t\trm: rm,\n\t\tstatus: &AutoRollStatusCache{},\n\t}\n\tsm, err := state_machine.New(arb, workdir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tarb.sm = sm\n\tcurrent := recent.CurrentRoll()\n\tif current != nil {\n\t\troll, err := arb.retrieveRoll(arb, current.Issue)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tarb.currentRoll = roll\n\t}\n\treturn arb, nil\n}", "func (o *RowerGroup) doAfterUpsertHooks(exec boil.Executor) (err error) {\n\tfor _, hook := range rowerGroupAfterUpsertHooks {\n\t\tif err := hook(exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r Retrier) Do(ctx context.Context, run func() error) error {\n\tcooldown := r.Cooldown()\n\n\tfor {\n\t\tif err := ctx.Err(); err != nil {\n\t\t\t// nolint:wrapcheck // no meaningful information can be added to this error\n\t\t\treturn err\n\t\t}\n\n\t\terr := run()\n\t\tif err != nil {\n\t\t\t// check to see if it's temporary.\n\t\t\tvar tempErr TemporaryError\n\t\t\tif ok := errors.As(err, &tempErr); ok && tempErr.Temporary() {\n\t\t\t\tdelay, err := cooldown() // nolint:govet // the shadow is intentional\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn pkgerrors.Wrap(err, \"sleeping during retry\")\n\t\t\t\t}\n\t\t\t\ttime.Sleep(delay)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// since it's not temporary, it can't be retried, so...\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n}", "func TestRetryDoInteralFunc(t *testing.T) {\n\texec := func() (error, interface{}){\n\t\tr := number.Random(0, 100000000)\n\t\tfmt.Println(\"r is\", r)\n\t\tif r < 100000000 / 2{\n\t\t\treturn errors.New(\"xx\"), nil\n\t\t}\n\t\treturn nil, map[string]string{\"abc\":\"wocao\"}\n\t}\n\terr, res, count := RetryDoInteralTime(exec, 5, 100)\n\tfmt.Printf(\"TestRetryDoInteralFunc error is %s res is %v count is %d\", err, res, count)\n}", "func Roll(n, d, b int, s *discordgo.Session, m *discordgo.MessageCreate) {\n\tresult := \"Rolled: [\"\n\ttotal := b\n\tfor i := 0; i < n-1; i++ {\n\t\tval := rand.Intn(d) + 1\n\t\tresult += strconv.Itoa(val) + \", \"\n\t\ttotal += val\n\t}\n\tval := rand.Intn(d) + 1\n\tresult += strconv.Itoa(val)\n\ttotal += val\n\tif b > 0 {\n\t\tresult += \"] +\" + strconv.Itoa(b)\n\t} else if b == 0 {\n\t\tresult += \"]\"\n\t} else {\n\t\tresult += \"] \" + strconv.Itoa(b)\n\t}\n\n\tresult += \" = \" + strconv.Itoa(total)\n\ts.ChannelMessageSend(m.ChannelID, result)\n}", "func Roll(number int, sided int) Result {\n\treturn roll(number, sided)\n}", "func calcApplyRowsHouseholderTransformation(mtxA [][]float64, c int, mtxY [][]float64, n int) {\n\tdenominator := calcRowsSumProduct(mtxA, c, mtxA, c, c, n)\n\tnumerator := calcRowsSumProduct(mtxA, c, mtxY, 0, c, n)\n\tfactor := 2 * (numerator / denominator)\n\tfor row := c; row < n; row++ {\n\t\tputDouble(mtxY, row, getDouble(mtxY, row)-factor*mtxA[c][row])\n\t}\n}", "func (r *AutoRoller) retrieveRoll(ctx context.Context, roll *autoroll.AutoRollIssue, rollingTo *revision.Revision) (codereview.RollImpl, error) {\n\treturn r.codereview.RetrieveRoll(ctx, roll, r.recent, rollingTo, r.rollFinished)\n}", "func (o *RentalRower) doAfterUpsertHooks(exec boil.Executor) (err error) {\n\tfor _, hook := range rentalRowerAfterUpsertHooks {\n\t\tif err := hook(exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestEnroll(t *testing.T) {\n\n\tfabricCAClient, err := NewFabricCAClient(org1, configImp, cryptoSuiteProvider)\n\tif err != nil {\n\t\tt.Fatalf(\"NewFabricCAClient return error: %v\", err)\n\t}\n\t_, _, err = fabricCAClient.Enroll(\"\", \"user1\")\n\tif err == nil {\n\t\tt.Fatalf(\"Enroll didn't return error\")\n\t}\n\tif err.Error() != \"enrollmentID required\" {\n\t\tt.Fatalf(\"Enroll didn't return right error\")\n\t}\n\t_, _, err = fabricCAClient.Enroll(\"test\", \"\")\n\tif err == nil {\n\t\tt.Fatalf(\"Enroll didn't return error\")\n\t}\n\tif err.Error() != \"enrollmentSecret required\" {\n\t\tt.Fatalf(\"Enroll didn't return right error\")\n\t}\n\t_, _, err = fabricCAClient.Enroll(\"enrollmentID\", \"enrollmentSecret\")\n\tif err != nil {\n\t\tt.Fatalf(\"fabricCAClient Enroll return error %v\", err)\n\t}\n\n\twrongConfigImp := mocks.NewMockConfig(wrongCAServerURL)\n\tfabricCAClient, err = NewFabricCAClient(org1, wrongConfigImp, cryptoSuiteProvider)\n\tif err != nil {\n\t\tt.Fatalf(\"NewFabricCAClient return error: %v\", err)\n\t}\n\t_, _, err = fabricCAClient.Enroll(\"enrollmentID\", \"enrollmentSecret\")\n\tif err == nil {\n\t\tt.Fatalf(\"Enroll didn't return error\")\n\t}\n\tif !strings.Contains(err.Error(), \"enroll failed\") {\n\t\tt.Fatalf(\"Expected error enroll failed. Got: %s\", err)\n\t}\n\n}", "func ExecuteTxGenericTest(ctx context.Context, framework WriteSkewTest) error {\n\tframework.Init(ctx)\n\t// wg is used as a barrier, blocking each transaction after it performs the\n\t// initial read until they both read.\n\tvar wg sync.WaitGroup\n\twg.Add(2)\n\trunTxn := func(iter *int) <-chan error {\n\t\terrCh := make(chan error, 1)\n\t\tgo func() {\n\t\t\t*iter = 0\n\t\t\terrCh <- framework.ExecuteTx(ctx, func(tx interface{}) (retErr error) {\n\t\t\t\tdefer func() {\n\t\t\t\t\tif retErr == nil {\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\t// Wrap the error so that we test the library's unwrapping.\n\t\t\t\t\tretErr = testError{cause: retErr}\n\t\t\t\t}()\n\n\t\t\t\t*iter++\n\t\t\t\tbal1, bal2, err := framework.GetBalances(ctx, tx)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\t// If this is the first iteration, wait for the other tx to also read.\n\t\t\t\tif *iter == 1 {\n\t\t\t\t\twg.Done()\n\t\t\t\t\twg.Wait()\n\t\t\t\t}\n\t\t\t\t// Now, subtract from one account and give to the other.\n\t\t\t\tif bal1 > bal2 {\n\t\t\t\t\tif err := framework.UpdateBalance(ctx, tx, 1, -100); err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tif err := framework.UpdateBalance(ctx, tx, 2, +100); err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tif err := framework.UpdateBalance(ctx, tx, 1, +100); err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tif err := framework.UpdateBalance(ctx, tx, 2, -100); err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t})\n\t\t}()\n\t\treturn errCh\n\t}\n\n\tvar iters1, iters2 int\n\ttxn1Err := runTxn(&iters1)\n\ttxn2Err := runTxn(&iters2)\n\tif err := <-txn1Err; err != nil {\n\t\treturn fmt.Errorf(\"expected success in txn1; got %s\", err)\n\t}\n\tif err := <-txn2Err; err != nil {\n\t\treturn fmt.Errorf(\"expected success in txn2; got %s\", err)\n\t}\n\tif iters1+iters2 <= 2 {\n\t\treturn fmt.Errorf(\"expected at least one retry between the competing transactions; \"+\n\t\t\t\"got txn1=%d, txn2=%d\", iters1, iters2)\n\t}\n\n\tvar bal1, bal2 int\n\terr := framework.ExecuteTx(ctx, func(txi interface{}) error {\n\t\tvar err error\n\t\tbal1, bal2, err = framework.GetBalances(ctx, txi)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tif bal1 != 100 || bal2 != 100 {\n\t\treturn fmt.Errorf(\"expected balances to be restored without error; \"+\n\t\t\t\"got acct1=%d, acct2=%d: %s\", bal1, bal2, err)\n\t}\n\treturn nil\n}", "func (r *Reconciler) doRollback(change *devicechange.DeviceChange) error {\n\tlog.Infof(\"Executing Rollback for %s\", change.ID)\n\tlog.Debug(change)\n\tdeltaChange, err := r.computeRollback(change)\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Rolling back %s with %v\", change.ID, deltaChange)\n\tlog.Debugf(\"%v\", change)\n\treturn r.translateAndSendChange(deltaChange)\n}", "func (gs *GameState) ApplyRoll(playerID xid.ID) error {\n\tif gs.Turn.Player == nil || *gs.Turn.Player != playerID {\n\t\treturn fmt.Errorf(\"You're not the rolling player\")\n\t}\n\troll := gs.NextRoll\n\tif roll == nil {\n\t\treturn fmt.Errorf(\"No roll ready\")\n\t}\n\n\tdefer gs.nextRoller()\n\n\tswitch roll.Action {\n\tcase ActionUp, ActionDown:\n\t\treturn gs.applyStockMove(roll)\n\tcase ActionDividend:\n\t\treturn gs.applyDividend(roll)\n\tdefault:\n\t\treturn fmt.Errorf(\"Unknown PlayerAction %v\", roll.Action)\n\t}\n}", "func (r *AutoRoller) Start(ctx context.Context, tickFrequency time.Duration) {\n\tsklog.Infof(\"Starting autoroller.\")\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(tickFrequency, func(_ context.Context) {\n\t\t// Explicitly ignore the passed-in context; this allows us to\n\t\t// continue running even if the context is canceled, which helps\n\t\t// to prevent errors due to interrupted syncs, etc.\n\t\tctx := context.Background()\n\t\tif err := r.Tick(ctx); err != nil {\n\t\t\t// Hack: we frequently get failures from GoB which trigger error-rate alerts.\n\t\t\t// These alerts are noise and sometimes hide real failures. If the error is\n\t\t\t// due to a sync failure, log it as a warning instead of an error. We'll rely\n\t\t\t// on the liveness alert in the case where we see persistent sync failures.\n\t\t\tif isSyncError(err) {\n\t\t\t\tsklog.Warningf(\"Failed to run autoroll: %s\", err)\n\t\t\t} else {\n\t\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t}, nil)\n\n\t// Update the current reviewers in a loop.\n\tlvReviewers := metrics2.NewLiveness(\"last_successful_reviewers_retrieval\", map[string]string{\"roller\": r.roller})\n\tcleanup.Repeat(30*time.Minute, func(ctx context.Context) {\n\t\temails := GetReviewers(r.client, r.cfg.RollerName, r.cfg.Reviewer, r.cfg.ReviewerBackup)\n\t\tr.emailsMtx.Lock()\n\t\tdefer r.emailsMtx.Unlock()\n\t\tr.emails = emails\n\n\t\tconfigCopies := replaceReviewersPlaceholder(r.cfg.Notifiers, emails)\n\t\tif err := r.notifier.ReloadConfigs(ctx, configCopies); err != nil {\n\t\t\tsklog.Errorf(\"Failed to reload configs: %s\", err)\n\t\t\treturn\n\t\t}\n\t\tlvReviewers.Reset()\n\t}, nil)\n\n\t// Handle requests for manual rolls.\n\tif r.cfg.SupportsManualRolls {\n\t\tlvManualRolls := metrics2.NewLiveness(\"last_successful_manual_roll_check\", map[string]string{\"roller\": r.roller})\n\t\tcleanup.Repeat(time.Minute, func(_ context.Context) {\n\t\t\t// Explicitly ignore the passed-in context; this allows\n\t\t\t// us to continue handling manual rolls even if the\n\t\t\t// context is canceled, which helps to prevent errors\n\t\t\t// due to interrupted syncs, etc.\n\t\t\tctx := context.Background()\n\t\t\tif err := r.handleManualRolls(ctx); err != nil {\n\t\t\t\tsklog.Error(err)\n\t\t\t} else {\n\t\t\t\tlvManualRolls.Reset()\n\t\t\t}\n\t\t}, nil)\n\t}\n}", "func DummyRoll(dieType int) int {\n\treturn dieType / 2\n}", "func (r *AutoRoller) makeRollResult(roll *autoroll.AutoRollIssue) string {\n\tif util.In(roll.Result, autoroll.DRY_RUN_RESULTS) {\n\t\tif roll.Result == autoroll.ROLL_RESULT_DRY_RUN_IN_PROGRESS {\n\t\t\treturn autoroll.ROLL_RESULT_DRY_RUN_FAILURE\n\t\t} else {\n\t\t\treturn roll.Result\n\t\t}\n\t}\n\treturn autoroll.ROLL_RESULT_FAILURE\n}", "func (s *Scaler) do() {\n\tvar (\n\t\twg sync.WaitGroup\n\t\ttr *http.Transport\n\t\tcli *http.Client\n\t)\n\ttr = &http.Transport{}\n\tcli = &http.Client{\n\t\tTransport: tr,\n\t\tTimeout: httpTimeout,\n\t}\n\n\tfor _, a := range s.Actions {\n\t\tgo func(a *model.ActionHTTP) {\n\t\t\twg.Add(1)\n\t\t\tdelay, _ := time.ParseDuration(a.Delay)\n\t\t\turl := a.URL.String()\n\t\t\terr := retry.Do(\n\t\t\t\tfunc() error {\n\t\t\t\t\t// TODO(kiennt): Check kind of action url -> Authen or not?\n\t\t\t\t\treq, err := http.NewRequest(a.Method, url, nil)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tresp, err := cli.Do(req)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t\tdefer resp.Body.Close()\n\t\t\t\t\treturn nil\n\t\t\t\t},\n\t\t\t\tretry.DelayType(func(n uint, config *retry.Config) time.Duration {\n\t\t\t\t\tvar f retry.DelayTypeFunc\n\t\t\t\t\tswitch a.DelayType {\n\t\t\t\t\tcase \"fixed\":\n\t\t\t\t\t\tf = retry.FixedDelay\n\t\t\t\t\tcase \"backoff\":\n\t\t\t\t\t\tf = retry.BackOffDelay\n\t\t\t\t\t}\n\t\t\t\t\treturn f(n, config)\n\t\t\t\t}),\n\t\t\t\tretry.Attempts(a.Attempts),\n\t\t\t\tretry.Delay(delay),\n\t\t\t\tretry.RetryIf(func(err error) bool {\n\t\t\t\t\treturn common.RetryableError(err)\n\t\t\t\t}),\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\tlevel.Error(s.logger).Log(\"msg\", \"Error doing scale action\", \"url\", url, \"err\", err)\n\t\t\t\texporter.ReportFailureScalerActionCounter(cluster.ClusterID, \"http\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\texporter.ReportSuccessScalerActionCounter(cluster.ClusterID, \"http\")\n\t\t\tlevel.Info(s.logger).Log(\"msg\", \"Sending request\",\n\t\t\t\t\"url\", url, \"method\", a.Method)\n\t\t\ts.alert.Fire(time.Now())\n\t\t\tdefer wg.Done()\n\t\t}(a)\n\t}\n\t// Wait until all actions were performed\n\twg.Wait()\n}", "func TestRoll(t *testing.T) {\n\ttype response struct {\n\t\tCode int\n\t\tBody string\n\t}\n\ttests := []struct {\n\t\tname string\n\t\targs string\n\t\twant response\n\t}{\n\t\t{\"Default roll\", \"\", response{http.StatusOK, ``}},\n\t\t{\"Valid query for sides\", \"?sides=4\", response{http.StatusOK, `\"sides\":4`}},\n\t\t{\"Invalid query for sides\", \"?sides=5\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for count\", \"?count=2\", response{http.StatusOK, `\"count\":2`}},\n\t\t{\"Invalid query for count\", \"?count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for sides, invalid for count\", \"?sides=4&count=0\", response{http.StatusNotAcceptable, `\"invalid count\"`}},\n\t\t{\"Valid query for count, invalid for sides\", \"?count=2&sides=1\", response{http.StatusNotAcceptable, `\"invalid sides\"`}},\n\t\t{\"Valid query for sides and count\", \"?sides=4&count=2\", response{http.StatusOK, `\"count\":2,\"sides\":4`}},\n\t}\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tr := gofight.New()\n\n\t\t\tr.GET(\"/api/v1/roll\"+tt.args).\n\t\t\t\tRun(router, func(r gofight.HTTPResponse, rq gofight.HTTPRequest) {\n\t\t\t\t\tif r.Code != tt.want.Code {\n\t\t\t\t\t\tt.Errorf(\"Handler returned wrong status code: got %v want %v\", r.Code, tt.want.Code)\n\t\t\t\t\t}\n\n\t\t\t\t\tif !bytes.Contains(r.Body.Bytes(), []byte(tt.want.Body)) {\n\t\t\t\t\t\tt.Errorf(\"Unexpected body returned.\\ngot %v\\nwant %v\", r.Body, tt.want.Body)\n\t\t\t\t\t}\n\t\t\t\t})\n\t\t})\n\t}\n}", "func (d *Dice) Roll() int {\n\t// We named this call \"roll\" so we can easily mock the specific call out in our tests\n\treturn d.rand.Intn(\"roll\", d.sides)\n}", "func (o *Rental) doAfterUpsertHooks(exec boil.Executor) (err error) {\n\tfor _, hook := range rentalAfterUpsertHooks {\n\t\tif err := hook(exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func enroll(orgName, enrollmentID string, sdk *fabsdk.FabricSDK, opts ...msp.EnrollmentOption) error {\n\tmspClient, err := msp.New(sdk.Context(), msp.WithOrg(orgName))\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn mspClient.Enroll(enrollmentID, opts...)\n}", "func (s *Scheduler) Run(root func()) {\n\tif s.active {\n\t\tpanic(\"nested Run call\")\n\t}\n\ts.active = true\n\tdefer func() { s.active = false }()\n\n\tcount = 0\n\tstartProgress()\n\tdefer stopProgress()\n\n\ts.Strategy.Reset()\n\tfor {\n\t\ts.run1(root)\n\n\t\tif !s.Strategy.Next() {\n\t\t\tbreak\n\t\t}\n\t\tcount++\n\t}\n}", "func (d *sqlDB) retry(operation retryFunc) (err error) {\n\tfor retries := 0; retries < maxRetries; retries++ {\n\t\tif err = operation(); err != nil {\n\t\t\t// cleanup\n\t\t\td.conn.Close()\n\n\t\t\t// reconnect and retry\n\t\t\ttime.Sleep(1000 * time.Millisecond)\n\t\t\td.connect()\n\t\t\tcontinue\n\t\t}\n\t}\n\treturn\n}", "func (r *EndpointsRolloutV1RestClient) AutoWatchRollout(ctx context.Context, options *api.ListWatchOptions) (kvstore.Watcher, error) {\n\tpath := r.instance + makeURIRolloutV1AutoWatchRolloutWatchOper(options)\n\tpath = strings.Replace(path, \"http://\", \"ws://\", 1)\n\tpath = strings.Replace(path, \"https://\", \"wss://\", 1)\n\tparams := apiutils.GetQueryStringFromListWatchOptions(options)\n\tif params != \"\" {\n\t\tpath = path + \"?\" + params\n\t}\n\theader := http.Header{}\n\tr.updateHTTPHeader(ctx, &header)\n\tdialer := websocket.DefaultDialer\n\tdialer.TLSClientConfig = &tls.Config{InsecureSkipVerify: true}\n\tconn, hresp, err := dialer.Dial(path, header)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to connect web socket to [%s](%s)[%+v]\", path, err, hresp)\n\t}\n\tbridgefn := func(lw *listerwatcher.WatcherClient) {\n\t\tfor {\n\t\t\tin := &AutoMsgRolloutWatchHelper{}\n\t\t\terr := conn.ReadJSON(in)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor _, e := range in.Events {\n\t\t\t\tev := kvstore.WatchEvent{\n\t\t\t\t\tType: kvstore.WatchEventType(e.Type),\n\t\t\t\t\tObject: e.Object,\n\t\t\t\t}\n\t\t\t\tselect {\n\t\t\t\tcase lw.OutCh <- &ev:\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tclose(lw.OutCh)\n\t\t\t\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tlw := listerwatcher.NewWatcherClient(nil, bridgefn)\n\tlw.Run()\n\tgo func() {\n\t\t<-ctx.Done()\n\t\tconn.WriteControl(websocket.CloseMessage, websocket.FormatCloseMessage(websocket.CloseNormalClosure, \"client closing\"), time.Now().Add(3*time.Second))\n\t}()\n\treturn lw, nil\n}", "func DoOpsInOneTxnWithRetry(cli *clientv3.Client, ops ...clientv3.Op) (*clientv3.TxnResponse, int64, error) {\n\tctx, cancel := context.WithTimeout(cli.Ctx(), DefaultRequestTimeout)\n\tdefer cancel()\n\ttctx := tcontext.NewContext(ctx, log.L())\n\tret, _, err := etcdDefaultTxnStrategy.Apply(tctx, etcdDefaultTxnRetryParam, func(t *tcontext.Context) (ret interface{}, err error) {\n\t\tresp, err := cli.Txn(ctx).Then(ops...).Commit()\n\t\tif err != nil {\n\t\t\treturn nil, errors.Trace(err)\n\t\t}\n\t\treturn resp, nil\n\t})\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tresp := ret.(*clientv3.TxnResponse)\n\treturn resp, resp.Header.Revision, nil\n}", "func TestRollDice(t *testing.T) {\n\tresult := Roll()\n\tfor i := 1; i == 20; i++ {\n\t\tif result > 6 || result < 0 {\n\t\t\tt.Errorf(\"the rolled result is %v\", result)\n\t\t}\n\t}\n}", "func rollDice(sides, count int) (result int, err error) {\n\td := chooseDice(sides)\n\tif d == nil {\n\t\treturn 0, invalidDice{}\n\t}\n\n\tfor i := 0; i < count; i++ {\n\t\tresult += d()\n\t}\n\n\treturn result, nil\n}", "func (r *AutoRoller) createNewRoll(ctx context.Context, from, to *revision.Revision, emails []string, dryRun, canary bool, manualRollRequester string) (rv *autoroll.AutoRollIssue, rvErr error) {\n\t// Track roll CL upload attempts vs failures.\n\tdefer func() {\n\t\tr.rollUploadAttempts.Inc(1)\n\t\tif rvErr == nil {\n\t\t\tr.rollUploadFailures.Reset()\n\t\t} else {\n\t\t\tr.rollUploadFailures.Inc(1)\n\t\t}\n\t}()\n\tr.statusMtx.RLock()\n\tvar revs []*revision.Revision\n\tfound := false\n\tfor _, rev := range r.notRolledRevs {\n\t\tif rev.Id == to.Id {\n\t\t\tfound = true\n\t\t}\n\t\tif found {\n\t\t\trevs = append(revs, rev)\n\t\t}\n\t}\n\tr.statusMtx.RUnlock()\n\n\tcommitMsg, err := r.commitMsgBuilder.Build(from, to, revs, emails, r.cfg.Contacts, canary, manualRollRequester)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tsklog.Infof(\"Creating new roll with commit message: \\n%s\", commitMsg)\n\tissueNum, err := r.rm.CreateNewRoll(ctx, from, to, revs, emails, dryRun, commitMsg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tissue := &autoroll.AutoRollIssue{\n\t\tIsDryRun: dryRun,\n\t\tIssue: issueNum,\n\t\tRollingFrom: from.Id,\n\t\tRollingTo: to.Id,\n\t}\n\treturn issue, nil\n}", "func PerformRollingUpgrade(client kubernetes.Interface, config util.Config, envarPostfix string, upgradeFuncs callbacks.RollingUpgradeFuncs) error {\n\titems := upgradeFuncs.ItemsFunc(client, config.Namespace)\n\tvar err error\n\tfor _, i := range items {\n\t\tcontainers := upgradeFuncs.ContainersFunc(i)\n\t\tresourceName := util.ToObjectMeta(i).Name\n\t\t// find correct annotation and update the resource\n\t\tannotationValue := util.ToObjectMeta(i).Annotations[config.Annotation]\n\t\tif annotationValue != \"\" {\n\t\t\tvalues := strings.Split(annotationValue, \",\")\n\t\t\tfor _, value := range values {\n\t\t\t\tif value == config.ResourceName {\n\t\t\t\t\tupdated := updateContainers(containers, value, config.SHAValue, envarPostfix)\n\t\t\t\t\tif !updated {\n\t\t\t\t\t\tlogrus.Warnf(\"Rolling upgrade failed because no container found to add environment variable in %s of type %s in namespace: %s\", resourceName, upgradeFuncs.ResourceType, config.Namespace)\n\t\t\t\t\t} else {\n\t\t\t\t\t\terr = upgradeFuncs.UpdateFunc(client, config.Namespace, i)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tlogrus.Errorf(\"Update for %s of type %s in namespace %s failed with error %v\", resourceName, upgradeFuncs.ResourceType, config.Namespace, err)\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tlogrus.Infof(\"Updated %s of type %s in namespace: %s \", resourceName, upgradeFuncs.ResourceType, config.Namespace)\n\t\t\t\t\t\t}\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn err\n}", "func (r *EndpointsRolloutV1RestClient) AutoUpdateRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\treturn nil, errors.New(\"not allowed\")\n}", "func (n *Sub) retry(uri *model.NotifyURL, msg string, source int) (err error) {\n\tlog.Info(\"Notify.retry do callback url(%v), msg(%s), source(%d)\", uri, msg, source)\n\tfor i := 0; i < _retry; i++ {\n\t\terr = n.clients.Post(context.TODO(), uri, msg)\n\t\tif err != nil {\n\t\t\ttime.Sleep(n.backoff.Backoff(i))\n\t\t\tcontinue\n\t\t} else {\n\t\t\tlog.Info(\"Notify.retry callback success group(%s), topic(%s), retry(%d), msg(%s), source(%d)\",\n\t\t\t\tn.w.Group, n.w.Topic, i, msg, source)\n\t\t\treturn\n\t\t}\n\t}\n\tif err != nil {\n\t\tlog.Error(\"Notify.retry callback error(%v), uri(%s), msg(%s), source(%d)\",\n\t\t\terr, uri, msg, source)\n\t}\n\treturn\n}", "func (t *Handler) runUserInteractions(teamCopy *apps_v1alpha.Team, teamChildNamespaceStr, ownerAuthority, teamOwner, teamOwnerName, operation string, enabled bool) {\n\t// This part creates the rolebindings for the users who participate in the team\n\tfor _, teamUser := range teamCopy.Spec.Users {\n\t\tuser, err := t.edgenetClientset.AppsV1alpha().Users(fmt.Sprintf(\"authority-%s\", teamUser.Authority)).Get(teamUser.Username, metav1.GetOptions{})\n\t\tif err == nil && user.Status.Active && user.Status.AUP {\n\t\t\tif operation == \"team-creation\" {\n\t\t\t\tregistration.CreateRoleBindingsByRoles(user.DeepCopy(), teamChildNamespaceStr, \"Team\")\n\t\t\t}\n\n\t\t\tif !(operation == \"team-creation\" && !enabled) {\n\t\t\t\tt.sendEmail(teamUser.Username, teamUser.Authority, ownerAuthority, teamCopy.GetNamespace(), teamCopy.GetName(), teamChildNamespaceStr, operation)\n\t\t\t}\n\t\t}\n\t}\n\t// To create the rolebindings for the users who are authority-admin and managers of the authority\n\tuserRaw, err := t.edgenetClientset.AppsV1alpha().Users(fmt.Sprintf(\"authority-%s\", ownerAuthority)).List(metav1.ListOptions{})\n\tif err == nil {\n\t\tfor _, userRow := range userRaw.Items {\n\t\t\tif userRow.Status.Active && userRow.Status.AUP && (containsRole(userRow.Spec.Roles, \"admin\") || containsRole(userRow.Spec.Roles, \"manager\")) {\n\t\t\t\tregistration.CreateRoleBindingsByRoles(userRow.DeepCopy(), teamChildNamespaceStr, \"Team\")\n\t\t\t}\n\t\t}\n\t}\n}", "func (r *AutoRoller) setDryRun(issue *autoroll.AutoRollIssue, dryRun bool) error {\n\tinfo, err := issue.ToGerritChangeInfo()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to convert issue to Gerrit ChangeInfo: %s\", err)\n\t}\n\tif dryRun {\n\t\tif err := r.rm.SendToGerritDryRun(info, \"\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tif err := r.rm.SendToGerritCQ(info, \"\"); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tupdated, err := r.retrieveRoll(issue.Issue)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn r.recent.Update(updated)\n}", "func ReconcileVerticalPodAutoscalers(ctx context.Context, namedGetters []NamedVerticalPodAutoscalerCreatorGetter, namespace string, client ctrlruntimeclient.Client, objectModifiers ...ObjectModifier) error {\n\tfor _, get := range namedGetters {\n\t\tname, create := get()\n\t\tcreateObject := VerticalPodAutoscalerObjectWrapper(create)\n\t\tcreateObject = createWithNamespace(createObject, namespace)\n\t\tcreateObject = createWithName(createObject, name)\n\n\t\tfor _, objectModifier := range objectModifiers {\n\t\t\tcreateObject = objectModifier(createObject)\n\t\t}\n\n\t\tif err := EnsureNamedObject(ctx, types.NamespacedName{Namespace: namespace, Name: name}, createObject, client, &autoscalingv1beta2.VerticalPodAutoscaler{}, false); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to ensure VerticalPodAutoscaler %s/%s: %v\", namespace, name, err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (e EndpointsRolloutV1Client) AutoWatchRollout(ctx context.Context, in *api.ListWatchOptions) (RolloutV1_AutoWatchRolloutClient, error) {\n\treturn e.Client.AutoWatchRollout(ctx, in)\n}", "func (e EndpointsRolloutV1Client) AutoUpdateRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoUpdateRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (r *RollupService) Rollup(name string) (err error) {\n\t// prevent nil db\n\tif r.db == nil {\n\t\treturn ErrNilDB\n\t}\n\n\t// prevent nil qs arr\n\tif r.tables == nil {\n\t\treturn ErrNilCreateTableQs\n\t}\n\n\t// prevent zero qs arr\n\tif len(r.tables) == 0 {\n\t\treturn ErrZeroCreateTableQs\n\t}\n\n\terr = r.rollupDB(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdbCreated := true\n\n\terr = r.rollupUser(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\tuserCreated := true\n\n\tcreatedDB, err := connect(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx, err := createdDB.BeginTx(context.Background(), nil)\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\t// ignore error, but should be handled\n\t\t\tif tx != nil {\n\t\t\t\ttx.Rollback()\n\t\t\t}\n\n\t\t\t// close created db if opened\n\t\t\tcreatedDB.Close()\n\n\t\t\tif dbCreated {\n\t\t\t\tr.rollbackDB(name)\n\t\t\t}\n\n\t\t\tif userCreated {\n\t\t\t\tr.rollbackUser(name)\n\t\t\t}\n\n\t\t\treturn\n\t\t}\n\n\t\ttx.Commit()\n\t}()\n\n\tfor _, table := range r.tables {\n\t\terr = r.rollupTableTx(name, table.String(), tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = r.rollupTablePrivilegesTx(name, table.Name, tx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif table.WithSeq {\n\t\t\terr = r.rollupIDSeqPrivilegesTx(name, table.Name, tx)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn\n}", "func (engine ssoEngineImpl) Enroll(authenticatedUser *authenticatedUser) (*common.AuthenticationResponse, error) {\n\n\treturn engine.generateAuthenticationResponse(authenticatedUser)\n}", "func (m *TaskRunner) rollback() {\n\tcount := len(m.rollbacks)\n\tif count == 0 {\n\t\tglog.Infof(\"Nothing to rollback\")\n\t\treturn\n\t}\n\n\t// execute the rollback tasks in reverse order\n\tfor i := count - 1; i >= 0; i-- {\n\t\t_, err := m.rollbacks[i].execute()\n\t\tif err != nil {\n\t\t\tglog.Warningf(\"Error during rollback: Task: '%#v' Error: '%s'\", m.rollbacks[i], err.Error())\n\t\t}\n\t}\n}", "func (r *Roller) onRevRoller(i interface{}, next func(g interface{})) {\n\tr.RevMunch(i)\n\tnext(nil)\n}", "func retry(fn executeJob, attempts int, logger *zap.SugaredLogger) (*common.CrawlResult, error) {\n\tvar (\n\t\tresult *common.CrawlResult\n\t\terr error\n\t)\n\n\tfor i := 0; i < attempts; i++ {\n\t\tresult, err = fn(logger)\n\t\tif err == nil {\n\t\t\treturn result, nil\n\t\t}\n\n\t\tlogger.Debugw(\"failed to execute job\", \"attempt\", i)\n\t\ttime.Sleep(time.Second)\n\t}\n\n\treturn result, err\n}", "func execInner(logger *zap.Logger, name string, stage interface{}) error {\n\tst := Config{}\n\tif err := mapstructure.Decode(stage, &st); err != nil {\n\t\treturn fmt.Errorf(\"ExecStage: unable to decode structure: %v\", err)\n\t}\n\n\treturn run(logger, name, st)\n}", "func (o *Employee) doAfterUpsertHooks(ctx context.Context, exec boil.ContextExecutor) (err error) {\n\tif boil.HooksAreSkipped(ctx) {\n\t\treturn nil\n\t}\n\n\tfor _, hook := range employeeAfterUpsertHooks {\n\t\tif err := hook(ctx, exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (d *Die) Reroll(ctx context.Context) error {\n\tif d == nil {\n\t\treturn ErrNilDie\n\t}\n\tif d.Result == nil {\n\t\treturn ErrUnrolled\n\t}\n\n\td.Result = nil\n\td.Rerolls++\n\t// reroll without reapplying all modifiers\n\treturn d.Roll(ctx)\n}", "func main() {\n\trand.Seed(time.Now().Unix())\n\tdice := NewDice(6, &randomock.Random{})\n\n\tfor roll := 0; roll < 10; roll++ {\n\t\tresult := dice.Roll()\n\t\tfmt.Printf(\"Rolling dice attempt %d.... got %d\\n\", roll, result)\n\t}\n}", "func RollDie() (int, error) {\n\tr, err := rand.Int(rand.Reader, sides)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(r.Int64()) + 1, nil\n}", "func (imw *innerMergeWorker) fetchNextInnerResult(ctx context.Context, task *lookUpMergeJoinTask) (beginRow chunk.Row, err error) {\n\ttask.innerResult = imw.ctx.GetSessionVars().GetNewChunkWithCapacity(retTypes(imw.innerExec), imw.ctx.GetSessionVars().MaxChunkSize, imw.ctx.GetSessionVars().MaxChunkSize, imw.innerExec.Base().AllocPool)\n\terr = Next(ctx, imw.innerExec, task.innerResult)\n\ttask.innerIter = chunk.NewIterator4Chunk(task.innerResult)\n\tbeginRow = task.innerIter.Begin()\n\treturn\n}", "func (dsc *ReconcileDaemonSet) rollingUpdate(ds *appsv1alpha1.DaemonSet, hash string) (delay time.Duration, err error) {\n\n\tif ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.StandardRollingUpdateType {\n\t\treturn delay, dsc.standardRollingUpdate(ds, hash)\n\t} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.SurgingRollingUpdateType {\n\t\treturn dsc.surgingRollingUpdate(ds, hash)\n\t\t//} else if ds.Spec.UpdateStrategy.RollingUpdate.Type == appsv1alpha1.InplaceRollingUpdateType {\n\t\t//\treturn dsc.inplaceRollingUpdate(ds, hash)\n\t} else {\n\t\tklog.Errorf(\"no matched RollingUpdate type\")\n\t}\n\treturn\n}", "func (jr *joinReader) mainLoop(ctx context.Context) error {\n\tprimaryKeyPrefix := sqlbase.MakeIndexKeyPrefix(&jr.desc, jr.index.ID)\n\n\tvar alloc sqlbase.DatumAlloc\n\tspans := make(roachpb.Spans, 0, joinReaderBatchSize)\n\n\ttxn := jr.flowCtx.txn\n\tif txn == nil {\n\t\tlog.Fatalf(ctx, \"joinReader outside of txn\")\n\t}\n\n\tlog.VEventf(ctx, 1, \"starting\")\n\tif log.V(1) {\n\t\tdefer log.Infof(ctx, \"exiting\")\n\t}\n\n\tspanToRows := make(map[string][]sqlbase.EncDatumRow)\n\tfor {\n\t\t// TODO(radu): figure out how to send smaller batches if the source has\n\t\t// a soft limit (perhaps send the batch out if we don't get a result\n\t\t// within a certain amount of time).\n\t\tfor spans = spans[:0]; len(spans) < joinReaderBatchSize; {\n\t\t\trow, meta := jr.input.Next()\n\t\t\tif meta != nil {\n\t\t\t\tif meta.Err != nil {\n\t\t\t\t\treturn meta.Err\n\t\t\t\t}\n\t\t\t\tif !emitHelper(ctx, &jr.out, nil /* row */, meta, jr.pushTrailingMeta, jr.input) {\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif row == nil {\n\t\t\t\tif len(spans) == 0 {\n\t\t\t\t\t// No fetching needed since we have collected no spans and\n\t\t\t\t\t// the input has signaled that no more records are coming.\n\t\t\t\t\tjr.out.Close()\n\t\t\t\t\treturn nil\n\t\t\t\t}\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tkey, err := jr.generateKey(row, &alloc, primaryKeyPrefix, jr.lookupCols)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tspan := roachpb.Span{\n\t\t\t\tKey: key,\n\t\t\t\tEndKey: key.PrefixEnd(),\n\t\t\t}\n\t\t\tif jr.isLookupJoin() {\n\t\t\t\tif spanToRows[key.String()] == nil {\n\t\t\t\t\tspans = append(spans, span)\n\t\t\t\t}\n\t\t\t\tspanToRows[key.String()] = append(spanToRows[key.String()], row)\n\t\t\t} else {\n\t\t\t\tspans = append(spans, span)\n\t\t\t}\n\t\t}\n\n\t\t// TODO(radu): we are consuming all results from a fetch before starting\n\t\t// the next batch. We could start the next batch early while we are\n\t\t// outputting rows.\n\t\tif earlyExit, err := jr.indexLookup(ctx, txn, spans, spanToRows); err != nil {\n\t\t\treturn err\n\t\t} else if earlyExit {\n\t\t\treturn nil\n\t\t}\n\n\t\tif len(spans) != joinReaderBatchSize {\n\t\t\t// This was the last batch.\n\t\t\tjr.pushTrailingMeta(ctx)\n\t\t\tjr.out.Close()\n\t\t\treturn nil\n\t\t}\n\t}\n}", "func (m *MockIDistributedEnforcer) UpdatePoliciesSelf(arg0 func() bool, arg1, arg2 string, arg3, arg4 [][]string) (bool, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"UpdatePoliciesSelf\", arg0, arg1, arg2, arg3, arg4)\n\tret0, _ := ret[0].(bool)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func rollingDeploy(clientset *kubernetes.Clientset, namespace, name, image string) error {\n\n\tfmt.Printf(\"rolling upgrade : %s \\n\", name)\n\n\tupdateClient := clientset.AppsV1().Deployments(namespace)\n\n\tretryErr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t// Retrieve the latest version of Deployment before attempting update\n\t\t// RetryOnConflict uses exponential backoff to avoid exhausting the apiserver\n\t\tresult, getErr := updateClient.Get(name, metav1.GetOptions{})\n\t\tif getErr != nil {\n\t\t\tpanic(fmt.Errorf(\"Failed to get latest version of Deployment: %v\", getErr))\n\t\t}\n\n\t\tresult.Spec.Template.Spec.Containers[0].Image = image // change nginx version\n\n\t\t_, updateErr := updateClient.Update(result)\n\n\t\treturn updateErr\n\t})\n\n\tif retryErr != nil {\n\t\treturn retryErr\n\t}\n\n\treturn nil\n}", "func InnerLoop(out1 *uint64, out2 *uint64, out3 *uint64, out4 *uint64, out5 *uint64, arg1 uint64, arg2 *[3]uint64, arg3 *[3]uint64) {\n\tx1 := (uint64((uint128((^arg1)) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x2 uint64\n\tcmovznzU64(&x2, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), arg1, (uint64((uint128((^arg1)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x3 uint64\n\tcmovznzU64(&x3, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), arg2[0], arg3[0])\n\tvar x4 uint64\n\tcmovznzU64(&x4, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), arg3[0], (uint64((uint128((^arg2[0])) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x5 uint1\n\tcmovznzU1(&x5, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), 0x1, 0x0)\n\tvar x6 uint1\n\tcmovznzU1(&x6, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), 0x0, 0x1)\n\tvar x7 uint64\n\tcmovznzU64(&x7, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), uint64(0x0), 0xffffffffffffffff)\n\tvar x8 uint1\n\tcmovznzU1(&x8, (uint1((x1 >> 63)) & (uint1(arg3[0]) & 0x1)), 0x1, 0x0)\n\tvar x9 uint64\n\tcmovznzU64(&x9, (uint1(x4) & 0x1), uint64(0x0), x3)\n\tvar x10 uint1\n\tcmovznzU1(&x10, (uint1(x4) & 0x1), 0x0, x5)\n\tvar x11 uint1\n\tcmovznzU1(&x11, (uint1(x4) & 0x1), 0x0, x6)\n\tx12 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x2))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x13 uint64\n\tcmovznzU64(&x13, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x2))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x2))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x14 uint64\n\tcmovznzU64(&x14, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), x3, (((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1)))\n\tvar x15 uint64\n\tcmovznzU64(&x15, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x3)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x16 uint64\n\tcmovznzU64(&x16, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64(x5) + uint64(x5)), (uint64((uint128(x7) + uint128(x10))) & 0xffffffffffffffff))\n\tvar x17 uint64\n\tcmovznzU64(&x17, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64(x6) + uint64(x6)), (uint64(x8) + uint64(x11)))\n\tvar x18 uint64\n\tcmovznzU64(&x18, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x7) + uint128(x10))) & 0xffffffffffffffff), (uint64((uint128((^(uint64(x5) + uint64(x5)))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x19 uint64\n\tcmovznzU64(&x19, (uint1((x12 >> 63)) & (uint1((((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x4) + uint128(x9))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64(x8) + uint64(x11)), (uint64((uint128((^(uint64(x6) + uint64(x6)))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x20 uint64\n\tcmovznzU64(&x20, (uint1(x15) & 0x1), uint64(0x0), x14)\n\tvar x21 uint64\n\tcmovznzU64(&x21, (uint1(x15) & 0x1), uint64(0x0), x16)\n\tvar x22 uint64\n\tcmovznzU64(&x22, (uint1(x15) & 0x1), uint64(0x0), x17)\n\tx23 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x13))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x24 uint64\n\tcmovznzU64(&x24, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x13))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x13))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x25 uint64\n\tcmovznzU64(&x25, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), x14, (((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1)))\n\tvar x26 uint64\n\tcmovznzU64(&x26, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x14)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x27 uint64\n\tcmovznzU64(&x27, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x16) + uint128(x16))) & 0xffffffffffffffff), (uint64((uint128(x18) + uint128(x21))) & 0xffffffffffffffff))\n\tvar x28 uint64\n\tcmovznzU64(&x28, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (x17 + x17), (uint64((uint128(x19) + uint128(x22))) & 0xffffffffffffffff))\n\tvar x29 uint64\n\tcmovznzU64(&x29, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x18) + uint128(x21))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x16) + uint128(x16))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x30 uint64\n\tcmovznzU64(&x30, (uint1((x23 >> 63)) & (uint1((((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x15) + uint128(x20))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x19) + uint128(x22))) & 0xffffffffffffffff), (uint64((uint128((^(x17 + x17))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x31 uint64\n\tcmovznzU64(&x31, (uint1(x26) & 0x1), uint64(0x0), x25)\n\tvar x32 uint64\n\tcmovznzU64(&x32, (uint1(x26) & 0x1), uint64(0x0), x27)\n\tvar x33 uint64\n\tcmovznzU64(&x33, (uint1(x26) & 0x1), uint64(0x0), x28)\n\tx34 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x24))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x35 uint64\n\tcmovznzU64(&x35, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x24))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x24))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x36 uint64\n\tcmovznzU64(&x36, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), x25, (((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1)))\n\tvar x37 uint64\n\tcmovznzU64(&x37, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x25)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x38 uint64\n\tcmovznzU64(&x38, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x27) + uint128(x27))) & 0xffffffffffffffff), (uint64((uint128(x29) + uint128(x32))) & 0xffffffffffffffff))\n\tvar x39 uint64\n\tcmovznzU64(&x39, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x28) + uint128(x28))) & 0xffffffffffffffff), (uint64((uint128(x30) + uint128(x33))) & 0xffffffffffffffff))\n\tvar x40 uint64\n\tcmovznzU64(&x40, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x29) + uint128(x32))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x27) + uint128(x27))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x41 uint64\n\tcmovznzU64(&x41, (uint1((x34 >> 63)) & (uint1((((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x26) + uint128(x31))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x30) + uint128(x33))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x28) + uint128(x28))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x42 uint64\n\tcmovznzU64(&x42, (uint1(x37) & 0x1), uint64(0x0), x36)\n\tvar x43 uint64\n\tcmovznzU64(&x43, (uint1(x37) & 0x1), uint64(0x0), x38)\n\tvar x44 uint64\n\tcmovznzU64(&x44, (uint1(x37) & 0x1), uint64(0x0), x39)\n\tx45 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x35))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x46 uint64\n\tcmovznzU64(&x46, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x35))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x35))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x47 uint64\n\tcmovznzU64(&x47, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), x36, (((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1)))\n\tvar x48 uint64\n\tcmovznzU64(&x48, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x36)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x49 uint64\n\tcmovznzU64(&x49, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x38) + uint128(x38))) & 0xffffffffffffffff), (uint64((uint128(x40) + uint128(x43))) & 0xffffffffffffffff))\n\tvar x50 uint64\n\tcmovznzU64(&x50, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x39) + uint128(x39))) & 0xffffffffffffffff), (uint64((uint128(x41) + uint128(x44))) & 0xffffffffffffffff))\n\tvar x51 uint64\n\tcmovznzU64(&x51, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x40) + uint128(x43))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x38) + uint128(x38))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x52 uint64\n\tcmovznzU64(&x52, (uint1((x45 >> 63)) & (uint1((((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x37) + uint128(x42))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x41) + uint128(x44))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x39) + uint128(x39))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x53 uint64\n\tcmovznzU64(&x53, (uint1(x48) & 0x1), uint64(0x0), x47)\n\tvar x54 uint64\n\tcmovznzU64(&x54, (uint1(x48) & 0x1), uint64(0x0), x49)\n\tvar x55 uint64\n\tcmovznzU64(&x55, (uint1(x48) & 0x1), uint64(0x0), x50)\n\tx56 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x46))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x57 uint64\n\tcmovznzU64(&x57, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x46))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x46))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x58 uint64\n\tcmovznzU64(&x58, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), x47, (((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1)))\n\tvar x59 uint64\n\tcmovznzU64(&x59, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x47)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x60 uint64\n\tcmovznzU64(&x60, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x49) + uint128(x49))) & 0xffffffffffffffff), (uint64((uint128(x51) + uint128(x54))) & 0xffffffffffffffff))\n\tvar x61 uint64\n\tcmovznzU64(&x61, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x50) + uint128(x50))) & 0xffffffffffffffff), (uint64((uint128(x52) + uint128(x55))) & 0xffffffffffffffff))\n\tvar x62 uint64\n\tcmovznzU64(&x62, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x51) + uint128(x54))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x49) + uint128(x49))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x63 uint64\n\tcmovznzU64(&x63, (uint1((x56 >> 63)) & (uint1((((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x48) + uint128(x53))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x52) + uint128(x55))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x50) + uint128(x50))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x64 uint64\n\tcmovznzU64(&x64, (uint1(x59) & 0x1), uint64(0x0), x58)\n\tvar x65 uint64\n\tcmovznzU64(&x65, (uint1(x59) & 0x1), uint64(0x0), x60)\n\tvar x66 uint64\n\tcmovznzU64(&x66, (uint1(x59) & 0x1), uint64(0x0), x61)\n\tx67 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x57))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x68 uint64\n\tcmovznzU64(&x68, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x57))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x57))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x69 uint64\n\tcmovznzU64(&x69, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), x58, (((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1)))\n\tvar x70 uint64\n\tcmovznzU64(&x70, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x58)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x71 uint64\n\tcmovznzU64(&x71, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x60) + uint128(x60))) & 0xffffffffffffffff), (uint64((uint128(x62) + uint128(x65))) & 0xffffffffffffffff))\n\tvar x72 uint64\n\tcmovznzU64(&x72, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x61) + uint128(x61))) & 0xffffffffffffffff), (uint64((uint128(x63) + uint128(x66))) & 0xffffffffffffffff))\n\tvar x73 uint64\n\tcmovznzU64(&x73, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x62) + uint128(x65))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x60) + uint128(x60))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x74 uint64\n\tcmovznzU64(&x74, (uint1((x67 >> 63)) & (uint1((((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x59) + uint128(x64))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x63) + uint128(x66))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x61) + uint128(x61))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x75 uint64\n\tcmovznzU64(&x75, (uint1(x70) & 0x1), uint64(0x0), x69)\n\tvar x76 uint64\n\tcmovznzU64(&x76, (uint1(x70) & 0x1), uint64(0x0), x71)\n\tvar x77 uint64\n\tcmovznzU64(&x77, (uint1(x70) & 0x1), uint64(0x0), x72)\n\tx78 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x68))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x79 uint64\n\tcmovznzU64(&x79, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x68))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x68))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x80 uint64\n\tcmovznzU64(&x80, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), x69, (((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1)))\n\tvar x81 uint64\n\tcmovznzU64(&x81, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x69)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x82 uint64\n\tcmovznzU64(&x82, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x71) + uint128(x71))) & 0xffffffffffffffff), (uint64((uint128(x73) + uint128(x76))) & 0xffffffffffffffff))\n\tvar x83 uint64\n\tcmovznzU64(&x83, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x72) + uint128(x72))) & 0xffffffffffffffff), (uint64((uint128(x74) + uint128(x77))) & 0xffffffffffffffff))\n\tvar x84 uint64\n\tcmovznzU64(&x84, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x73) + uint128(x76))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x71) + uint128(x71))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x85 uint64\n\tcmovznzU64(&x85, (uint1((x78 >> 63)) & (uint1((((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x70) + uint128(x75))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x74) + uint128(x77))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x72) + uint128(x72))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x86 uint64\n\tcmovznzU64(&x86, (uint1(x81) & 0x1), uint64(0x0), x80)\n\tvar x87 uint64\n\tcmovznzU64(&x87, (uint1(x81) & 0x1), uint64(0x0), x82)\n\tvar x88 uint64\n\tcmovznzU64(&x88, (uint1(x81) & 0x1), uint64(0x0), x83)\n\tx89 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x79))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x90 uint64\n\tcmovznzU64(&x90, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x79))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x79))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x91 uint64\n\tcmovznzU64(&x91, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), x80, (((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1)))\n\tvar x92 uint64\n\tcmovznzU64(&x92, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x80)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x93 uint64\n\tcmovznzU64(&x93, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x82) + uint128(x82))) & 0xffffffffffffffff), (uint64((uint128(x84) + uint128(x87))) & 0xffffffffffffffff))\n\tvar x94 uint64\n\tcmovznzU64(&x94, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x83) + uint128(x83))) & 0xffffffffffffffff), (uint64((uint128(x85) + uint128(x88))) & 0xffffffffffffffff))\n\tvar x95 uint64\n\tcmovznzU64(&x95, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x84) + uint128(x87))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x82) + uint128(x82))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x96 uint64\n\tcmovznzU64(&x96, (uint1((x89 >> 63)) & (uint1((((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x81) + uint128(x86))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x85) + uint128(x88))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x83) + uint128(x83))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x97 uint64\n\tcmovznzU64(&x97, (uint1(x92) & 0x1), uint64(0x0), x91)\n\tvar x98 uint64\n\tcmovznzU64(&x98, (uint1(x92) & 0x1), uint64(0x0), x93)\n\tvar x99 uint64\n\tcmovznzU64(&x99, (uint1(x92) & 0x1), uint64(0x0), x94)\n\tx100 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x90))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x101 uint64\n\tcmovznzU64(&x101, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x90))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x90))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x102 uint64\n\tcmovznzU64(&x102, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), x91, (((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1)))\n\tvar x103 uint64\n\tcmovznzU64(&x103, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x91)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x104 uint64\n\tcmovznzU64(&x104, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x93) + uint128(x93))) & 0xffffffffffffffff), (uint64((uint128(x95) + uint128(x98))) & 0xffffffffffffffff))\n\tvar x105 uint64\n\tcmovznzU64(&x105, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x94) + uint128(x94))) & 0xffffffffffffffff), (uint64((uint128(x96) + uint128(x99))) & 0xffffffffffffffff))\n\tvar x106 uint64\n\tcmovznzU64(&x106, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x95) + uint128(x98))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x93) + uint128(x93))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x107 uint64\n\tcmovznzU64(&x107, (uint1((x100 >> 63)) & (uint1((((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x92) + uint128(x97))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x96) + uint128(x99))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x94) + uint128(x94))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x108 uint64\n\tcmovznzU64(&x108, (uint1(x103) & 0x1), uint64(0x0), x102)\n\tvar x109 uint64\n\tcmovznzU64(&x109, (uint1(x103) & 0x1), uint64(0x0), x104)\n\tvar x110 uint64\n\tcmovznzU64(&x110, (uint1(x103) & 0x1), uint64(0x0), x105)\n\tx111 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x101))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x112 uint64\n\tcmovznzU64(&x112, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x101))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x101))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x113 uint64\n\tcmovznzU64(&x113, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), x102, (((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1)))\n\tvar x114 uint64\n\tcmovznzU64(&x114, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x102)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x115 uint64\n\tcmovznzU64(&x115, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x104) + uint128(x104))) & 0xffffffffffffffff), (uint64((uint128(x106) + uint128(x109))) & 0xffffffffffffffff))\n\tvar x116 uint64\n\tcmovznzU64(&x116, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x105) + uint128(x105))) & 0xffffffffffffffff), (uint64((uint128(x107) + uint128(x110))) & 0xffffffffffffffff))\n\tvar x117 uint64\n\tcmovznzU64(&x117, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x106) + uint128(x109))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x104) + uint128(x104))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x118 uint64\n\tcmovznzU64(&x118, (uint1((x111 >> 63)) & (uint1((((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x103) + uint128(x108))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x107) + uint128(x110))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x105) + uint128(x105))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x119 uint64\n\tcmovznzU64(&x119, (uint1(x114) & 0x1), uint64(0x0), x113)\n\tvar x120 uint64\n\tcmovznzU64(&x120, (uint1(x114) & 0x1), uint64(0x0), x115)\n\tvar x121 uint64\n\tcmovznzU64(&x121, (uint1(x114) & 0x1), uint64(0x0), x116)\n\tx122 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x112))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x123 uint64\n\tcmovznzU64(&x123, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x112))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x112))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x124 uint64\n\tcmovznzU64(&x124, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), x113, (((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1)))\n\tvar x125 uint64\n\tcmovznzU64(&x125, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x113)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x126 uint64\n\tcmovznzU64(&x126, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x115) + uint128(x115))) & 0xffffffffffffffff), (uint64((uint128(x117) + uint128(x120))) & 0xffffffffffffffff))\n\tvar x127 uint64\n\tcmovznzU64(&x127, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x116) + uint128(x116))) & 0xffffffffffffffff), (uint64((uint128(x118) + uint128(x121))) & 0xffffffffffffffff))\n\tvar x128 uint64\n\tcmovznzU64(&x128, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x117) + uint128(x120))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x115) + uint128(x115))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x129 uint64\n\tcmovznzU64(&x129, (uint1((x122 >> 63)) & (uint1((((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x114) + uint128(x119))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x118) + uint128(x121))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x116) + uint128(x116))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x130 uint64\n\tcmovznzU64(&x130, (uint1(x125) & 0x1), uint64(0x0), x124)\n\tvar x131 uint64\n\tcmovznzU64(&x131, (uint1(x125) & 0x1), uint64(0x0), x126)\n\tvar x132 uint64\n\tcmovznzU64(&x132, (uint1(x125) & 0x1), uint64(0x0), x127)\n\tx133 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x123))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x134 uint64\n\tcmovznzU64(&x134, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x123))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x123))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x135 uint64\n\tcmovznzU64(&x135, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), x124, (((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1)))\n\tvar x136 uint64\n\tcmovznzU64(&x136, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x124)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x137 uint64\n\tcmovznzU64(&x137, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x126) + uint128(x126))) & 0xffffffffffffffff), (uint64((uint128(x128) + uint128(x131))) & 0xffffffffffffffff))\n\tvar x138 uint64\n\tcmovznzU64(&x138, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x127) + uint128(x127))) & 0xffffffffffffffff), (uint64((uint128(x129) + uint128(x132))) & 0xffffffffffffffff))\n\tvar x139 uint64\n\tcmovznzU64(&x139, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x128) + uint128(x131))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x126) + uint128(x126))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x140 uint64\n\tcmovznzU64(&x140, (uint1((x133 >> 63)) & (uint1((((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x125) + uint128(x130))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x129) + uint128(x132))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x127) + uint128(x127))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x141 uint64\n\tcmovznzU64(&x141, (uint1(x136) & 0x1), uint64(0x0), x135)\n\tvar x142 uint64\n\tcmovznzU64(&x142, (uint1(x136) & 0x1), uint64(0x0), x137)\n\tvar x143 uint64\n\tcmovznzU64(&x143, (uint1(x136) & 0x1), uint64(0x0), x138)\n\tx144 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x134))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x145 uint64\n\tcmovznzU64(&x145, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x134))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x134))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x146 uint64\n\tcmovznzU64(&x146, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), x135, (((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1)))\n\tvar x147 uint64\n\tcmovznzU64(&x147, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x135)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x148 uint64\n\tcmovznzU64(&x148, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x137) + uint128(x137))) & 0xffffffffffffffff), (uint64((uint128(x139) + uint128(x142))) & 0xffffffffffffffff))\n\tvar x149 uint64\n\tcmovznzU64(&x149, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x138) + uint128(x138))) & 0xffffffffffffffff), (uint64((uint128(x140) + uint128(x143))) & 0xffffffffffffffff))\n\tvar x150 uint64\n\tcmovznzU64(&x150, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x139) + uint128(x142))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x137) + uint128(x137))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x151 uint64\n\tcmovznzU64(&x151, (uint1((x144 >> 63)) & (uint1((((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x136) + uint128(x141))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x140) + uint128(x143))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x138) + uint128(x138))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x152 uint64\n\tcmovznzU64(&x152, (uint1(x147) & 0x1), uint64(0x0), x146)\n\tvar x153 uint64\n\tcmovznzU64(&x153, (uint1(x147) & 0x1), uint64(0x0), x148)\n\tvar x154 uint64\n\tcmovznzU64(&x154, (uint1(x147) & 0x1), uint64(0x0), x149)\n\tx155 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x145))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x156 uint64\n\tcmovznzU64(&x156, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x145))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x145))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x157 uint64\n\tcmovznzU64(&x157, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), x146, (((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1)))\n\tvar x158 uint64\n\tcmovznzU64(&x158, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x146)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x159 uint64\n\tcmovznzU64(&x159, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x148) + uint128(x148))) & 0xffffffffffffffff), (uint64((uint128(x150) + uint128(x153))) & 0xffffffffffffffff))\n\tvar x160 uint64\n\tcmovznzU64(&x160, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x149) + uint128(x149))) & 0xffffffffffffffff), (uint64((uint128(x151) + uint128(x154))) & 0xffffffffffffffff))\n\tvar x161 uint64\n\tcmovznzU64(&x161, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x150) + uint128(x153))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x148) + uint128(x148))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x162 uint64\n\tcmovznzU64(&x162, (uint1((x155 >> 63)) & (uint1((((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x147) + uint128(x152))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x151) + uint128(x154))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x149) + uint128(x149))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x163 uint64\n\tcmovznzU64(&x163, (uint1(x158) & 0x1), uint64(0x0), x157)\n\tvar x164 uint64\n\tcmovznzU64(&x164, (uint1(x158) & 0x1), uint64(0x0), x159)\n\tvar x165 uint64\n\tcmovznzU64(&x165, (uint1(x158) & 0x1), uint64(0x0), x160)\n\tx166 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x156))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x167 uint64\n\tcmovznzU64(&x167, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x156))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x156))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x168 uint64\n\tcmovznzU64(&x168, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), x157, (((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1)))\n\tvar x169 uint64\n\tcmovznzU64(&x169, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x157)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x170 uint64\n\tcmovznzU64(&x170, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x159) + uint128(x159))) & 0xffffffffffffffff), (uint64((uint128(x161) + uint128(x164))) & 0xffffffffffffffff))\n\tvar x171 uint64\n\tcmovznzU64(&x171, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x160) + uint128(x160))) & 0xffffffffffffffff), (uint64((uint128(x162) + uint128(x165))) & 0xffffffffffffffff))\n\tvar x172 uint64\n\tcmovznzU64(&x172, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x161) + uint128(x164))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x159) + uint128(x159))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x173 uint64\n\tcmovznzU64(&x173, (uint1((x166 >> 63)) & (uint1((((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x158) + uint128(x163))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x162) + uint128(x165))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x160) + uint128(x160))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x174 uint64\n\tcmovznzU64(&x174, (uint1(x169) & 0x1), uint64(0x0), x168)\n\tvar x175 uint64\n\tcmovznzU64(&x175, (uint1(x169) & 0x1), uint64(0x0), x170)\n\tvar x176 uint64\n\tcmovznzU64(&x176, (uint1(x169) & 0x1), uint64(0x0), x171)\n\tx177 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x167))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x178 uint64\n\tcmovznzU64(&x178, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x167))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x167))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x179 uint64\n\tcmovznzU64(&x179, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), x168, (((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1)))\n\tvar x180 uint64\n\tcmovznzU64(&x180, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x168)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x181 uint64\n\tcmovznzU64(&x181, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x170) + uint128(x170))) & 0xffffffffffffffff), (uint64((uint128(x172) + uint128(x175))) & 0xffffffffffffffff))\n\tvar x182 uint64\n\tcmovznzU64(&x182, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x171) + uint128(x171))) & 0xffffffffffffffff), (uint64((uint128(x173) + uint128(x176))) & 0xffffffffffffffff))\n\tvar x183 uint64\n\tcmovznzU64(&x183, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x172) + uint128(x175))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x170) + uint128(x170))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x184 uint64\n\tcmovznzU64(&x184, (uint1((x177 >> 63)) & (uint1((((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x169) + uint128(x174))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x173) + uint128(x176))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x171) + uint128(x171))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x185 uint64\n\tcmovznzU64(&x185, (uint1(x180) & 0x1), uint64(0x0), x179)\n\tvar x186 uint64\n\tcmovznzU64(&x186, (uint1(x180) & 0x1), uint64(0x0), x181)\n\tvar x187 uint64\n\tcmovznzU64(&x187, (uint1(x180) & 0x1), uint64(0x0), x182)\n\tx188 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x178))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x189 uint64\n\tcmovznzU64(&x189, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x178))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x178))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x190 uint64\n\tcmovznzU64(&x190, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), x179, (((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1)))\n\tvar x191 uint64\n\tcmovznzU64(&x191, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x179)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x192 uint64\n\tcmovznzU64(&x192, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x181) + uint128(x181))) & 0xffffffffffffffff), (uint64((uint128(x183) + uint128(x186))) & 0xffffffffffffffff))\n\tvar x193 uint64\n\tcmovznzU64(&x193, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x182) + uint128(x182))) & 0xffffffffffffffff), (uint64((uint128(x184) + uint128(x187))) & 0xffffffffffffffff))\n\tvar x194 uint64\n\tcmovznzU64(&x194, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x183) + uint128(x186))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x181) + uint128(x181))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x195 uint64\n\tcmovznzU64(&x195, (uint1((x188 >> 63)) & (uint1((((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x180) + uint128(x185))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x184) + uint128(x187))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x182) + uint128(x182))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x196 uint64\n\tcmovznzU64(&x196, (uint1(x191) & 0x1), uint64(0x0), x190)\n\tvar x197 uint64\n\tcmovznzU64(&x197, (uint1(x191) & 0x1), uint64(0x0), x192)\n\tvar x198 uint64\n\tcmovznzU64(&x198, (uint1(x191) & 0x1), uint64(0x0), x193)\n\tx199 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x189))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x200 uint64\n\tcmovznzU64(&x200, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x189))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x189))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x201 uint64\n\tcmovznzU64(&x201, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), x190, (((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1)))\n\tvar x202 uint64\n\tcmovznzU64(&x202, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x190)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x203 uint64\n\tcmovznzU64(&x203, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x192) + uint128(x192))) & 0xffffffffffffffff), (uint64((uint128(x194) + uint128(x197))) & 0xffffffffffffffff))\n\tvar x204 uint64\n\tcmovznzU64(&x204, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x193) + uint128(x193))) & 0xffffffffffffffff), (uint64((uint128(x195) + uint128(x198))) & 0xffffffffffffffff))\n\tvar x205 uint64\n\tcmovznzU64(&x205, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x194) + uint128(x197))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x192) + uint128(x192))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x206 uint64\n\tcmovznzU64(&x206, (uint1((x199 >> 63)) & (uint1((((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x191) + uint128(x196))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x195) + uint128(x198))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x193) + uint128(x193))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x207 uint64\n\tcmovznzU64(&x207, (uint1(x202) & 0x1), uint64(0x0), x201)\n\tvar x208 uint64\n\tcmovznzU64(&x208, (uint1(x202) & 0x1), uint64(0x0), x203)\n\tvar x209 uint64\n\tcmovznzU64(&x209, (uint1(x202) & 0x1), uint64(0x0), x204)\n\tx210 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x200))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x211 uint64\n\tcmovznzU64(&x211, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x200))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x200))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x212 uint64\n\tcmovznzU64(&x212, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), x201, (((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1)))\n\tvar x213 uint64\n\tcmovznzU64(&x213, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x201)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x214 uint64\n\tcmovznzU64(&x214, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x203) + uint128(x203))) & 0xffffffffffffffff), (uint64((uint128(x205) + uint128(x208))) & 0xffffffffffffffff))\n\tvar x215 uint64\n\tcmovznzU64(&x215, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x204) + uint128(x204))) & 0xffffffffffffffff), (uint64((uint128(x206) + uint128(x209))) & 0xffffffffffffffff))\n\tvar x216 uint64\n\tcmovznzU64(&x216, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x205) + uint128(x208))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x203) + uint128(x203))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x217 uint64\n\tcmovznzU64(&x217, (uint1((x210 >> 63)) & (uint1((((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x202) + uint128(x207))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x206) + uint128(x209))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x204) + uint128(x204))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x218 uint64\n\tcmovznzU64(&x218, (uint1(x213) & 0x1), uint64(0x0), x212)\n\tvar x219 uint64\n\tcmovznzU64(&x219, (uint1(x213) & 0x1), uint64(0x0), x214)\n\tvar x220 uint64\n\tcmovznzU64(&x220, (uint1(x213) & 0x1), uint64(0x0), x215)\n\tx221 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x211))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x222 uint64\n\tcmovznzU64(&x222, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x211))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x211))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x223 uint64\n\tcmovznzU64(&x223, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), x212, (((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1)))\n\tvar x224 uint64\n\tcmovznzU64(&x224, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x212)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x225 uint64\n\tcmovznzU64(&x225, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x214) + uint128(x214))) & 0xffffffffffffffff), (uint64((uint128(x216) + uint128(x219))) & 0xffffffffffffffff))\n\tvar x226 uint64\n\tcmovznzU64(&x226, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x215) + uint128(x215))) & 0xffffffffffffffff), (uint64((uint128(x217) + uint128(x220))) & 0xffffffffffffffff))\n\tvar x227 uint64\n\tcmovznzU64(&x227, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x216) + uint128(x219))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x214) + uint128(x214))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x228 uint64\n\tcmovznzU64(&x228, (uint1((x221 >> 63)) & (uint1((((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x213) + uint128(x218))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x217) + uint128(x220))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x215) + uint128(x215))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x229 uint64\n\tcmovznzU64(&x229, (uint1(x224) & 0x1), uint64(0x0), x223)\n\tvar x230 uint64\n\tcmovznzU64(&x230, (uint1(x224) & 0x1), uint64(0x0), x225)\n\tvar x231 uint64\n\tcmovznzU64(&x231, (uint1(x224) & 0x1), uint64(0x0), x226)\n\tx232 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x222))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x233 uint64\n\tcmovznzU64(&x233, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x222))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x222))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x234 uint64\n\tcmovznzU64(&x234, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), x223, (((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1)))\n\tvar x235 uint64\n\tcmovznzU64(&x235, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x223)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x236 uint64\n\tcmovznzU64(&x236, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x225) + uint128(x225))) & 0xffffffffffffffff), (uint64((uint128(x227) + uint128(x230))) & 0xffffffffffffffff))\n\tvar x237 uint64\n\tcmovznzU64(&x237, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x226) + uint128(x226))) & 0xffffffffffffffff), (uint64((uint128(x228) + uint128(x231))) & 0xffffffffffffffff))\n\tvar x238 uint64\n\tcmovznzU64(&x238, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x227) + uint128(x230))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x225) + uint128(x225))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x239 uint64\n\tcmovznzU64(&x239, (uint1((x232 >> 63)) & (uint1((((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x224) + uint128(x229))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x228) + uint128(x231))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x226) + uint128(x226))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x240 uint64\n\tcmovznzU64(&x240, (uint1(x235) & 0x1), uint64(0x0), x234)\n\tvar x241 uint64\n\tcmovznzU64(&x241, (uint1(x235) & 0x1), uint64(0x0), x236)\n\tvar x242 uint64\n\tcmovznzU64(&x242, (uint1(x235) & 0x1), uint64(0x0), x237)\n\tx243 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x233))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x244 uint64\n\tcmovznzU64(&x244, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x233))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x233))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x245 uint64\n\tcmovznzU64(&x245, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), x234, (((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1)))\n\tvar x246 uint64\n\tcmovznzU64(&x246, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x234)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x247 uint64\n\tcmovznzU64(&x247, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x236) + uint128(x236))) & 0xffffffffffffffff), (uint64((uint128(x238) + uint128(x241))) & 0xffffffffffffffff))\n\tvar x248 uint64\n\tcmovznzU64(&x248, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x237) + uint128(x237))) & 0xffffffffffffffff), (uint64((uint128(x239) + uint128(x242))) & 0xffffffffffffffff))\n\tvar x249 uint64\n\tcmovznzU64(&x249, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x238) + uint128(x241))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x236) + uint128(x236))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x250 uint64\n\tcmovznzU64(&x250, (uint1((x243 >> 63)) & (uint1((((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x235) + uint128(x240))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x239) + uint128(x242))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x237) + uint128(x237))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x251 uint64\n\tcmovznzU64(&x251, (uint1(x246) & 0x1), uint64(0x0), x245)\n\tvar x252 uint64\n\tcmovznzU64(&x252, (uint1(x246) & 0x1), uint64(0x0), x247)\n\tvar x253 uint64\n\tcmovznzU64(&x253, (uint1(x246) & 0x1), uint64(0x0), x248)\n\tx254 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x244))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x255 uint64\n\tcmovznzU64(&x255, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x244))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x244))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x256 uint64\n\tcmovznzU64(&x256, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), x245, (((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1)))\n\tvar x257 uint64\n\tcmovznzU64(&x257, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x245)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x258 uint64\n\tcmovznzU64(&x258, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x247) + uint128(x247))) & 0xffffffffffffffff), (uint64((uint128(x249) + uint128(x252))) & 0xffffffffffffffff))\n\tvar x259 uint64\n\tcmovznzU64(&x259, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x248) + uint128(x248))) & 0xffffffffffffffff), (uint64((uint128(x250) + uint128(x253))) & 0xffffffffffffffff))\n\tvar x260 uint64\n\tcmovznzU64(&x260, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x249) + uint128(x252))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x247) + uint128(x247))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x261 uint64\n\tcmovznzU64(&x261, (uint1((x254 >> 63)) & (uint1((((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x246) + uint128(x251))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x250) + uint128(x253))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x248) + uint128(x248))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x262 uint64\n\tcmovznzU64(&x262, (uint1(x257) & 0x1), uint64(0x0), x256)\n\tvar x263 uint64\n\tcmovznzU64(&x263, (uint1(x257) & 0x1), uint64(0x0), x258)\n\tvar x264 uint64\n\tcmovznzU64(&x264, (uint1(x257) & 0x1), uint64(0x0), x259)\n\tx265 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x255))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x266 uint64\n\tcmovznzU64(&x266, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x255))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x255))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x267 uint64\n\tcmovznzU64(&x267, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), x256, (((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1)))\n\tvar x268 uint64\n\tcmovznzU64(&x268, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x256)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x269 uint64\n\tcmovznzU64(&x269, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x258) + uint128(x258))) & 0xffffffffffffffff), (uint64((uint128(x260) + uint128(x263))) & 0xffffffffffffffff))\n\tvar x270 uint64\n\tcmovznzU64(&x270, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x259) + uint128(x259))) & 0xffffffffffffffff), (uint64((uint128(x261) + uint128(x264))) & 0xffffffffffffffff))\n\tvar x271 uint64\n\tcmovznzU64(&x271, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x260) + uint128(x263))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x258) + uint128(x258))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x272 uint64\n\tcmovznzU64(&x272, (uint1((x265 >> 63)) & (uint1((((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x257) + uint128(x262))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x261) + uint128(x264))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x259) + uint128(x259))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x273 uint64\n\tcmovznzU64(&x273, (uint1(x268) & 0x1), uint64(0x0), x267)\n\tvar x274 uint64\n\tcmovznzU64(&x274, (uint1(x268) & 0x1), uint64(0x0), x269)\n\tvar x275 uint64\n\tcmovznzU64(&x275, (uint1(x268) & 0x1), uint64(0x0), x270)\n\tx276 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x266))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x277 uint64\n\tcmovznzU64(&x277, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x266))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x266))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x278 uint64\n\tcmovznzU64(&x278, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), x267, (((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1)))\n\tvar x279 uint64\n\tcmovznzU64(&x279, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x267)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x280 uint64\n\tcmovznzU64(&x280, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x269) + uint128(x269))) & 0xffffffffffffffff), (uint64((uint128(x271) + uint128(x274))) & 0xffffffffffffffff))\n\tvar x281 uint64\n\tcmovznzU64(&x281, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x270) + uint128(x270))) & 0xffffffffffffffff), (uint64((uint128(x272) + uint128(x275))) & 0xffffffffffffffff))\n\tvar x282 uint64\n\tcmovznzU64(&x282, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x271) + uint128(x274))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x269) + uint128(x269))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x283 uint64\n\tcmovznzU64(&x283, (uint1((x276 >> 63)) & (uint1((((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x268) + uint128(x273))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x272) + uint128(x275))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x270) + uint128(x270))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x284 uint64\n\tcmovznzU64(&x284, (uint1(x279) & 0x1), uint64(0x0), x278)\n\tvar x285 uint64\n\tcmovznzU64(&x285, (uint1(x279) & 0x1), uint64(0x0), x280)\n\tvar x286 uint64\n\tcmovznzU64(&x286, (uint1(x279) & 0x1), uint64(0x0), x281)\n\tx287 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x277))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x288 uint64\n\tcmovznzU64(&x288, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x277))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x277))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x289 uint64\n\tcmovznzU64(&x289, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), x278, (((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1)))\n\tvar x290 uint64\n\tcmovznzU64(&x290, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x278)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x291 uint64\n\tcmovznzU64(&x291, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x280) + uint128(x280))) & 0xffffffffffffffff), (uint64((uint128(x282) + uint128(x285))) & 0xffffffffffffffff))\n\tvar x292 uint64\n\tcmovznzU64(&x292, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x281) + uint128(x281))) & 0xffffffffffffffff), (uint64((uint128(x283) + uint128(x286))) & 0xffffffffffffffff))\n\tvar x293 uint64\n\tcmovznzU64(&x293, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x282) + uint128(x285))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x280) + uint128(x280))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x294 uint64\n\tcmovznzU64(&x294, (uint1((x287 >> 63)) & (uint1((((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x279) + uint128(x284))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x283) + uint128(x286))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x281) + uint128(x281))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x295 uint64\n\tcmovznzU64(&x295, (uint1(x290) & 0x1), uint64(0x0), x289)\n\tvar x296 uint64\n\tcmovznzU64(&x296, (uint1(x290) & 0x1), uint64(0x0), x291)\n\tvar x297 uint64\n\tcmovznzU64(&x297, (uint1(x290) & 0x1), uint64(0x0), x292)\n\tx298 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x288))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x299 uint64\n\tcmovznzU64(&x299, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x288))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x288))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x300 uint64\n\tcmovznzU64(&x300, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), x289, (((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1)))\n\tvar x301 uint64\n\tcmovznzU64(&x301, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x289)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x302 uint64\n\tcmovznzU64(&x302, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x291) + uint128(x291))) & 0xffffffffffffffff), (uint64((uint128(x293) + uint128(x296))) & 0xffffffffffffffff))\n\tvar x303 uint64\n\tcmovznzU64(&x303, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x292) + uint128(x292))) & 0xffffffffffffffff), (uint64((uint128(x294) + uint128(x297))) & 0xffffffffffffffff))\n\tvar x304 uint64\n\tcmovznzU64(&x304, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x293) + uint128(x296))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x291) + uint128(x291))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x305 uint64\n\tcmovznzU64(&x305, (uint1((x298 >> 63)) & (uint1((((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x290) + uint128(x295))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x294) + uint128(x297))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x292) + uint128(x292))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x306 uint64\n\tcmovznzU64(&x306, (uint1(x301) & 0x1), uint64(0x0), x300)\n\tvar x307 uint64\n\tcmovznzU64(&x307, (uint1(x301) & 0x1), uint64(0x0), x302)\n\tvar x308 uint64\n\tcmovznzU64(&x308, (uint1(x301) & 0x1), uint64(0x0), x303)\n\tx309 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x299))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x310 uint64\n\tcmovznzU64(&x310, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x299))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x299))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x311 uint64\n\tcmovznzU64(&x311, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), x300, (((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1)))\n\tvar x312 uint64\n\tcmovznzU64(&x312, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x300)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x313 uint64\n\tcmovznzU64(&x313, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x302) + uint128(x302))) & 0xffffffffffffffff), (uint64((uint128(x304) + uint128(x307))) & 0xffffffffffffffff))\n\tvar x314 uint64\n\tcmovznzU64(&x314, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x303) + uint128(x303))) & 0xffffffffffffffff), (uint64((uint128(x305) + uint128(x308))) & 0xffffffffffffffff))\n\tvar x315 uint64\n\tcmovznzU64(&x315, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x304) + uint128(x307))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x302) + uint128(x302))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x316 uint64\n\tcmovznzU64(&x316, (uint1((x309 >> 63)) & (uint1((((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x301) + uint128(x306))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x305) + uint128(x308))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x303) + uint128(x303))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x317 uint64\n\tcmovznzU64(&x317, (uint1(x312) & 0x1), uint64(0x0), x311)\n\tvar x318 uint64\n\tcmovznzU64(&x318, (uint1(x312) & 0x1), uint64(0x0), x313)\n\tvar x319 uint64\n\tcmovznzU64(&x319, (uint1(x312) & 0x1), uint64(0x0), x314)\n\tx320 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x310))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x321 uint64\n\tcmovznzU64(&x321, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x310))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x310))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x322 uint64\n\tcmovznzU64(&x322, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), x311, (((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1)))\n\tvar x323 uint64\n\tcmovznzU64(&x323, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x311)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x324 uint64\n\tcmovznzU64(&x324, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x313) + uint128(x313))) & 0xffffffffffffffff), (uint64((uint128(x315) + uint128(x318))) & 0xffffffffffffffff))\n\tvar x325 uint64\n\tcmovznzU64(&x325, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x314) + uint128(x314))) & 0xffffffffffffffff), (uint64((uint128(x316) + uint128(x319))) & 0xffffffffffffffff))\n\tvar x326 uint64\n\tcmovznzU64(&x326, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x315) + uint128(x318))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x313) + uint128(x313))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x327 uint64\n\tcmovznzU64(&x327, (uint1((x320 >> 63)) & (uint1((((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x312) + uint128(x317))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x316) + uint128(x319))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x314) + uint128(x314))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x328 uint64\n\tcmovznzU64(&x328, (uint1(x323) & 0x1), uint64(0x0), x322)\n\tvar x329 uint64\n\tcmovznzU64(&x329, (uint1(x323) & 0x1), uint64(0x0), x324)\n\tvar x330 uint64\n\tcmovznzU64(&x330, (uint1(x323) & 0x1), uint64(0x0), x325)\n\tx331 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x321))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x332 uint64\n\tcmovznzU64(&x332, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x321))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x321))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x333 uint64\n\tcmovznzU64(&x333, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), x322, (((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1)))\n\tvar x334 uint64\n\tcmovznzU64(&x334, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x322)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x335 uint64\n\tcmovznzU64(&x335, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x324) + uint128(x324))) & 0xffffffffffffffff), (uint64((uint128(x326) + uint128(x329))) & 0xffffffffffffffff))\n\tvar x336 uint64\n\tcmovznzU64(&x336, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x325) + uint128(x325))) & 0xffffffffffffffff), (uint64((uint128(x327) + uint128(x330))) & 0xffffffffffffffff))\n\tvar x337 uint64\n\tcmovznzU64(&x337, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x326) + uint128(x329))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x324) + uint128(x324))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x338 uint64\n\tcmovznzU64(&x338, (uint1((x331 >> 63)) & (uint1((((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x323) + uint128(x328))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x327) + uint128(x330))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x325) + uint128(x325))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x339 uint64\n\tcmovznzU64(&x339, (uint1(x334) & 0x1), uint64(0x0), x333)\n\tvar x340 uint64\n\tcmovznzU64(&x340, (uint1(x334) & 0x1), uint64(0x0), x335)\n\tvar x341 uint64\n\tcmovznzU64(&x341, (uint1(x334) & 0x1), uint64(0x0), x336)\n\tx342 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x332))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x343 uint64\n\tcmovznzU64(&x343, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x332))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x332))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x344 uint64\n\tcmovznzU64(&x344, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), x333, (((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1)))\n\tvar x345 uint64\n\tcmovznzU64(&x345, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x333)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x346 uint64\n\tcmovznzU64(&x346, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x335) + uint128(x335))) & 0xffffffffffffffff), (uint64((uint128(x337) + uint128(x340))) & 0xffffffffffffffff))\n\tvar x347 uint64\n\tcmovznzU64(&x347, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x336) + uint128(x336))) & 0xffffffffffffffff), (uint64((uint128(x338) + uint128(x341))) & 0xffffffffffffffff))\n\tvar x348 uint64\n\tcmovznzU64(&x348, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x337) + uint128(x340))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x335) + uint128(x335))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x349 uint64\n\tcmovznzU64(&x349, (uint1((x342 >> 63)) & (uint1((((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x334) + uint128(x339))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x338) + uint128(x341))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x336) + uint128(x336))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x350 uint64\n\tcmovznzU64(&x350, (uint1(x345) & 0x1), uint64(0x0), x344)\n\tvar x351 uint64\n\tcmovznzU64(&x351, (uint1(x345) & 0x1), uint64(0x0), x346)\n\tvar x352 uint64\n\tcmovznzU64(&x352, (uint1(x345) & 0x1), uint64(0x0), x347)\n\tx353 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x343))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x354 uint64\n\tcmovznzU64(&x354, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x343))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x343))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x355 uint64\n\tcmovznzU64(&x355, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), x344, (((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1)))\n\tvar x356 uint64\n\tcmovznzU64(&x356, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x344)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x357 uint64\n\tcmovznzU64(&x357, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x346) + uint128(x346))) & 0xffffffffffffffff), (uint64((uint128(x348) + uint128(x351))) & 0xffffffffffffffff))\n\tvar x358 uint64\n\tcmovznzU64(&x358, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x347) + uint128(x347))) & 0xffffffffffffffff), (uint64((uint128(x349) + uint128(x352))) & 0xffffffffffffffff))\n\tvar x359 uint64\n\tcmovznzU64(&x359, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x348) + uint128(x351))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x346) + uint128(x346))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x360 uint64\n\tcmovznzU64(&x360, (uint1((x353 >> 63)) & (uint1((((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x345) + uint128(x350))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x349) + uint128(x352))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x347) + uint128(x347))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x361 uint64\n\tcmovznzU64(&x361, (uint1(x356) & 0x1), uint64(0x0), x355)\n\tvar x362 uint64\n\tcmovznzU64(&x362, (uint1(x356) & 0x1), uint64(0x0), x357)\n\tvar x363 uint64\n\tcmovznzU64(&x363, (uint1(x356) & 0x1), uint64(0x0), x358)\n\tx364 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x354))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x365 uint64\n\tcmovznzU64(&x365, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x354))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x354))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x366 uint64\n\tcmovznzU64(&x366, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), x355, (((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1)))\n\tvar x367 uint64\n\tcmovznzU64(&x367, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x355)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x368 uint64\n\tcmovznzU64(&x368, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x357) + uint128(x357))) & 0xffffffffffffffff), (uint64((uint128(x359) + uint128(x362))) & 0xffffffffffffffff))\n\tvar x369 uint64\n\tcmovznzU64(&x369, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x358) + uint128(x358))) & 0xffffffffffffffff), (uint64((uint128(x360) + uint128(x363))) & 0xffffffffffffffff))\n\tvar x370 uint64\n\tcmovznzU64(&x370, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x359) + uint128(x362))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x357) + uint128(x357))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x371 uint64\n\tcmovznzU64(&x371, (uint1((x364 >> 63)) & (uint1((((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x356) + uint128(x361))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x360) + uint128(x363))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x358) + uint128(x358))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x372 uint64\n\tcmovznzU64(&x372, (uint1(x367) & 0x1), uint64(0x0), x366)\n\tvar x373 uint64\n\tcmovznzU64(&x373, (uint1(x367) & 0x1), uint64(0x0), x368)\n\tvar x374 uint64\n\tcmovznzU64(&x374, (uint1(x367) & 0x1), uint64(0x0), x369)\n\tx375 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x365))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x376 uint64\n\tcmovznzU64(&x376, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x365))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x365))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x377 uint64\n\tcmovznzU64(&x377, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), x366, (((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1)))\n\tvar x378 uint64\n\tcmovznzU64(&x378, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x366)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x379 uint64\n\tcmovznzU64(&x379, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x368) + uint128(x368))) & 0xffffffffffffffff), (uint64((uint128(x370) + uint128(x373))) & 0xffffffffffffffff))\n\tvar x380 uint64\n\tcmovznzU64(&x380, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x369) + uint128(x369))) & 0xffffffffffffffff), (uint64((uint128(x371) + uint128(x374))) & 0xffffffffffffffff))\n\tvar x381 uint64\n\tcmovznzU64(&x381, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x370) + uint128(x373))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x368) + uint128(x368))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x382 uint64\n\tcmovznzU64(&x382, (uint1((x375 >> 63)) & (uint1((((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x367) + uint128(x372))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x371) + uint128(x374))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x369) + uint128(x369))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x383 uint64\n\tcmovznzU64(&x383, (uint1(x378) & 0x1), uint64(0x0), x377)\n\tvar x384 uint64\n\tcmovznzU64(&x384, (uint1(x378) & 0x1), uint64(0x0), x379)\n\tvar x385 uint64\n\tcmovznzU64(&x385, (uint1(x378) & 0x1), uint64(0x0), x380)\n\tx386 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x376))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x387 uint64\n\tcmovznzU64(&x387, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x376))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x376))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x388 uint64\n\tcmovznzU64(&x388, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), x377, (((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1)))\n\tvar x389 uint64\n\tcmovznzU64(&x389, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x377)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x390 uint64\n\tcmovznzU64(&x390, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x379) + uint128(x379))) & 0xffffffffffffffff), (uint64((uint128(x381) + uint128(x384))) & 0xffffffffffffffff))\n\tvar x391 uint64\n\tcmovznzU64(&x391, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x380) + uint128(x380))) & 0xffffffffffffffff), (uint64((uint128(x382) + uint128(x385))) & 0xffffffffffffffff))\n\tvar x392 uint64\n\tcmovznzU64(&x392, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x381) + uint128(x384))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x379) + uint128(x379))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x393 uint64\n\tcmovznzU64(&x393, (uint1((x386 >> 63)) & (uint1((((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x378) + uint128(x383))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x382) + uint128(x385))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x380) + uint128(x380))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x394 uint64\n\tcmovznzU64(&x394, (uint1(x389) & 0x1), uint64(0x0), x388)\n\tvar x395 uint64\n\tcmovznzU64(&x395, (uint1(x389) & 0x1), uint64(0x0), x390)\n\tvar x396 uint64\n\tcmovznzU64(&x396, (uint1(x389) & 0x1), uint64(0x0), x391)\n\tx397 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x387))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x398 uint64\n\tcmovznzU64(&x398, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x387))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x387))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x399 uint64\n\tcmovznzU64(&x399, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), x388, (((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1)))\n\tvar x400 uint64\n\tcmovznzU64(&x400, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x388)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x401 uint64\n\tcmovznzU64(&x401, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x390) + uint128(x390))) & 0xffffffffffffffff), (uint64((uint128(x392) + uint128(x395))) & 0xffffffffffffffff))\n\tvar x402 uint64\n\tcmovznzU64(&x402, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x391) + uint128(x391))) & 0xffffffffffffffff), (uint64((uint128(x393) + uint128(x396))) & 0xffffffffffffffff))\n\tvar x403 uint64\n\tcmovznzU64(&x403, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x392) + uint128(x395))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x390) + uint128(x390))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x404 uint64\n\tcmovznzU64(&x404, (uint1((x397 >> 63)) & (uint1((((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x389) + uint128(x394))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x393) + uint128(x396))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x391) + uint128(x391))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x405 uint64\n\tcmovznzU64(&x405, (uint1(x400) & 0x1), uint64(0x0), x399)\n\tvar x406 uint64\n\tcmovznzU64(&x406, (uint1(x400) & 0x1), uint64(0x0), x401)\n\tvar x407 uint64\n\tcmovznzU64(&x407, (uint1(x400) & 0x1), uint64(0x0), x402)\n\tx408 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x398))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x409 uint64\n\tcmovznzU64(&x409, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x398))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x398))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x410 uint64\n\tcmovznzU64(&x410, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), x399, (((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1)))\n\tvar x411 uint64\n\tcmovznzU64(&x411, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x399)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x412 uint64\n\tcmovznzU64(&x412, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x401) + uint128(x401))) & 0xffffffffffffffff), (uint64((uint128(x403) + uint128(x406))) & 0xffffffffffffffff))\n\tvar x413 uint64\n\tcmovznzU64(&x413, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x402) + uint128(x402))) & 0xffffffffffffffff), (uint64((uint128(x404) + uint128(x407))) & 0xffffffffffffffff))\n\tvar x414 uint64\n\tcmovznzU64(&x414, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x403) + uint128(x406))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x401) + uint128(x401))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x415 uint64\n\tcmovznzU64(&x415, (uint1((x408 >> 63)) & (uint1((((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x400) + uint128(x405))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x404) + uint128(x407))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x402) + uint128(x402))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x416 uint64\n\tcmovznzU64(&x416, (uint1(x411) & 0x1), uint64(0x0), x410)\n\tvar x417 uint64\n\tcmovznzU64(&x417, (uint1(x411) & 0x1), uint64(0x0), x412)\n\tvar x418 uint64\n\tcmovznzU64(&x418, (uint1(x411) & 0x1), uint64(0x0), x413)\n\tx419 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x409))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x420 uint64\n\tcmovznzU64(&x420, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x409))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x409))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x421 uint64\n\tcmovznzU64(&x421, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), x410, (((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1)))\n\tvar x422 uint64\n\tcmovznzU64(&x422, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x410)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x423 uint64\n\tcmovznzU64(&x423, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x412) + uint128(x412))) & 0xffffffffffffffff), (uint64((uint128(x414) + uint128(x417))) & 0xffffffffffffffff))\n\tvar x424 uint64\n\tcmovznzU64(&x424, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x413) + uint128(x413))) & 0xffffffffffffffff), (uint64((uint128(x415) + uint128(x418))) & 0xffffffffffffffff))\n\tvar x425 uint64\n\tcmovznzU64(&x425, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x414) + uint128(x417))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x412) + uint128(x412))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x426 uint64\n\tcmovznzU64(&x426, (uint1((x419 >> 63)) & (uint1((((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x411) + uint128(x416))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x415) + uint128(x418))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x413) + uint128(x413))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x427 uint64\n\tcmovznzU64(&x427, (uint1(x422) & 0x1), uint64(0x0), x421)\n\tvar x428 uint64\n\tcmovznzU64(&x428, (uint1(x422) & 0x1), uint64(0x0), x423)\n\tvar x429 uint64\n\tcmovznzU64(&x429, (uint1(x422) & 0x1), uint64(0x0), x424)\n\tx430 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x420))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x431 uint64\n\tcmovznzU64(&x431, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x420))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x420))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x432 uint64\n\tcmovznzU64(&x432, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), x421, (((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1)))\n\tvar x433 uint64\n\tcmovznzU64(&x433, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x421)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x434 uint64\n\tcmovznzU64(&x434, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x423) + uint128(x423))) & 0xffffffffffffffff), (uint64((uint128(x425) + uint128(x428))) & 0xffffffffffffffff))\n\tvar x435 uint64\n\tcmovznzU64(&x435, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x424) + uint128(x424))) & 0xffffffffffffffff), (uint64((uint128(x426) + uint128(x429))) & 0xffffffffffffffff))\n\tvar x436 uint64\n\tcmovznzU64(&x436, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x425) + uint128(x428))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x423) + uint128(x423))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x437 uint64\n\tcmovznzU64(&x437, (uint1((x430 >> 63)) & (uint1((((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x422) + uint128(x427))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x426) + uint128(x429))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x424) + uint128(x424))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x438 uint64\n\tcmovznzU64(&x438, (uint1(x433) & 0x1), uint64(0x0), x432)\n\tvar x439 uint64\n\tcmovznzU64(&x439, (uint1(x433) & 0x1), uint64(0x0), x434)\n\tvar x440 uint64\n\tcmovznzU64(&x440, (uint1(x433) & 0x1), uint64(0x0), x435)\n\tx441 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x431))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x442 uint64\n\tcmovznzU64(&x442, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x431))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x431))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x443 uint64\n\tcmovznzU64(&x443, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), x432, (((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1)))\n\tvar x444 uint64\n\tcmovznzU64(&x444, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x432)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x445 uint64\n\tcmovznzU64(&x445, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x434) + uint128(x434))) & 0xffffffffffffffff), (uint64((uint128(x436) + uint128(x439))) & 0xffffffffffffffff))\n\tvar x446 uint64\n\tcmovznzU64(&x446, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x435) + uint128(x435))) & 0xffffffffffffffff), (uint64((uint128(x437) + uint128(x440))) & 0xffffffffffffffff))\n\tvar x447 uint64\n\tcmovznzU64(&x447, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x436) + uint128(x439))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x434) + uint128(x434))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x448 uint64\n\tcmovznzU64(&x448, (uint1((x441 >> 63)) & (uint1((((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x433) + uint128(x438))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x437) + uint128(x440))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x435) + uint128(x435))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x449 uint64\n\tcmovznzU64(&x449, (uint1(x444) & 0x1), uint64(0x0), x443)\n\tvar x450 uint64\n\tcmovznzU64(&x450, (uint1(x444) & 0x1), uint64(0x0), x445)\n\tvar x451 uint64\n\tcmovznzU64(&x451, (uint1(x444) & 0x1), uint64(0x0), x446)\n\tx452 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x442))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x453 uint64\n\tcmovznzU64(&x453, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x442))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x442))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x454 uint64\n\tcmovznzU64(&x454, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), x443, (((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1)))\n\tvar x455 uint64\n\tcmovznzU64(&x455, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x443)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x456 uint64\n\tcmovznzU64(&x456, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x445) + uint128(x445))) & 0xffffffffffffffff), (uint64((uint128(x447) + uint128(x450))) & 0xffffffffffffffff))\n\tvar x457 uint64\n\tcmovznzU64(&x457, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x446) + uint128(x446))) & 0xffffffffffffffff), (uint64((uint128(x448) + uint128(x451))) & 0xffffffffffffffff))\n\tvar x458 uint64\n\tcmovznzU64(&x458, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x447) + uint128(x450))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x445) + uint128(x445))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x459 uint64\n\tcmovznzU64(&x459, (uint1((x452 >> 63)) & (uint1((((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x444) + uint128(x449))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x448) + uint128(x451))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x446) + uint128(x446))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x460 uint64\n\tcmovznzU64(&x460, (uint1(x455) & 0x1), uint64(0x0), x454)\n\tvar x461 uint64\n\tcmovznzU64(&x461, (uint1(x455) & 0x1), uint64(0x0), x456)\n\tvar x462 uint64\n\tcmovznzU64(&x462, (uint1(x455) & 0x1), uint64(0x0), x457)\n\tx463 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x453))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x464 uint64\n\tcmovznzU64(&x464, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x453))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x453))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x465 uint64\n\tcmovznzU64(&x465, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), x454, (((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1)))\n\tvar x466 uint64\n\tcmovznzU64(&x466, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x454)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x467 uint64\n\tcmovznzU64(&x467, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x456) + uint128(x456))) & 0xffffffffffffffff), (uint64((uint128(x458) + uint128(x461))) & 0xffffffffffffffff))\n\tvar x468 uint64\n\tcmovznzU64(&x468, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x457) + uint128(x457))) & 0xffffffffffffffff), (uint64((uint128(x459) + uint128(x462))) & 0xffffffffffffffff))\n\tvar x469 uint64\n\tcmovznzU64(&x469, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x458) + uint128(x461))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x456) + uint128(x456))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x470 uint64\n\tcmovznzU64(&x470, (uint1((x463 >> 63)) & (uint1((((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x455) + uint128(x460))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x459) + uint128(x462))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x457) + uint128(x457))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x471 uint64\n\tcmovznzU64(&x471, (uint1(x466) & 0x1), uint64(0x0), x465)\n\tvar x472 uint64\n\tcmovznzU64(&x472, (uint1(x466) & 0x1), uint64(0x0), x467)\n\tvar x473 uint64\n\tcmovznzU64(&x473, (uint1(x466) & 0x1), uint64(0x0), x468)\n\tx474 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x464))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x475 uint64\n\tcmovznzU64(&x475, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x464))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x464))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x476 uint64\n\tcmovznzU64(&x476, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), x465, (((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1)))\n\tvar x477 uint64\n\tcmovznzU64(&x477, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x465)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x478 uint64\n\tcmovznzU64(&x478, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x467) + uint128(x467))) & 0xffffffffffffffff), (uint64((uint128(x469) + uint128(x472))) & 0xffffffffffffffff))\n\tvar x479 uint64\n\tcmovznzU64(&x479, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x468) + uint128(x468))) & 0xffffffffffffffff), (uint64((uint128(x470) + uint128(x473))) & 0xffffffffffffffff))\n\tvar x480 uint64\n\tcmovznzU64(&x480, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x469) + uint128(x472))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x467) + uint128(x467))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x481 uint64\n\tcmovznzU64(&x481, (uint1((x474 >> 63)) & (uint1((((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x466) + uint128(x471))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x470) + uint128(x473))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x468) + uint128(x468))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x482 uint64\n\tcmovznzU64(&x482, (uint1(x477) & 0x1), uint64(0x0), x476)\n\tvar x483 uint64\n\tcmovznzU64(&x483, (uint1(x477) & 0x1), uint64(0x0), x478)\n\tvar x484 uint64\n\tcmovznzU64(&x484, (uint1(x477) & 0x1), uint64(0x0), x479)\n\tx485 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x475))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x486 uint64\n\tcmovznzU64(&x486, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x475))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x475))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x487 uint64\n\tcmovznzU64(&x487, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), x476, (((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1)))\n\tvar x488 uint64\n\tcmovznzU64(&x488, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x476)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x489 uint64\n\tcmovznzU64(&x489, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x478) + uint128(x478))) & 0xffffffffffffffff), (uint64((uint128(x480) + uint128(x483))) & 0xffffffffffffffff))\n\tvar x490 uint64\n\tcmovznzU64(&x490, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x479) + uint128(x479))) & 0xffffffffffffffff), (uint64((uint128(x481) + uint128(x484))) & 0xffffffffffffffff))\n\tvar x491 uint64\n\tcmovznzU64(&x491, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x480) + uint128(x483))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x478) + uint128(x478))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x492 uint64\n\tcmovznzU64(&x492, (uint1((x485 >> 63)) & (uint1((((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x477) + uint128(x482))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x481) + uint128(x484))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x479) + uint128(x479))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x493 uint64\n\tcmovznzU64(&x493, (uint1(x488) & 0x1), uint64(0x0), x487)\n\tvar x494 uint64\n\tcmovznzU64(&x494, (uint1(x488) & 0x1), uint64(0x0), x489)\n\tvar x495 uint64\n\tcmovznzU64(&x495, (uint1(x488) & 0x1), uint64(0x0), x490)\n\tx496 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x486))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x497 uint64\n\tcmovznzU64(&x497, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x486))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x486))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x498 uint64\n\tcmovznzU64(&x498, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), x487, (((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1)))\n\tvar x499 uint64\n\tcmovznzU64(&x499, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x487)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x500 uint64\n\tcmovznzU64(&x500, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x489) + uint128(x489))) & 0xffffffffffffffff), (uint64((uint128(x491) + uint128(x494))) & 0xffffffffffffffff))\n\tvar x501 uint64\n\tcmovznzU64(&x501, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x490) + uint128(x490))) & 0xffffffffffffffff), (uint64((uint128(x492) + uint128(x495))) & 0xffffffffffffffff))\n\tvar x502 uint64\n\tcmovznzU64(&x502, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x491) + uint128(x494))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x489) + uint128(x489))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x503 uint64\n\tcmovznzU64(&x503, (uint1((x496 >> 63)) & (uint1((((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x488) + uint128(x493))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x492) + uint128(x495))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x490) + uint128(x490))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x504 uint64\n\tcmovznzU64(&x504, (uint1(x499) & 0x1), uint64(0x0), x498)\n\tvar x505 uint64\n\tcmovznzU64(&x505, (uint1(x499) & 0x1), uint64(0x0), x500)\n\tvar x506 uint64\n\tcmovznzU64(&x506, (uint1(x499) & 0x1), uint64(0x0), x501)\n\tx507 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x497))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x508 uint64\n\tcmovznzU64(&x508, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x497))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x497))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x509 uint64\n\tcmovznzU64(&x509, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), x498, (((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1)))\n\tvar x510 uint64\n\tcmovznzU64(&x510, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x498)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x511 uint64\n\tcmovznzU64(&x511, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x500) + uint128(x500))) & 0xffffffffffffffff), (uint64((uint128(x502) + uint128(x505))) & 0xffffffffffffffff))\n\tvar x512 uint64\n\tcmovznzU64(&x512, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x501) + uint128(x501))) & 0xffffffffffffffff), (uint64((uint128(x503) + uint128(x506))) & 0xffffffffffffffff))\n\tvar x513 uint64\n\tcmovznzU64(&x513, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x502) + uint128(x505))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x500) + uint128(x500))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x514 uint64\n\tcmovznzU64(&x514, (uint1((x507 >> 63)) & (uint1((((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x499) + uint128(x504))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x503) + uint128(x506))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x501) + uint128(x501))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x515 uint64\n\tcmovznzU64(&x515, (uint1(x510) & 0x1), uint64(0x0), x509)\n\tvar x516 uint64\n\tcmovznzU64(&x516, (uint1(x510) & 0x1), uint64(0x0), x511)\n\tvar x517 uint64\n\tcmovznzU64(&x517, (uint1(x510) & 0x1), uint64(0x0), x512)\n\tx518 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x508))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x519 uint64\n\tcmovznzU64(&x519, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x508))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x508))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x520 uint64\n\tcmovznzU64(&x520, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), x509, (((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1)))\n\tvar x521 uint64\n\tcmovznzU64(&x521, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x509)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x522 uint64\n\tcmovznzU64(&x522, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x511) + uint128(x511))) & 0xffffffffffffffff), (uint64((uint128(x513) + uint128(x516))) & 0xffffffffffffffff))\n\tvar x523 uint64\n\tcmovznzU64(&x523, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x512) + uint128(x512))) & 0xffffffffffffffff), (uint64((uint128(x514) + uint128(x517))) & 0xffffffffffffffff))\n\tvar x524 uint64\n\tcmovznzU64(&x524, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x513) + uint128(x516))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x511) + uint128(x511))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x525 uint64\n\tcmovznzU64(&x525, (uint1((x518 >> 63)) & (uint1((((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x510) + uint128(x515))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x514) + uint128(x517))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x512) + uint128(x512))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x526 uint64\n\tcmovznzU64(&x526, (uint1(x521) & 0x1), uint64(0x0), x520)\n\tvar x527 uint64\n\tcmovznzU64(&x527, (uint1(x521) & 0x1), uint64(0x0), x522)\n\tvar x528 uint64\n\tcmovznzU64(&x528, (uint1(x521) & 0x1), uint64(0x0), x523)\n\tx529 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x519))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x530 uint64\n\tcmovznzU64(&x530, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x519))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x519))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x531 uint64\n\tcmovznzU64(&x531, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), x520, (((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1)))\n\tvar x532 uint64\n\tcmovznzU64(&x532, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x520)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x533 uint64\n\tcmovznzU64(&x533, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x522) + uint128(x522))) & 0xffffffffffffffff), (uint64((uint128(x524) + uint128(x527))) & 0xffffffffffffffff))\n\tvar x534 uint64\n\tcmovznzU64(&x534, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x523) + uint128(x523))) & 0xffffffffffffffff), (uint64((uint128(x525) + uint128(x528))) & 0xffffffffffffffff))\n\tvar x535 uint64\n\tcmovznzU64(&x535, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x524) + uint128(x527))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x522) + uint128(x522))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x536 uint64\n\tcmovznzU64(&x536, (uint1((x529 >> 63)) & (uint1((((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x521) + uint128(x526))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x525) + uint128(x528))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x523) + uint128(x523))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x537 uint64\n\tcmovznzU64(&x537, (uint1(x532) & 0x1), uint64(0x0), x531)\n\tvar x538 uint64\n\tcmovznzU64(&x538, (uint1(x532) & 0x1), uint64(0x0), x533)\n\tvar x539 uint64\n\tcmovznzU64(&x539, (uint1(x532) & 0x1), uint64(0x0), x534)\n\tx540 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x530))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x541 uint64\n\tcmovznzU64(&x541, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x530))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x530))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x542 uint64\n\tcmovznzU64(&x542, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), x531, (((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1)))\n\tvar x543 uint64\n\tcmovznzU64(&x543, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x531)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x544 uint64\n\tcmovznzU64(&x544, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x533) + uint128(x533))) & 0xffffffffffffffff), (uint64((uint128(x535) + uint128(x538))) & 0xffffffffffffffff))\n\tvar x545 uint64\n\tcmovznzU64(&x545, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x534) + uint128(x534))) & 0xffffffffffffffff), (uint64((uint128(x536) + uint128(x539))) & 0xffffffffffffffff))\n\tvar x546 uint64\n\tcmovznzU64(&x546, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x535) + uint128(x538))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x533) + uint128(x533))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x547 uint64\n\tcmovznzU64(&x547, (uint1((x540 >> 63)) & (uint1((((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x532) + uint128(x537))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x536) + uint128(x539))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x534) + uint128(x534))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x548 uint64\n\tcmovznzU64(&x548, (uint1(x543) & 0x1), uint64(0x0), x542)\n\tvar x549 uint64\n\tcmovznzU64(&x549, (uint1(x543) & 0x1), uint64(0x0), x544)\n\tvar x550 uint64\n\tcmovznzU64(&x550, (uint1(x543) & 0x1), uint64(0x0), x545)\n\tx551 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x541))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x552 uint64\n\tcmovznzU64(&x552, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x541))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x541))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x553 uint64\n\tcmovznzU64(&x553, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), x542, (((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1)))\n\tvar x554 uint64\n\tcmovznzU64(&x554, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x542)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x555 uint64\n\tcmovznzU64(&x555, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x544) + uint128(x544))) & 0xffffffffffffffff), (uint64((uint128(x546) + uint128(x549))) & 0xffffffffffffffff))\n\tvar x556 uint64\n\tcmovznzU64(&x556, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x545) + uint128(x545))) & 0xffffffffffffffff), (uint64((uint128(x547) + uint128(x550))) & 0xffffffffffffffff))\n\tvar x557 uint64\n\tcmovznzU64(&x557, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x546) + uint128(x549))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x544) + uint128(x544))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x558 uint64\n\tcmovznzU64(&x558, (uint1((x551 >> 63)) & (uint1((((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x543) + uint128(x548))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x547) + uint128(x550))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x545) + uint128(x545))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x559 uint64\n\tcmovznzU64(&x559, (uint1(x554) & 0x1), uint64(0x0), x553)\n\tvar x560 uint64\n\tcmovznzU64(&x560, (uint1(x554) & 0x1), uint64(0x0), x555)\n\tvar x561 uint64\n\tcmovznzU64(&x561, (uint1(x554) & 0x1), uint64(0x0), x556)\n\tx562 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x552))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x563 uint64\n\tcmovznzU64(&x563, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x552))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x552))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x564 uint64\n\tcmovznzU64(&x564, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), x553, (((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1)))\n\tvar x565 uint64\n\tcmovznzU64(&x565, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x553)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x566 uint64\n\tcmovznzU64(&x566, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x555) + uint128(x555))) & 0xffffffffffffffff), (uint64((uint128(x557) + uint128(x560))) & 0xffffffffffffffff))\n\tvar x567 uint64\n\tcmovznzU64(&x567, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x556) + uint128(x556))) & 0xffffffffffffffff), (uint64((uint128(x558) + uint128(x561))) & 0xffffffffffffffff))\n\tvar x568 uint64\n\tcmovznzU64(&x568, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x557) + uint128(x560))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x555) + uint128(x555))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x569 uint64\n\tcmovznzU64(&x569, (uint1((x562 >> 63)) & (uint1((((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x554) + uint128(x559))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x558) + uint128(x561))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x556) + uint128(x556))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x570 uint64\n\tcmovznzU64(&x570, (uint1(x565) & 0x1), uint64(0x0), x564)\n\tvar x571 uint64\n\tcmovznzU64(&x571, (uint1(x565) & 0x1), uint64(0x0), x566)\n\tvar x572 uint64\n\tcmovznzU64(&x572, (uint1(x565) & 0x1), uint64(0x0), x567)\n\tx573 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x563))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x574 uint64\n\tcmovznzU64(&x574, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x563))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x563))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x575 uint64\n\tcmovznzU64(&x575, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), x564, (((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1)))\n\tvar x576 uint64\n\tcmovznzU64(&x576, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x564)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x577 uint64\n\tcmovznzU64(&x577, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x566) + uint128(x566))) & 0xffffffffffffffff), (uint64((uint128(x568) + uint128(x571))) & 0xffffffffffffffff))\n\tvar x578 uint64\n\tcmovznzU64(&x578, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x567) + uint128(x567))) & 0xffffffffffffffff), (uint64((uint128(x569) + uint128(x572))) & 0xffffffffffffffff))\n\tvar x579 uint64\n\tcmovznzU64(&x579, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x568) + uint128(x571))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x566) + uint128(x566))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x580 uint64\n\tcmovznzU64(&x580, (uint1((x573 >> 63)) & (uint1((((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x565) + uint128(x570))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x569) + uint128(x572))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x567) + uint128(x567))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x581 uint64\n\tcmovznzU64(&x581, (uint1(x576) & 0x1), uint64(0x0), x575)\n\tvar x582 uint64\n\tcmovznzU64(&x582, (uint1(x576) & 0x1), uint64(0x0), x577)\n\tvar x583 uint64\n\tcmovznzU64(&x583, (uint1(x576) & 0x1), uint64(0x0), x578)\n\tx584 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x574))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x585 uint64\n\tcmovznzU64(&x585, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x574))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x574))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x586 uint64\n\tcmovznzU64(&x586, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), x575, (((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1)))\n\tvar x587 uint64\n\tcmovznzU64(&x587, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x575)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x588 uint64\n\tcmovznzU64(&x588, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x577) + uint128(x577))) & 0xffffffffffffffff), (uint64((uint128(x579) + uint128(x582))) & 0xffffffffffffffff))\n\tvar x589 uint64\n\tcmovznzU64(&x589, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x578) + uint128(x578))) & 0xffffffffffffffff), (uint64((uint128(x580) + uint128(x583))) & 0xffffffffffffffff))\n\tvar x590 uint64\n\tcmovznzU64(&x590, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x579) + uint128(x582))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x577) + uint128(x577))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x591 uint64\n\tcmovznzU64(&x591, (uint1((x584 >> 63)) & (uint1((((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x576) + uint128(x581))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x580) + uint128(x583))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x578) + uint128(x578))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x592 uint64\n\tcmovznzU64(&x592, (uint1(x587) & 0x1), uint64(0x0), x586)\n\tvar x593 uint64\n\tcmovznzU64(&x593, (uint1(x587) & 0x1), uint64(0x0), x588)\n\tvar x594 uint64\n\tcmovznzU64(&x594, (uint1(x587) & 0x1), uint64(0x0), x589)\n\tx595 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x585))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x596 uint64\n\tcmovznzU64(&x596, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x585))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x585))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x597 uint64\n\tcmovznzU64(&x597, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), x586, (((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1)))\n\tvar x598 uint64\n\tcmovznzU64(&x598, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x586)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x599 uint64\n\tcmovznzU64(&x599, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x588) + uint128(x588))) & 0xffffffffffffffff), (uint64((uint128(x590) + uint128(x593))) & 0xffffffffffffffff))\n\tvar x600 uint64\n\tcmovznzU64(&x600, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x589) + uint128(x589))) & 0xffffffffffffffff), (uint64((uint128(x591) + uint128(x594))) & 0xffffffffffffffff))\n\tvar x601 uint64\n\tcmovznzU64(&x601, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x590) + uint128(x593))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x588) + uint128(x588))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x602 uint64\n\tcmovznzU64(&x602, (uint1((x595 >> 63)) & (uint1((((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x587) + uint128(x592))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x591) + uint128(x594))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x589) + uint128(x589))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x603 uint64\n\tcmovznzU64(&x603, (uint1(x598) & 0x1), uint64(0x0), x597)\n\tvar x604 uint64\n\tcmovznzU64(&x604, (uint1(x598) & 0x1), uint64(0x0), x599)\n\tvar x605 uint64\n\tcmovznzU64(&x605, (uint1(x598) & 0x1), uint64(0x0), x600)\n\tx606 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x596))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x607 uint64\n\tcmovznzU64(&x607, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x596))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x596))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x608 uint64\n\tcmovznzU64(&x608, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), x597, (((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1)))\n\tvar x609 uint64\n\tcmovznzU64(&x609, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x597)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x610 uint64\n\tcmovznzU64(&x610, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x599) + uint128(x599))) & 0xffffffffffffffff), (uint64((uint128(x601) + uint128(x604))) & 0xffffffffffffffff))\n\tvar x611 uint64\n\tcmovznzU64(&x611, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x600) + uint128(x600))) & 0xffffffffffffffff), (uint64((uint128(x602) + uint128(x605))) & 0xffffffffffffffff))\n\tvar x612 uint64\n\tcmovznzU64(&x612, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x601) + uint128(x604))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x599) + uint128(x599))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x613 uint64\n\tcmovznzU64(&x613, (uint1((x606 >> 63)) & (uint1((((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x598) + uint128(x603))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x602) + uint128(x605))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x600) + uint128(x600))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x614 uint64\n\tcmovznzU64(&x614, (uint1(x609) & 0x1), uint64(0x0), x608)\n\tvar x615 uint64\n\tcmovznzU64(&x615, (uint1(x609) & 0x1), uint64(0x0), x610)\n\tvar x616 uint64\n\tcmovznzU64(&x616, (uint1(x609) & 0x1), uint64(0x0), x611)\n\tx617 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x607))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x618 uint64\n\tcmovznzU64(&x618, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x607))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x607))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x619 uint64\n\tcmovznzU64(&x619, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), x608, (((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1)))\n\tvar x620 uint64\n\tcmovznzU64(&x620, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x608)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x621 uint64\n\tcmovznzU64(&x621, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x610) + uint128(x610))) & 0xffffffffffffffff), (uint64((uint128(x612) + uint128(x615))) & 0xffffffffffffffff))\n\tvar x622 uint64\n\tcmovznzU64(&x622, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x611) + uint128(x611))) & 0xffffffffffffffff), (uint64((uint128(x613) + uint128(x616))) & 0xffffffffffffffff))\n\tvar x623 uint64\n\tcmovznzU64(&x623, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x612) + uint128(x615))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x610) + uint128(x610))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x624 uint64\n\tcmovznzU64(&x624, (uint1((x617 >> 63)) & (uint1((((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x609) + uint128(x614))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x613) + uint128(x616))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x611) + uint128(x611))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x625 uint64\n\tcmovznzU64(&x625, (uint1(x620) & 0x1), uint64(0x0), x619)\n\tvar x626 uint64\n\tcmovznzU64(&x626, (uint1(x620) & 0x1), uint64(0x0), x621)\n\tvar x627 uint64\n\tcmovznzU64(&x627, (uint1(x620) & 0x1), uint64(0x0), x622)\n\tx628 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x618))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x629 uint64\n\tcmovznzU64(&x629, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x618))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x618))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x630 uint64\n\tcmovznzU64(&x630, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), x619, (((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1)))\n\tvar x631 uint64\n\tcmovznzU64(&x631, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x619)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x632 uint64\n\tcmovznzU64(&x632, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x621) + uint128(x621))) & 0xffffffffffffffff), (uint64((uint128(x623) + uint128(x626))) & 0xffffffffffffffff))\n\tvar x633 uint64\n\tcmovznzU64(&x633, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x622) + uint128(x622))) & 0xffffffffffffffff), (uint64((uint128(x624) + uint128(x627))) & 0xffffffffffffffff))\n\tvar x634 uint64\n\tcmovznzU64(&x634, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x623) + uint128(x626))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x621) + uint128(x621))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x635 uint64\n\tcmovznzU64(&x635, (uint1((x628 >> 63)) & (uint1((((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x620) + uint128(x625))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x624) + uint128(x627))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x622) + uint128(x622))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x636 uint64\n\tcmovznzU64(&x636, (uint1(x631) & 0x1), uint64(0x0), x630)\n\tvar x637 uint64\n\tcmovznzU64(&x637, (uint1(x631) & 0x1), uint64(0x0), x632)\n\tvar x638 uint64\n\tcmovznzU64(&x638, (uint1(x631) & 0x1), uint64(0x0), x633)\n\tx639 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x629))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x640 uint64\n\tcmovznzU64(&x640, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x629))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x629))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x641 uint64\n\tcmovznzU64(&x641, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), x630, (((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1)))\n\tvar x642 uint64\n\tcmovznzU64(&x642, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x630)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x643 uint64\n\tcmovznzU64(&x643, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x632) + uint128(x632))) & 0xffffffffffffffff), (uint64((uint128(x634) + uint128(x637))) & 0xffffffffffffffff))\n\tvar x644 uint64\n\tcmovznzU64(&x644, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x633) + uint128(x633))) & 0xffffffffffffffff), (uint64((uint128(x635) + uint128(x638))) & 0xffffffffffffffff))\n\tvar x645 uint64\n\tcmovznzU64(&x645, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x634) + uint128(x637))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x632) + uint128(x632))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x646 uint64\n\tcmovznzU64(&x646, (uint1((x639 >> 63)) & (uint1((((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x631) + uint128(x636))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x635) + uint128(x638))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x633) + uint128(x633))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x647 uint64\n\tcmovznzU64(&x647, (uint1(x642) & 0x1), uint64(0x0), x641)\n\tvar x648 uint64\n\tcmovznzU64(&x648, (uint1(x642) & 0x1), uint64(0x0), x643)\n\tvar x649 uint64\n\tcmovznzU64(&x649, (uint1(x642) & 0x1), uint64(0x0), x644)\n\tx650 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x640))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x651 uint64\n\tcmovznzU64(&x651, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x640))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x640))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x652 uint64\n\tcmovznzU64(&x652, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), x641, (((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1)))\n\tvar x653 uint64\n\tcmovznzU64(&x653, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x641)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x654 uint64\n\tcmovznzU64(&x654, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x643) + uint128(x643))) & 0xffffffffffffffff), (uint64((uint128(x645) + uint128(x648))) & 0xffffffffffffffff))\n\tvar x655 uint64\n\tcmovznzU64(&x655, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x644) + uint128(x644))) & 0xffffffffffffffff), (uint64((uint128(x646) + uint128(x649))) & 0xffffffffffffffff))\n\tvar x656 uint64\n\tcmovznzU64(&x656, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x645) + uint128(x648))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x643) + uint128(x643))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x657 uint64\n\tcmovznzU64(&x657, (uint1((x650 >> 63)) & (uint1((((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x642) + uint128(x647))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x646) + uint128(x649))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x644) + uint128(x644))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x658 uint64\n\tcmovznzU64(&x658, (uint1(x653) & 0x1), uint64(0x0), x652)\n\tvar x659 uint64\n\tcmovznzU64(&x659, (uint1(x653) & 0x1), uint64(0x0), x654)\n\tvar x660 uint64\n\tcmovznzU64(&x660, (uint1(x653) & 0x1), uint64(0x0), x655)\n\tx661 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x651))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x662 uint64\n\tcmovznzU64(&x662, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x651))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x651))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x663 uint64\n\tcmovznzU64(&x663, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), x652, (((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1)))\n\tvar x664 uint64\n\tcmovznzU64(&x664, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x652)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x665 uint64\n\tcmovznzU64(&x665, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x654) + uint128(x654))) & 0xffffffffffffffff), (uint64((uint128(x656) + uint128(x659))) & 0xffffffffffffffff))\n\tvar x666 uint64\n\tcmovznzU64(&x666, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x655) + uint128(x655))) & 0xffffffffffffffff), (uint64((uint128(x657) + uint128(x660))) & 0xffffffffffffffff))\n\tvar x667 uint64\n\tcmovznzU64(&x667, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x656) + uint128(x659))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x654) + uint128(x654))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x668 uint64\n\tcmovznzU64(&x668, (uint1((x661 >> 63)) & (uint1((((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x653) + uint128(x658))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x657) + uint128(x660))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x655) + uint128(x655))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x669 uint64\n\tcmovznzU64(&x669, (uint1(x664) & 0x1), uint64(0x0), x663)\n\tvar x670 uint64\n\tcmovznzU64(&x670, (uint1(x664) & 0x1), uint64(0x0), x665)\n\tvar x671 uint64\n\tcmovznzU64(&x671, (uint1(x664) & 0x1), uint64(0x0), x666)\n\tx672 := (uint64((uint128((^(uint64((uint128(0x1) + uint128(x662))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff)\n\tvar x673 uint64\n\tcmovznzU64(&x673, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(0x1) + uint128(x662))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(0x1) + uint128(x662))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x674 uint64\n\tcmovznzU64(&x674, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1)), (uint64((uint128((^x663)) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x675 uint64\n\tcmovznzU64(&x675, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x665) + uint128(x665))) & 0xffffffffffffffff), (uint64((uint128(x667) + uint128(x670))) & 0xffffffffffffffff))\n\tvar x676 uint64\n\tcmovznzU64(&x676, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x666) + uint128(x666))) & 0xffffffffffffffff), (uint64((uint128(x668) + uint128(x671))) & 0xffffffffffffffff))\n\tvar x677 uint64\n\tcmovznzU64(&x677, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x667) + uint128(x670))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x665) + uint128(x665))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x678 uint64\n\tcmovznzU64(&x678, (uint1((x672 >> 63)) & (uint1((((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) & 0x8000000000000000) | ((uint64((uint128(x664) + uint128(x669))) & 0xffffffffffffffff) >> 1))) & 0x1)), (uint64((uint128(x668) + uint128(x671))) & 0xffffffffffffffff), (uint64((uint128((^(uint64((uint128(x666) + uint128(x666))) & 0xffffffffffffffff))) + uint128(0x1))) & 0xffffffffffffffff))\n\tvar x679 uint64\n\tcmovznzU64(&x679, (uint1(x674) & 0x1), uint64(0x0), x675)\n\tvar x680 uint64\n\tcmovznzU64(&x680, (uint1(x674) & 0x1), uint64(0x0), x676)\n\tx681 := (uint64((uint128(0x1) + uint128(x673))) & 0xffffffffffffffff)\n\tx682 := (uint64((uint128(x675) + uint128(x675))) & 0xffffffffffffffff)\n\tx683 := (uint64((uint128(x676) + uint128(x676))) & 0xffffffffffffffff)\n\tx684 := (uint64((uint128(x677) + uint128(x679))) & 0xffffffffffffffff)\n\tx685 := (uint64((uint128(x678) + uint128(x680))) & 0xffffffffffffffff)\n\t*out1 = x681\n\t*out2 = x682\n\t*out3 = x683\n\t*out4 = x684\n\t*out5 = x685\n}", "func (r *AutoRoller) retrieveRoll(issueNum int64) (*autoroll.AutoRollIssue, error) {\n\tvar a *autoroll.AutoRollIssue\n\tinfo, err := r.gerrit.GetIssueProperties(issueNum)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to get issue properties: %s\", err)\n\t}\n\ta, err = autoroll.FromGerritChangeInfo(info, r.rm.FullChildHash, r.rollIntoAndroid)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to convert issue format: %s\", err)\n\t}\n\ttryResults, err := autoroll.GetTryResultsFromGerrit(r.gerrit, a)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to retrieve try results: %s\", err)\n\t}\n\ta.TryResults = tryResults\n\treturn a, nil\n}", "func (d Die) Roll() int {\n\treturn rand.Intn(int(d)) + 1\n}", "func (f *FabricCAClientImpl) ReEnroll(identity *Identity) (*Identity, error) {\n\n\tif identity == nil || identity.EnrollmentId() == \"\" {\n\t\treturn nil, ErrCertificateEmpty\n\t}\n\n\t// create new cert and send it to CA for signing\n\tkey, err := f.Crypto.GenerateKey()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcsr, err := f.Crypto.CreateCertificateRequest(identity.EnrollmentId(), key)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turl := fmt.Sprintf(\"%s/api/v1/reenroll\", f.Url)\n\n\tcrm, err := json.Marshal(CertificateRequest{CR: string(csr)})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err := http.NewRequest(\"POST\", url, bytes.NewBuffer(crm))\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\ttoken, err := f.createAuthToken(identity, crm)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"authorization\", token)\n\tvar tr *http.Transport\n\tif f.Transport == nil {\n\t\ttr = &http.Transport{\n\t\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: f.SkipTLSVerification},\n\t\t}\n\t} else {\n\t\ttr = f.Transport\n\t}\n\n\thttpClient := &http.Client{Transport: tr}\n\n\tresp, err := httpClient.Do(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tenrResp := new(enrollmentResponse)\n\tif err := json.Unmarshal(body, enrResp); err != nil {\n\t\treturn nil, err\n\t}\n\tif !enrResp.Success {\n\t\treturn nil, ErrEnrollment\n\t}\n\trawCert, err := base64.StdEncoding.DecodeString(enrResp.Result.Cert)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta, _ := pem.Decode(rawCert)\n\tcert, err := x509.ParseCertificate(a.Bytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Identity{Certificate: cert, PrivateKey: key}, nil\n}", "func DoFunc(retries int, fn Func, sleeps ...Sleep) error {\n\tif retries <= 0 {\n\t\treturn nil\n\t}\n\n\ttotal := retries\n\nLBBEGIN:\n\n\tif err := fn(); err != nil {\n\t\tretries--\n\t\tif retries == 0 {\n\t\t\treturn ErrMaxRetries\n\t\t}\n\n\t\tif len(sleeps) > 0 {\n\t\t\tif d := sleeps[0](total - retries); d > 0 {\n\t\t\t\ttime.Sleep(d)\n\t\t\t}\n\t\t}\n\n\t\tgoto LBBEGIN\n\t}\n\n\treturn nil\n}", "func (s *sink) handleRetryBatches(retry seqRecBatches) {\n\tvar needsMetaUpdate bool\n\tretry.onEachFirstBatchWhileBatchOwnerLocked(func(batch seqRecBatch) {\n\t\t// If we fail all records here, we likely will have out of\n\t\t// order seq nums; hopefully the client user does not stop\n\t\t// on data loss, since this is not truly data loss.\n\t\tif batch.isTimedOut(s.cl.cfg.recordTimeout) {\n\t\t\tbatch.owner.lockedFailAllRecords(ErrRecordTimeout)\n\t\t} else if batch.tries == s.cl.cfg.retries {\n\t\t\terr := errors.New(\"record failed after being retried too many times\")\n\t\t\tbatch.owner.lockedFailAllRecords(err)\n\t\t}\n\t\tbatch.owner.resetBatchDrainIdx()\n\t\tbatch.owner.failing = true\n\t\tneedsMetaUpdate = true\n\t})\n\tif needsMetaUpdate {\n\t\ts.cl.triggerUpdateMetadata()\n\t}\n}", "func rollingStateful(clientset *kubernetes.Clientset, namespace, name, image string) error {\n\n\tfmt.Printf(\"rolling upgrade : %s \\n\", name)\n\n\tupdateClient := clientset.AppsV1().StatefulSets(namespace)\n\n\tretryErr := retry.RetryOnConflict(retry.DefaultRetry, func() error {\n\t\t// Retrieve the latest version of Deployment before attempting update\n\t\t// RetryOnConflict uses exponential backoff to avoid exhausting the apiserver\n\t\tresult, getErr := updateClient.Get(name, metav1.GetOptions{})\n\t\tif getErr != nil {\n\t\t\tpanic(fmt.Errorf(\"Failed to get latest version of Deployment: %v\", getErr))\n\t\t}\n\n\t\tresult.Spec.Template.Spec.Containers[0].Image = image // change nginx version\n\n\t\t_, updateErr := updateClient.Update(result)\n\n\t\treturn updateErr\n\t})\n\n\tif retryErr != nil {\n\t\treturn retryErr\n\t}\n\n\treturn nil\n}", "func (m *MockCAClient) Enroll(arg0 *api.EnrollmentRequest) error {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Enroll\", arg0)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (c *ReconnectingSFTPClient) do(fun func(SFTPClient) error) error {\n\tclient, err := c.getClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = fun(client)\n\tif err == nil {\n\t\treturn nil\n\t}\n\n\t// We already have an error, so we don't report if handleErr\n\t// returns an error.\n\tc.handleError(client, err)\n\treturn err\n}", "func (r *AutoRoller) Start(tickFrequency, repoFrequency time.Duration, ctx context.Context) {\n\tsklog.Infof(\"Starting autoroller.\")\n\trepo_manager.Start(r.rm, repoFrequency, ctx)\n\tlv := metrics2.NewLiveness(\"last_successful_autoroll_tick\")\n\tgo util.RepeatCtx(tickFrequency, ctx, func() {\n\t\tif err := r.Tick(); err != nil {\n\t\t\tsklog.Errorf(\"Failed to run autoroll: %s\", err)\n\t\t} else {\n\t\t\tlv.Reset()\n\t\t}\n\t})\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ctx.Done():\n\t\t\t\tutil.LogErr(r.recent.Close())\n\t\t\t\tutil.LogErr(r.modeHistory.Close())\n\t\t\tdefault:\n\t\t\t}\n\t\t}\n\t}()\n}", "func Loop(action Action, strategies ...Strategy) error {\n\tfor i := uint(1); ; i++ {\n\t\terr := action()\n\t\tif err == nil {\n\t\t\ti = 0\n\t\t\tcontinue\n\t\t}\n\n\t\tfor _, s := range strategies {\n\t\t\tif shouldRetry := s(i, err); !shouldRetry {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n}", "func Runner(restorepath string) int {\n\tconsulClient := &consul.Consul{Client: *consul.Client()}\n\n\tconf := config.ParseConfig(false)\n\n\tlog.Printf(\"[DEBUG] Starting restore of %s/%s\", conf.S3Bucket, restorepath)\n\tdoWork(conf, consulClient, restorepath)\n\treturn 0\n}", "func createSOWalkInCustomerAutoInvoice(so *model.SalesOrder, si *model.SalesInvoice) (e error) {\n\tvar fr *model.FinanceRevenue\n\t// ubah status sales order invoice menjadi 'finished'\n\tso.InvoiceStatus = \"finished\"\n\t// ubah status sales invoice manjadi 'finished'\n\tsi.DocumentStatus = \"finished\"\n\n\t// simpan sales order dan sales invoice\n\tif e = so.Save(); e == nil {\n\t\tif e = si.Save(); e == nil {\n\t\t\t// buat finance revenue dari sales invoice\n\t\t\tif fr, e = createFinanceRevenue(si); e == nil {\n\t\t\t\t// jumlahkan semua total revenue dari sales invoice\n\t\t\t\tif _, e = sumTotalRevenuedSalesInvoice(si.ID); e == nil {\n\t\t\t\t\t// approve-kan finance revenue yang telah dibuat\n\t\t\t\t\te = approveRevenue(fr)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (b *batch) run() {\n\tb.db.batchMu.Lock()\n\tb.timer.Stop()\n\t// Make sure no new work is added to this batch, but don't break\n\t// other batches.\n\tif b.db.batch == b {\n\t\tb.db.batch = nil\n\t}\n\tb.db.batchMu.Unlock()\n\nretry:\n\tfor len(b.calls) > 0 {\n\t\tvar failIdx = -1\n\t\terr := b.db.Update(func(tx *Tx) error {\n\t\t\tfor i, c := range b.calls {\n\t\t\t\tif err := safelyCall(c.fn, tx); err != nil {\n\t\t\t\t\tfailIdx = i\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\tif failIdx >= 0 {\n\t\t\t// take the failing transaction out of the batch. it's\n\t\t\t// safe to shorten b.calls here because db.batch no longer\n\t\t\t// points to us, and we hold the mutex anyway.\n\t\t\tc := b.calls[failIdx]\n\t\t\tb.calls[failIdx], b.calls = b.calls[len(b.calls)-1], b.calls[:len(b.calls)-1]\n\t\t\t// tell the submitter re-run it solo, continue with the rest of the batch\n\t\t\tc.err <- trySolo\n\t\t\tcontinue retry\n\t\t}\n\n\t\t// pass success, or bolt internal errors, to all callers\n\t\tfor _, c := range b.calls {\n\t\t\tc.err <- err\n\t\t}\n\t\tbreak retry\n\t}\n}", "func (b *batch) run() {\n\tb.db.batchMu.Lock()\n\tb.timer.Stop()\n\t// Make sure no new work is added to this batch, but don't break\n\t// other batches.\n\tif b.db.batch == b {\n\t\tb.db.batch = nil\n\t}\n\tb.db.batchMu.Unlock()\n\nretry:\n\tfor len(b.calls) > 0 {\n\t\tvar failIdx = -1\n\t\terr := b.db.Update(func(tx *Tx) error {\n\t\t\tfor i, c := range b.calls {\n\t\t\t\tif err := safelyCall(c.fn, tx); err != nil {\n\t\t\t\t\tfailIdx = i\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\tif failIdx >= 0 {\n\t\t\t// take the failing transaction out of the batch. it's\n\t\t\t// safe to shorten b.calls here because db.batch no longer\n\t\t\t// points to us, and we hold the mutex anyway.\n\t\t\tc := b.calls[failIdx]\n\t\t\tb.calls[failIdx], b.calls = b.calls[len(b.calls)-1], b.calls[:len(b.calls)-1]\n\t\t\t// tell the submitter re-run it solo, continue with the rest of the batch\n\t\t\tc.err <- trySolo\n\t\t\tcontinue retry\n\t\t}\n\n\t\t// pass success, or bolt internal errors, to all callers\n\t\tfor _, c := range b.calls {\n\t\t\tc.err <- err\n\t\t}\n\t\tbreak retry\n\t}\n}", "func (r *Rollback) Run(name string) (*release.Release, error) {\n\tr.cfg.Log(\"preparing rollback of %s\", name)\n\tcurrentRelease, targetRelease, err := r.prepareRollback(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !r.DryRun {\n\t\tr.cfg.Log(\"creating rolled back release for %s\", name)\n\t\tif err := r.cfg.Releases.Create(targetRelease); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tr.cfg.Log(\"performing rollback of %s\", name)\n\tres, err := r.performRollback(currentRelease, targetRelease)\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\tif !r.DryRun {\n\t\tr.cfg.Log(\"updating status for rolled back release for %s\", name)\n\t\tif err := r.cfg.Releases.Update(targetRelease); err != nil {\n\t\t\treturn res, err\n\t\t}\n\t}\n\n\treturn res, nil\n}", "func (s *store) doRebalance(tasks []ShardRestoreTask) {\n\tfor _, task := range tasks {\n\t\t// apply\n\t\tif len(task.CurrentOwners) == 0 {\n\t\t\ts.logger.Warn(fmt.Sprintf(\"shard %d current owners is zero.\", task.ShardID))\n\t\t\tcontinue\n\t\t}\n\t\ts.logger.Info(fmt.Sprintf(\"restore shard %d from %d to %v \",\n\t\t\ttask.ShardID, task.CurrentOwners[0].NodeID, task.NewOwners))\n\t\tvar newOwnerNodeIDs []uint64\n\t\tfor _, owner := range task.NewOwners{\n\t\t\tnewOwnerNodeIDs = append(newOwnerNodeIDs, owner.NodeID)\n\t\t}\n\n\t\tc := &internal.UpdateShardCommand{\n\t\t\tID: proto.Uint64(task.ShardID),\n\t\t\tNewOwnerNodeIDs: newOwnerNodeIDs,\n\t\t}\n\t\ttyp := internal.Command_UpdateShardCommand\n\t\tcmd := &internal.Command{ Type: &typ}\n\t\tif err := proto.SetExtension(cmd, internal.E_UpdateShardCommand_Command, c); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tb, err := proto.Marshal(cmd)\n\t\tif err != nil {\n\t\t\ts.logger.Error(err.Error())\n\t\t}\n\t\ts.apply(b)\n\n\t\tfor _, newOwner := range task.NewOwners {\n\t\t\ts.restoreShard(task.Database, task.RetentionPolicy,\n\t\t\t\ttask.NewOwnerAddress[newOwner.NodeID],\n\t\t\t\ttask.ShardID, task.CurrentOwners[0].NodeID,\n\t\t\t\tnewOwner.NodeID)\n\t\t}\n\t\ts.logger.Info(\"do rebalance done.\")\n\n\t}\n}", "func (ti *testItem) retry() {\n\t// release it to make it available for somebody else to try later:\n\t<-ti.take\n}", "func roll(sides int) int {\n\tmin, max := 1, sides\n\troll := min + seededRand.Intn(max-min+1)\n\treturn roll\n}", "func (r *EndpointsRolloutV1RestClient) AutoGetRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tpath := makeURIRolloutV1AutoGetRolloutGetOper(in)\n\tif r.bufferId != \"\" {\n\t\tpath = strings.Replace(path, \"/configs\", \"/staging/\"+r.bufferId, 1)\n\t}\n\treq, err := r.getHTTPRequest(ctx, in, \"GET\", path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := r.client.Do(req.WithContext(ctx))\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"request failed (%s)\", err)\n\t}\n\tdefer resp.Body.Close()\n\tret, err := decodeHTTPrespRolloutV1AutoGetRollout(ctx, resp)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn ret.(*Rollout), err\n}", "func (c *Context) ROLL(ci, mr operand.Op) {\n\tc.addinstruction(x86.ROLL(ci, mr))\n}", "func (w *FabricSDKWrapper) ReEnrollUser(enrollmentID string, orgName string) error {\n\tctxProvider := w.sdk.Context()\n\tmspClient, err := msp.New(ctxProvider, msp.WithOrg(orgName))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn mspClient.Reenroll(enrollmentID)\n}", "func (c *Client) Enroll(req *api.EnrollmentRequest, saveOnDisk bool) (*api.EnrollmentResponse, error) {\n\tlog.Debugf(\"Enrolling %+v\", req)\n\n\terr := c.Init()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.handleX509Enroll(req, saveOnDisk)\n}", "func RunConsumerUpon(conn *Connection, qName string, prefetchCount, prefetchSize, maxRetries, workers int,\n\thandler MessageHandler) {\n\tfor id := 0; id < workers; id++ {\n\t\tc := newRetryableConsumer(id, qName, prefetchCount, prefetchSize, handler, maxRetries)\n\t\tc.run(conn)\n\t}\n\tconn.start()\n}", "func (sm *storageMgr) handleRollback(cmd Message) {\n\n\tsm.supvCmdch <- &MsgSuccess{}\n\n\t// During rollback, some of the snapshot stats get reset\n\t// or updated by slice. Therefore, serialise rollback and\n\t// retrieving stats from slice to avoid any inconsistency\n\t// in stats\n\tsm.statsLock.Lock()\n\tdefer sm.statsLock.Unlock()\n\n\tstreamId := cmd.(*MsgRollback).GetStreamId()\n\trollbackTs := cmd.(*MsgRollback).GetRollbackTs()\n\tkeyspaceId := cmd.(*MsgRollback).GetKeyspaceId()\n\tsessionId := cmd.(*MsgRollback).GetSessionId()\n\n\tlogging.Infof(\"StorageMgr::handleRollback %v %v rollbackTs %v\", streamId, keyspaceId, rollbackTs)\n\n\tvar err error\n\tvar restartTs *common.TsVbuuid\n\tvar rollbackToZero bool\n\n\tindexInstMap := sm.indexInstMap.Get()\n\tindexPartnMap := sm.indexPartnMap.Get()\n\t//for every index managed by this indexer\n\tfor idxInstId, partnMap := range indexPartnMap {\n\t\tidxInst := indexInstMap[idxInstId]\n\n\t\t//if this keyspace in stream needs to be rolled back\n\t\tif idxInst.Defn.KeyspaceId(idxInst.Stream) == keyspaceId &&\n\t\t\tidxInst.Stream == streamId &&\n\t\t\tidxInst.State != common.INDEX_STATE_DELETED {\n\n\t\t\trestartTs, err = sm.rollbackIndex(streamId,\n\t\t\t\tkeyspaceId, rollbackTs, idxInstId, partnMap, restartTs)\n\n\t\t\tif err != nil {\n\t\t\t\tsm.supvRespch <- &MsgRollbackDone{streamId: streamId,\n\t\t\t\t\tkeyspaceId: keyspaceId,\n\t\t\t\t\terr: err,\n\t\t\t\t\tsessionId: sessionId}\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif restartTs == nil {\n\t\t\t\terr = sm.rollbackAllToZero(streamId, keyspaceId)\n\t\t\t\tif err != nil {\n\t\t\t\t\tsm.supvRespch <- &MsgRollbackDone{streamId: streamId,\n\t\t\t\t\t\tkeyspaceId: keyspaceId,\n\t\t\t\t\t\terr: err,\n\t\t\t\t\t\tsessionId: sessionId}\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\trollbackToZero = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\tgo func() {\n\t\t// Notify all scan waiters for indexes in this keyspaceId\n\t\t// and stream with error\n\t\tstats := sm.stats.Get()\n\t\twaitersMap := sm.waitersMap.Get()\n\t\tfor idxInstId, wc := range waitersMap {\n\t\t\tidxInst := sm.indexInstMap.Get()[idxInstId]\n\t\t\tidxStats := stats.indexes[idxInst.InstId]\n\t\t\tif idxInst.Defn.KeyspaceId(idxInst.Stream) == keyspaceId &&\n\t\t\t\tidxInst.Stream == streamId {\n\t\t\t\twc.Lock()\n\t\t\t\tfor _, w := range wc.waiters {\n\t\t\t\t\tw.Error(ErrIndexRollback)\n\t\t\t\t\tif idxStats != nil {\n\t\t\t\t\t\tidxStats.numSnapshotWaiters.Add(-1)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\twc.waiters = nil\n\t\t\t\twc.Unlock()\n\t\t\t}\n\t\t}\n\t}()\n\n\tsm.updateIndexSnapMap(sm.indexPartnMap.Get(), streamId, keyspaceId)\n\n\tkeyspaceStats := sm.stats.GetKeyspaceStats(streamId, keyspaceId)\n\tif keyspaceStats != nil {\n\t\tkeyspaceStats.numRollbacks.Add(1)\n\t\tif rollbackToZero {\n\t\t\tkeyspaceStats.numRollbacksToZero.Add(1)\n\t\t}\n\t}\n\n\tif restartTs != nil {\n\t\t//for pre 7.0 index snapshots, the manifestUID needs to be set to epoch\n\t\trestartTs.SetEpochManifestUIDIfEmpty()\n\t\trestartTs = sm.validateRestartTsVbuuid(keyspaceId, restartTs)\n\t}\n\n\tsm.supvRespch <- &MsgRollbackDone{streamId: streamId,\n\t\tkeyspaceId: keyspaceId,\n\t\trestartTs: restartTs,\n\t\tsessionId: sessionId,\n\t}\n}", "func (e EndpointsRolloutV1Client) AutoGetRollout(ctx context.Context, in *Rollout) (*Rollout, error) {\n\tresp, err := e.AutoGetRolloutEndpoint(ctx, in)\n\tif err != nil {\n\t\treturn &Rollout{}, err\n\t}\n\treturn resp.(*Rollout), nil\n}", "func (app Application) applyRbacOnPipelineRunner() error {\n\tserviceAccount, err := app.applyPipelineServiceAccount()\n\tif err != nil {\n\t\tlogger.Errorf(\"Failed to apply service account needed by pipeline. %v\", err)\n\t\treturn err\n\t}\n\n\terr = app.givePipelineAccessToRR(serviceAccount, \"radix-pipeline-rr\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn app.givePipelineAccessToAppNamespace(serviceAccount)\n}", "func (e *Executor) Run() { e.loop() }", "func (re *RequestExecutor) DoWithRetries(req *Request, retryCount int) (*http.Response, error) {\n\t// Always rewind the request body when non-nil.\n\tif req.Body != nil {\n\t\tbody, err := req.body()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif c, ok := body.(io.ReadCloser); ok {\n\t\t\treq.Body = c\n\t\t} else {\n\t\t\treq.Body = ioutil.NopCloser(body)\n\t\t}\n\t}\n\n\tresp, err := re.httpClient.Do(req.Request)\n\tmaxRetries := int(re.config.MaxRetries)\n\tbo := re.config.BackoffEnabled\n\n\tif (err != nil || isTooMany(resp)) && retryCount < maxRetries {\n\t\tif resp != nil {\n\t\t\t// retrying so we must drain the body\n\t\t\ttryDrainBody(resp.Body)\n\t\t}\n\n\t\tif isTooMany(resp) {\n\t\t\t// Using an exponential back off method with no jitter for simplicity.\n\t\t\tif bo {\n\t\t\t\tBackoff(backoffDuration(retryCount, re.config.MinWait, re.config.MaxWait))\n\t\t\t}\n\t\t}\n\t\tretryCount++\n\n\t\tresp, err = re.DoWithRetries(req, retryCount)\n\t}\n\n\treturn resp, err\n}", "func TestEnroll(t *testing.T) {\n\tInitConfigForMsp()\n\tclient := fabric_sdk.NewClient()\n\tks := &sw.FileBasedKeyStore{}\n\tif err := ks.Init(nil, config.GetKeyStorePath(), false); err != nil {\n\t\tt.Fatalf(\"Failed initializing key store [%s]\", err)\n\t}\n\n\tcryptoSuite, err := bccspFactory.GetBCCSP(&bccspFactory.SwOpts{Ephemeral_: true, SecLevel: config.GetSecurityLevel(),\n\t\tHashFamily: config.GetSecurityAlgorithm(), KeyStore: ks})\n\tif err != nil {\n\t\tt.Fatalf(\"Failed getting ephemeral software-based BCCSP [%s]\", err)\n\t}\n\tclient.SetCryptoSuite(cryptoSuite)\n\tstateStore, err := kvs.CreateNewFileKeyValueStore(\"/tmp/enroll_user\")\n\tif err != nil {\n\t\tt.Fatalf(\"CreateNewFileKeyValueStore return error[%s]\", err)\n\t}\n\tclient.SetStateStore(stateStore)\n\n\tmsps, err := msp.NewMSPServices(config.GetMspURL(), config.GetMspClientPath())\n\tif err != nil {\n\t\tt.Fatalf(\"NewMSPServices return error: %v\", err)\n\t}\n\tkey, cert, err := msps.Enroll(\"testUser2\", \"user2\")\n\tif err != nil {\n\t\tt.Fatalf(\"Enroll return error: %v\", err)\n\t}\n\tif key == nil {\n\t\tt.Fatalf(\"private key return from Enroll is nil\")\n\t}\n\tif cert == nil {\n\t\tt.Fatalf(\"cert return from Enroll is nil\")\n\t}\n\n\tcertPem, _ := pem.Decode(cert)\n\tif err != nil {\n\t\tt.Fatalf(\"pem Decode return error: %v\", err)\n\t}\n\n\tcert509, err := x509.ParseCertificate(certPem.Bytes)\n\tif err != nil {\n\t\tt.Fatalf(\"x509 ParseCertificate return error: %v\", err)\n\t}\n\tif cert509.Subject.CommonName != \"testUser2\" {\n\t\tt.Fatalf(\"CommonName in x509 cert is not the enrollmentID\")\n\t}\n\n\tkeyPem, _ := pem.Decode(key)\n\tif err != nil {\n\t\tt.Fatalf(\"pem Decode return error: %v\", err)\n\t}\n\tuser := fabric_sdk.NewUser(\"testUser2\")\n\tk, err := client.GetCryptoSuite().KeyImport(keyPem.Bytes, &bccsp.ECDSAPrivateKeyImportOpts{Temporary: false})\n\tif err != nil {\n\t\tt.Fatalf(\"KeyImport return error: %v\", err)\n\t}\n\tuser.SetPrivateKey(k)\n\tuser.SetEnrollmentCertificate(cert)\n\terr = client.SetUserContext(user, false)\n\tif err != nil {\n\t\tt.Fatalf(\"client.SetUserContext return error: %v\", err)\n\t}\n\tuser, err = client.GetUserContext(\"testUser2\")\n\tif err != nil {\n\t\tt.Fatalf(\"client.GetUserContext return error: %v\", err)\n\t}\n\tif user == nil {\n\t\tt.Fatalf(\"client.GetUserContext return nil\")\n\t}\n\n}", "func RetryDuring(attempts int, timeout time.Duration, sleep time.Duration, myfunc func() error) error {\n\tt0 := time.Now()\n\n\tvar err error\n\tfor i := 0; ; i++ {\n\t\tif err = myfunc(); err == nil {\n\t\t\tglog.V(4).Infof(\"[retry-%d/%d] success\", i+1, attempts)\n\t\t\treturn nil\n\t\t}\n\n\t\tif _, skipRetry := err.(*ErrorSkipRetry); skipRetry {\n\t\t\terr = fmt.Errorf(\"failing without retries: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\tglog.V(4).Infof(\"[retry-%d/%d] Warning %v\", i+1, attempts, err)\n\t\tif i >= (attempts - 1) {\n\t\t\tbreak\n\t\t}\n\n\t\tif timeout > 0 {\n\t\t\tif delta := time.Now().Sub(t0); delta > timeout {\n\t\t\t\terr = fmt.Errorf(\"failed after %d attepmts (during %v) last error: %v\", i+1, delta, err)\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\tif sleep > 0 {\n\t\t\ttime.Sleep(sleep)\n\t\t}\n\t}\n\n\terr = fmt.Errorf(\"failed after %d attepmts, last error: %v\", attempts, err)\n\treturn err\n}", "func (t *Retrier) Do(ctx context.Context, action ActionFunc) error {\n\tvar actionErr error\n\tfor attempts := uint(0); attempts < t.maxAttempts; attempts++ {\n\t\t//sleep for a bit to avoid bombarding the requested resource. The backoff func should return 0 for the first attempt\n\t\ttime.Sleep(t.backoff(ctx, attempts))\n\n\t\t//check if the context was cancelled\n\t\tif IsContextDone(ctx) && !t.ignoreCtx {\n\t\t\treturn ErrContextCanceled\n\t\t}\n\n\t\terr, retriable := action()\n\t\tif err == nil {\n\t\t\treturn nil //success\n\t\t}\n\n\t\tt.onError(err) //allow the user to handle/log the error\n\n\t\t//it can happen that the context is canceled during the request\n\t\tif IsCanceledContextError(err) && !t.ignoreCtx {\n\t\t\treturn t.errorHandler(ErrContextCanceled, err)\n\t\t}\n\n\t\tif !retriable {\n\t\t\treturn t.errorHandler(ErrRequestNotRetriable, err)\n\t\t}\n\n\t\tactionErr = err\n\t}\n\n\treturn t.errorHandler(ErrNotSuccessful, actionErr)\n}", "func (o *Subscriber) doAfterUpsertHooks(ctx context.Context, exec boil.ContextExecutor) (err error) {\n\tfor _, hook := range subscriberAfterUpsertHooks {\n\t\tif err := hook(ctx, exec, o); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}" ]
[ "0.60126114", "0.48023757", "0.474493", "0.472079", "0.44610307", "0.44345552", "0.4426037", "0.44184694", "0.43913168", "0.43720782", "0.43568626", "0.43513268", "0.43281728", "0.43149644", "0.42825097", "0.42096183", "0.4195741", "0.41598186", "0.4157024", "0.41542852", "0.41475776", "0.41407564", "0.41112512", "0.41012916", "0.40813786", "0.40481365", "0.40333334", "0.40284222", "0.40118328", "0.4011502", "0.40099952", "0.40089408", "0.3999752", "0.3983021", "0.39722484", "0.39586982", "0.39529312", "0.39424926", "0.393923", "0.39370495", "0.3927457", "0.3923348", "0.3922919", "0.39126533", "0.39092043", "0.3904339", "0.3898979", "0.3898801", "0.38958043", "0.38953796", "0.3892277", "0.3891307", "0.38890573", "0.38779727", "0.38495442", "0.38477975", "0.384611", "0.38386425", "0.3831882", "0.38317308", "0.38309416", "0.38265046", "0.381818", "0.38160077", "0.38132212", "0.3811772", "0.3807983", "0.38070098", "0.37979454", "0.37877426", "0.37809783", "0.37778386", "0.37766558", "0.37742355", "0.37712714", "0.3749945", "0.37403634", "0.37390643", "0.3738761", "0.37381074", "0.37358052", "0.37358052", "0.37341753", "0.3732958", "0.37300482", "0.37279886", "0.37266177", "0.37261364", "0.37256905", "0.37241203", "0.37227362", "0.3716445", "0.37031016", "0.3701021", "0.3693382", "0.36926425", "0.36906183", "0.36904395", "0.36876693", "0.36865622" ]
0.7432249
0
ReadPointer tries to read LFS pointer data from the reader
func ReadPointer(reader io.Reader) (Pointer, error) { buf := make([]byte, blobSizeCutoff) n, err := io.ReadFull(reader, buf) if err != nil && err != io.ErrUnexpectedEOF { return Pointer{}, err } buf = buf[:n] return ReadPointerFromBuffer(buf) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ReadPointerFile(reader io.Reader) (*models.LFSMetaObject, *[]byte) {\n\tif !setting.LFS.StartServer {\n\t\treturn nil, nil\n\t}\n\n\tbuf := make([]byte, 1024)\n\tn, _ := reader.Read(buf)\n\tbuf = buf[:n]\n\n\tif isTextFile := base.IsTextFile(buf); !isTextFile {\n\t\treturn nil, nil\n\t}\n\n\treturn IsPointerFile(&buf), &buf\n}", "func Read(fd int32, p unsafe.Pointer, n int32) int32", "func ReadPointerFromBuffer(buf []byte) (Pointer, error) {\n\tvar p Pointer\n\n\theadString := string(buf)\n\tif !strings.HasPrefix(headString, MetaFileIdentifier) {\n\t\treturn p, ErrMissingPrefix\n\t}\n\n\tsplitLines := strings.Split(headString, \"\\n\")\n\tif len(splitLines) < 3 {\n\t\treturn p, ErrInvalidStructure\n\t}\n\n\toid := strings.TrimPrefix(splitLines[1], MetaFileOidPrefix)\n\tif len(oid) != 64 || !oidPattern.MatchString(oid) {\n\t\treturn p, ErrInvalidOIDFormat\n\t}\n\tsize, err := strconv.ParseInt(strings.TrimPrefix(splitLines[2], \"size \"), 10, 64)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\n\tp.Oid = oid\n\tp.Size = size\n\n\treturn p, nil\n}", "func ReadLedgerMarker(row *sql.Row, a *LedgerMarker) error {\n\terr := row.Scan(&a.LMID, &a.LID, &a.BID, &a.RAID, &a.RID, &a.TCID, &a.Dt, &a.Balance, &a.State, &a.CreateTS, &a.CreateBy, &a.LastModTime, &a.LastModBy)\n\tSkipSQLNoRowsError(&err)\n\treturn err\n}", "func (s *SeekerWrapper) ReadAt(p []byte, offset int64) (int, error) { return s.s.ReadAt(p, offset) }", "func readll() int64 {\n\treturn _readInt64()\n}", "func readll() int64 {\n\treturn _readInt64()\n}", "func Pread64(t *kernel.Task, args arch.SyscallArguments) (uintptr, *kernel.SyscallControl, error) {\n\tfd := args[0].Int()\n\taddr := args[1].Pointer()\n\tsize := args[2].SizeT()\n\toffset := args[3].Int64()\n\n\tfile := t.GetFile(fd)\n\tif file == nil {\n\t\treturn 0, nil, syserror.EBADF\n\t}\n\tdefer file.DecRef()\n\n\t// Check that the offset is legitimate.\n\tif offset < 0 {\n\t\treturn 0, nil, syserror.EINVAL\n\t}\n\n\t// Is reading at an offset supported?\n\tif !file.Flags().Pread {\n\t\treturn 0, nil, syserror.ESPIPE\n\t}\n\n\t// Check that the file is readable.\n\tif !file.Flags().Read {\n\t\treturn 0, nil, syserror.EBADF\n\t}\n\n\t// Check that the size is legitimate.\n\tsi := int(size)\n\tif si < 0 {\n\t\treturn 0, nil, syserror.EINVAL\n\t}\n\n\t// Get the destination of the read.\n\tdst, err := t.SingleIOSequence(addr, si, usermem.IOOpts{\n\t\tAddressSpaceActive: true,\n\t})\n\tif err != nil {\n\t\treturn 0, nil, err\n\t}\n\n\tn, err := preadv(t, file, dst, offset)\n\tt.IOUsage().AccountReadSyscall(n)\n\treturn uintptr(n), nil, handleIOError(t, n != 0, err, kernel.ERESTARTSYS, \"pread64\", file)\n}", "func (m *wasiSnapshotPreview1Impl) fdPread(pfd wasiFd, piovs list, poffset wasiFilesize) (rv wasiSize, err wasiErrno) {\n\tf, err := m.files.getFile(pfd, wasiRightsFdRead)\n\tif err != wasiErrnoSuccess {\n\t\treturn 0, err\n\t}\n\n\tn, ferr := f.Pread(m.buffers(wasiIovecArray(piovs)), int64(poffset))\n\tif ferr != nil {\n\t\treturn n, fileErrno(ferr)\n\t}\n\treturn n, wasiErrnoSuccess\n}", "func (b *buffer) read(loc Location) unsafe.Pointer {\n\treturn atomic.LoadPointer(b.index(loc.index()))\n}", "func (r *lockedSource) read(p []byte, readVal *int64, readPos *int8) (n int, err error) {\n\tr.lk.Lock()\n\tn, err = read(p, r.src.Int63, readVal, readPos)\n\tr.lk.Unlock()\n\treturn\n}", "func (m *Mmap) GetPointer() int64 {\n\treturn m.FilePointer\n}", "func readPointerCaptureState(ctx context.Context, d *ui.Device) (*pointerCaptureState, error) {\n\tview := d.Object(ui.ID(motioninput.Package + \":id/pointer_capture_state\"))\n\tif err := view.WaitForExists(ctx, 5*time.Second); err != nil {\n\t\treturn nil, err\n\t}\n\ttext, err := view.GetText(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar state pointerCaptureState\n\tif err := json.Unmarshal([]byte(text), &state); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &state, nil\n}", "func (wpr *Wrapper) Read(p []byte) (n int, err error) {\n\tn, err = wpr.ReadAt(p, 0)\n\twpr.L -= int64(n)\n\twpr.O += int64(n)\n\twpr.O %= wpr.N\n\treturn n, err\n}", "func (w *wrapper) Read(path string, buff []byte, ofst int64, fd uint64) int {\n\tfh, ok := w.getFileDescriptor(fd)\n\tif !ok {\n\t\treturn -fuse.EINVAL\n\t}\n\tn, err := fh.ReadAt(buff, ofst)\n\tif n > 0 || err == io.EOF {\n\t\treturn n\n\t}\n\treturn convertError(err)\n}", "func readll() int64 {\n\treturn readInt64()\n}", "func (m *wasiSnapshotPreview1Impl) pathReadlink(pfd wasiFd, ppath list, pbuf pointer, pbufLen wasiSize) (rv wasiSize, err wasiErrno) {\n\tpath, err := m.loadPath(ppath)\n\tif err != wasiErrnoSuccess {\n\t\treturn 0, err\n\t}\n\n\tdir, err := m.files.getDirectory(pfd, wasiRightsPathReadlink)\n\tif err != wasiErrnoSuccess {\n\t\treturn 0, err\n\t}\n\n\tdest, ferr := dir.ReadLink(path)\n\tif ferr != nil {\n\t\treturn 0, fileErrno(ferr)\n\t}\n\treturn wasiSize(copy(m.slice(pbuf, pbufLen), dest)), wasiErrnoSuccess\n}", "func readPosData(posDataFilename string) (*dataFile, error) {\n data := dataFile{}\n\n infile, err := os.Open(posDataFilename)\n defer infile.Close()\n if err != nil {\n return nil, fmt.Errorf(\"can't open %s: %v\", posDataFilename, err)\n }\n r := bufio.NewReader(infile)\n if (r == nil) {\n return nil, fmt.Errorf(\"can't read %s: %v\" + posDataFilename, err)\n }\n\n var readerr error = nil\n for ; readerr == nil ; {\n bytebuf, readerr := r.ReadBytes('\\n')\n if readerr != nil && readerr != io.EOF {\n panic(readerr)\n }\n if len(bytebuf) == 0 {\n break;\n }\n line := string(bytebuf)\n if line[0:2] == \" \" {\n // comment line\n continue\n }\n fields := strings.SplitN(strings.TrimSpace(line), \" \", -1)\n synset_offset, _ := strconv.Atoi(fields[0])\n lex_filenum, _ := strconv.Atoi(fields[1])\n ss_type := oneCharPosTagToPosId(fields[2])\n w_cnt64, _ := strconv.ParseInt(fields[3], 16, 0)\n w_cnt := int(w_cnt64)\n words := make([]string, w_cnt)\n lex_ids := make([]int, w_cnt)\n fieldIndex := 4\n for i := 0; i < w_cnt; i++ {\n words[i] = readStoredLemma(fields[fieldIndex])\n fieldIndex++\n lex_id64, _ := strconv.ParseInt(fields[fieldIndex], 16, 0)\n lex_ids[i] = int(lex_id64)\n fieldIndex++\n }\n p_cnt, _ := strconv.Atoi(fields[fieldIndex])\n fieldIndex++\n pointers := make([]RelationshipEdge, p_cnt)\n for i := 0; i < p_cnt; i++ {\n pointer_type, symbolFound := RELATIONSHIP_POINTER_SYMBOLS[fields[fieldIndex]]\n if !symbolFound {\n panic(fmt.Sprintf(\"could not handle relationship symbol %s in line <<%v>>, file %s\", fields[fieldIndex], line, posDataFilename))\n }\n fieldIndex++\n synset_offset, _ := strconv.Atoi(fields[fieldIndex])\n fieldIndex++\n pos := oneCharPosTagToPosId(fields[fieldIndex])\n fieldIndex++\n\n src_wordnum64, _ := strconv.ParseInt(fields[fieldIndex][0:2], 16, 0)\n dest_wordnum64, _ := strconv.ParseInt(fields[fieldIndex][2:4], 16, 0)\n fieldIndex++\n src_word_num := int(src_wordnum64)\n dest_word_num := int(dest_wordnum64)\n pointers[i] = RelationshipEdge {\n RelationshipType: pointer_type,\n SynsetOffset: synset_offset,\n PartOfSpeech: pos,\n SourceWordNumber: src_word_num,\n TargetWordNumber: dest_word_num,\n }\n }\n // skip data.verb frames\n\n pipeIndex := strings.LastIndex(line, \"|\")\n var gloss string\n if pipeIndex >= 0 {\n gloss = strings.TrimSpace(line[pipeIndex + 2:])\n } else {\n gloss = \"\"\n }\n\n data[synset_offset] = Synset {\n SynsetOffset: synset_offset,\n LexographerFilenum: lex_filenum,\n PartOfSpeech: ss_type,\n Words: words,\n LexIds: lex_ids,\n Relationships: pointers,\n Gloss: gloss,\n }\n }\n\n return &data, nil\n}", "func (m *wasiSnapshotPreview1Impl) fdRead(pfd wasiFd, piovs list) (rv wasiSize, err wasiErrno) {\n\tf, err := m.files.getFile(pfd, wasiRightsFdRead)\n\tif err != wasiErrnoSuccess {\n\t\treturn 0, err\n\t}\n\n\tn, ferr := f.Readv(m.buffers(wasiIovecArray(piovs)))\n\tif ferr != nil {\n\t\treturn n, fileErrno(ferr)\n\t}\n\treturn n, wasiErrnoSuccess\n}", "func (fr *FileReader) readerForOffset(off int64) (io.ReadCloser, error) {\n\tif off < 0 {\n\t\tpanic(\"negative offset\")\n\t}\n\tif off >= fr.size {\n\t\treturn eofReader, nil\n\t}\n\toffRemain := off\n\tparts := fr.ss.Parts\n\tfor len(parts) > 0 && parts[0].Size <= uint64(offRemain) {\n\t\toffRemain -= int64(parts[0].Size)\n\t\tparts = parts[1:]\n\t}\n\tif len(parts) == 0 {\n\t\treturn eofReader, nil\n\t}\n\tp0 := parts[0]\n\tvar rsc blobref.ReadSeekCloser\n\tvar err error\n\tswitch {\n\tcase p0.BlobRef != nil && p0.BytesRef != nil:\n\t\treturn nil, fmt.Errorf(\"part illegally contained both a blobRef and bytesRef\")\n\tcase p0.BlobRef == nil && p0.BytesRef == nil:\n\t\treturn &nZeros{int(p0.Size - uint64(offRemain))}, nil\n\tcase p0.BlobRef != nil:\n\t\trsc, _, err = fr.fetcher.Fetch(p0.BlobRef)\n\tcase p0.BytesRef != nil:\n\t\trsc, err = NewFileReader(fr.fetcher, p0.BytesRef)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toffRemain += int64(p0.Offset)\n\tif offRemain > 0 {\n\t\tnewPos, err := rsc.Seek(offRemain, os.SEEK_SET)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif newPos != offRemain {\n\t\t\tpanic(\"Seek didn't work\")\n\t\t}\n\t}\n\treturn struct {\n\t\tio.Reader\n\t\tio.Closer\n\t}{\n\t\tio.LimitReader(rsc, int64(p0.Size)),\n\t\trsc,\n\t}, nil\n}", "func (*FileSystemBase) Readlink(path string) (int, string) {\n\treturn -ENOSYS, \"\"\n}", "func (fsys *FS) Readlink(path string) (errc int, linkPath string) {\n\tdefer fs.Trace(path, \"\")(\"linkPath=%q, errc=%d\", &linkPath, &errc)\n\treturn -fuse.ENOSYS, \"\"\n}", "func ReadFromManagedReader(oid uint64, data []byte) (int, int) {\n\tri, ok := GetManagedObject(oid)\n\tif !ok {\n\t\tpanic(\"failed to get reader\")\n\t}\n\tr := ri.(io.Reader)\n\tn, err := r.Read(data)\n\tif err != nil {\n\t\tif err == io.EOF {\n\t\t\treturn 0, 0\n\t\t}\n\t\tplog.Errorf(\"got err %+v when reading from the snapshot reader\", err)\n\t\treturn -1, getErrnoFromError(err)\n\t}\n\treturn n, 0\n}", "func GeneratePointer(content io.Reader) (Pointer, error) {\n\th := sha256.New()\n\tc, err := io.Copy(h, content)\n\tif err != nil {\n\t\treturn Pointer{}, err\n\t}\n\tsum := h.Sum(nil)\n\treturn Pointer{Oid: hex.EncodeToString(sum), Size: c}, nil\n}", "func (vm *VirtualMachine) handleRead(quad quads.Quad) {\n\treader := bufio.NewReader(os.Stdin)\n\tbytes, _ := reader.ReadBytes('\\n')\n\tstr := strings.TrimSpace(string(bytes))\n\n\tvar memblock Memory\n\tvar addr int\n\tif strings.Contains(quad.Result.ID(), \"ptr_\") {\n\t\tmemblock = vm.getMemBlockForAddr(quad.Result.GetAddr())\n\t\taddrFloat, ok := memblock.Get(quad.Result.GetAddr()).(float64)\n\t\tif !ok {\n\t\t\tlog.Fatalf(\"Error: (handleRead) couldn't cast %v to float64\",\n\t\t\t\tmemblock.Get(int(addr)))\n\t\t}\n\n\t\taddr = int(addrFloat)\n\t\tauxElement := quads.NewElement(addr, quad.Result.ID(), quad.Result.Type(), \"\")\n\t\tmemblock = vm.getMemBlockForElement(auxElement)\n\t} else {\n\t\tmemblock = vm.getMemBlockForElement(quad.Result)\n\t\taddr = quad.Result.GetAddr()\n\t}\n\n\tswitch quad.Result.Type() {\n\tcase constants.TYPEINT:\n\t\tfloatVal, err := strconv.ParseFloat(str, 64)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Warning: read to int expects int\")\n\t\t}\n\n\t\terr = memblock.Set(float64(int(floatVal)), addr)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (handleRead) %s\", err)\n\t\t}\n\n\tcase constants.TYPEFLOAT:\n\t\tfloatVal, err := strconv.ParseFloat(str, 64)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Warning: read to float expects float\")\n\t\t}\n\t\terr = memblock.Set(floatVal, addr)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (handleRead) %s\", err)\n\t\t}\n\n\tcase constants.TYPECHAR:\n\t\tif len(str) > 0 {\n\t\t\truneVal := str[0]\n\t\t\terr := memblock.Set(runeVal, addr)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatalf(\"Error: (handleRead) %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Println(\"Warning: read to char expects a single char\")\n\t\t}\n\n\tcase constants.TYPEBOOL:\n\t\tboolVal, err := strconv.ParseBool(str)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"Warning: read to bool expects bool\")\n\t\t}\n\t\terr = memblock.Set(boolVal, addr)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (handleRead) %s\", err)\n\t\t}\n\t}\n}", "func (r *trackingreader) Read(b []byte) (int, error) {\n\tn, err := r.Reader.Read(b)\n\tr.pos += int64(n)\n\treturn n, err\n}", "func (o *LargeObject) Read(p []byte) (int, error) {\n\tvar res []byte\n\terr := o.tx.QueryRow(o.ctx, \"select loread($1, $2)\", o.fd, len(p)).Scan(&res)\n\tcopy(p, res)\n\tif err != nil {\n\t\treturn len(res), err\n\t}\n\n\tif len(res) < len(p) {\n\t\terr = io.EOF\n\t}\n\treturn len(res), err\n}", "func (t *Tailer) GetReadOffset() int64 {\n\treturn atomic.LoadInt64(&t.readOffset)\n}", "func (p *Stream) GetReadPos() int {\n\treturn p.readSeg*streamBlockSize + p.readIndex\n}", "func (reader *SnapshotStreamReader) Read(p []byte) (n int, err error) {\n\tif reader.readPtr == len(reader.buf) {\n\t\t// No more data in buf, do another read\n\t\theader, data, err := reader.recv()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tif *header != *reader.header {\n\t\t\treturn 0, errorDifferentHeader\n\t\t}\n\n\t\tif !reader.partcb(header) {\n\t\t\treturn 0, errorSnapshotFromStaleLeader\n\t\t}\n\n\t\treader.buf = data\n\t\treader.readPtr = 0\n\t}\n\n\tn = copy(p, reader.buf[reader.readPtr:])\n\treader.readPtr += n\n\n\treturn n, nil\n}", "func (v *VFS) Readlink(name string) (string, error) {\n\tf, err := v.FileSystem.Open(name)\n\tif nil != err {\n\t\treturn \"\", err\n\t}\n\tb, err := ioutil.ReadAll(f)\n\tif nil != err {\n\t\treturn \"\", err\n\t}\n\treturn string(b), nil\n}", "func (w *wrapper) Readlink(path string) (int, string) {\n\tif sfs, ok := w.underlying.(billy.Symlink); ok {\n\t\tfn, err := sfs.Readlink(path)\n\t\tif err != nil {\n\t\t\treturn convertError(err), \"\"\n\t\t}\n\t\treturn 0, fn\n\t}\n\treturn -fuse.ENOSYS, \"\"\n}", "func(this*wrapper)Read(p[]byte)(int,error){\nreturn this.f.Read(p)\n}", "func (s *Session) read(key *Key, offset uint64, size uint64) (data unsafe.Pointer, dataSize uint64, err error) {\n\tatomic.AddUint64(&cReads, 1)\n\n\tio_attr := C.struct_dnet_io_attr{\n\t\tparent: key.id.id, id: key.id.id, _type: key.id._type,\n\t\toffset: C.uint64_t(offset), size: C.uint64_t(size),\n\t}\n\n\tvar cflags C.uint64_t\n\tvar errp C.int\n\tdata = C.dnet_read_data_wait(s.session, &key.id, &io_attr, cflags, &errp)\n\tdataSize = uint64(io_attr.size)\n\tif errp != 0 {\n\t\terr = Error(errp)\n\t}\n\treturn\n}", "func (f *FakeFileSystem) Readlink(name string) (string, error) {\n\treturn f.ReadlinkName, f.ReadlinkError\n}", "func (d *dataHandlerWays) ReadNode(n gosmparse.Node) {\n\n}", "func (b *Buffer) Read(data []byte, c Cursor) (n int, next Cursor, err error) {\n\tb.mu.RLock()\n\tdefer b.mu.RUnlock()\n\n\tseq, offset := c.seq, c.offset\n\n\tif seq >= b.nextSeq || offset > b.last {\n\t\treturn 0, next, ErrNotArrived\n\t}\n\n\tf := b.frame(offset)\n\tif f.size() == 0 || f.seq() != seq {\n\t\treturn b.readFirst(data)\n\t}\n\n\treturn b.readOffset(data, offset)\n}", "func (r *progressReader) Read(p []byte) (n int, err error) {\n\tr.lastRead = time.Now()\n\tlei, err := (*(r.reader)).Read(p)\n\tr.pos += int64(lei)\n\treturn lei, err\n}", "func (fr *fieldReader) readBlock(fileOff int64) error {\n\tfb := &fr.fb\n\n\tfr.rio.Seek(recordio.ItemLocation{uint64(fileOff), 0})\n\tif !fr.rio.Scan() {\n\t\terr := fr.rio.Err()\n\t\tif err == nil {\n\t\t\terr = errors.Errorf(\"Failed to read a block at offset %d\", fileOff)\n\t\t}\n\t\treturn err\n\t}\n\tfb.buf = fr.rio.Get().([]byte)\n\tvar err error\n\tfb.header, err = readBlockHeader(&fb.buf)\n\treturn err\n}", "func (h *DevicePluginHandlerImpl) readCheckpoint() error {\n\tfilepath := h.devicePluginManager.CheckpointFile()\n\tcontent, err := ioutil.ReadFile(filepath)\n\tif err != nil && !os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"failed to read checkpoint file %q: %v\", filepath, err)\n\t}\n\tglog.V(2).Infof(\"Read checkpoint file %s\\n\", filepath)\n\tvar data checkpointData\n\tif err := json.Unmarshal(content, &data); err != nil {\n\t\treturn fmt.Errorf(\"failed to unmarshal checkpoint data: %v\", err)\n\t}\n\tfor _, entry := range data.Entries {\n\t\tglog.V(2).Infof(\"Get checkpoint entry: %v %v %v %v\\n\", entry.PodUID, entry.ContainerName, entry.ResourceName, entry.DeviceID)\n\t\tif h.allocatedDevices[entry.ResourceName] == nil {\n\t\t\th.allocatedDevices[entry.ResourceName] = make(podDevices)\n\t\t}\n\t\th.allocatedDevices[entry.ResourceName].insert(entry.PodUID, entry.ContainerName, entry.DeviceID)\n\t}\n\treturn nil\n}", "func (fs *Mysqlfs) Readlink(link string) (string, error) {\n\tf, err := fs.storage.GetFile(link)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif f == nil {\n\t\treturn \"\", os.ErrNotExist\n\t}\n\n\tif !isSymlink(f.Mode) {\n\t\treturn \"\", &os.PathError{\n\t\t\tOp: \"readlink\",\n\t\t\tPath: link,\n\t\t\tErr: fmt.Errorf(\"not a symlink\"),\n\t\t}\n\t}\n\n\treturn string(f.Content), nil\n}", "func (r *chunkReader) Read(data []byte) (int, error) {\n\tbytesToRead := len(data)\n\tr.l.Debug(\"Start cafs reader Read\", zap.Int(\"length\", bytesToRead))\n\n\tif r.lastChunk && r.rdr == nil {\n\t\treturn 0, io.EOF\n\t}\n\tfor {\n\t\tkey := r.keys[r.idx]\n\t\tif r.rdr == nil {\n\t\t\trdr, err := r.fs.Get(context.Background(), r.pather(key))\n\t\t\tif err != nil {\n\t\t\t\treturn r.readSoFar, err\n\t\t\t}\n\t\t\tr.rdr = rdr\n\t\t}\n\n\t\tn, errRead := r.rdr.Read(data[r.readSoFar:])\n\n\t\tdefer func() {\n\t\t\tif r.MetricsEnabled() && errRead == nil {\n\t\t\t\tr.m.Volume.Blobs.IncBlob(\"read\")\n\t\t\t\tr.m.Volume.Blobs.Size(int64(n), \"read\")\n\t\t\t}\n\t\t\tr.l.Debug(\"End cafs reader Read\", zap.Int(\"length\", bytesToRead))\n\t\t}()\n\n\t\tr.currLeaf = append(r.currLeaf, data[r.readSoFar:r.readSoFar+n]...)\n\t\tif errRead != nil {\n\t\t\tr.rdr.Close() // TODO(fred): nice - why are we ignoring errors here?\n\t\t\tr.readSoFar += n\n\t\t\tif errRead == io.EOF { // we reached the end of the stream for this key\n\t\t\t\tr.idx++\n\t\t\t\tr.rdr = nil\n\t\t\t\tr.lastChunk = r.idx == len(r.keys)\n\t\t\t\tif r.withVerifyHash {\n\t\t\t\t\tnodeOffset := r.idx\n\t\t\t\t\tisLastNode := false\n\n\t\t\t\t\t// NOTE: we follow the checksumming scheme adopted by the writer.\n\t\t\t\t\t// The writer behaves in a way a bit unexpected here: not only offets don't start at zero\n\t\t\t\t\t// as one might expect, but the last node is not flagged as the last one\n\t\t\t\t\t// when the content size is aligned with the leaf size.\n\t\t\t\t\tif r.lastChunk && uint32(len(r.currLeaf)) != r.leafSize {\n\t\t\t\t\t\tnodeOffset--\n\t\t\t\t\t\tisLastNode = true\n\t\t\t\t\t}\n\t\t\t\t\tr.l.Debug(\"cafs reader Read: hash verification\", zap.Stringer(\"key\", key))\n\t\t\t\t\tif err := r.verifyHash(key, r.currLeaf, nodeOffset, isLastNode); err != nil {\n\t\t\t\t\t\treturn 0, err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif r.lastChunk { // this was the last chunk, so also EOF for this hash\n\t\t\t\t\tif n == bytesToRead {\n\t\t\t\t\t\treturn n, nil\n\t\t\t\t\t}\n\t\t\t\t\treturn r.readSoFar, io.EOF\n\t\t\t\t}\n\t\t\t\t// move on to the next key\n\t\t\t\tr.currLeaf = make([]byte, 0)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn n, errRead\n\t\t}\n\t\t// we filled up the entire byte slice but still have data remaining in the reader,\n\t\t// we should move on to receive the next buffer\n\t\tr.readSoFar += n\n\t\tif r.readSoFar >= bytesToRead {\n\t\t\tr.readSoFar = 0\n\t\t\t// return without error\n\t\t\treturn bytesToRead, nil\n\t\t}\n\t}\n}", "func (image *Image2D) GetDataPointer() unsafe.Pointer {\n\treturn gl.Ptr(image.data)\n}", "func (bp *Proc) Read(ctx context.Context, file *fs.File, dst usermem.IOSequence, offset int64) (int64, error) {\n\treturn 0, syserror.EOPNOTSUPP\n}", "func (f unixFileHandle) ReadAt(b []byte, offset int64) (n int, err error) {\n\tn, err = syscall.Pread(syscallFd(f), b, offset)\n\terr = handleSyscallError(err)\n\tif n == 0 && len(b) > 0 && err == nil {\n\t\terr = io.EOF\n\t}\n\treturn\n}", "func LZRead(unnamed0 INT, unnamed1 LPSTR, unnamed2 INT) INT {\n\tret1 := syscall3(lZRead, 3,\n\t\tuintptr(unnamed0),\n\t\tuintptr(unsafe.Pointer(unnamed1)),\n\t\tuintptr(unnamed2))\n\treturn INT(ret1)\n}", "func (f *ClientFD) ReadLinkAt(ctx context.Context) (string, error) {\n\treq := ReadLinkAtReq{FD: f.fd}\n\tvar resp ReadLinkAtResp\n\tctx.UninterruptibleSleepStart(false)\n\terr := f.client.SndRcvMessage(ReadLinkAt, uint32(req.SizeBytes()), req.MarshalUnsafe, resp.UnmarshalBytes, nil)\n\tctx.UninterruptibleSleepFinish(false)\n\treturn string(resp.Target), err\n}", "func (fr *fieldReader) readNextBlock() bool {\n\tif len(fr.blocks) == 0 {\n\t\treturn false\n\t}\n\taddr := fr.blocks[0]\n\tfr.blocks = fr.blocks[1:]\n\n\t// Read and uncompress the recordio block.\n\tif err := fr.readBlock(int64(addr.FileOffset)); err != nil {\n\t\tfr.err.Set(err)\n\t\treturn false\n\t}\n\t// Set up the read pointers\n\tfb := &fr.fb\n\tlimitOffset := uint32(len(fb.buf))\n\tdoassert(fb.header.Offset <= fb.header.BlobOffset)\n\tdoassert(fb.header.BlobOffset <= limitOffset)\n\tfb.reset(addr,\n\t\tfb.buf[fb.header.Offset:fb.header.BlobOffset],\n\t\tfb.buf[fb.header.BlobOffset:limitOffset])\n\tvlog.VI(2).Infof(\"%v: Read block %+v, %d remaining\", fr.label, addr, len(fr.blocks))\n\treturn true\n}", "func (l *localLinker) Read(ctx context.Context, req *pbd.ReadRequest) (*pbd.ReadResponse, error) {\n\tif val, ok := l.store[req.Key]; ok {\n\t\treturn &pbd.ReadResponse{Payload: val}, nil\n\t}\n\treturn nil, status.Error(codes.InvalidArgument, fmt.Sprintf(\"Unable to locate %v\", req.Key))\n}", "func (reader *embedFileReader) ReadAt(p []byte, off int64) (int, error) {\n\treturn 0, ErrNotImplemented\n}", "func (lf *logFile) read(buf []byte, offset int64) error {\n\tnbr, err := lf.fd.ReadAt(buf, offset)\n\ty.NumReads.Add(1)\n\ty.NumBytesRead.Add(int64(nbr))\n\treturn err\n}", "func (f *FileObject) Read(p []byte) (n int, err error) {\n\tif f.loc == f.length {\n\t\treturn 0, nil\n\t}\n\treturn f.obj.Read(p)\n}", "func (*FileSystemBase) Read(path string, buff []byte, ofst int64, fh uint64) int {\n\treturn -ENOSYS\n}", "func (r *regulator) Reader(ctx context.Context, path string, offset int64) (io.ReadCloser, error) {\n\tr.enter()\n\tdefer r.exit()\n\n\treturn r.StorageDriver.Reader(ctx, path, offset)\n}", "func (o *ODirectReader) Read(buf []byte) (n int, err error) {\n\tif o.err != nil && (len(o.buf) == 0 || !o.seenRead) {\n\t\treturn 0, o.err\n\t}\n\tif o.buf == nil {\n\t\tif o.SmallFile {\n\t\t\to.bufp = ODirectPoolSmall.Get().(*[]byte)\n\t\t} else {\n\t\t\to.bufp = ODirectPoolLarge.Get().(*[]byte)\n\t\t}\n\t}\n\tif !o.seenRead {\n\t\to.buf = *o.bufp\n\t\tn, err = o.File.Read(o.buf)\n\t\tif err != nil && err != io.EOF {\n\t\t\tif isSysErrInvalidArg(err) {\n\t\t\t\tif err = disk.DisableDirectIO(o.File); err != nil {\n\t\t\t\t\to.err = err\n\t\t\t\t\treturn n, err\n\t\t\t\t}\n\t\t\t\tn, err = o.File.Read(o.buf)\n\t\t\t}\n\t\t\tif err != nil && err != io.EOF {\n\t\t\t\to.err = err\n\t\t\t\treturn n, err\n\t\t\t}\n\t\t}\n\t\tif n == 0 {\n\t\t\t// err is likely io.EOF\n\t\t\to.err = err\n\t\t\treturn n, err\n\t\t}\n\t\to.err = err\n\t\to.buf = o.buf[:n]\n\t\to.seenRead = true\n\t}\n\tif len(buf) >= len(o.buf) {\n\t\tn = copy(buf, o.buf)\n\t\to.seenRead = false\n\t\treturn n, o.err\n\t}\n\tn = copy(buf, o.buf)\n\to.buf = o.buf[n:]\n\t// There is more left in buffer, do not return any EOF yet.\n\treturn n, nil\n}", "func (p *Process) read(addr uintptr, ptr interface{}) error {\n\t// Reflection magic!\n\tv := reflect.ValueOf(ptr)\n\tdataAddr := getDataAddr(v)\n\tdataSize := getDataSize(v)\n\n\t// Open the file mapped process memory.\n\tmem, err := os.Open(fmt.Sprintf(\"/proc/%d/mem\", p.PID))\n\tdefer mem.Close()\n\tif err != nil {\n\t\treturn errors.New(fmt.Sprintf(\"Error opening /proc/%d/mem. Are you root?\", p.PID))\n\t}\n\n\t// Create a buffer and read data into it.\n\tdataBuf := make([]byte, dataSize)\n\tn, err := mem.ReadAt(dataBuf, int64(addr))\n\tif n != int(dataSize) {\n\t\treturn errors.New(fmt.Sprintf(\"Tried to read %d bytes, actually read %d bytes\\n\", dataSize, n))\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\t// Unsafely cast to []byte to copy data into.\n\tbuf := (*[]byte)(unsafe.Pointer(&reflect.SliceHeader{\n\t\tData: dataAddr,\n\t\tLen: int(dataSize),\n\t\tCap: int(dataSize),\n\t}))\n\tcopy(*buf, dataBuf)\n\treturn nil\n}", "func (z *zstdReader) Read(p []byte) (n int, err error) {\n\treturn z.r.Read(p)\n}", "func (a *reader) Read(p []byte) (n int, err error) {\n\tif a.err != nil {\n\t\treturn 0, a.err\n\t}\n\t// Swap buffer and maybe return error\n\terr = a.fill()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Copy what we can\n\tn = copy(p, a.cur.buffer())\n\ta.cur.inc(n)\n\n\tif a.cur.isEmpty() {\n\t\t// Return current, so a fetch can start.\n\t\tif a.cur != nil {\n\t\t\t// If at end of buffer, return any error, if present\n\t\t\ta.err = a.cur.err\n\t\t\ta.reuse <- a.cur\n\t\t\ta.cur = nil\n\t\t}\n\t\treturn n, a.err\n\t}\n\treturn n, nil\n}", "func (ll *LogLink) Read() (*buffer.VectorisedView, error) {\n\tselect {\n\tcase event := <-ll.readPoller.Cr:\n\t\tif event.err != nil {\n\t\t\tlog.Printf(\"Read: Poll Error: %v\", event.err)\n\t\t\tbreak\n\t\t}\n\n\t\t// Unmarshal\n\t\tvar packetLog PacketLog\n\t\terr := json.Unmarshal(event.data, &packetLog)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\th := make([]byte, 1024)\n\t\tn, err := base64.StdEncoding.Decode(h, []byte(packetLog.Header))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\theader := buffer.NewViewFromBytes(h[:n])\n\n\t\tj := make([]byte, ll.ep.MTU())\n\t\tm, err := base64.StdEncoding.Decode(j, []byte(packetLog.Payload))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tpayload := buffer.NewViewFromBytes(j[:m])\n\n\t\tvv := buffer.NewVectorisedView(n+m, []buffer.View{header, payload})\n\t\treturn &vv, nil\n\t}\n\treturn nil, nil\n}", "func (r *Reader) ReadLink(path string) (string, error) {\n\treturn r.rootFS.readLink(path)\n}", "func (r *Dev) read(blockAddr byte) ([]byte, error) {\n\tdata, _, err := r.preAccess(blockAddr, commands.PICC_READ)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(data) != 16 {\n\t\treturn nil, wrapf(\"expected 16 bytes, actual %d\", len(data))\n\t}\n\treturn data, nil\n}", "func readOutPoint(r io.Reader, op *OutPoint) error {\n\t_, err := io.ReadFull(r, op.Hash[:])\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn serialization.ReadUint32(r, &op.Index)\n}", "func (l *Log) Read(off uint64) (*api.Record, error) {\n\tl.mu.Lock()\n\tdefer l.mu.Unlock()\n\tvar s *segment\n\tfor _, segment := range l.segments {\n\t\tif segment.baseOffset <= segment.nextOffset {\n\t\t\ts = segment\n\t\t\tbreak\n\t\t}\n\t}\n\tif s == nil || s.nextOffset <= off {\n\t\treturn nil, api.ErrOffsetOutOfRange{Offset: off}\n\t}\n\treturn s.Read(off)\n}", "func (r *RedirectNode) LeftFD() int { return r.rmap.lfd }", "func IsPointerFile(buf *[]byte) *models.LFSMetaObject {\n\tif !setting.LFS.StartServer {\n\t\treturn nil\n\t}\n\n\theadString := string(*buf)\n\tif !strings.HasPrefix(headString, models.LFSMetaFileIdentifier) {\n\t\treturn nil\n\t}\n\n\tsplitLines := strings.Split(headString, \"\\n\")\n\tif len(splitLines) < 3 {\n\t\treturn nil\n\t}\n\n\toid := strings.TrimPrefix(splitLines[1], models.LFSMetaFileOidPrefix)\n\tsize, err := strconv.ParseInt(strings.TrimPrefix(splitLines[2], \"size \"), 10, 64)\n\tif len(oid) != 64 || err != nil {\n\t\treturn nil\n\t}\n\n\tcontentStore := &ContentStore{ObjectStorage: storage.LFS}\n\tmeta := &models.LFSMetaObject{Oid: oid, Size: size}\n\texist, err := contentStore.Exists(meta)\n\tif err != nil || !exist {\n\t\treturn nil\n\t}\n\n\treturn meta\n}", "func (f *File) Readlink(ctx context.Context, req *fuse.ReadlinkRequest) (string, error) {\n\tino := f.info.Inode\n\tinfo, err := f.super.InodeGet(ino)\n\tif err != nil {\n\t\tlog.LogErrorf(\"Readlink: ino(%v) err(%v)\", ino, err)\n\t\treturn \"\", ParseError(err)\n\t}\n\tlog.LogDebugf(\"TRACE Readlink: ino(%v) target(%v)\", ino, string(info.Target))\n\treturn string(info.Target), nil\n}", "func (c closeWrapper) Read(p []byte) (n int, err error) {\n\treturn c.d.Read(p)\n}", "func (wk *WebmKeeper) handleRandomAccessPoint(el *edtd.Elem) error {\n\tswitch el.Name {\n\tcase \"Cluster\":\n\t\twk.lastCluster = wk.body.Len()\n\t\twk.resetTracking()\n\t\treturn nil\n\n\tcase \"SimpleBlock\", \"Block\":\n\n\t\ttrackNumber, keyframe, err := webm.AsKeyBlock(el)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif trackNumber > 2 || trackNumber < 1 {\n\t\t\treturn fmt.Errorf(\"impossible track number %d\", trackNumber)\n\t\t}\n\t\ttrackNumber--\n\n\t\twk.incTrackCount(trackNumber)\n\n\t\tif keyframe {\n\t\t\twk.trackBlockKeyframe[trackNumber] = true\n\t\t\tb := make([]byte, wk.body.Len() - wk.lastCluster)\n\t\t\tcopy(b, wk.body.Bytes()[wk.lastCluster:])\n\t\t\twk.body.Reset()\n\t\t\twk.body.Write(b)\n\t\t\twk.lastCluster = 0\n\t\t}\n\n\t\treturn nil\n\n\tdefault:\n\t\treturn nil\n\t}\n}", "func readlinkat(directory int, path string, buffer []byte) (int, error) {\n\t// Extract a raw pointer to the path bytes.\n\tvar pathBytes *byte\n\tpathBytes, err := unix.BytePtrFromString(path)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Extract a raw pointer to the buffer bytes.\n\tvar bufferBytes *byte\n\tif len(buffer) > 0 {\n\t\tbufferBytes = &buffer[0]\n\t}\n\n\t// Perform the system call.\n\tn, _, errnoErr := sysvicall6(uintptr(unsafe.Pointer(&procReadlinkat)), 4, uintptr(directory), uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(bufferBytes)), uintptr(len(buffer)), 0, 0)\n\tif errnoErr != 0 {\n\t\treturn 0, errnoErr\n\t}\n\n\t// Success.\n\treturn int(n), nil\n}", "func (p *pipe) readFrom(r io.Reader) (read int64, failure error) {\n\tfor {\n\t\t// Wait until some space frees up\n\t\tsafeFree, err := p.inputWait()\n\t\tif err != nil {\n\t\t\treturn read, err\n\t\t}\n\t\t// Try to fill the buffer either till the reader position, or the end\n\t\tlimit := p.inPos + safeFree\n\t\tif limit > p.size {\n\t\t\tlimit = p.size\n\t\t}\n\t\tnr, err := r.Read(p.buffer[p.inPos:limit])\n\t\tread += int64(nr)\n\n\t\t// Update the pipe input state and handle any occurred errors\n\t\tp.inputAdvance(nr)\n\t\tif err == io.EOF {\n\t\t\treturn read, nil\n\t\t}\n\t\tif err != nil {\n\t\t\treturn read, err\n\t\t}\n\t}\n}", "func (rh *readHandle) ReadAt(ctx context.Context, p []byte, off int64) (n int, err error) {\n\trh.g.add(ctx, Event{\n\t\tOp: ReadOp,\n\t\tFileNum: rh.fileNum,\n\t\tHandleID: rh.handleID,\n\t\tOffset: off,\n\t\tSize: int64(len(p)),\n\t})\n\treturn rh.rh.ReadAt(ctx, p, off)\n}", "func dataPtr(ptr uint64) uint64 {\n\toffset := uint64(1) // The first block is the tree head.\n\n\t// Every 8 blocks we have 1 first-level block containing the hashes of the\n\t// previous 8 data blocks. Then every 64 blocks, we have 1 second-level\n\t// block containing the hashes of the previous 8 first-level blocks. And so\n\t// on...\n\tn := uint64(8)\n\tfor level := uint64(0); level < 21; level++ {\n\t\toffset += ptr / n\n\t\tn = 8 * n\n\t}\n\n\treturn ptr + offset\n}", "func (dd *dictDecoder) availRead() int {\n\treturn dd.wrPos - dd.rdPos\n}", "func (np *NamePointer) DecodeRLP(s *rlp.Stream) (err error) {\n\tvar blob []byte\n\tnpRLP := &namePointerRLP{}\n\n\tif blob, err = s.Raw(); err != nil {\n\t\treturn\n\t}\n\tif err = rlp.DecodeBytes(blob, npRLP); err != nil {\n\t\treturn\n\t}\n\t_, ID, err := readIDTag(npRLP.Pointer)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tnp.Key = npRLP.Key\n\tnp.Pointer = ID\n\treturn err\n}", "func (c *Client) Read(path gfs.Path, offset gfs.Offset, data []byte) (n int, err error) {\n\tvar f gfs.GetFileInfoReply\n\terr = util.Call(c.master, \"Master.RPCGetFileInfo\", gfs.GetFileInfoArg{path}, &f)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\tif int64(offset/gfs.MaxChunkSize) > f.Chunks {\n\t\treturn -1, fmt.Errorf(\"read offset exceeds file size\")\n\t}\n\n\tpos := 0\n\tfor pos < len(data) {\n\t\tindex := gfs.ChunkIndex(offset / gfs.MaxChunkSize)\n\t\tchunkOffset := offset % gfs.MaxChunkSize\n\n\t\tif int64(index) >= f.Chunks {\n\t\t\terr = gfs.Error{gfs.ReadEOF, \"EOF over chunks\"}\n\t\t\tbreak\n\t\t}\n\n\t\tvar handle gfs.ChunkHandle\n\t\thandle, err = c.GetChunkHandle(path, index)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tvar n int\n\t\t//wait := time.NewTimer(gfs.ClientTryTimeout)\n\t\t//loop:\n\t\tfor {\n\t\t\t//select {\n\t\t\t//case <-wait.C:\n\t\t\t// err = gfs.Error{gfs.Timeout, \"Read Timeout\"}\n\t\t\t// break loop\n\t\t\t//default:\n\t\t\t//}\n\t\t\tn, err = c.ReadChunk(handle, chunkOffset, data[pos:])\n\t\t\tif err == nil || err.(gfs.Error).Code == gfs.ReadEOF {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tlog.Warning(\"Read \", handle, \" connection error, try again: \", err)\n\t\t}\n\n\t\toffset += gfs.Offset(n)\n\t\tpos += n\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif err != nil && err.(gfs.Error).Code == gfs.ReadEOF {\n\t\treturn pos, io.EOF\n\t} else {\n\t\treturn pos, err\n\t}\n}", "func (r *StreamReader) Read(p []byte) (int, error) {\n\tif !r.initiated {\n\t\tpanic(\"ReaderStream not created via NewReaderStream\")\n\t}\n\tvar ok bool\n\tr.stripEmpty()\n\tfor !r.closed && len(r.current) == 0 {\n\t\tif r.first {\n\t\t\tr.first = false\n\t\t} else {\n\t\t\tr.done <- true\n\t\t}\n\t\tif r.current, ok = <-r.reassembled; ok {\n\t\t\tr.stripEmpty()\n\t\t} else {\n\t\t\tr.closed = true\n\t\t}\n\t}\n\tif len(r.current) > 0 {\n\t\tcurrent := &r.current[0]\n\t\tif r.LossErrors && !r.lossReported && current.Skip != 0 {\n\t\t\tr.lossReported = true\n\t\t\treturn 0, DataLost\n\t\t}\n\t\tlength := copy(p, current.Bytes)\n\t\tcurrent.Bytes = current.Bytes[length:]\n\t\treturn length, nil\n\t}\n\treturn 0, io.EOF\n}", "func readBlock(reader io.ReadSeeker) (*Block, int64, error) {\n\t// Protect function with lock since it modifies reader state\n\tm.Lock()\n\tdefer m.Unlock()\n\n\toffset, err := reader.Seek(0, 1)\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\theaderData := make([]byte, HEADER_SIZE)\n\tn, err := reader.Read(headerData)\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\tif n != HEADER_SIZE {\n\t\treturn nil, offset, NotEnoughDataErr\n\t}\n\n\tblockSize := binary.LittleEndian.Uint32(headerData)\n\tblockBuffer := make([]byte, blockSize)\n\tn, err = reader.Read(blockBuffer)\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\tif uint32(n) != blockSize {\n\t\treturn nil, offset, NotEnoughDataErr\n\t}\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\tvar block = &Block{}\n\n\terr = json.Unmarshal(blockBuffer, block)\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\t// Set reader to begin of next block or EOF\n\t_, err = reader.Seek(DIGEST_SIZE, 1)\n\n\tif err != nil {\n\t\treturn nil, offset, err\n\t}\n\n\treturn block, offset, nil\n}", "func ReadLedgerMarkers(rows *sql.Rows, a *LedgerMarker) error {\n\treturn rows.Scan(&a.LMID, &a.LID, &a.BID, &a.RAID, &a.RID, &a.TCID, &a.Dt, &a.Balance, &a.State, &a.CreateTS, &a.CreateBy, &a.LastModTime, &a.LastModBy)\n}", "func (handle *Handle) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(handle.x)\n}", "func (wpr *Wrapper) ReadAt(p []byte, off int64) (n int, err error) {\n\twrap := NewWrapReader(wpr.rwa, wpr.O+off, wpr.N)\n\tr := io.LimitReader(wrap, wpr.L-off)\n\treturn r.Read(p)\n}", "func ReadFromReader(reader io.ReaderAt, offset uint32, blockSize int) ([]byte, uint32, error) {\n\tb, err := ReadFromReader64(reader, uint64(offset*PAD), blockSize)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tnextOffset := (offset + ((uint32(16+len(b)))+PAD-1)/PAD)\n\treturn b, uint32(nextOffset), nil\n}", "func ReadFromReader(reader io.ReaderAt, offset uint32, blockSize int) ([]byte, uint32, error) {\n\tb, err := ReadFromReader64(reader, uint64(offset*PAD), blockSize)\n\tif err != nil {\n\t\treturn nil, 0, err\n\t}\n\tnextOffset := (offset + ((uint32(16+len(b)))+PAD-1)/PAD)\n\treturn b, uint32(nextOffset), nil\n}", "func LoadPointer(addr *unsafe.Pointer) (val unsafe.Pointer)", "func TestMockReadSeekerSeeks(t *testing.T) {\n\tvar reader = NewMockReadSeeker(&[]byte{0x01, 0x02, 0x03, 0x04, 0x05, 0x06, 0x07})\n\tvar dataSize int64 = 7\n\n\tvar count = 0\n\n\t// Single byte buffer.\n\tvar byteBuffer []byte = []byte{0x00}\n\n\t/*\n\t * 0 - Relative to start of file \n\t */\n\n\t// Seek from the start of the file to the last byte.\n\tsook, err := reader.Seek(dataSize-1, 0)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != dataSize-1 {\n\t\tt.Fatal(\"Expected to return \", dataSize-1, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif count != 1 {\n\t\tt.Fatal(\"Expected to read 1 byte, got \", count)\n\t}\n\n\tif byteBuffer[0] != 0x07 {\n\t\tt.Fatal(\"Expected 0x07 got \", byteBuffer)\n\t}\n\n\t// Seek from the start of the file to the 3rd byte.\n\tsook, err = reader.Seek(2, 0)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 2 {\n\t\tt.Fatal(\"Expected to return \", 2, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x03 {\n\t\tt.Fatal(\"Expected 0x03 got \", byteBuffer)\n\t}\n\n\t// Seek from the start of the file to the first byte.\n\tsook, err = reader.Seek(0, 0)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 0 {\n\t\tt.Fatal(\"Expected to return \", 0, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x01 {\n\t\tt.Fatal(\"Expected 0x01 got \", byteBuffer)\n\t}\n\n\t/*\n\t * 1 - Relative to current position\n\t */\n\n\t// Seek from the current position to the same place.\n\n\t// Get in the middle\n\tsook, err = reader.Seek(4, 0)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 4 {\n\t\tt.Fatal(\"Expected to return \", 4, \" got \", sook)\n\t}\n\n\t// Seek same place relative to current.\n\tsook, err = reader.Seek(0, 1)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 4 {\n\t\tt.Fatal(\"Expected to return \", 4, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x05 {\n\t\tt.Fatal(\"Expected 0x05 got \", byteBuffer)\n\t}\n\n\t// Seek forward a byte\n\tsook, err = reader.Seek(1, 1)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 6 {\n\t\tt.Fatal(\"Expected to return \", 6, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x07 {\n\t\tt.Fatal(\"Expected 0x07 got \", byteBuffer)\n\t}\n\n\t/*\n\t * 2 - Relative to end of file\n\t */\n\n\t// Seek from the current position to the same place.\n\n\t// Get to the end.\n\t// Seek same place relative to end\n\tsook, err = reader.Seek(0, 2)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 7 {\n\t\tt.Fatal(\"Expected to return \", 7, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x07 {\n\t\tt.Fatal(\"Expected 0x07 got \", byteBuffer)\n\t}\n\n\t// Seek back a byte\n\tsook, err = reader.Seek(1, 2)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif sook != 6 {\n\t\tt.Fatal(\"Expected to return \", 6, \" got \", sook)\n\t}\n\n\tcount, err = reader.Read(byteBuffer)\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif byteBuffer[0] != 0x07 {\n\t\tt.Fatal(\"Expected 0x07 got \", byteBuffer)\n\t}\n}", "func Read(addr int64, data UintN) error {\n\treturn pathRead(memPath, addr, data)\n}", "func (f *realFile) Read(p []byte) (n int, err error) { return f.file.Read(p) }", "func readFull(r io.Reader, p []byte) (int, error) {\n\tcur := 0\n\tfor cur < len(p) {\n\t\tamt, err := r.Read(p[cur:])\n\t\tcur += amt\n\t\tif err != nil {\n\t\t\treturn cur, err\n\t\t}\n\t}\n\treturn cur, nil\n}", "func(this*Window)ReadAddr()(begin int,end int,err error){\nf,err:=this.File(\"addr\")\nif err!=nil{\nreturn\n}\nif _,err= f.Seek(0,0);err!=nil{\nreturn\n}\n_,err= fmt.Fscanf(f,\"%d %d\",&begin,&end)\nreturn\n}", "func Preadv2(t *kernel.Task, args arch.SyscallArguments) (uintptr, *kernel.SyscallControl, error) {\n\t// While the syscall is\n\t// preadv2(int fd, struct iovec* iov, int iov_cnt, off_t offset, int flags)\n\t// the linux internal call\n\t// (https://elixir.bootlin.com/linux/v4.18/source/fs/read_write.c#L1248)\n\t// splits the offset argument into a high/low value for compatibility with\n\t// 32-bit architectures. The flags argument is the 5th argument.\n\n\tfd := args[0].Int()\n\taddr := args[1].Pointer()\n\tiovcnt := int(args[2].Int())\n\toffset := args[3].Int64()\n\tflags := int(args[5].Int())\n\n\tfile := t.GetFile(fd)\n\tif file == nil {\n\t\treturn 0, nil, syserror.EBADF\n\t}\n\tdefer file.DecRef()\n\n\t// Check that the offset is legitimate.\n\tif offset < -1 {\n\t\treturn 0, nil, syserror.EINVAL\n\t}\n\n\t// Is reading at an offset supported?\n\tif offset > -1 && !file.Flags().Pread {\n\t\treturn 0, nil, syserror.ESPIPE\n\t}\n\n\t// Check that the file is readable.\n\tif !file.Flags().Read {\n\t\treturn 0, nil, syserror.EBADF\n\t}\n\n\t// Check flags field.\n\t// Note: gVisor does not implement the RWF_HIPRI feature, but the flag is\n\t// accepted as a valid flag argument for preadv2.\n\tif flags&^linux.RWF_VALID != 0 {\n\t\treturn 0, nil, syserror.EOPNOTSUPP\n\t}\n\n\t// Read the iovecs that specify the destination of the read.\n\tdst, err := t.IovecsIOSequence(addr, iovcnt, usermem.IOOpts{\n\t\tAddressSpaceActive: true,\n\t})\n\tif err != nil {\n\t\treturn 0, nil, err\n\t}\n\n\t// If preadv2 is called with an offset of -1, readv is called.\n\tif offset == -1 {\n\t\tn, err := readv(t, file, dst)\n\t\tt.IOUsage().AccountReadSyscall(n)\n\t\treturn uintptr(n), nil, handleIOError(t, n != 0, err, kernel.ERESTARTSYS, \"preadv2\", file)\n\t}\n\n\tn, err := preadv(t, file, dst, offset)\n\tt.IOUsage().AccountReadSyscall(n)\n\treturn uintptr(n), nil, handleIOError(t, n != 0, err, kernel.ERESTARTSYS, \"preadv2\", file)\n}", "func (i *index) Read(indexOff int64) (off uint32, indexPos uint64, err error) {\n\tif i.size == 0 {\n\t\treturn 0, 0, io.EOF\n\t}\n\n\trealIndexOff := uint32(indexOff)\n\tif indexOff == -1 {\n\t\trealIndexOff = uint32((i.size / entWidth) - 1)\n\t}\n\n\tindexPos = uint64(realIndexOff) * entWidth\n\tif i.size < indexPos+entWidth {\n\t\treturn 0, 0, io.EOF\n\t}\n\n\tdataOff := enc.Uint32(i.mmap[indexPos : indexPos+offWidth])\n\tdataPos := enc.Uint64(i.mmap[indexPos+offWidth : indexPos+entWidth])\n\treturn dataOff, dataPos, nil\n}", "func (m *intCodeMachine) resolveRead(pos int) int {\n\t// resolve against current op\n\toperand := m.ram[m.ip+pos]\n\tf := pFlag(m.op, pos)\n\tif f == 0 {\n\t\toperand = m.ram[operand]\n\t}\n\tif f == 2 {\n\t\toperand = m.ram[m.rb+operand]\n\t}\n\treturn operand\n}", "func (rwc *noPIReadWriteCloser) Read(p []byte) (n int, err error) {\n\tn, err = rwc.ReadWriteCloser.Read(rwc.rBuffer)\n\tif err == nil && n >= 4 {\n\t\tcopy(p, rwc.rBuffer[4:n])\n\t\tn -= 4\n\t}\n\treturn\n}", "func (in *InBuffer) ReadFloat64LE() float64 {\n\treturn math.Float64frombits(in.ReadUint64LE())\n}", "func (in *InBuffer) ReadFloat64LE() float64 {\n\treturn math.Float64frombits(in.ReadUint64LE())\n}", "func TestReadOneRecord(t *testing.T) {\n\tbuffer := bytes.NewBufferString(s19TestFile)\n\treader := Open(buffer)\n\tif reader != nil {\n\t\trec, err := reader.Next()\n\t\tif err != nil {\n\t\t\tt.Fatal(err)\n\t\t}\n\t\tif rec == nil {\n\t\t\tt.Fatalf(\"Returned record is nil\")\n\t\t}\n\t\tif rec.Address() != 0x400 {\n\t\t\tt.Errorf(\"Address mismatch expected: 0x400 got: %v\", rec.Address())\n\t\t}\n\t\tdata := rec.Data()\n\t\tif bytes.Compare(data, rec.Data()[:len(data)]) != 0 {\n\t\t\tt.Errorf(\"Data mismatch\")\n\t\t}\n\t} else {\n\t\tt.Fatal(\"Open call failed\")\n\t}\n}", "func (m *mockMatrix) Read(p []byte) (n int, err error) {\n\tswitch m.state {\n\tcase stateSwap:\n\t\tif len(p) == 0 {\n\t\t\treturn 0, fmt.Errorf(\"insufficient capacity\")\n\t\t} else {\n\t\t\tp[0] = 'F'\n\t\t\tm.count = 0\n\t\t\tm.state = stateCmd\n\t\t\treturn 1, nil\n\t\t}\n\tcase stateClose:\n\t\treturn 0, fmt.Errorf(\"stream is closed\")\n\tdefault:\n\t\treturn 0, nil\n\t}\n}", "func (decoder *Decoder) ReadRef() (interface{}, error){\n code, err := decoder.read()\n if err != nil {\n return nil, err\n }\n if code != 0x51 {\n return nil, errors.New(\"readRef: unexpected code\")\n }\n refId, err := decoder.ReadInt()\n if err != nil {\n return nil, err\n }\n ret, ok := decoder.refMap[refId]\n if !ok {\n return nil, errors.New(\"readRef: unexpected ref\")\n }\n decoder.success()\n return ret, nil\n}", "func (decoder *EbpfDecoder) ReadAmountBytes() int {\n\treturn decoder.cursor\n}", "func (s *HandleTable) FindPointer(key *Slice, hash uint32) **LRUHandle {\n var ptr **LRUHandle = &s.list_[hash & (s.length_ - 1)]\n for (*ptr != nil) && ((*ptr).hash != hash || key.NotEqual((*ptr).key())) {\n ptr = &(*ptr).next_hash\n }\n return ptr\n}", "func (r *Reader) SkipLong() {\n\tvar offset int8\n\tfor r.Error == nil {\n\t\tif offset == maxLongBufSize {\n\t\t\treturn\n\t\t}\n\n\t\tb := r.readByte()\n\t\tif b&0x80 == 0 {\n\t\t\tbreak\n\t\t}\n\t\toffset++\n\t}\n}" ]
[ "0.7145606", "0.6092693", "0.57624817", "0.5569016", "0.5522428", "0.5497365", "0.5497365", "0.5490105", "0.5487384", "0.54738414", "0.54729754", "0.5465387", "0.54471433", "0.5444666", "0.538409", "0.53686434", "0.5305466", "0.5305383", "0.5300171", "0.5293714", "0.52786624", "0.52743304", "0.5245505", "0.52371264", "0.52012664", "0.5133126", "0.51298827", "0.51152635", "0.5114761", "0.50886345", "0.50689787", "0.5057347", "0.5045742", "0.5019539", "0.5012655", "0.5003613", "0.50023264", "0.50010866", "0.4999248", "0.4989547", "0.49871212", "0.49863625", "0.4982429", "0.49785984", "0.4977925", "0.49757037", "0.49742833", "0.49641818", "0.49419403", "0.49414197", "0.4940444", "0.49327794", "0.4926058", "0.49235117", "0.4921109", "0.49209502", "0.49184632", "0.49064928", "0.4905257", "0.48975036", "0.48932964", "0.48895872", "0.48885912", "0.48856503", "0.4879053", "0.48740923", "0.48658162", "0.48601845", "0.48601723", "0.48564047", "0.4854462", "0.48318166", "0.48300856", "0.48287252", "0.48212248", "0.4809144", "0.480776", "0.48038155", "0.48030016", "0.479821", "0.47905302", "0.47905302", "0.4785425", "0.47824183", "0.47745633", "0.4772391", "0.47660148", "0.4762411", "0.47622466", "0.4762115", "0.47584736", "0.47476435", "0.47465792", "0.47465792", "0.4739946", "0.47392553", "0.47374746", "0.47358537", "0.47292605", "0.4728039" ]
0.73345095
0
ReadPointerFromBuffer will return a pointer if the provided byte slice is a pointer file or an error otherwise.
func ReadPointerFromBuffer(buf []byte) (Pointer, error) { var p Pointer headString := string(buf) if !strings.HasPrefix(headString, MetaFileIdentifier) { return p, ErrMissingPrefix } splitLines := strings.Split(headString, "\n") if len(splitLines) < 3 { return p, ErrInvalidStructure } oid := strings.TrimPrefix(splitLines[1], MetaFileOidPrefix) if len(oid) != 64 || !oidPattern.MatchString(oid) { return p, ErrInvalidOIDFormat } size, err := strconv.ParseInt(strings.TrimPrefix(splitLines[2], "size "), 10, 64) if err != nil { return p, err } p.Oid = oid p.Size = size return p, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func ReadPointer(reader io.Reader) (Pointer, error) {\n\tbuf := make([]byte, blobSizeCutoff)\n\tn, err := io.ReadFull(reader, buf)\n\tif err != nil && err != io.ErrUnexpectedEOF {\n\t\treturn Pointer{}, err\n\t}\n\tbuf = buf[:n]\n\n\treturn ReadPointerFromBuffer(buf)\n}", "func NewPtrBuffer(aSlice interface{}) *Buffer {\n aSliceValue := sliceValue(aSlice, true)\n return &Buffer{\n buffer: aSliceValue,\n handler: overwriteNilPtrHandler{\n creater: newCreaterFunc(nil, aSliceValue.Type())}}\n}", "func NewKeyPointerFromBuffer(buffer []byte) KeyPointer {\n\tkey := [16]byte{}\n\tcopy(key[:], buffer[0:16])\n\tx := KeyPointer{\n\t\tKey: Key(key),\n\t\tBlockAddress: sliceToUint64(buffer[16:24]),\n\t}\n\treturn x\n}", "func ReadPointerFile(reader io.Reader) (*models.LFSMetaObject, *[]byte) {\n\tif !setting.LFS.StartServer {\n\t\treturn nil, nil\n\t}\n\n\tbuf := make([]byte, 1024)\n\tn, _ := reader.Read(buf)\n\tbuf = buf[:n]\n\n\tif isTextFile := base.IsTextFile(buf); !isTextFile {\n\t\treturn nil, nil\n\t}\n\n\treturn IsPointerFile(&buf), &buf\n}", "func ReadBuffer(src uint32) {\n\tsyscall.Syscall(gpReadBuffer, 1, uintptr(src), 0, 0)\n}", "func BytesPointer(data []byte) unsafe.Pointer {\n\tif len(data) == 0 {\n\t\treturn nil\n\t}\n\treturn unsafe.Pointer(&data[0])\n}", "func GetBufferPointerv(target uint32, pname uint32, params *unsafe.Pointer) {\n\tsyscall.Syscall(gpGetBufferPointerv, 3, uintptr(target), uintptr(pname), uintptr(unsafe.Pointer(params)))\n}", "func LoadPointer(addr *unsafe.Pointer) (val unsafe.Pointer)", "func getPtr(b []byte) uintptr {\n\n\tswitch ptrOffset {\n\tcase 8:\n\t\treturn uintptr(b[7]) | uintptr(b[6])<<8 | uintptr(b[5])<<16 | uintptr(b[4])<<24 | uintptr(b[3])<<32 | uintptr(b[2])<<40 | uintptr(b[1])<<48 | uintptr(b[0])<<56\n\tcase 4:\n\t\treturn uintptr(b[3]) | uintptr(b[2])<<8 | uintptr(b[1])<<16 | uintptr(b[0])<<24\n\n\t}\n\tpanic(\"Invalid size detected\")\n}", "func GetNamedBufferPointerv(buffer uint32, pname uint32, params *unsafe.Pointer) {\n\tsyscall.Syscall(gpGetNamedBufferPointerv, 3, uintptr(buffer), uintptr(pname), uintptr(unsafe.Pointer(params)))\n}", "func (sb *SeekableBuffer) Read(p []byte) (n int, err error) {\n\tdefer func() {\n\t\tif state := recover(); state != nil {\n\t\t\terr = state.(error)\n\t\t}\n\t}()\n\n\tif sb.position >= len64(sb.data) {\n\t\treturn 0, io.EOF\n\t}\n\n\tn = copy(p, sb.data[sb.position:])\n\tsb.position += int64(n)\n\n\treturn n, nil\n\n}", "func (c *poolConn) ReadUnsafeBuffer() error {\n\n\t//judge whether the buffer can be moved\n\tif c.buffer.index < c.buffer.size/2 {\n\t\treturn errors.New(\"The bytes need to move is too long!\")\n\t}\n\n\tj := 0\n\tfor i := c.buffer.index; i < c.buffer.size; i, j = i+1, j+1 {\n\t\tc.buffer.realBuffer[j] = c.buffer.realBuffer[i]\n\t}\n\tc.buffer.index = 0\n\treturn c.ReadTcpBlockLink(j)\n}", "func (f *FixedBuffer) ReadFrom() (int, error) {\n\treturn f.r.Read(f.buf)\n}", "func getMemorySliceFromPointer(p unsafe.Pointer, length int) []byte {\n\treturn *(*[]byte)(unsafe.Pointer(&sliceHeader{\n\t\tData: p,\n\t\tLen: length,\n\t\tCap: length,\n\t}))\n}", "func (f unixFileHandle) ReadAt(b []byte, offset int64) (n int, err error) {\n\tn, err = syscall.Pread(syscallFd(f), b, offset)\n\terr = handleSyscallError(err)\n\tif n == 0 && len(b) > 0 && err == nil {\n\t\terr = io.EOF\n\t}\n\treturn\n}", "func GetDirectBufferAddress(env *C.JNIEnv, buf C.jobject) unsafe.Pointer {\n\treturn C._GoJniGetDirectBufferAddress(env, buf)\n}", "func NewReadSeekCloserBuffer(rd ReadSeekCloser, buffers [][]byte) (res ReadSeekCloser, err error) {\n\treader, err := NewReadCloserBuffer(rd, buffers)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//Not checking for result as the input interface guarantees it's seekable\n\tres, _ = reader.(ReadSeekCloser)\n\treturn\n}", "func Read(fd int32, p unsafe.Pointer, n int32) int32", "func IsPointerFile(buf *[]byte) *models.LFSMetaObject {\n\tif !setting.LFS.StartServer {\n\t\treturn nil\n\t}\n\n\theadString := string(*buf)\n\tif !strings.HasPrefix(headString, models.LFSMetaFileIdentifier) {\n\t\treturn nil\n\t}\n\n\tsplitLines := strings.Split(headString, \"\\n\")\n\tif len(splitLines) < 3 {\n\t\treturn nil\n\t}\n\n\toid := strings.TrimPrefix(splitLines[1], models.LFSMetaFileOidPrefix)\n\tsize, err := strconv.ParseInt(strings.TrimPrefix(splitLines[2], \"size \"), 10, 64)\n\tif len(oid) != 64 || err != nil {\n\t\treturn nil\n\t}\n\n\tcontentStore := &ContentStore{ObjectStorage: storage.LFS}\n\tmeta := &models.LFSMetaObject{Oid: oid, Size: size}\n\texist, err := contentStore.Exists(meta)\n\tif err != nil || !exist {\n\t\treturn nil\n\t}\n\n\treturn meta\n}", "func GetBufferPointerv(target uint32, pname uint32, params *unsafe.Pointer) {\n\tC.glowGetBufferPointerv(gpGetBufferPointerv, (C.GLenum)(target), (C.GLenum)(pname), params)\n}", "func GetBufferPointerv(target uint32, pname uint32, params *unsafe.Pointer) {\n\tC.glowGetBufferPointerv(gpGetBufferPointerv, (C.GLenum)(target), (C.GLenum)(pname), params)\n}", "func FileReadAt(f *os.File, b []byte, off int64) (int, error)", "func GetBufferPointerv(target uint32, pname uint32, params *unsafe.Pointer) {\n C.glowGetBufferPointerv(gpGetBufferPointerv, (C.GLenum)(target), (C.GLenum)(pname), params)\n}", "func FromBytePtr(buffer [][]*byte, startRow, startCol, rows, cols int) ([][]*byte, error) {\n\tview := make([][]*byte, rows)\n\tif len(buffer) < startRow+rows {\n\t\treturn nil, fmt.Errorf(\"matrix has less rows than asked: %d x %d\", len(buffer), startRow+rows)\n\t}\n\tfor i := range view {\n\t\tif len(buffer[startRow+i]) < startCol+cols {\n\t\t\treturn nil, fmt.Errorf(\"row %d has less cols than asked: %d x %d\", i, len(buffer[startRow+i]), startCol+cols)\n\t\t}\n\t\tview[i] = buffer[startRow+i][startCol : startCol+cols : startCol+cols]\n\t}\n\treturn view, nil\n}", "func getPalettesBufferPointer() *uint8 {\n\treturn &palettes[0]\n}", "func (bb *BytesBuffer) Read(p []byte) (n int, err error) {\n\treturn bb.reader.Read(p)\n}", "func (m *Mmap) GetPointer() int64 {\n\treturn m.FilePointer\n}", "func BytePointer(b []byte) *uint8 {\n\tif len(b) > 0 {\n\t\treturn &b[0]\n\t} else {\n\t\treturn nil\n\t}\n}", "func ReadFrom(reader io.Reader, buffer *alloc.Buffer) (*alloc.Buffer, error) {\n\tif buffer == nil {\n\t\tbuffer = alloc.NewBuffer()\n\t}\n\tnBytes, err := reader.Read(buffer.Value)\n\tbuffer.Slice(0, nBytes)\n\treturn buffer, err\n}", "func MapBuffer(target uint32, access uint32) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpMapBuffer, 2, uintptr(target), uintptr(access), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func readlinkat(directory int, path string, buffer []byte) (int, error) {\n\t// Extract a raw pointer to the path bytes.\n\tvar pathBytes *byte\n\tpathBytes, err := unix.BytePtrFromString(path)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Extract a raw pointer to the buffer bytes.\n\tvar bufferBytes *byte\n\tif len(buffer) > 0 {\n\t\tbufferBytes = &buffer[0]\n\t}\n\n\t// Perform the system call.\n\tn, _, errnoErr := sysvicall6(uintptr(unsafe.Pointer(&procReadlinkat)), 4, uintptr(directory), uintptr(unsafe.Pointer(pathBytes)), uintptr(unsafe.Pointer(bufferBytes)), uintptr(len(buffer)), 0, 0)\n\tif errnoErr != 0 {\n\t\treturn 0, errnoErr\n\t}\n\n\t// Success.\n\treturn int(n), nil\n}", "func NewSeekableBuffer() *SeekableBuffer {\n\tdata := make([]byte, 0)\n\n\treturn &SeekableBuffer{\n\t\tdata: data,\n\t}\n}", "func (b *Buffer) Read(p []byte) (n int, err error) {\n\tbuf := b.Bytes()\n\tif len(buf) == 0 {\n\t\tif len(p) == 0 {\n\t\t\treturn 0, nil\n\t\t}\n\t\treturn 0, io.EOF\n\t}\n\n\tn = copy(p, buf)\n\treturn n, nil\n}", "func (s *SeekerWrapper) ReadAt(p []byte, offset int64) (int, error) { return s.s.ReadAt(p, offset) }", "func ReadBytes(buffer []byte, offset int, size int) []byte {\n return buffer[offset:offset + size]\n}", "func (this *Data) Ptr() uintptr {\n\treturn this.buf\n}", "func GetNamedBufferPointerv(buffer uint32, pname uint32, params *unsafe.Pointer) {\n\tC.glowGetNamedBufferPointerv(gpGetNamedBufferPointerv, (C.GLuint)(buffer), (C.GLenum)(pname), params)\n}", "func GetNamedBufferPointerv(buffer uint32, pname uint32, params *unsafe.Pointer) {\n\tC.glowGetNamedBufferPointerv(gpGetNamedBufferPointerv, (C.GLuint)(buffer), (C.GLenum)(pname), params)\n}", "func NewReaderBuffer(buf []byte) ByteBuffer {\n\treturn newReaderByteBuffer(buf)\n}", "func (s *safeBuffer) Read(p []byte) (int, error) {\n\ts.Lock()\n\tdefer s.Unlock()\n\treturn s.buf.Read(p)\n}", "func (obj *BlobManager) GetBlobItemFromPointer(ctx context.Context, parent string, name string) (*BlobItem, *minipointer.Pointer, error) {\n\tpointerObj, pointerErr := obj.pointerMgr.GetPointer(ctx, obj.MakeBlobId(parent, name), minipointer.TypePointer)\n\tif pointerErr != nil {\n\t\tif obj.pointerMgr.IsMemcachedOnly() == false {\n\t\t\treturn nil, nil, pointerErr\n\t\t} else {\n\t\t\to, e := obj.GetBlobItemFromQuery(ctx, parent, name)\n\t\t\tvar p *minipointer.Pointer = nil\n\t\t\tif e == nil {\n\t\t\t\tp, _ = obj.SavePointer(ctx, o)\n\t\t\t}\n\t\t\treturn o, p, e\n\t\t}\n\t}\n\tretObj, retErr := obj.GetBlobItem(ctx, parent, name, pointerObj.GetSign())\n\treturn retObj, pointerObj, retErr\n}", "func NewPtrPageBuffer(aSlice interface{}, desiredPageNo int) *PageBuffer {\n aSliceValue := sliceValue(aSlice, true)\n return newPageBuffer(\n aSliceValue,\n desiredPageNo,\n overwriteNilPtrHandler{\n creater: newCreaterFunc(nil, aSliceValue.Type())})\n}", "func getBufferFromInputResource(inputPath string) ([]byte, error) {\n\n\turl, err := url.Parse(inputPath)\n\tif err != nil {\n\t\treturn nil, errors.New(\"Error parsing input file\")\n\t}\n\tif url.Scheme == \"http\" || url.Scheme == \"https\" {\n\t\tres, err := http.Get(inputPath)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdefer res.Body.Close()\n\t\treturn ioutil.ReadAll(res.Body)\n\t} else if url.Scheme == \"ftp\" {\n\t\treturn nil, errors.New(\"ftp not supported yet\")\n\n\t} else {\n\t\treturn ioutil.ReadFile(inputPath)\n\t}\n}", "func (s *stream) Ptr() unsafe.Pointer {\n\treturn (unsafe.Pointer)(s)\n}", "func BytesToPointer(b []byte) *uint8 {\n\treturn &b[0]\n}", "func GeneratePointer(content io.Reader) (Pointer, error) {\n\th := sha256.New()\n\tc, err := io.Copy(h, content)\n\tif err != nil {\n\t\treturn Pointer{}, err\n\t}\n\tsum := h.Sum(nil)\n\treturn Pointer{Oid: hex.EncodeToString(sum), Size: c}, nil\n}", "func (b *Buffer) PeekAsByteSlice(length int) []byte {\n\treturn b.buf[b.readerIndex:b.readerIndex+length]\n}", "func (k Querier) CheckpointBuffer(c context.Context, req *types.QueryCheckpointBufferRequest) (*types.QueryCheckpointBufferResponse, error) {\n\tif req == nil {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"empty request\")\n\t}\n\n\tctx := sdk.UnwrapSDKContext(c)\n\tres, err := k.GetCheckpointFromBuffer(ctx)\n\tif err != nil {\n\t\treturn nil, types.ErrNoCheckpointBufferFound\n\t}\n\n\tif res == nil {\n\t\treturn nil, types.ErrNoCheckpointBufferFound\n\t}\n\n\treturn &types.QueryCheckpointBufferResponse{CheckpointBuffer: res}, nil\n}", "func (c *poolConn) ReadBuffer(size int) ([]byte, error) {\n\tif c.mustRead == true {\n\t\terr := c.ReadTcpBlock()\n\t\tif err != nil {\n\t\t\tc.err = err\n\t\t\treturn nil, err\n\t\t}\n\t\tc.buffer.index = 0\n\t\tc.mustRead = false\n\t}\n\n\t//if size < c.buffer.size-c.buffer.index, normal stitching\n\t//if c.buffer.size-c.buffer.index < size < c.buffer.capacity-c.buffer.size+c.buffer.index, move usable data in buffer to front\n\t//if size > c.buffer.capacity, directly read the specified size\n\tif size+2 <= c.buffer.size-c.buffer.index {\n\n\t\tif c.buffer.realBuffer[c.buffer.index+size] == '\\r' && c.buffer.realBuffer[c.buffer.index+size+1] == '\\n' {\n\t\t\tcpy_index := c.buffer.index\n\t\t\tc.buffer.index = c.buffer.index + size + 2\n\t\t\tif c.buffer.index >= c.buffer.size {\n\t\t\t\tc.mustRead = true\n\t\t\t}\n\t\t\treturn c.buffer.realBuffer[cpy_index: cpy_index+size], nil\n\t\t} else {\n\t\t\treturn nil, errors.New(\"ReadBuffer is read wrong!\")\n\t\t}\n\t} else if size+2 <= c.buffer.capacity-c.buffer.size+c.buffer.index {\n\t\tc.ReadUnsafeBuffer()\n\t\tif c.buffer.realBuffer[c.buffer.index+size] == '\\r' && c.buffer.realBuffer[c.buffer.index+size+1] == '\\n' {\n\t\t\tc.buffer.index = c.buffer.index + size + 2\n\t\t\tif c.buffer.index >= c.buffer.size {\n\t\t\t\tc.mustRead = true\n\t\t\t}\n\t\t\treturn c.buffer.realBuffer[0:size], nil\n\t\t} else {\n\t\t\treturn nil, errors.New(\"ReadBuffer is read wrong!\")\n\t\t}\n\n\t} else {\n\t\tvar err error\n\t\tbigBuffer := make([]byte, size+2)\n\t\tcopy(bigBuffer, c.buffer.realBuffer[c.buffer.index:])\n\n\t\t//Make the results right , when the BigSize < buffer.capacity\n\t\tif len(bigBuffer) > c.buffer.size-c.buffer.index {\n\t\t\tbigBuffer, err = c.ReadTcpBigBlockLink(bigBuffer)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\n\t\t//judge weather the bigBuffer is right\n\t\tif bigBuffer[size] == '\\r' && bigBuffer[size+1] == '\\n' {\n\t\t\tc.buffer.index = c.buffer.index + size + 2\n\t\t\tif c.buffer.index >= c.buffer.size {\n\t\t\t\tc.mustRead = true\n\t\t\t}\n\t\t\treturn bigBuffer[:size], nil\n\t\t} else {\n\t\t\treturn nil, errors.New(\"bigBuffer is read wrong!\")\n\t\t}\n\t}\n}", "func ReadBuffer(src uint32) {\n\tC.glowReadBuffer(gpReadBuffer, (C.GLenum)(src))\n}", "func ReadBuffer(src uint32) {\n\tC.glowReadBuffer(gpReadBuffer, (C.GLenum)(src))\n}", "func IsBuffer(buffer uint32) bool {\n\tret, _, _ := syscall.Syscall(gpIsBuffer, 1, uintptr(buffer), 0, 0)\n\treturn ret != 0\n}", "func (b *buffer) read(loc Location) unsafe.Pointer {\n\treturn atomic.LoadPointer(b.index(loc.index()))\n}", "func (g *GLTF) loadBuffer(bufIdx int) ([]byte, error) {\n\n\t// Check if provided buffer index is valid\n\tif bufIdx < 0 || bufIdx >= len(g.Buffers) {\n\t\treturn nil, fmt.Errorf(\"invalid buffer index\")\n\t}\n\tbufData := &g.Buffers[bufIdx]\n\t// Return cached if available\n\tif bufData.cache != nil {\n\t\tlog.Debug(\"Fetching Buffer %d (cached)\", bufIdx)\n\t\treturn bufData.cache, nil\n\t}\n\tlog.Debug(\"Loading Buffer %d\", bufIdx)\n\n\t// If buffer URI use the chunk data field\n\tif bufData.Uri == \"\" {\n\t\treturn g.data, nil\n\t}\n\n\t// Checks if buffer URI is a data URI\n\tvar data []byte\n\tvar err error\n\tif isDataURL(bufData.Uri) {\n\t\tdata, err = loadDataURL(bufData.Uri)\n\t} else {\n\t\t// Try to load buffer from file\n\t\tdata, err = g.loadFileBytes(bufData.Uri)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Checks data length\n\tif len(data) != bufData.ByteLength {\n\t\treturn nil, fmt.Errorf(\"buffer:%d read data length:%d expected:%d\", bufIdx, len(data), bufData.ByteLength)\n\t}\n\t// Cache buffer data\n\tg.Buffers[bufIdx].cache = data\n\tlog.Debug(\"cache data:%v\", len(bufData.cache))\n\treturn data, nil\n}", "func ptrFromHandle_Slice_byte(h CGoHandle) *[]byte {\n\tp := gopyh.VarFromHandle((gopyh.CGoHandle)(h), \"[]byte\")\n\tif p == nil {\n\t\treturn nil\n\t}\n\treturn p.(*[]byte)\n}", "func (image *Image2D) GetDataPointer() unsafe.Pointer {\n\treturn gl.Ptr(image.data)\n}", "func getWasmMemoryBufferPointer() *uint8 {\n\treturn &buffer[0]\n}", "func (mmap *File) GetMemoryPtr() unsafe.Pointer {\n\treturn mmap.mmap\n}", "func (p *Process) read(addr uintptr, ptr interface{}) error {\n\t// Reflection magic!\n\tv := reflect.ValueOf(ptr)\n\tdataAddr := getDataAddr(v)\n\tdataSize := getDataSize(v)\n\n\t// Open the file mapped process memory.\n\tmem, err := os.Open(fmt.Sprintf(\"/proc/%d/mem\", p.PID))\n\tdefer mem.Close()\n\tif err != nil {\n\t\treturn errors.New(fmt.Sprintf(\"Error opening /proc/%d/mem. Are you root?\", p.PID))\n\t}\n\n\t// Create a buffer and read data into it.\n\tdataBuf := make([]byte, dataSize)\n\tn, err := mem.ReadAt(dataBuf, int64(addr))\n\tif n != int(dataSize) {\n\t\treturn errors.New(fmt.Sprintf(\"Tried to read %d bytes, actually read %d bytes\\n\", dataSize, n))\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\t// Unsafely cast to []byte to copy data into.\n\tbuf := (*[]byte)(unsafe.Pointer(&reflect.SliceHeader{\n\t\tData: dataAddr,\n\t\tLen: int(dataSize),\n\t\tCap: int(dataSize),\n\t}))\n\tcopy(*buf, dataBuf)\n\treturn nil\n}", "func (ite *ifdTagEnumerator) ReadBuffer(n int) (buf []byte, err error) {\n\tif n > len(ite.exifReader.rawBuffer) {\n\t\treturn nil, ErrDataLength\n\t}\n\t// Read from underlying exifReader io.ReaderAt interface\n\tn, err = ite.exifReader.ReadAt(ite.exifReader.rawBuffer[:n], int64(ite.offset+ite.ifdOffset))\n\n\tite.offset += uint32(n) // Update reader offset\n\n\treturn ite.exifReader.rawBuffer[:n], err\n}", "func (dw downloadBuffer) ReadFrom(r io.Reader) (int64, error) {\n\tvar n int64\n\tfor len(dw.buf) > 0 {\n\t\tread, err := io.ReadFull(r, dw.buf[0])\n\n\t\tif err == io.ErrUnexpectedEOF || err == io.EOF {\n\t\t\tn += int64(read)\n\t\t\treturn n, nil\n\t\t}\n\n\t\tif err != nil {\n\t\t\treturn n, err\n\t\t}\n\n\t\tdw.buf = dw.buf[1:]\n\t\tn += int64(read)\n\t}\n\treturn n, nil\n}", "func newBuffer(r io.Reader, offset int64) *buffer {\n\treturn &buffer{\n\t\tr: r,\n\t\toffset: offset,\n\t\tbuf: make([]byte, 0, 4096),\n\t\tallowObjptr: true,\n\t\tallowStream: true,\n\t}\n}", "func ChunkFromBuffer(buf arrow.TableBuffer) Chunk {\n\treturn Chunk{buf: buf}\n}", "func LoadThumbnailFromBuffer(buf []byte, width, height int, crop Interesting, size Size, params *ImportParams) (*ImageRef, error) {\n\tstartupIfNeeded()\n\n\tvipsImage, format, err := vipsThumbnailFromBuffer(buf, width, height, crop, size, params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tref := newImageRef(vipsImage, format, format, buf)\n\n\tgovipsLog(\"govips\", LogLevelDebug, fmt.Sprintf(\"created imageref %p\", ref))\n\treturn ref, nil\n}", "func ReadByte(buffer []byte, offset int) byte {\n return buffer[offset]\n}", "func (r *SnpReportRespABI) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(r)\n}", "func BytesSlicePtr(src [][]byte) []*[]byte {\n\tdst := make([]*[]byte, len(src))\n\tfor i := 0; i < len(src); i++ {\n\t\tdst[i] = &(src[i])\n\t}\n\treturn dst\n}", "func (b *SafeBuffer) Read(p []byte) (n int, err error) {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.Read(p)\n}", "func (cb *Buffer) Read(buf []byte) (int, error) {\n\tif buf == nil || len(buf) == 0 {\n\t\treturn 0, fmt.Errorf(\"Target buffer is null or empty\")\n\t}\n\n\ttoRead := min(len(buf), cb.ReadAvailability())\n\n\tlBytes := min(cb.rpos, toRead)\n\tcopy(buf[toRead-lBytes:toRead], cb.buffer[cb.rpos-lBytes:cb.rpos])\n\n\tif toRead > lBytes {\n\t\trBytes := toRead - lBytes\n\t\tcopy(buf[:rBytes], cb.buffer[len(cb.buffer)-rBytes:len(cb.buffer)])\n\t\tcb.rpos = len(cb.buffer) - rBytes\n\t} else {\n\t\tcb.rpos -= lBytes\n\t}\n\n\tcb.full = false\n\treturn toRead, nil\n}", "func NewBuffer(length int) *Buffer {\n\treturn &Buffer{\n\t\titems: make([]unsafe.Pointer, length),\n\t}\n}", "func (b *Buffer) Read(p []byte) (n int, err error) {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.Read(p)\n}", "func (s PointBuffer) Get(idx int) PointPtr {\n\tinBounds := idx >= 0 && idx < int(s.len)\n\tif !inBounds {\n\t\tpanic(fmt.Errorf(\n\t\t\t\"runtime error: index out of range [%d] with length %d\",\n\t\t\tidx, s.len,\n\t\t))\n\t}\n\tvar tVar Point\n\ttSize := unsafe.Sizeof(tVar)\n\ttype internalPtr struct {\n\t\toffset uintptr\n\t\tbucketIdx uint8\n\t\tarenaMask uint16\n\t}\n\tcurrentPtr := *(*internalPtr)(unsafe.Pointer(&s.data))\n\tnewPtr := internalPtr{\n\t\toffset: currentPtr.offset + uintptr(idx*int(tSize)),\n\t\tbucketIdx: currentPtr.bucketIdx,\n\t\tarenaMask: currentPtr.arenaMask,\n\t}\n\treturn PointPtr{\n\t\tptr: *(*arena.Ptr)(unsafe.Pointer(&newPtr)),\n\t}\n}", "func NewSeekableBufferWithBytes(originalData []byte) *SeekableBuffer {\n\tdata := make([]byte, len(originalData))\n\tcopy(data, originalData)\n\n\treturn &SeekableBuffer{\n\t\tdata: data,\n\t}\n}", "func (tf *Temp) ReadAt(buffer []byte, off int64) (int, error) {\n\ttf.Lock()\n\tdefer tf.Unlock()\n\n\tread, err := tf.file.ReadAt(buffer, off)\n\treturn read, ex.New(err)\n}", "func (r *SnpReportReqABI) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(r)\n}", "func (rh *readHandle) ReadAt(ctx context.Context, p []byte, off int64) (n int, err error) {\n\trh.g.add(ctx, Event{\n\t\tOp: ReadOp,\n\t\tFileNum: rh.fileNum,\n\t\tHandleID: rh.handleID,\n\t\tOffset: off,\n\t\tSize: int64(len(p)),\n\t})\n\treturn rh.rh.ReadAt(ctx, p, off)\n}", "func (jbobject *JavaNioCharBuffer) Read(a JavaNioCharBufferInterface) (int, error) {\n\tconv_a := javabind.NewGoToJavaCallable()\n\tif err := conv_a.Convert(a); err != nil {\n\t\tpanic(err)\n\t}\n\tjret, err := jbobject.CallMethod(javabind.GetEnv(), \"read\", javabind.Int, conv_a.Value().Cast(\"java/nio/CharBuffer\"))\n\tif err != nil {\n\t\tvar zero int\n\t\treturn zero, err\n\t}\n\tconv_a.CleanUp()\n\treturn jret.(int), nil\n}", "func (pstFile *File) Read(outputBufferSize int, offset int) ([]byte, error) {\n\tinputFile, err := os.Open(pstFile.Filepath)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t_, err = inputFile.Seek(int64(offset), 0)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinputReader := bufio.NewReader(inputFile)\n\n\toutputBuffer := make([]byte, outputBufferSize)\n\n\t_, err = inputReader.Read(outputBuffer)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = inputFile.Close()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn outputBuffer, nil\n}", "func (c *PacketConn) ReadFrom(p []byte) (int, net.Addr, error) {\n\treturn c.buffer.ReadFrom(p)\n}", "func NewLocalBuffer(b bytes.Buffer) *LocalBuffer { return &LocalBuffer{b: b} }", "func PointerToSlice(ptr interface{}) interface{} {\n\tptrVal := reflect.ValueOf(ptr)\n\tptrTy := reflect.TypeOf(ptr)\n\tif ptrTy.Kind() != reflect.Ptr {\n\t\tpanic(\"given value is not a pointer\")\n\t}\n\n\telemTy := ptrTy.Elem()\n\taddr := ptrVal.Elem().UnsafeAddr()\n\thdr := reflect.SliceHeader{\n\t\tData: addr,\n\t\tLen: 1,\n\t\tCap: 1,\n\t}\n\tslicePtrVal := reflect.NewAt(reflect.SliceOf(elemTy), unsafe.Pointer(&hdr))\n\n\t// Dummy extra reference to ptrVal after we've built slicePtrVal\n\t// to ensure that our original value doesn't get collected before\n\t// we get a chance to return.\n\tif ptrVal.Interface() == nil {\n\t\t// should never happen.\n\t\tpanic(\"ptrVal.Interface() returned nil\")\n\t}\n\n\treturn slicePtrVal.Elem().Interface()\n}", "func (b *Buffer) ReadFirst(data []byte) (n int, next Cursor, err error) {\n\tb.mu.RLock()\n\tdefer b.mu.RUnlock()\n\treturn b.readOffset(data, b.first)\n}", "func (pk PacketBufferPtr) ToBuffer() buffer.Buffer {\n\tb := pk.buf.Clone()\n\tb.TrimFront(int64(pk.headerOffset()))\n\treturn b\n}", "func (handle *Handle) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(handle.x)\n}", "func (dw downloadDestinationBuffer) ReadFrom(r io.Reader) (int64, error) {\n\tvar n int64\n\tfor len(dw.buf) > 0 {\n\t\tread, err := io.ReadFull(r, dw.buf[0])\n\t\tif err != nil {\n\t\t\treturn n, err\n\t\t}\n\t\tdw.buf = dw.buf[1:]\n\t\tn += int64(read)\n\t}\n\treturn n, nil\n}", "func (b *Buffer) Read(data []byte, c Cursor) (n int, next Cursor, err error) {\n\tb.mu.RLock()\n\tdefer b.mu.RUnlock()\n\n\tseq, offset := c.seq, c.offset\n\n\tif seq >= b.nextSeq || offset > b.last {\n\t\treturn 0, next, ErrNotArrived\n\t}\n\n\tf := b.frame(offset)\n\tif f.size() == 0 || f.seq() != seq {\n\t\treturn b.readFirst(data)\n\t}\n\n\treturn b.readOffset(data, offset)\n}", "func (reader *embedFileReader) ReadAt(p []byte, off int64) (int, error) {\n\treturn 0, ErrNotImplemented\n}", "func MapNamedBuffer(buffer uint32, access uint32) unsafe.Pointer {\n\tret, _, _ := syscall.Syscall(gpMapNamedBuffer, 2, uintptr(buffer), uintptr(access), 0)\n\treturn (unsafe.Pointer)(ret)\n}", "func (p *TempFileProc) ReadSlice(delim byte) (line []byte, err error) {\n\treturn p.Buf.ReadSlice(delim)\n}", "func FromBuffer(buffer []byte) (File, error) {\n\ttotalSize := len(buffer)\n\ttotalSize64 := uint64(totalSize)\n\treader := bytes.NewReader(buffer)\n\thdrRiff, err := readHeaderRIFF(reader, totalSize64)\n\triffChunkId := hdrRiff.ChunkID\n\n\t/*\n\t * Check if RIFF header was successfully read.\n\t */\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\thdrDataSize := &dataSizeHeader{}\n\n\t\t/*\n\t\t * If this is an 'RF64' or 'BW64' file, read data size header.\n\t\t */\n\t\tif riffChunkId == ID_RIFF64 || riffChunkId == ID_BW64 {\n\t\t\thdrDataSize, err = readHeaderDataSize(reader, totalSize64)\n\n\t\t\t/*\n\t\t\t * If data size header was successfully read, skip over optional table entries.\n\t\t\t */\n\t\t\tif err != nil {\n\t\t\t\tmsg := err.Error()\n\t\t\t\treturn nil, fmt.Errorf(\"Failed to read data size chunk: %s\", msg)\n\t\t\t} else {\n\t\t\t\tnumEntries := hdrDataSize.TableLength\n\t\t\t\tnumEntries64 := uint64(numEntries)\n\t\t\t\tbytesToSkip := LENGTH_DATASIZE_TABLE_ENTRIES * numEntries64\n\t\t\t\terr := skipData(reader, bytesToSkip)\n\n\t\t\t\t/*\n\t\t\t\t * Check if we successfully skipped over the table entries.\n\t\t\t\t */\n\t\t\t\tif err != nil {\n\t\t\t\t\tmsg := err.Error()\n\t\t\t\t\treturn nil, fmt.Errorf(\"Failed to skip over data size table entries: %s\", msg)\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t}\n\n\t\terr := skipToChunk(reader, ID_FORMAT)\n\n\t\t/*\n\t\t * Check if we successfully arrived at the format chunk.\n\t\t */\n\t\tif err != nil {\n\t\t\tmsg := err.Error()\n\t\t\treturn nil, fmt.Errorf(\"Failed to locate format chunk: %s\", msg)\n\t\t} else {\n\t\t\thdrFormat, err := readHeaderFormat(reader)\n\n\t\t\t/*\n\t\t\t * Check if format header was successfully read.\n\t\t\t */\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t} else {\n\t\t\t\tbitDepth := hdrFormat.BitDepth\n\t\t\t\tsampleFormat := hdrFormat.AudioFormat\n\t\t\t\terr := skipToChunk(reader, ID_DATA)\n\n\t\t\t\t/*\n\t\t\t\t * Check if we successfully arrived at the data chunk.\n\t\t\t\t */\n\t\t\t\tif err != nil {\n\t\t\t\t\tmsg := err.Error()\n\t\t\t\t\treturn nil, fmt.Errorf(\"Failed to locate data chunk: %s\", msg)\n\t\t\t\t} else {\n\t\t\t\t\thdrData, err := readHeaderData(reader, totalSize64)\n\t\t\t\t\tchunkSize32 := hdrData.ChunkSize\n\t\t\t\t\tchunkSize64 := uint64(chunkSize32)\n\n\t\t\t\t\t/*\n\t\t\t\t\t * If this is an 'RF64' or 'BW64' file, take chunk size from data size header.\n\t\t\t\t\t */\n\t\t\t\t\tif riffChunkId == ID_RIFF64 || riffChunkId == ID_BW64 {\n\t\t\t\t\t\tchunkSize64 = hdrDataSize.SizeData\n\t\t\t\t\t}\n\n\t\t\t\t\t/*\n\t\t\t\t\t * Check if data header was successfully read.\n\t\t\t\t\t */\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\treturn nil, err\n\t\t\t\t\t} else {\n\t\t\t\t\t\tsampleData := make([]byte, chunkSize64)\n\t\t\t\t\t\t_, err = reader.Read(sampleData)\n\n\t\t\t\t\t\t/*\n\t\t\t\t\t\t * Check if sample data was read.\n\t\t\t\t\t\t */\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\tmsg := err.Error()\n\t\t\t\t\t\t\treturn nil, fmt.Errorf(\"Failed to read sample data: %s\", msg)\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tsamples, err := bytesToSamples(sampleData, sampleFormat, bitDepth)\n\n\t\t\t\t\t\t\t/*\n\t\t\t\t\t\t\t * Check if sample data was decoded.\n\t\t\t\t\t\t\t */\n\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\tmsg := err.Error()\n\t\t\t\t\t\t\t\treturn nil, fmt.Errorf(\"Failed to decode sample data: %s\", msg)\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\tchannelCount := hdrFormat.ChannelCount\n\t\t\t\t\t\t\t\tchannels := samplesToChannels(samples, channelCount)\n\n\t\t\t\t\t\t\t\t/*\n\t\t\t\t\t\t\t\t * Create a new data structure representing the contents of the wave file.\n\t\t\t\t\t\t\t\t */\n\t\t\t\t\t\t\t\tfile := fileStruct{\n\t\t\t\t\t\t\t\t\tbitDepth: bitDepth,\n\t\t\t\t\t\t\t\t\tsampleFormat: sampleFormat,\n\t\t\t\t\t\t\t\t\tsampleRate: hdrFormat.SampleRate,\n\t\t\t\t\t\t\t\t\tchannels: channels,\n\t\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t\t\treturn &file, nil\n\t\t\t\t\t\t\t}\n\n\t\t\t\t\t\t}\n\n\t\t\t\t\t}\n\n\t\t\t\t}\n\n\t\t\t}\n\n\t\t}\n\n\t}\n\n}", "func NewPointerFromSignalArgument(obj SignalArgument) Pointer {\n\treturn Pointer{base: unsafe.Pointer(obj.getBase())}\n}", "func NamedFramebufferReadBuffer(framebuffer uint32, src uint32) {\n\tsyscall.Syscall(gpNamedFramebufferReadBuffer, 2, uintptr(framebuffer), uintptr(src), 0)\n}", "func (recv *Value) GetPointer() uintptr {\n\tretC := C.g_value_get_pointer((*C.GValue)(recv.native))\n\tretGo := (uintptr)(unsafe.Pointer(retC))\n\n\treturn retGo\n}", "func GetFixedByteSliceFromBuf(buf []byte, curPos int, byteSlice []byte) (nextPos int, err error) {\n\tnextPos, err = getFixedByteSliceFromBuf(buf, curPos, byteSlice)\n\treturn\n}", "func (r *RingBuffer) Read(p []byte) (n int, err error) {\n\tif len(p) == 0 {\n\t\treturn 0, nil\n\t}\n\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\tn, err = r.read(p)\n\treturn\n}", "func NewBuffer(m []byte, skip, size int64) (*Buffer, error) {\n\tb := &Buffer{\n\t\toffset: skip,\n\t\tsize: size,\n\t\tdata: m,\n\t}\n\treturn b, nil\n}", "func pointerOf(n interface{}) uintptr {\n\tif n == nil {\n\t\treturn 0\n\t}\n\tv := reflect.ValueOf(n)\n\tif v.IsNil() {\n\t\treturn 0\n\t}\n\treturn v.Pointer()\n}", "func (r *bytesReader) ReadAt(b []byte, offset int64) (n int, err error) {\n\tif offset < 0 {\n\t\treturn 0, errors.New(\"buffer.bytesReader.ReadAt: negative offset\")\n\t}\n\tif offset >= int64(len(r.bs)) {\n\t\treturn 0, io.EOF\n\t}\n\tn = copy(b, r.bs[offset:])\n\tif n < len(b) {\n\t\terr = io.EOF\n\t}\n\treturn\n}", "func FileGetCbDataAsBoolPointer(data FileCbData) *bool {\n\tpdata, ok := data.(*bool)\n\tif !ok {\n\t\t// This never happens.\n\t\tpanic(\"invalid type\")\n\t}\n\treturn pdata\n}", "func (b Byte) Ptr() *byte {\n\tif !b.Valid {\n\t\treturn nil\n\t}\n\treturn &b.Byte\n}" ]
[ "0.65911293", "0.6003083", "0.59482414", "0.5822819", "0.5704803", "0.5696935", "0.56339794", "0.5532788", "0.5454545", "0.53906405", "0.536567", "0.5354107", "0.5321199", "0.5296994", "0.52856916", "0.52807784", "0.52178603", "0.5209607", "0.51778156", "0.5169711", "0.5169711", "0.5155652", "0.5154801", "0.50991714", "0.5070371", "0.5064656", "0.5044021", "0.50251013", "0.5012039", "0.50077206", "0.4999579", "0.49984774", "0.49941412", "0.49858376", "0.4978382", "0.49768826", "0.4972543", "0.4972543", "0.4960916", "0.4953336", "0.49517867", "0.49516127", "0.4946865", "0.49249506", "0.48940518", "0.48907664", "0.48881054", "0.4867627", "0.48672312", "0.48652005", "0.48652005", "0.484371", "0.4837868", "0.48323873", "0.48255143", "0.48203632", "0.4809864", "0.48063523", "0.47855887", "0.47810185", "0.47732785", "0.47571552", "0.47511488", "0.47482565", "0.4736684", "0.47329167", "0.47298592", "0.47262484", "0.47242343", "0.4705914", "0.4705525", "0.47004935", "0.46999422", "0.4693859", "0.46844113", "0.46814325", "0.46735427", "0.46692908", "0.4666303", "0.46555442", "0.4653083", "0.4651888", "0.46484476", "0.46437094", "0.4622518", "0.461688", "0.46146265", "0.45990404", "0.4593678", "0.4592788", "0.45899457", "0.4585769", "0.45846528", "0.45825046", "0.457744", "0.45703384", "0.45669076", "0.45656067", "0.4562999", "0.45622826" ]
0.7550936
0
IsValid checks if the pointer has a valid structure. It doesn't check if the pointedtocontent exists.
func (p Pointer) IsValid() bool { if len(p.Oid) != 64 { return false } if !oidPattern.MatchString(p.Oid) { return false } if p.Size < 0 { return false } return true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (v *Value) Valid() bool {\n return v.data != nil\n}", "func (handle Handle) IsValid() bool {\n\treturn handle != ^Handle(0)\n}", "func (p Pin) IsValid() bool {\n\treturn p.h != 0\n}", "func (p Point) Is_valid() (C.int){\n\treturn (C.crypto_core_ed25519_is_valid_point((*C.uchar)(&p.Val[0])))\n}", "func (pos *Position) IsValid() bool { return pos.Line > 0 }", "func (pos *Position) IsValid() bool { return pos.Line > 0 }", "func (p Pos) IsValid() bool {\n\treturn p != NoPos\n}", "func (p Pos) IsValid() bool {\n\treturn p != NoPos\n}", "func (pos *Position) IsValid() bool {\n\treturn pos.Line > 0\n}", "func (p Position) Valid() bool {\n\treturn p.Start <= p.End\n}", "func (msg messageStruct) IsValid() bool {\n\treturn !(msg.move == nil && msg.update == nil && !msg.skip && !msg.join)\n}", "func (g *Georeference) IsValid() bool {\n\treturn g.Point.IsValid()\n}", "func (l *Link) Valid() bool {\n\tswitch {\n\tcase l.DeletedAt != nil:\n\t\treturn false\n\tcase l.ViewsLimit > 0 && l.Views >= l.ViewsLimit:\n\t\treturn false\n\tdefault:\n\t\treturn true\n\t}\n}", "func (l Location) IsValid() bool {\n\tif l.X < 0 || l.X > Size-1 || l.Y < 0 || l.Y > Size-1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (h *FriendlyHost) Valid() bool {\n\treturn svchost.IsValid(h.Raw)\n}", "func (t *Type) IsPtr() bool", "func (p Page) IsValid() bool {\n\treturn p.valid\n}", "func (p Position) Valid() bool {\n\tok := func(i int) bool { return 1 <= i && i <= 8 }\n\treturn ok(p[0]) && ok(p[1])\n}", "func IsPosValid(posX, posY, length, width int) bool {\n\treturn posX >= 0 && posY >= 0 && posX < length && posY < width\n}", "func (p Position) IsValid() bool {\n\tif p < PointGuard || p > Center {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (p *Point) IsValid() bool {\n\treturn IsLon(p.Lon) && IsLat(p.Lat)\n}", "func CheckValid(data Dstream) bool {\n\n\tdata.Reset()\n\tnames := data.Names()\n\n\tfor c := 0; data.Next(); c++ {\n\t\tn0 := ilen(data.GetPos(0))\n\t\tfor j := 1; j < len(names); j++ {\n\t\t\tn1 := ilen(data.GetPos(j))\n\t\t\tif n1 != n0 {\n\t\t\t\tmsg := fmt.Sprintf(\"Length mismatch in chunk %d: len(%s) = %d, len(%s) = %d\\n\",\n\t\t\t\t\tc, names[0], n0, names[j], n1)\n\t\t\t\t_, _ = io.WriteString(os.Stderr, msg)\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\n\tdata.Reset()\n\n\treturn true\n}", "func (x *fastReflection_Metadata) IsValid() bool {\n\treturn x != nil\n}", "func validatorStruct(doc interface{}) bool {\n\tval := reflect.ValueOf(doc)\n\tif val.Kind() == reflect.Ptr && !val.IsNil() {\n\t\tval = val.Elem()\n\t}\n\tif val.Kind() != reflect.Struct || val.Type() == reflect.TypeOf(time.Time{}) {\n\t\treturn false\n\t}\n\treturn true\n}", "func (x *fastReflection_AddressBytesToStringResponse) IsValid() bool {\n\treturn x != nil\n}", "func (b *Block) IsValid() bool {\n\treturn b.valid\n}", "func (c Client) IsValidData() (ok bool, err error) {\n\tv := reflect.ValueOf(c.data)\n\tif v.Kind() == reflect.Ptr {\n\t\tv = v.Elem()\n\t}\n\t//If not a struct its a problem now every data has to be a Struct\n\tif v.Kind() != reflect.Struct {\n\t\terr = fmt.Errorf(assets.String.PackageInconsistancy+\" [only accepts structs; got (%+[1]v)%[1]T]\", v)\n\t\treturn\n\t}\n\n\t//Now pass through all the values and make sure all the mandatory fields are set\n\ttp := v.Type()\n\tfor i := 0; i < tp.NumField(); i++ {\n\t\tfield := tp.Field(i)\n\t\temptyValue := helpers.IsEmptyValue(v.Field(i))\n\t\tif emptyValue {\n\t\t\tif tagv := field.Tag.Get(\"mandatory\"); tagv == \"true\" {\n\t\t\t\terr = fmt.Errorf(assets.String.MandatoryFieldUnavailable, field.Name)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\tok = true\n\treturn\n}", "func (i *ObjectIterator) IsValid() bool {\n\treturn i.position < i.size\n}", "func Valid(data []byte) (ok bool, start int) {\n n := len(data)\n if n == 0 {\n return false, -1\n }\n s := rt.Mem2Str(data)\n p := 0\n m := types.NewStateMachine()\n ret := native.ValidateOne(&s, &p, m)\n types.FreeStateMachine(m)\n\n if ret < 0 {\n return false, p-1\n }\n\n /* check for trailing spaces */\n for ;p < n; p++ {\n if (types.SPACE_MASK & (1 << data[p])) == 0 {\n return false, p\n }\n }\n\n return true, ret\n}", "func (p *Page) Valid() bool {\n\tif p.Limit > 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func (t Type) IsValid() bool {\n\treturn valid_start < t && t < valid_end\n}", "func (x *fastReflection_AddressStringToBytesResponse) IsValid() bool {\n\treturn x != nil\n}", "func (x *fastReflection_Params) IsValid() bool {\n\treturn x != nil\n}", "func checkStructPtr(data interface{}) error {\n\tval := reflect.ValueOf(data)\n\tind := reflect.Indirect(val)\n\tindType := ind.Type()\n\tif val.Kind() != reflect.Ptr || ind.Kind() != reflect.Struct {\n\t\treturn errors.New(\"Argument must be a struct pointer\")\n\t}\n\n\tif _, ok := indType.FieldByName(\"ID\"); !ok {\n\t\treturn errors.New(\"Struct must have an ID field\")\n\t}\n\tif f, _ := indType.FieldByName(\"ID\"); f.Type != reflect.TypeOf(int64(1.0)) {\n\t\treturn errors.New(\"Struct ID field must be of type `int64`\")\n\t}\n\treturn nil\n}", "func checkStructPtr(data interface{}) error {\n\tval := reflect.ValueOf(data)\n\tind := reflect.Indirect(val)\n\tindType := ind.Type()\n\tif val.Kind() != reflect.Ptr || ind.Kind() != reflect.Struct {\n\t\treturn errors.New(\"Argument must be a struct pointer\")\n\t}\n\n\tif _, ok := indType.FieldByName(\"ID\"); !ok {\n\t\treturn errors.New(\"Struct must have an ID field\")\n\t}\n\tif f, _ := indType.FieldByName(\"ID\"); f.Type != reflect.TypeOf(int64(1.0)) {\n\t\treturn errors.New(\"Struct ID field must be of type `int64`\")\n\t}\n\treturn nil\n}", "func (field Field) IsPointer() bool {\n\t_, ok := TypeDeepest(field.TypeValue.Type).(*types.Pointer)\n\treturn ok\n}", "func (ds *DynamicStruct) IsPtr() bool {\n\treturn ds.isPtr\n}", "func (x *fastReflection_PositionalArgDescriptor) IsValid() bool {\n\treturn x != nil\n}", "func (m *MaybeError[E, A]) IsValid() bool {\n\tpresent := 0\n\tfor _, v := range []interface{}{m.Success, m.Failure, m.Error} {\n\t\tif v != nil {\n\t\t\tpresent++\n\t\t}\n\t}\n\treturn present == 1\n}", "func (c cell) IsValid() bool {\n\treturn c.contentHandler != nil\n}", "func (t *Type) IsUnsafePtr() bool", "func (m *Metadata) Valid() bool {\n\tif m.ProbeCC == \"\" {\n\t\treturn false\n\t}\n\tif m.ProbeASN == \"\" {\n\t\treturn false\n\t}\n\tif m.Platform == \"\" {\n\t\treturn false\n\t}\n\tif m.SoftwareName == \"\" {\n\t\treturn false\n\t}\n\tif m.SoftwareVersion == \"\" {\n\t\treturn false\n\t}\n\tif len(m.SupportedTests) < 1 {\n\t\treturn false\n\t}\n\tswitch m.Platform {\n\tcase \"ios\", \"android\":\n\t\tif m.DeviceToken == \"\" {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (fr *Frame) Valid() bool {\n\treturn false\n}", "func (kind Kind) IsValid() bool {\n\treturn kind&Invalid == 0\n}", "func (x *fastReflection_AddressBytesToStringRequest) IsValid() bool {\n\treturn x != nil\n}", "func (p *Packet) IsValid() bool {\n\t// TODO\n\t// . 0x0 cmd is handled in protocodes\n\tif p.Code == 0 {\n\t\treturn false\n\t}\n\treturn IsValidCommand(p.Code)\n}", "func (a DataAsset) IsValid() (bool, error) {\n\tif len([]byte(a)) > byteSizeData {\n\t\treturn false, errors.New(\"data length exceeds maximum payload size\")\n\t}\n\treturn true, nil\n}", "func (x *fastReflection_DuplicateVoteEvidence) IsValid() bool {\n\treturn x != nil\n}", "func (x *fastReflection_Output) IsValid() bool {\n\treturn x != nil\n}", "func TestNotPointer(t *testing.T) {\n\tvalidator := NewValidator(mocks.FakeNewLogger(nil))\n\td := testStruct{\n\t\tPercent: 0,\n\t\tPort: 8080,\n\t\tIpPort: \"127.0.0.1\",\n\t}\n\n\tassert.False(t, validator.Validate(d))\n}", "func (q *QueryTop) IsValid() (bool, error) {\n\treturn true, nil\n}", "func (x *fastReflection_AddressStringToBytesRequest) IsValid() bool {\n\treturn x != nil\n}", "func (p Picture) ValidPicture() bool {\n\tok, err := govalidator.ValidateStruct(p)\n\treturn err == nil && ok\n}", "func (v *BytesArePresent) IsValid(errors *validate.Errors) {\n\tif len(v.Field) == 0 {\n\t\terrors.Add(GenerateKey(v.Name), fmt.Sprintf(\"%s can not be blank.\", v.Name))\n\t}\n}", "func (c cursor) Valid() bool {\n\texpiration := time.Now().Add(cursorValidDuration)\n\tif c.CreatedAt.After(expiration) {\n\t\treturn false\n\t}\n\treturn true\n}", "func (x *fastReflection_EvidenceList) IsValid() bool {\n\treturn x != nil\n}", "func IsPointer(a cc.Type) bool { return a.RawDeclarator().PointerOpt != nil }", "func (x *fastReflection_Evidence) IsValid() bool {\n\treturn x != nil\n}", "func (o KeystoresAliasesSelfSignedCertCertsInfoCertInfoOutput) IsValid() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KeystoresAliasesSelfSignedCertCertsInfoCertInfo) *string { return v.IsValid }).(pulumi.StringPtrOutput)\n}", "func (*Component_Memory) IsYANGGoStruct() {}", "func (ps Pool) Valid() error {\n\tif ps.IsEmpty() {\n\t\treturn errors.New(\"pool asset cannot be empty\")\n\t}\n\treturn nil\n}", "func (o KeystoresAliasesKeyCertFileCertsInfoCertInfoOutput) IsValid() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KeystoresAliasesKeyCertFileCertsInfoCertInfo) *string { return v.IsValid }).(pulumi.StringPtrOutput)\n}", "func (pe *ProposalEntry) IsDataValid() (bool, error) {\n\treturn true, nil\n\t// Cannot have both `text` and `externalRef` field\n\tif pe.Proposal.Text != \"\" && pe.Proposal.ExternalRef.Href != \"\" {\n\t\treturn false, fmt.Errorf(\"cannot have both 'text' and 'externalRef' fields\")\n\t}\n\n\t// TODO: What else makes it invalid?\n\n\treturn true, nil\n}", "func (geom Geometry) IsValid() bool {\n\tval := C.OGR_G_IsValid(geom.cval)\n\treturn val != 0\n}", "func (f Fields) Valid() bool {\n\treturn f.Len() == 0\n}", "func (t *Type) IsPtrShaped() bool", "func (h *Header) Valid() bool {\n\treturn h.File.Valid() && h.Format.Valid() && h.Data.ID == 0x61746164\n}", "func (x *fastReflection_Input) IsValid() bool {\n\treturn x != nil\n}", "func IsPtr(value interface{}) bool {\n\treturn kindOf(value) == reflect.Ptr\n}", "func (ct ContentType) IsValid() error {\n\tswitch ct {\n\tcase Collection, Credential, DIDResolutionResponse, Metadata, Connection, Key:\n\t\treturn nil\n\t}\n\n\treturn fmt.Errorf(\"invalid content type '%s', supported types are %s\", ct,\n\t\t[]ContentType{Collection, Credential, DIDResolutionResponse, Metadata, Connection, Key})\n}", "func (it *Iterator) Valid() bool { return it.item != nil }", "func (block *Block) IsValid() bool {\n\tflagBits := util.BytesToBitField(block.Flags)\n\thashes := make([][]byte, len(block.Hashes))\n\tfor i, hash := range block.Hashes {\n\t\thashes[i] = make([]byte, len(hash))\n\t\tcopy(hashes[i], hash)\n\t\tutil.ReverseByteArray(hashes[i])\n\t}\n\ttree := NewTree(int(block.Total))\n\ttree.PopulateTree(flagBits, hashes)\n\treturn bytes.Equal(util.ReverseByteArray(tree.Root()), block.MerkleRoot[:])\n}", "func (v *Validator) IsValid(target interface{}) error{\n\tif target == nil {\n\t\treturn fmt.Errorf(\"valid target can not be nil\")\n\t}\n\n\t// 检查对象类型必须为struct或者struct pointer\n\ttType := reflect.TypeOf(target)\n\tif tType.Kind() == reflect.Ptr {\n\t\ttType = tType.Elem()\n\t}\n\tif tType.Kind() != reflect.Struct {\n\t\treturn fmt.Errorf(\"valid target model must be struct or struct pointer\")\n\t}\n\ttValue := reflect.Indirect(reflect.ValueOf(target))\n\n\tvar resError = &ValidError{}\n\tfor i := 0; i < tValue.NumField(); i++ {\n\t\tfield := tType.Field(i)\n\t\tif tagStr, ok := field.Tag.Lookup(validatorTag); ok {\n\t\t\tif err := v.check(tagStr, &field, tValue.Field(i)); err != nil {\n\t\t\t\tresError.errors = append(resError.errors, err)\n\t\t\t\tif !v.CheckAll {\n\t\t\t\t\treturn resError\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tif len(resError.errors) == 0 {\n\t\treturn nil\n\t}\n\treturn resError\n}", "func (ct ContentType) IsValid() error {\n\tswitch ct {\n\tcase Collection, Credential, DIDDocument, Metadata, Connection:\n\t\treturn nil\n\t}\n\n\treturn fmt.Errorf(\"invalid content type '%s', supported types are %s\", ct,\n\t\t[]ContentType{Collection, Credential, DIDDocument, Metadata, Connection})\n}", "func (x *fastReflection_MsgCommunityPoolSpend) IsValid() bool {\n\treturn x != nil\n}", "func (iter *SliceIterator) IsValid() bool {\n\tif iter.position >= 0 && iter.position < iter.s.Len() {\n\t\treturn true\n\t}\n\treturn false\n}", "func IsPointer(data interface{}) bool {\n\treturn typeIs(data, reflect.Ptr)\n}", "func (m fsMetaV1) IsValid() bool {\n\treturn isFSMetaValid(m.Version, m.Format)\n}", "func (pt MDTurbo) Validate() bool {\n\tif pt.Magic != 52426 {\n\t\treturn false\n\t}\n\tif pt.Partitions1[0].Start != 256 {\n\t\treturn false\n\t}\n\treturn true\n}", "func isValidAddress(addr int64) bool {\n\tif addr >= 0 && addr%getBinaryNodeSize() == 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func (n *Node) IsValid() bool {\n\n\t// obviously a port number of zero won't work\n\tif n.TcpAddr.Port == 0 {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func IsValidJSON(content []byte) bool {\n\tvar result map[string]interface{}\n\treturn json.Unmarshal(content, &result) == nil\n}", "func (gpm *GenericPoolManager) IsValid(fsvc *fscache.FuncSvc) bool {\n\tfor _, obj := range fsvc.KubernetesObjects {\n\t\tif strings.ToLower(obj.Kind) == \"pod\" {\n\t\t\tpod, err := gpm.kubernetesClient.CoreV1().Pods(obj.Namespace).Get(obj.Name, metav1.GetOptions{})\n\t\t\tif err == nil && utils.IsReadyPod(pod) {\n\t\t\t\t// Normally, the address format is http://[pod-ip]:[port], however, if the\n\t\t\t\t// Istio is enabled the address format changes to http://[svc-name]:[port].\n\t\t\t\t// So if the Istio is enabled and pod is in ready state, we return true directly;\n\t\t\t\t// Otherwise, we need to ensure that the address contains pod ip.\n\t\t\t\tif gpm.enableIstio ||\n\t\t\t\t\t(!gpm.enableIstio && strings.Contains(fsvc.Address, pod.Status.PodIP)) {\n\t\t\t\t\tgpm.logger.Debug(\"valid address\",\n\t\t\t\t\t\tzap.String(\"address\", fsvc.Address),\n\t\t\t\t\t\tzap.Any(\"function\", fsvc.Function),\n\t\t\t\t\t\tzap.String(\"executor\", string(fsvc.Executor)),\n\t\t\t\t\t)\n\t\t\t\t\treturn true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (box BoundingBox) IsValid() bool {\n\treturn box.Min.X < box.Max.X && box.Min.Y < box.Max.Y && box.Min.Z < box.Max.Z\n}", "func IsValid(h H3Index) bool {\n\treturn C.h3IsValid(h) == 1\n}", "func IsValid(h H3Index) bool {\n\treturn C.h3IsValid(h) == 1\n}", "func (v *Validation) Valid() bool {\n\treturn len(*v) == 0\n}", "func checkStructSlicePtr(data interface{}) error {\n\tval := reflect.ValueOf(data)\n\tind := reflect.Indirect(val)\n\tindType := ind.Type()\n\tif val.Kind() != reflect.Ptr ||\n\t\tind.Kind() != reflect.Slice ||\n\t\tindType.Elem().Kind() != reflect.Ptr ||\n\t\tindType.Elem().Elem().Kind() != reflect.Struct {\n\n\t\treturn errors.New(\"Argument must be a pointer to a slice of struct pointers\")\n\t}\n\tstructType := indType.Elem().Elem()\n\n\tif _, ok := structType.FieldByName(\"ID\"); !ok {\n\t\treturn errors.New(\"Struct must have an ID field\")\n\t}\n\tif f, _ := structType.FieldByName(\"ID\"); f.Type != reflect.TypeOf(int64(1.0)) {\n\t\treturn errors.New(\"Struct ID field must be of type `int64`\")\n\t}\n\treturn nil\n}", "func checkStructSlicePtr(data interface{}) error {\n\tval := reflect.ValueOf(data)\n\tind := reflect.Indirect(val)\n\tindType := ind.Type()\n\tif val.Kind() != reflect.Ptr ||\n\t\tind.Kind() != reflect.Slice ||\n\t\tindType.Elem().Kind() != reflect.Ptr ||\n\t\tindType.Elem().Elem().Kind() != reflect.Struct {\n\n\t\treturn errors.New(\"Argument must be a pointer to a slice of struct pointers\")\n\t}\n\tstructType := indType.Elem().Elem()\n\n\tif _, ok := structType.FieldByName(\"ID\"); !ok {\n\t\treturn errors.New(\"Struct must have an ID field\")\n\t}\n\tif f, _ := structType.FieldByName(\"ID\"); f.Type != reflect.TypeOf(int64(1.0)) {\n\t\treturn errors.New(\"Struct ID field must be of type `int64`\")\n\t}\n\treturn nil\n}", "func (p *PCache) IsValid(kind string, meta *api.ObjectMeta) bool {\n\tif obj := p.Get(kind, meta); obj != nil {\n\t\tp.RLock()\n\t\tvalidateFn := p.kindOptsMap[kind].Validator\n\t\tp.RUnlock()\n\t\tvalid, _ := validateFn(obj)\n\t\treturn valid\n\t}\n\treturn false\n}", "func Valid(data []byte) bool {\n\treturn checkValid(data, &scanner{}) == nil\n}", "func (d *BinaryValue) IsValid() int32 {\n\treturn int32(C.gocef_binary_value_is_valid(d.toNative(), d.is_valid))\n}", "func (c FileDescriptor) IsValid() bool {\n\treturn capnp.Client(c).IsValid()\n}", "func ValueTypeIsValid(v reflect.Value, depth int) (valid bool) {\n\n\tif depth > 1 {\n\t\tif v.Kind() == reflect.Slice {\n\t\t\treturn false\n\t\t}\n\t}\n\n\tswitch v.Kind() {\n\tcase reflect.Uintptr, reflect.Chan, reflect.Struct, reflect.Invalid:\n\t\treturn false\n\tcase reflect.Ptr:\n\t\tv = reflect.Indirect(v)\n\t\treturn ValueTypeIsValid(v, depth + 1)\n\t}\n\treturn true\n}", "func (nt Time) Valid() bool {\n\treturn !nt.IsZero()\n}", "func (c *Cursor) Valid() bool {\n\treturn c.pos >= c.start && c.pos < c.end\n}", "func (x *fastReflection_ServiceCommandDescriptor) IsValid() bool {\n\treturn x != nil\n}", "func (d Definition) IsValid(args ...interface{}) bool {\n\tif d.Inputs == nil {\n\t\treturn true\n\t}\n\tif len(args) != len(d.Inputs) {\n\t\treturn false\n\t}\n\tfor i, a := range args {\n\t\tswitch x := d.Inputs[i].(type) {\n\t\tcase reflect.Type:\n\t\t\txv := reflect.ValueOf(x)\n\t\t\tav := reflect.ValueOf(a)\n\t\t\tif !xv.IsNil() {\n\t\t\t\tif !av.IsValid() {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t\tif ak := reflect.TypeOf(a).Kind(); (ak == reflect.Chan || ak == reflect.Func || ak == reflect.Map || ak == reflect.Ptr || ak == reflect.Slice) && av.IsNil() {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t\tif !av.Type().AssignableTo(x) {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\tcase reflect.Kind:\n\t\t\txv := reflect.ValueOf(x)\n\t\t\tav := reflect.ValueOf(a)\n\t\t\tif xv.IsValid() || !xv.IsNil() {\n\t\t\t\tif !av.IsValid() {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t\tif xv.IsValid() && av.Type().Kind() != x {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn true\n}", "func (m *socatManager) ValidHost(host string) bool {\n\t_, err := getSource(host)\n\treturn err == nil\n}", "func (t *Tensor) IsValid() bool {\n\tif DimProduct(t.Dims) != int64(len(t.Buffer)/int(DtypeSize[t.Dtype])) {\n\t\treturn false\n\t}\n\treturn true\n}" ]
[ "0.587822", "0.58659065", "0.5818778", "0.57707906", "0.5719109", "0.5719109", "0.5714936", "0.5714936", "0.5713461", "0.56701905", "0.56682885", "0.5663861", "0.5629885", "0.558988", "0.55898464", "0.5557078", "0.55483496", "0.55470836", "0.5545316", "0.5544507", "0.55180955", "0.55024755", "0.5487519", "0.54580754", "0.5444703", "0.54299116", "0.54244816", "0.5422236", "0.54186", "0.5409078", "0.53807014", "0.53595066", "0.5349123", "0.5345405", "0.5345405", "0.53441536", "0.5341501", "0.5320186", "0.53193927", "0.5315895", "0.53100336", "0.52574074", "0.52552867", "0.52538574", "0.52535814", "0.5241609", "0.5230499", "0.52257234", "0.5225635", "0.52026117", "0.51986164", "0.5193196", "0.5189435", "0.5183566", "0.5179577", "0.5178944", "0.5176953", "0.5174685", "0.51677585", "0.5166106", "0.5158714", "0.5156799", "0.5143625", "0.51354927", "0.51351213", "0.51351154", "0.51257", "0.51245755", "0.5121022", "0.5120792", "0.51203173", "0.5117915", "0.51176506", "0.51175976", "0.51161784", "0.51138794", "0.51097465", "0.5108155", "0.51074266", "0.5106523", "0.51044077", "0.50984466", "0.50951624", "0.50863826", "0.50856704", "0.50856704", "0.5084246", "0.5084134", "0.5084134", "0.50832856", "0.5079345", "0.5079142", "0.5079139", "0.5077511", "0.5075039", "0.50744617", "0.5073041", "0.506702", "0.50668144", "0.506477" ]
0.73442924
0
StringContent returns the string representation of the pointer
func (p Pointer) StringContent() string { return fmt.Sprintf("%s\n%s%s\nsize %d\n", MetaFileIdentifier, MetaFileOidPrefix, p.Oid, p.Size) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p JSPointer) String() string {\n\treturn p.raw\n}", "func (o FileContentBufferPtrOutput) Content() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FileContentBuffer) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Content\n\t}).(pulumi.StringPtrOutput)\n}", "func (e *Element) ContentString() string {\n\treturn string(e.content)\n}", "func (j JSONMetaPrimitiveNode) ContentString(_ int) string {\n\treturn fmt.Sprintf(\"%v\", j.Value)\n}", "func (n *NotCommentToken) ContentString() string {\n\treturn string(n.Content())\n}", "func (jm JSONMeta) ContentString(_ int) string {\n\treturn fmt.Sprintf(\"%s => %v\", jm.Path(), jm.Parent())\n}", "func (c *CommentBlock) ContentString() string {\n\treturn string(c.content)\n}", "func (o FileContentBufferOutput) Content() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FileContentBuffer) *string { return v.Content }).(pulumi.StringPtrOutput)\n}", "func (o BlobOutput) SourceContent() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Blob) pulumi.StringPtrOutput { return v.SourceContent }).(pulumi.StringPtrOutput)\n}", "func (o RegistryTaskEncodedStepPtrOutput) ValueContent() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RegistryTaskEncodedStep) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ValueContent\n\t}).(pulumi.StringPtrOutput)\n}", "func (o FileContentBufferResponsePtrOutput) Content() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FileContentBufferResponse) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Content\n\t}).(pulumi.StringPtrOutput)\n}", "func (s *StringPointerValue) String() string {\n\tif s.stringPtr == nil {\n\t\treturn \"\"\n\t}\n\treturn *s.stringPtr\n}", "func (b *BlankLine) ContentString() string {\n\treturn \"\"\n}", "func (m *ChatMessageAttachment) GetContent()(*string) {\n val, err := m.GetBackingStore().Get(\"content\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func some_content() (as_string string, as_bytes []byte) {\n\tcontent_counter++\n\tas_string = fmt.Sprintf(\"<<<%d>>>\", content_counter)\n\tas_bytes = []byte(as_string)\n\treturn\n}", "func (o ServicePolicyPtrOutput) XmlContent() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ServicePolicy) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.XmlContent\n\t}).(pulumi.StringPtrOutput)\n}", "func (j JSONMetaPrimitiveNode) String() string {\n\treturn fmt.Sprintf(\"(%q -> %s)\", j.Path(), j.ContentString(0))\n}", "func (response *BaseResponse) GetContentString() string {\n\treturn response.contentString\n}", "func (o RegistryTaskEncodedStepOutput) ValueContent() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v RegistryTaskEncodedStep) *string { return v.ValueContent }).(pulumi.StringPtrOutput)\n}", "func (s *fakeString) String() string {\n\treturn s.content\n}", "func ptrToString(ptr uint32, size uint32) string {\n\t// Get a slice view of the underlying bytes in the stream. We use SliceHeader, not StringHeader\n\t// as it allows us to fix the capacity to what was allocated.\n\treturn *(*string)(unsafe.Pointer(&reflect.SliceHeader{\n\t\tData: uintptr(ptr),\n\t\tLen: uintptr(size), // Tinygo requires these as uintptrs even if they are int fields.\n\t\tCap: uintptr(size), // ^^ See https://github.com/tinygo-org/tinygo/issues/1284\n\t}))\n}", "func ToStringPointer(s string) *string {\n\treturn &s\n}", "func (obj *content) String() string {\n\tif obj.IsNormal() {\n\t\treturn fmt.Sprintf(protocolPattern, NormalProtocol, protocolSeparator, obj.Normal().String())\n\t}\n\n\treturn fmt.Sprintf(protocolPattern, TorProtocol, protocolSeparator, obj.Tor().String())\n}", "func (w *Reflow) String() string {\n\treturn w.buf.String()\n}", "func (r *Document) Content() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"content\"])\n}", "func (c *C) String() PtrType {\n\t// we represent it differently\n\treturn c.e.PtrT(c.e.Go().Byte())\n}", "func GetStringContent(m proto.Message) string {\n\treturn m.(*StringContentType).Value\n}", "func (c *clipboard) Content() string {\n\treturn glfw.GetClipboardString()\n}", "func (w *Wrap) String() string {\n\treturn w.buf.String()\n}", "func (j JSONMetaContainerNode) ContentString(level int) string {\n\tif level > 1 {\n\t\treturn \"{...}\"\n\t}\n\n\tvar o []string\n\tfor k, v := range j.Children {\n\t\tvs := \"null\"\n\t\tif v != nil {\n\t\t\tvs = v.ContentString(level + 1)\n\t\t}\n\t\to = append(o, fmt.Sprintf(\"%s:%v\", k, vs))\n\t}\n\treturn fmt.Sprintf(\"{%s}\", strings.Join(o, \", \"))\n}", "func GetContentString(url string) (string, error) {\n\tb, err := GetContent(url)\n\treturn string(b), err\n}", "func (p *ParseData) Content() string {\n\treturn p.content\n}", "func (s ViewContent) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o ApiImportPtrOutput) ContentValue() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApiImport) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.ContentValue\n\t}).(pulumi.StringPtrOutput)\n}", "func (r *Relationships) Content() string {\n\tcontent, err := DefaultEncode(r)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn content\n}", "func (b *BufferPool) GetContent() []byte {\n\treturn b.Data\n}", "func (e *ExprGetElementPtr) String() string {\n\treturn fmt.Sprintf(\"%s %s\", e.Type(), e.Ident())\n}", "func (this Comment) GetContent() string {\n\tif len(this) == 0 {\n\t\treturn \"\"\n\t}\n\ts := []byte(this)\n\tif isLineComment(string(this)) {\n\t\treturn string(s[2 : len(s)-1])\n\t}\n\treturn string(s[2 : len(s)-2])\n}", "func (this *SIPMessage) GetContent() string {\n\tif this.messageContentObject != nil {\n\t\treturn this.messageContentObject.(string)\n\t} else if this.messageContentBytes != nil {\n\t\treturn string(this.messageContentBytes)\n\t} else if this.messageContent != \"\" {\n\t\treturn this.messageContent\n\t} else {\n\t\treturn \"\"\n\t}\n}", "func (r *URIRef) Content() Content {\n\tif r.bytes == nil {\n\t\treturn NilContent{}\n\t}\n\tbc := ByteContent(r.bytes)\n\treturn &bc\n}", "func (m *CallTranscript) GetContent()([]byte) {\n val, err := m.GetBackingStore().Get(\"content\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]byte)\n }\n return nil\n}", "func ptrString(toPtr string) *string {\n\treturn &toPtr\n}", "func (c contentMessage) String() string {\n\tmessage := console.Colorize(\"Time\", fmt.Sprintf(\"[%s]\", c.Time.Format(printDate)))\n\tmessage += console.Colorize(\"Size\", fmt.Sprintf(\"%7s\", strings.Join(strings.Fields(humanize.IBytes(uint64(c.Size))), \"\")))\n\tfileDesc := \"\"\n\n\tif c.VersionID != \"\" {\n\t\tfileDesc += console.Colorize(\"VersionID\", \" \"+c.VersionID) + console.Colorize(\"VersionOrd\", fmt.Sprintf(\" v%d\", c.VersionOrd))\n\t\tif c.IsDeleteMarker {\n\t\t\tfileDesc += console.Colorize(\"DEL\", \" DEL\")\n\t\t} else {\n\t\t\tfileDesc += console.Colorize(\"PUT\", \" PUT\")\n\t\t}\n\t}\n\n\tfileDesc += \" \" + c.Key\n\n\tif c.Filetype == \"folder\" {\n\t\tmessage += console.Colorize(\"Dir\", fileDesc)\n\t} else {\n\t\tmessage += console.Colorize(\"File\", fileDesc)\n\t}\n\treturn message\n}", "func (o *Learningshareablecontentobject) String() string {\n \n \n \n \n \n \n \n \n \n\n j, _ := json.Marshal(o)\n str, _ := strconv.Unquote(strings.Replace(strconv.Quote(string(j)), `\\\\u`, `\\u`, -1))\n\n return str\n}", "func (o ServicePolicyOutput) XmlContent() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServicePolicy) *string { return v.XmlContent }).(pulumi.StringPtrOutput)\n}", "func String(value string) PartContent {\n\treturn stringContent(value)\n}", "func PointerToString(p *uint8, size int) string {\n\treturn *(*string)(unsafe.Pointer(&reflect.SliceHeader{\n\t\tData: uintptr(unsafe.Pointer(p)),\n\t\tLen: size,\n\t\tCap: size,\n\t}))\n}", "func (s GetReferenceMetadataOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o *Patchcontentpositionproperties) String() string {\n \n \n \n \n\n j, _ := json.Marshal(o)\n str, _ := strconv.Unquote(strings.Replace(strconv.Quote(string(j)), `\\\\u`, `\\u`, -1))\n\n return str\n}", "func (p *Packet) Content() []byte {\n\treturn p._pkt.body.anon0[:p._pkt.body.length]\n}", "func StringContent(s string) proto.Message {\n\treturn &StringContentType{Value: s}\n}", "func (o *SimpleStringWeb) GetContent() string {\n\tif o == nil || o.Content == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Content\n}", "func (P *Parser) stringifyContent(content interface{}) (string, error) {\n\tswitch c := content.(type) {\n\tcase string:\n\t\treturn c, nil\n\tcase []byte:\n\t\treturn string(c), nil\n\tcase *bytes.Buffer:\n\t\tif c != nil {\n\t\t\treturn c.String(), nil\n\t\t}\n\tcase io.Reader:\n\t\tvar buf bytes.Buffer\n\t\tif _, err := io.Copy(&buf, c); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\treturn buf.String(), nil\n\t}\n\treturn \"\", fmt.Errorf(\"unsupported content type %T\", content)\n}", "func (j JSONMetaArrayNode) ContentString(level int) string {\n\tif level > 1 {\n\t\treturn \"[...]\"\n\t}\n\n\tvar o []string\n\tfor _, v := range j.Items {\n\t\tvs := \"null\"\n\t\tif v != nil {\n\t\t\tvs = v.ContentString(level + 1)\n\t\t}\n\t\to = append(o, vs)\n\t}\n\n\treturn fmt.Sprintf(\"[%s]\", strings.Join(o, \", \"))\n}", "func StringPtr(literal string) *string {\n\treturn &literal\n}", "func (c Status) StringPtr() *string {\n\tv := string(c)\n\treturn &v\n}", "func (obj *codeMatch) Content() string {\n\treturn obj.content\n}", "func (j JSONMetaContainerNode) String() string {\n\treturn fmt.Sprintf(\"(%q -> %s)\", j.Path(), j.ContentString(0))\n}", "func bufferToString(buffer *bytes.Buffer, unsafePtr *bool) string {\n defer buffer.Reset()//ensure buffer is reset\n if !*unsafePtr {\n return buffer.String()\n }\n bb := buffer.Bytes()\n s := *(*string)(unsafe.Pointer(&bb))\n return s\n}", "func (s HubContentInfo) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (b *Blob) GetContent() string {\n\tif b == nil || b.Content == nil {\n\t\treturn \"\"\n\t}\n\treturn *b.Content\n}", "func (m *ChatMessageAttachment) GetContentUrl()(*string) {\n val, err := m.GetBackingStore().Get(\"contentUrl\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.(*string)\n }\n return nil\n}", "func (obj *Value) GetPointer() string {\n\treturn obj.Candy().Guify(\"g_value_get_pointer\", obj).String()\n}", "func (u UUID) StringPtr() *string {\n\tif len(u) != 16 {\n\t\treturn stringPtr(\"\")\n\t}\n\tvar buf [36]byte\n\tu.encodeHex(buf[:])\n\treturn stringPtr(string(buf[:]))\n}", "func (s GetDatasetContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (n *NotCommentToken) Content() []byte {\n\treturn n.c\n}", "func (o LookupPolicyResultOutput) Content() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupPolicyResult) string { return v.Content }).(pulumi.StringOutput)\n}", "func (ab About) String() string {\n\treturn string(ab.Body)\n}", "func (s ViewInputContent) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (i contentElementType) String() string {\n\tout, err := json.Marshal(i)\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn string(out)\n}", "func (this *Reference) String() string {\n\treturn this.At.String() + this.Name\n}", "func strptr(value string) *string {\n\treturn &value\n}", "func (ob *PyObject) String() string {\n\tif ob.rawptr == nil {\n\t\treturn \"\"\n\t}\n\trepr := C.PyObject_Str(ob.rawptr)\n\tif repr == nil {\n\t\treturn \"\"\n\t}\n\tdefer C.Py_DecRef(repr)\n\ts := C.PyUnicode_AsEncodedString(repr, encoding, codecErrors)\n\tif s == nil {\n\t\treturn \"invalid Unicode string\"\n\t}\n\tdefer C.Py_DecRef(s)\n\treturn C.GoString(C.PyBytes_AsString(s))\n}", "func (m *WorkbookCommentReply) GetContent()(*string) {\n return m.content\n}", "func (statics AssestStruct) GetContent(name string) string {\n\ts, err := statics.GetAssestFile(name)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn s.Content\n}", "func (b *Builder) String(s string) value.Pointer {\n\treturn b.constantMemory.writeString(s)\n}", "func (sp *SmartPosterPayload) String() string {\n\tstr := \"\\n\"\n\tstr += sp.Message.String()\n\treturn str\n}", "func (s DeleteDatasetContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DeleteDatasetContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s *schema) Content() []byte {\n\treturn s.content\n}", "func (c *HostConfig) Content() string {\n\treturn c.ContentBuffer().String()\n}", "func (s UpdateContactFlowContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (c contentMessage) String() string {\n\tmessage := console.Colorize(\"Time\", fmt.Sprintf(\"[%s] \", c.Time.Format(printDate)))\n\tmessage = message + console.Colorize(\"Size\", fmt.Sprintf(\"%7s \", strings.Join(strings.Fields(humanize.IBytes(uint64(c.Size))), \"\")))\n\tmessage = func() string {\n\t\tif c.Filetype == \"folder\" {\n\t\t\treturn message + console.Colorize(\"Dir\", c.Key)\n\t\t}\n\t\treturn message + console.Colorize(\"File\", c.Key)\n\t}()\n\treturn message\n}", "func (me TxsdPresentationAttributesTextContentElementsTextAnchor) String() string {\n\treturn xsdt.String(me).String()\n}", "func (s VirtualNodeRef) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o SchemaOutput) Content() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Schema) pulumi.StringOutput { return v.Content }).(pulumi.StringOutput)\n}", "func cel_Source_Content(_handle CGoHandle) *C.char {\n\tvifc, __err := gopyh.VarFromHandleTry((gopyh.CGoHandle)(_handle), \"cel.Source\")\n\tif __err != nil {\n\t\treturn C.CString(\"\")\n\t}\n\treturn C.CString(vifc.(cel.Source).Content())\n\n}", "func (p *Part) String() string {\n\tobjStr := C.g_mime_object_to_string(p.asGMimeObject(), nil)\n\tdefer C.g_free(C.gpointer(unsafe.Pointer(objStr)))\n\treturn strings.TrimSpace(C.GoString(objStr))\n}", "func (s CreateDatasetContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeHubContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (p *Proxy) String() string {\n\treturn string(p.lastResponseBody)\n}", "func (b *SafeBuffer) String() string {\n\tb.m.RLock()\n\tdefer b.m.RUnlock()\n\treturn b.b.String()\n}", "func (s ImportHubContentOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (e *Element) String() string {\n\tswitch e.Type {\n\tcase DocumentType:\n\t\treturn e.makeDocumentContent()\n\tcase TagType:\n\t\treturn e.makeTagContent()\n\tcase TextType:\n\t\treturn e.makeTextContent()\n\t}\n\treturn \"\"\n}", "func (s String) Ptr() *string {\n\tif !s.Valid || s.Null {\n\t\treturn nil\n\t}\n\treturn &s.String\n}", "func StringPointer(s string) *string {\n\treturn &s\n}", "func StringPointer(s string) *string {\n\treturn &s\n}", "func StringPointer(s string) *string {\n\treturn &s\n}", "func StringPointer(s string) *string {\n\treturn &s\n}", "func parseString(content []byte) string {\n return string(content)\n}" ]
[ "0.65256065", "0.652558", "0.64234", "0.6413323", "0.627286", "0.6232988", "0.62002516", "0.6170929", "0.61600256", "0.6067722", "0.5998857", "0.5981334", "0.5805871", "0.5782421", "0.57737684", "0.5763933", "0.57556957", "0.5751003", "0.5747597", "0.5746763", "0.57434756", "0.5739089", "0.57016146", "0.569444", "0.56862724", "0.5672143", "0.5671681", "0.5670149", "0.56680447", "0.5659907", "0.5644866", "0.56306845", "0.56251365", "0.5622702", "0.55594134", "0.55278623", "0.55152667", "0.5480137", "0.54764616", "0.54662526", "0.5462021", "0.54567045", "0.5447718", "0.5439403", "0.5437766", "0.54360974", "0.5435951", "0.5434044", "0.54334706", "0.54256", "0.5414315", "0.54107654", "0.5402824", "0.5395022", "0.53918254", "0.5383478", "0.53649366", "0.5364007", "0.5363494", "0.5352719", "0.5334597", "0.5328568", "0.532357", "0.5322399", "0.5312136", "0.52926874", "0.5283649", "0.52803236", "0.527938", "0.52791303", "0.5274007", "0.5259328", "0.52584326", "0.5257268", "0.52550703", "0.52526474", "0.5250345", "0.5242326", "0.5242326", "0.5241255", "0.52336174", "0.5231376", "0.52272743", "0.5223177", "0.52154416", "0.5210543", "0.5207753", "0.5203553", "0.5195075", "0.51848704", "0.51826936", "0.5176286", "0.5175852", "0.5166845", "0.51636434", "0.51592255", "0.51592255", "0.51592255", "0.51592255", "0.5154303" ]
0.8267372
0
RelativePath returns the relative storage path of the pointer
func (p Pointer) RelativePath() string { if len(p.Oid) < 5 { return p.Oid } return path.Join(p.Oid[0:2], p.Oid[2:4], p.Oid[4:]) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (cfp *FsPool) GetRelativePath(fileIndex int64) string {\n\treturn cfp.container.Files[fileIndex].Path\n}", "func relativePath(storePath, shardPath string) (string, error) {\n\tpath, err := filepath.Abs(storePath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"store abs path: %s\", err)\n\t}\n\n\tfp, err := filepath.Abs(shardPath)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"file abs path: %s\", err)\n\t}\n\n\tname, err := filepath.Rel(path, fp)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"file rel path: %s\", err)\n\t}\n\n\treturn name, nil\n}", "func (iter *Iteration) RelativePath() string {\n\treturn filepath.Join(iter.Dir, iter.TrackID, iter.Problem) + string(filepath.Separator)\n}", "func (s *Store) ShardRelativePath(id uint64) (string, error) {\n\tshard := s.Shard(id)\n\tif shard == nil {\n\t\treturn \"\", fmt.Errorf(\"shard %d doesn't exist on this server\", id)\n\t}\n\treturn relativePath(s.path, shard.path)\n}", "func (o LocalCopyPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *LocalCopy) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (m *Module) relativePath(p *types.Package) string {\n\treturn strings.TrimPrefix(p.Path(), m.Path)\n}", "func (c *CacheHandle) RelativeFile() string {\n\treturn path.Join(c.Digest[0:2], c.Digest[2:4], c.FileName())\n}", "func RelativeDir(relativePath string) string {\n\t_, f, _, _ := runtime.Caller(1)\n\treturn filepath.Join(filepath.Dir(f), relativePath)\n}", "func Relative(base, complete string) string {\n\trel, err := filepath.Rel(base, complete)\n\tif err != nil {\n\t\tFatalf(\"Diff Path %s, %s: %s\", base, complete, err)\n\t}\n\n\t// special case\n\tif rel == \".\" {\n\t\trel = \"\"\n\t}\n\n\treturn rel\n}", "func (s *BashScript) RelativePath() string {\n\treturn s.Path\n}", "func (i *Image) RelativePath() string {\n\t// Convert the gallery ID to a string\n\tgalleryID := fmt.Sprintf(\"%v\", i.GalleryID)\n\treturn filepath.ToSlash(filepath.Join(\"images\", \"galleries\", galleryID, i.Filename))\n}", "func GetSelfFilePath(relativePath string) string {\n\treturn path.Join(ProjectSelfPath, relativePath)\n}", "func (p *Partition) GetFullPath() string {\n\treturn p.Device + p.Name\n}", "func (b *Binary) LocalPath() string {\n\treturn b.file.LocalPath()\n}", "func (pool *PackagePool) RelativePath(filename string, hashMD5 string) (string, error) {\n\tfilename = filepath.Base(filename)\n\tif filename == \".\" || filename == \"/\" {\n\t\treturn \"\", fmt.Errorf(\"filename %s is invalid\", filename)\n\t}\n\n\tif len(hashMD5) < 4 {\n\t\treturn \"\", fmt.Errorf(\"unable to compute pool location for filename %v, MD5 is missing\", filename)\n\t}\n\n\treturn filepath.Join(hashMD5[0:2], hashMD5[2:4], filename), nil\n}", "func RelativePath(relativepath string, optionalpaths ...string) string {\n\t_, thisFile, _, _ := runtime.Caller(1)\n\tlocalDir := filepath.Dir(thisFile)\n\n\t// If we have optional paths, join them to the relativepath\n\tif len(optionalpaths) > 0 {\n\t\tpaths := []string{relativepath}\n\t\tpaths = append(paths, optionalpaths...)\n\t\trelativepath = filepath.Join(paths...)\n\t}\n\tresult, err := filepath.Abs(filepath.Join(localDir, relativepath))\n\tif err != nil {\n\t\t// I'm allowing this for 1 reason only: It's fatal if the path\n\t\t// supplied is wrong as it's only used internally in Wails. If we get\n\t\t// that path wrong, we should know about it immediately. The other reason is\n\t\t// that it cuts down a ton of unnecassary error handling.\n\t\tpanic(err)\n\t}\n\treturn result\n}", "func (o TransferJobTransferSpecAzureBlobStorageDataSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecAzureBlobStorageDataSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (df DataFile) LocalPath() string {\n\treturn df.pathFrom\n}", "func GetFullPath(relativePath string) string {\n\tcurrentPath, _ := os.Getwd()\n\tfullPathTemplate := path.Join(currentPath, relativePath)\n\treturn fullPathTemplate\n}", "func (f File) LocalPath() string {\n\treturn path.Join(f.Public, f.VPath())\n}", "func path(relPath string) string {\n\tif filepath.IsAbs(relPath) {\n\t\treturn relPath\n\t}\n\n\treturn filepath.Join(basepath, relPath)\n}", "func (pr *PathResolver) RelativePath(filename, baseDirectory string) string {\n\tif filepath.IsAbs(filename) && filepath.IsAbs(baseDirectory) {\n\t\toffset := baseDirectory\n\t\tif strings.HasSuffix(baseDirectory, string(pr.FileSeparator())) {\n\t\t\toffset = baseDirectory[:len(baseDirectory)-1]\n\t\t}\n\t\tfilename = filename[len(offset)+1:]\n\t}\n\treturn filename\n}", "func Path(relPath string) string {\n\tif filepath.IsAbs(relPath) {\n\t\treturn relPath\n\t}\n\n\treturn filepath.Join(basepath, relPath)\n}", "func (i *IndexBuilder) relativeDir() string {\n\trelativeDir := i.cfg.RelativeDir\n\tif relativeDir == \"/\" {\n\t\treturn relativeDir\n\t}\n\treturn fmt.Sprintf(\"/%s/\", relativeDir)\n}", "func relToPath(path, workingDir string) (string, error) {\n\tworkingDir = util.ConsistentFilepath(workingDir)\n\tpath = util.ConsistentFilepath(path)\n\n\tif !filepath.IsAbs(path) {\n\t\treturn path, nil\n\t}\n\n\tif workingDir == \"\" || !strings.HasPrefix(path, workingDir) {\n\t\treturn path, errNotRelativeToWorkingDir\n\t}\n\n\trelPath, err := filepath.Rel(workingDir, path)\n\tif err != nil {\n\t\treturn path, errors.Wrap(err, errNotRelativeToWorkingDir.Error())\n\t}\n\n\treturn util.ConsistentFilepath(relPath), nil\n}", "func Path(rel string) string {\n\tif filepath.IsAbs(rel) {\n\t\treturn rel\n\t}\n\n\tv, err := goPackagePath(\"google.golang.org/grpc/testdata\")\n\tif err != nil {\n\t\tlog.Fatalf(\"Error finding google.golang.org/grpc/testdata directory: %v\", err)\n\t}\n\n\treturn filepath.Join(v, rel)\n}", "func GetRelativePath(path, base string) (final string, err error) {\n\tif filepath.IsAbs(path) && base == \"\" {\n\t\treturn \"\", errors.New(\"source: missing base directory\")\n\t}\n\tname := filepath.Clean(path)\n\tbase = filepath.Clean(base)\n\n\tname, err = filepath.Rel(base, name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif strings.HasSuffix(filepath.FromSlash(path), FilePathSeparator) && !strings.HasSuffix(name, FilePathSeparator) {\n\t\tname += FilePathSeparator\n\t}\n\treturn name, nil\n}", "func (o ApplicationStatusSyncComparedToSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusSyncComparedToSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (o LocalCopyOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopy) string { return v.Path }).(pulumi.StringOutput)\n}", "func (e *Entry) realPath() string {\n\tparts := make([]string, 1, len(e.path)+1)\n\tparts[0] = e.root.realPath\n\tparts = append(parts, e.path...)\n\treturn filepath.Join(parts...)\n}", "func (o ApplicationOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (util copyHandlerUtil) relativePathToRoot(rootPath, filePath string, pathSep byte) string {\n\tif len(rootPath) == 0 {\n\t\treturn filePath\n\t}\n\tresult := strings.Replace(filePath, rootPath, \"\", 1)\n\tif len(result) > 0 && result[0] == pathSep {\n\t\tresult = result[1:]\n\t}\n\treturn result\n}", "func getAbsolutePath(p string) (string, error) {\n\troot, err := git.RootDir()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(root, p), nil\n}", "func (o TransferJobTransferSpecAzureBlobStorageDataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecAzureBlobStorageDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func RelativeDocPath(source, target string) string {\n\tvar result []string\n\n\tsep := string(os.PathSeparator)\n\tdname, _ := path.Split(source)\n\tfor i := 0; i < strings.Count(dname, sep); i++ {\n\t\tresult = append(result, \"..\")\n\t}\n\tresult = append(result, target)\n\treturn strings.Join(result, sep)\n}", "func GetLocalPath(projectData *project.Project, file string) string {\n\treturn filepath.Clean(filepath.Join(projectData.GetBuildDir(), filepath.FromSlash(file)))\n}", "func GetRelativePath(path []*NodeId) (relatives []*NodeId) {\n\tfor _, n := range path {\n\t\trelatives = append(relatives, n.GetRelative())\n\n\t}\n\treturn relatives\n}", "func (md *MetaData) RelativeDateDirectory() string {\n\tt := md.PreferredDate()\n\n\treturn filepath.Join(fmt.Sprintf(\"%d\", t.Year()), fmt.Sprintf(\"%02d-%s\", t.Month(), shortmonths[t.Month()-1]), fmt.Sprintf(\"%d\", t.Day()))\n\n}", "func getRelativePathValue(i *ini.File, section, key, vdefault, docroot string) string {\n\tp := i.Section(section).Key(key).MustString(vdefault)\n\th := p\n\tif !path.IsAbs(p) {\n\t\th = path.Join(docroot, p)\n\t}\n\treturn filepath.FromSlash(h)\n}", "func (o ApplicationStatusHistorySourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusHistorySource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func GetPath(relpath string) string {\n\trp, err := filepath.Abs(relpath)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error retrieving path \\\"%s\\\".\\n\", relpath)\n\t\tpanic(err)\n\t}\n\n\treturn rp\n}", "func Realpath(path string) string {\n\tp, _ := filepath.Abs(path)\n\treturn p\n}", "func (fs *BaseFs) MakePathRelative(filename string) (string, string) {\n\tfor _, sfs := range fs.FileSystems() {\n\t\tif sfs.Contains(filename) {\n\t\t\tif s, found := sfs.MakePathRelative(filename); found {\n\t\t\t\treturn sfs.Name, s\n\t\t\t}\n\t\t}\n\t}\n\t// May be a static file.\n\tif s := fs.MakeStaticPathRelative(filename); s != \"\" {\n\t\treturn files.ComponentFolderStatic, s\n\t}\n\t// Fall back to relative to the working dir.\n\tif strings.HasPrefix(filename, fs.workingDir) {\n\t\treturn \"\", strings.TrimPrefix(filename, fs.workingDir)\n\t}\n\treturn \"\", \"\"\n}", "func (r *RepoRef) LocalPath() string {\n\tlocalPath, err := r.localPath()\n\tif err != nil {\n\t\tlog.WithError(err).\n\t\t\tWithField(\"name\", r.Name).\n\t\t\tPanic(\"failed to determine local path for repository\")\n\t}\n\n\treturn localPath\n}", "func (r *RelativePath) FullPath() string {\n\treturn \"/\" + strings.Join(r.stack, \"/\")\n}", "func (o ApplicationOperationSyncSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperationSyncSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func makeRelative(child, parent string) string {\n\treturn strings.TrimPrefix(child, dirName(parent))\n}", "func GetPathRelativeToConfig(configuration *viper.Viper, key string) string {\n\tconfigFile := configuration.ConfigFileUsed()\n\tp := configuration.GetString(key)\n\tif p == \"\" || filepath.IsAbs(p) {\n\t\treturn p\n\t}\n\treturn filepath.Clean(filepath.Join(filepath.Dir(configFile), p))\n}", "func relativePath(target, path string) string {\n\tsplit := strings.Split(path, \"/\")\n\tfor i, s := range split {\n\t\tif s == target {\n\t\t\treturn strings.Join(split[i:], \"/\")\n\t\t}\n\t}\n\tfmt.Println(\"WARN: Failed to \")\n\treturn \"\"\n}", "func (s *Site) RelativePath(path string) string {\n\tif s.themeDir != \"\" {\n\t\tif rel, err := filepath.Rel(s.themeDir, path); err == nil {\n\t\t\treturn rel\n\t\t}\n\t}\n\treturn utils.MustRel(s.cfg.Source, path)\n}", "func (g git) GetRelativeLocation(currentDir string) (string, error) {\n\tcmd := exec.Exec(currentDir, \"git\", \"rev-parse\", \"--show-prefix\")\n\tif cmd.Err != nil {\n\t\treturn \"\", cmd.Err\n\t}\n\treturn cmd.StdOut, nil\n}", "func (fi *fileInfo) Path() string {\n\treturn fi.fullPath\n}", "func (o TransferJobTransferSpecAwsS3DataSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecAwsS3DataSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (dir *Dir) RelPath() string {\n\trel, err := filepath.Rel(dir.repoBase, dir.Path)\n\tif dir.repoBase == \"\" || err != nil {\n\t\treturn dir.BaseName()\n\t}\n\treturn rel\n}", "func (dir *Dir) RelPath() string {\n\trel, err := filepath.Rel(dir.repoBase, dir.Path)\n\tif dir.repoBase == \"\" || err != nil {\n\t\treturn dir.BaseName()\n\t}\n\treturn rel\n}", "func fullPath(path string) (string, error) {\n\tpath = strings.Replace(path, \"~\", os.Getenv(\"HOME\"), 1)\n\n\treturn filepath.Abs(path)\n}", "func MakeRelativeReference(space *storageProvider.StorageSpace, relativePath string, spacesDavRequest bool) *storageProvider.Reference {\n\tif space.Opaque == nil || space.Opaque.Map == nil || space.Opaque.Map[\"path\"] == nil || space.Opaque.Map[\"path\"].Decoder != \"plain\" {\n\t\treturn nil // not mounted\n\t}\n\tspacePath := string(space.Opaque.Map[\"path\"].Value)\n\trelativeSpacePath := \".\"\n\tif strings.HasPrefix(relativePath, spacePath) {\n\t\trelativeSpacePath = utils.MakeRelativePath(strings.TrimPrefix(relativePath, spacePath))\n\t} else if spacesDavRequest {\n\t\trelativeSpacePath = utils.MakeRelativePath(relativePath)\n\t}\n\treturn &storageProvider.Reference{\n\t\tResourceId: space.Root,\n\t\tPath: relativeSpacePath,\n\t}\n}", "func (o *Object) filePath() string {\n\treturn o.fs.filePath(o.remote)\n}", "func (m *mountPoint) Path() string {\n\tif m.Volume != nil {\n\t\treturn m.Volume.Path()\n\t}\n\n\treturn m.Source\n}", "func (e *Entry) fullPath() string {\n\tparts := make([]string, 1, len(e.path)+1)\n\tparts[0] = e.root.path\n\tparts = append(parts, e.path...)\n\treturn filepath.Join(parts...)\n}", "func RelativeToCwd(relativePath string) (string, error) {\n\tcwd, err := os.Getwd()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(cwd, relativePath), nil\n}", "func (copyHandlerUtil) getRelativePath(rootPath, filePath string) string {\n\t// root path contains the entire absolute path to the root directory, so we need to take away everything except the root directory from filePath\n\t// example: rootPath = \"/dir1/dir2/dir3\" filePath = \"/dir1/dir2/dir3/file1.txt\" result = \"dir3/file1.txt\" scrubAway=\"/dir1/dir2/\"\n\tif len(rootPath) == 0 {\n\t\treturn filePath\n\t}\n\tresult := filePath\n\n\t// replace the path separator in filepath with AZCOPY_PATH_SEPARATOR\n\t// this replacement is required to handle the windows filepath\n\tfilePath = strings.Replace(filePath, common.OS_PATH_SEPARATOR, common.AZCOPY_PATH_SEPARATOR_STRING, -1)\n\tvar scrubAway string\n\t// test if root path finishes with a /, if yes, ignore it\n\tif rootPath[len(rootPath)-1:] == common.AZCOPY_PATH_SEPARATOR_STRING {\n\t\tscrubAway = rootPath[:strings.LastIndex(rootPath[:len(rootPath)-1], common.AZCOPY_PATH_SEPARATOR_STRING)+1]\n\t} else {\n\t\t// +1 because we want to include the / at the end of the dir\n\t\tscrubAway = rootPath[:strings.LastIndex(rootPath, common.AZCOPY_PATH_SEPARATOR_STRING)+1]\n\t}\n\n\tresult = strings.Replace(filePath, scrubAway, \"\", 1)\n\n\treturn result\n}", "func (o DocumentationPartLocationPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *DocumentationPartLocation) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (o TransferJobTransferSpecGcsDataSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecGcsDataSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (dw *DrawingWand) PathMoveToRelative(x, y float64) {\n\tC.MagickDrawPathMoveToRelative(dw.dw, C.double(x), C.double(y))\n}", "func (e *Entry) tempPath() string {\n\tparts := make([]string, 1, len(e.path)+1)\n\tparts[0] = e.root.path\n\tparts = append(parts, e.path[:len(e.path)-1]...)\n\tparts = append(parts, e.tempName())\n\treturn filepath.Join(parts...)\n}", "func RealPath(file string) (string, error) {\n\tif path.IsAbs(file) {\n\t\treturn file, nil\n\t}\n\twd, err := os.Getwd()\n\treturn path.Join(wd, file), err\n}", "func (o ApplicationStatusSyncComparedToSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSyncComparedToSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (r *Repo) Path() (path string) {\n\tif r.IsRaw() {\n\t\treturn r.GitDir\n\t}\n\treturn r.WorkDir\n}", "func (p *SeriesPartition) Path() string { return p.path }", "func pathRelDir(path string) (string, error) {\n\tpath, err := filepath.Abs(path)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\twd, err := os.Getwd()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tpath, err = filepath.Rel(wd, path)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// If Rel returned \".\", fix it to empty string which will eventually mutate to \"./\".\n\tif path == \".\" {\n\t\tpath = \"\"\n\t}\n\t// Add a \"./\" prefix.\n\tif !strings.HasPrefix(path, \"./\") {\n\t\tpath = \"./\" + path\n\t}\n\treturn path, nil\n}", "func DeviceFullPath(repository, id, file string) string {\n\treturn filepath.Join(repository, id, file)\n}", "func GetAbsoluteFilePath(relFilePath string, sheet string) (result string, err error) {\n\tdir, err := os.Getwd()\n\tif err != nil {\n\t\tlog.Fatal(sheet, \" : \", err)\n\t}\n\tabsFilePath := filepath.Join(dir, relFilePath)\n\n\treturn absFilePath, nil\n\n}", "func (o ApplicationStatusSyncComparedToSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSyncComparedToSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (c *vertexCollection) relPath() string {\n\tescapedName := pathEscape(c.name)\n\treturn path.Join(c.g.relPath(), \"vertex\", escapedName)\n}", "func (this *YamlFile) GetFullPath() string {\n\treturn this.fullpath\n}", "func (pool *PackagePool) Path(filename string, hashMD5 string) (string, error) {\n\trelative, err := pool.RelativePath(filename, hashMD5)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(pool.rootPath, relative), nil\n}", "func (h *HttpParentProxy) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/http/parent_proxy/%s\", h.Reference)\n}", "func (o ApplicationOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (d *SourceFilesystem) MakePathRelative(filename string) (string, bool) {\n\tfor _, dir := range d.Dirs {\n\t\tmeta := dir.(hugofs.FileMetaInfo).Meta()\n\t\tcurrentPath := meta.Filename\n\n\t\tif strings.HasPrefix(filename, currentPath) {\n\t\t\trel := strings.TrimPrefix(filename, currentPath)\n\t\t\tif mp := meta.Path; mp != \"\" {\n\t\t\t\trel = filepath.Join(mp, rel)\n\t\t\t}\n\t\t\treturn strings.TrimPrefix(rel, filePathSeparator), true\n\t\t}\n\t}\n\treturn \"\", false\n}", "func (f File) Path() string {\n\treturn string(f)\n}", "func (o IopingSpecVolumeVolumeSourceNfsPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceNfs) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (o LocalCopyResponseOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopyResponse) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o ApplicationStatusOperationStateOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func RelPath(source, path string) string {\n\tsrcDir := filepath.Dir(source)\n\tif srcDir != \".\" {\n\t\treturn filepath.Join(srcDir, path)\n\t}\n\treturn path\n}", "func getFullPath(params *input.Params) string {\n\treturn fmt.Sprintf(FULL_PATH, params.ProjectName)\n}", "func (o ApplicationStatusOperationStateSyncResultSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (metadata EventMetadata) GetPath() (string, error) {\n\tpath, err := os.Readlink(\n\t\tfilepath.Join(\n\t\t\tProcFsFdInfo,\n\t\t\tstrconv.FormatUint(\n\t\t\t\tuint64(metadata.Fd),\n\t\t\t\t10,\n\t\t\t),\n\t\t),\n\t)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"fanotify: %w\", err)\n\t}\n\n\treturn path, nil\n}", "func (o ApplicationStatusOperationStateOperationSyncSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (v *VFile) RealPath() string {\n\treturn filepath.Join(v.BaseDir, v.ShadowDir, v.FileName)\n}", "func realPath(file string) string {\n\tif string(file[0]) == \"/\" {\n\t\treturn file\n\t}\n\n\tif string(file[0]) != \"/\" {\n\t\tfile = \"/\" + file\n\t}\n\n\t_, filename, _, _ := runtime.Caller(3)\n\tdir := path.Join(path.Dir(filename), file)\n\n\tif _, err := os.Stat(dir); err == nil && strings.HasSuffix(dir, file) {\n\t\treturn dir\n\t}\n\n\tcurrent, err := os.Getwd()\n\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tdir = file\n\n\tif strings.HasSuffix(dir, current) {\n\t\treturn dir\n\t}\n\n\treturn current + dir\n}", "func RelativeToMetric(p string) string {\n\tp = path.Clean(p)\n\tp = strings.Replace(p, \".wsp\", \"\", 1)\n\treturn strings.Replace(p, \"/\", \".\", -1)\n}", "func (e *entry) Path() string {\n\treturn e.path\n}", "func (h *HttpLocalSite) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/http/local_site/%s\", h.Reference)\n}", "func (o ApplicationSpecSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationSpecSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (cd ChildData) Path() string {\n\treturn cd.path\n}", "func (p *TempFileProc) FullPath() string {\n\treturn filepath.Join(p.Dir, p.Name)\n}", "func (p *Part) RemotePath(prefix string) string {\n\tfor strings.HasSuffix(prefix, \"/\") {\n\t\tprefix = prefix[:len(prefix)-1]\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%016X_%016X_%016X\", prefix, p.Path, p.FileSize, p.Offset, p.Size)\n}", "func MakeRelative(path, base string) (string, error) {\n\tif len(path) > 0 {\n\t\trel, err := filepath.Rel(base, path)\n\t\tif err != nil {\n\t\t\treturn path, err\n\t\t}\n\t\treturn rel, nil\n\t}\n\treturn path, nil\n}", "func (o ApplicationStatusHistorySourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusHistorySource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}" ]
[ "0.68203026", "0.67226255", "0.668174", "0.66008323", "0.645827", "0.6438866", "0.6362144", "0.6352666", "0.63249975", "0.6230617", "0.6208116", "0.61719966", "0.6137963", "0.6112202", "0.6104346", "0.6102327", "0.6061354", "0.60527205", "0.60439134", "0.60378504", "0.6030129", "0.6008691", "0.60065293", "0.5958633", "0.5950721", "0.5902145", "0.5894866", "0.58674884", "0.58617234", "0.5860046", "0.5845014", "0.58102745", "0.5785807", "0.5781603", "0.5777749", "0.57739884", "0.57718784", "0.57491744", "0.57259893", "0.5708269", "0.5705983", "0.570407", "0.56823915", "0.5677113", "0.56626135", "0.5655148", "0.5642354", "0.5642137", "0.56169784", "0.56123644", "0.56098336", "0.56016695", "0.56013435", "0.5598679", "0.5598679", "0.5582693", "0.5561831", "0.5554911", "0.5545322", "0.55421597", "0.5531991", "0.55303717", "0.5513313", "0.5504977", "0.54968333", "0.54880774", "0.54863024", "0.5479264", "0.5479162", "0.5473482", "0.5468219", "0.5468093", "0.54665303", "0.5461065", "0.54581577", "0.5454401", "0.5453078", "0.54510874", "0.54472506", "0.54463786", "0.5445741", "0.54450035", "0.54311645", "0.54284626", "0.54231215", "0.5420888", "0.54207516", "0.5417104", "0.5415653", "0.5405045", "0.54031146", "0.53941685", "0.5394101", "0.53875154", "0.53863144", "0.53843707", "0.5381394", "0.53800386", "0.53667164", "0.5364135" ]
0.746149
0
GeneratePointer generates a pointer for arbitrary content
func GeneratePointer(content io.Reader) (Pointer, error) { h := sha256.New() c, err := io.Copy(h, content) if err != nil { return Pointer{}, err } sum := h.Sum(nil) return Pointer{Oid: hex.EncodeToString(sum), Size: c}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewPointer(elem Type) *Pointer { return &Pointer{base: elem} }", "func Pointer(v interface{}, q string) Proxy {\n\treturn pointer(New(v), q)\n}", "func Pointer(v int64) *int64 {\n\treturn helpy.Pointer(v)\n}", "func (gen *DataGen) Ptr() unsafe.Pointer {\n\treturn unsafe.Pointer(gen)\n}", "func (d *decoder) pointer(name string, value reflect.Value, def string) error {\n\tif !d.getGlobalProvider().Get(name).HasValue() {\n\t\treturn nil\n\t}\n\n\tif value.IsNil() {\n\t\tvalue.Set(reflect.New(value.Type().Elem()))\n\t}\n\n\treturn d.unmarshal(name, value.Elem(), def)\n}", "func DefinePointer(htype *IHType) Pointer {\n\treturn Pointer{htype: htype}\n}", "func execNewPointer(_ int, p *gop.Context) {\n\targs := p.GetArgs(1)\n\tret := types.NewPointer(args[0].(types.Type))\n\tp.Ret(1, ret)\n}", "func (obj *Value) GetPointer() string {\n\treturn obj.Candy().Guify(\"g_value_get_pointer\", obj).String()\n}", "func (r *snpUserGuestRequestConversion) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(&r.abi)\n}", "func (prg Program) Pointer(ctx gl.Context, a gl.Attrib, size int) {\n\tctx.EnableVertexAttribArray(a)\n\tctx.VertexAttribPointer(a, size, gl.FLOAT, false, 0, 0)\n}", "func TexCoordPointer(size int32, xtype uint32, stride int32, pointer unsafe.Pointer) {\n C.glowTexCoordPointer(gpTexCoordPointer, (C.GLint)(size), (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func NewPointerFromPropertyAttributes(obj PropertyAttributes) Pointer {\n\treturn Pointer{base: unsafe.Pointer(obj.getBase())}\n}", "func (g GoType) PointerMethod() Expr {\n\tif g == GoBytes {\n\t\treturn \"Bytes\"\n\t}\n\treturn Expr(strings.ToUpper(string(g[:1])) + string(g[1:]))\n}", "func (handle *Handle) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(handle.x)\n}", "func (node *SimpleNode) Pointer() string {\n\tif node == nil {\n\t\treturn \"\"\n\t}\n\n\treturn node.pointer\n}", "func (m PinPolicy) Pointer() *PinPolicy {\n\treturn &m\n}", "func TexCoordPointer(size int32, xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tC.glowTexCoordPointer(gpTexCoordPointer, (C.GLint)(size), (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func TestPointerTypes(t *testing.T) {\n\n}", "func NewPointer(conn *websocket.Conn, pointingSession *PointingSession, name string) *Pointer {\n\treturn &Pointer{\n\t\tconn: conn,\n\t\tsend: make(chan []byte),\n\t\t// the pointer has knows about the pointing session this is to BroadCast to everyone in that session\n\t\tpointingSession: pointingSession,\n\t\tName: name,\n\t}\n}", "func createPointer() *float64 {\n\tmyFloat := 98.5\n\treturn &myFloat\n}", "func (r *SnpDerivedKeyReqABI) Pointer() unsafe.Pointer { return unsafe.Pointer(r) }", "func (r *SnpDerivedKeyRespABI) Pointer() unsafe.Pointer { return unsafe.Pointer(r) }", "func (m PartitionUsage) Pointer() *PartitionUsage {\n\treturn &m\n}", "func NewPointerField(size int) []byte {\n\tdata := make([]byte, size+1)\n\tdata[0] = byte(size)\n\tfor i := 1; i < size+1; i++ {\n\t\tdata[i] = 0xFF\n\t}\n\treturn data\n}", "func NewPointerFromMethodArg(obj MethodArg) Pointer {\n\treturn Pointer{base: unsafe.Pointer(obj.getBase())}\n}", "func pointerOf(n interface{}) uintptr {\n\tif n == nil {\n\t\treturn 0\n\t}\n\tv := reflect.ValueOf(n)\n\tif v.IsNil() {\n\t\treturn 0\n\t}\n\treturn v.Pointer()\n}", "func LoadPointer(addr *unsafe.Pointer) (val unsafe.Pointer)", "func (b *Builder) StorePointer(idx value.PointerIndex, ptr value.Pointer) {\n\tb.instructions = append(b.instructions,\n\t\tasm.Push{Value: ptr},\n\t\tasm.Store{Destination: idx},\n\t)\n\trng := memory.Range{\n\t\tBase: uint64(idx) * uint64(b.memoryLayout.GetPointer().GetSize()),\n\t\tSize: uint64(b.memoryLayout.GetPointer().GetSize()),\n\t}\n\tinterval.Merge(&b.pointerMemory, rng.Span(), true)\n}", "func (m EveroutePackageOrderByInput) Pointer() *EveroutePackageOrderByInput {\n\treturn &m\n}", "func (m *Mmap) GetPointer() int64 {\n\treturn m.FilePointer\n}", "func NormalPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n C.glowNormalPointer(gpNormalPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func TexCoordPointer(size int32, xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tsyscall.Syscall6(gpTexCoordPointer, 4, uintptr(size), uintptr(xtype), uintptr(stride), uintptr(pointer), 0, 0)\n}", "func (m DeployOrderByInput) Pointer() *DeployOrderByInput {\n\treturn &m\n}", "func (b *WString) Pointer() *uint16 {\n\tif b.empty() {\n\t\treturn nil\n\t}\n\treturn &b.b[0]\n}", "func NewPtrPageBuffer(aSlice interface{}, desiredPageNo int) *PageBuffer {\n aSliceValue := sliceValue(aSlice, true)\n return newPageBuffer(\n aSliceValue,\n desiredPageNo,\n overwriteNilPtrHandler{\n creater: newCreaterFunc(nil, aSliceValue.Type())})\n}", "func testPointer() string {\n\tvar myVariable int = 10\n\tvar myPointer *int = &myVariable\n\tfmt.Println(*myPointer, myVariable)\n\treturn \"Run out\"\n}", "func (s *slice) Pointer() cu.DevicePtr { return s.ptr_ }", "func DataGenFromPtr(ptr unsafe.Pointer) *DataGen {\n\treturn (*DataGen)(ptr)\n}", "func idPointer(id interface{}) (p *interface{}) {\n\tif id != nil {\n\t\tp = &id\n\t}\n\treturn\n}", "func (m ZoneOrderByInput) Pointer() *ZoneOrderByInput {\n\treturn &m\n}", "func (t *Dense) Pointer() unsafe.Pointer {\n\treturn t.data\n}", "func (r *SnpExtendedReportReqABI) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(r)\n}", "func ReadPointer(reader io.Reader) (Pointer, error) {\n\tbuf := make([]byte, blobSizeCutoff)\n\tn, err := io.ReadFull(reader, buf)\n\tif err != nil && err != io.ErrUnexpectedEOF {\n\t\treturn Pointer{}, err\n\t}\n\tbuf = buf[:n]\n\n\treturn ReadPointerFromBuffer(buf)\n}", "func (w *Writer) Ptr(index int64)", "func getCodePointer(value *reflect.Value) unsafe.Pointer {\n\tp := getInternalPtrFromValue(value)\n\tif p != nil {\n\t\tp = *(*unsafe.Pointer)(p)\n\t}\n\treturn p\n}", "func (obj *Value) SetPointer(v sugar.CandyWrapper) {\n\tobj.Candy().Guify(\"g_value_set_pointer\", obj, v)\n}", "func FogCoordPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tC.glowFogCoordPointer(gpFogCoordPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func (m OrganizationOrderByInput) Pointer() *OrganizationOrderByInput {\n\treturn &m\n}", "func NewPointerFromSignalArgument(obj SignalArgument) Pointer {\n\treturn Pointer{base: unsafe.Pointer(obj.getBase())}\n}", "func returnPointer() *int {\n\ti := 5\n\treturn &i\n}", "func (r *SnpReportReqABI) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(r)\n}", "func FogCoordPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n C.glowFogCoordPointer(gpFogCoordPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func NewPointerButton(text string, onTapped func()) *PointerButton {\n\tbtn := &PointerButton{}\n\tbtn.ExtendBaseWidget(btn)\n\tbtn.Text = text\n\tbtn.OnTapped = onTapped\n\treturn btn\n}", "func (indx AttribLocation) AttribPointerInternal(size uint, t gl.Enum, normalized bool, stride int, offset uintptr) {\n\tgl.VertexAttribPointer(gl.Uint(indx), gl.Int(size), gl.Enum(t), gl.GLBool(normalized), gl.Sizei(stride), gl.Pointer(offset))\n}", "func NormalPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tC.glowNormalPointer(gpNormalPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func (m VipType) Pointer() *VipType {\n\treturn &m\n}", "func IndexPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n C.glowIndexPointer(gpIndexPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func (m IsolationPolicyOrderByInput) Pointer() *IsolationPolicyOrderByInput {\n\treturn &m\n}", "func (c *Compiler) trackPointer(value llvm.Value) {\n\tif value.Type() != c.i8ptrType {\n\t\tvalue = c.builder.CreateBitCast(value, c.i8ptrType, \"\")\n\t}\n\tc.createRuntimeCall(\"trackPointer\", []llvm.Value{value}, \"\")\n}", "func (e *encoder) marshalPointer(t reflect.Type, v reflect.Value, n nestedTypeData) error {\n\tswitch t.Elem().Kind() {\n\tcase reflect.Slice:\n\t\treturn e.marshalVector(t.Elem(), v.Elem(), n)\n\tcase reflect.String:\n\t\treturn e.marshalString(v.Elem(), n)\n\tcase reflect.Struct:\n\t\treturn e.marshalStructOrUnionPointer(t, v)\n\t}\n\treturn newValueError(ErrInvalidPointerType, t.Name())\n}", "func FogCoordPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tsyscall.Syscall(gpFogCoordPointer, 3, uintptr(xtype), uintptr(stride), uintptr(pointer))\n}", "func (*BasicTargetPointerRule) Build(gen Generator, ctx *MethodContext, sourceID *xtype.JenID, source, target *xtype.Type) ([]jen.Code, *xtype.JenID, *Error) {\n\tname := ctx.Name(target.ID())\n\n\tstmt, id, err := gen.Build(ctx, sourceID, source, target.PointerInner)\n\tif err != nil {\n\t\treturn nil, nil, err.Lift(&Path{\n\t\t\tSourceID: \"*\",\n\t\t\tSourceType: source.T.String(),\n\t\t\tTargetID: \"*\",\n\t\t\tTargetType: target.PointerInner.T.String(),\n\t\t})\n\t}\n\tstmt = append(stmt, jen.Id(name).Op(\":=\").Add(id.Code))\n\tnewID := jen.Op(\"&\").Id(name)\n\n\treturn stmt, xtype.OtherID(newID), err\n}", "func (c *Compiler) emitPointerPack(values []llvm.Value) llvm.Value {\n\treturn llvmutil.EmitPointerPack(c.builder, c.mod, c.Config, values)\n}", "func (r *SnpReportRespABI) Pointer() unsafe.Pointer {\n\treturn unsafe.Pointer(r)\n}", "func VertexPointer(size int32, xtype uint32, stride int32, pointer unsafe.Pointer) {\n C.glowVertexPointer(gpVertexPointer, (C.GLint)(size), (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func StructPtr(rt reflect.Type, gens map[string]gopter.Gen) gopter.Gen {\n\tif rt.Kind() == reflect.Ptr {\n\t\trt = rt.Elem()\n\t}\n\n\tbuildPtrType := reflect.FuncOf([]reflect.Type{rt}, []reflect.Type{reflect.PtrTo(rt)}, false)\n\tunbuildPtrType := reflect.FuncOf([]reflect.Type{reflect.PtrTo(rt)}, []reflect.Type{rt}, false)\n\n\tbuildPtrFunc := reflect.MakeFunc(buildPtrType, func(args []reflect.Value) []reflect.Value {\n\t\tsp := reflect.New(rt)\n\t\tsp.Elem().Set(args[0])\n\t\treturn []reflect.Value{sp}\n\t})\n\tunbuildPtrFunc := reflect.MakeFunc(unbuildPtrType, func(args []reflect.Value) []reflect.Value {\n\t\treturn []reflect.Value{args[0].Elem()}\n\t})\n\n\treturn gopter.DeriveGen(\n\t\tbuildPtrFunc.Interface(),\n\t\tunbuildPtrFunc.Interface(),\n\t\tStruct(rt, gens),\n\t)\n}", "func getPointer(n *int) {\n *n = *n * *n\n}", "func (m IPType) Pointer() *IPType {\n\treturn &m\n}", "func (m VMFirmware) Pointer() *VMFirmware {\n\treturn &m\n}", "func NewPointer() *link {\n\treturn new(link)\n}", "func GetPointerv(pname uint32, params *unsafe.Pointer) {\n C.glowGetPointerv(gpGetPointerv, (C.GLenum)(pname), params)\n}", "func generatePtrs(structFields []*r.Field) []interface{} {\n\tptrs := make([]interface{}, len(structFields))\n\tfor i, sf := range structFields {\n\t\tif sf == nil {\n\t\t\tptrs[i] = &ignoreFieldType{}\n\t\t\tcontinue\n\t\t}\n\n\t\tval := reflect.New(sf.Type())\n\t\tptrs[i] = val.Interface()\n\t}\n\treturn ptrs\n}", "func VertexAttribPointer(index uint32, size int32, xtype uint32, normalized bool, stride int32, pointer unsafe.Pointer) {\n C.glowVertexAttribPointer(gpVertexAttribPointer, (C.GLuint)(index), (C.GLint)(size), (C.GLenum)(xtype), (C.GLboolean)(boolToInt(normalized)), (C.GLsizei)(stride), pointer)\n}", "func DBPointer(ns string, ptr primitive.ObjectID) Val {\n\tdbptr := primitive.DBPointer{DB: ns, Pointer: ptr}\n\treturn Val{t: bsontype.DBPointer, primitive: dbptr}\n}", "func (image *Image2D) GetDataPointer() unsafe.Pointer {\n\treturn gl.Ptr(image.data)\n}", "func (v *ApiVisitor) VisitPointerType(ctx *api.PointerTypeContext) interface{} {\n\tnameExpr := v.newExprWithTerminalNode(ctx.ID())\n\treturn &Pointer{\n\t\tPointerExpr: v.newExprWithText(ctx.GetText(), ctx.GetStar().GetLine(), ctx.GetStar().GetColumn(), ctx.GetStar().GetStart(), ctx.ID().GetSymbol().GetStop()),\n\t\tStar: v.newExprWithToken(ctx.GetStar()),\n\t\tName: nameExpr,\n\t}\n}", "func NewPointerFromSignal(obj Signal) Pointer {\n\treturn Pointer{base: unsafe.Pointer(obj.getBase())}\n}", "func FromPointer[T any, TT *T](val *T) (zero T) {\n\tif val == nil {\n\t\treturn zero\n\t}\n\treturn *val\n}", "func (m VMVolumeOrderByInput) Pointer() *VMVolumeOrderByInput {\n\treturn &m\n}", "func NewPtrBuffer(aSlice interface{}) *Buffer {\n aSliceValue := sliceValue(aSlice, true)\n return &Buffer{\n buffer: aSliceValue,\n handler: overwriteNilPtrHandler{\n creater: newCreaterFunc(nil, aSliceValue.Type())}}\n}", "func ReadPointerFromBuffer(buf []byte) (Pointer, error) {\n\tvar p Pointer\n\n\theadString := string(buf)\n\tif !strings.HasPrefix(headString, MetaFileIdentifier) {\n\t\treturn p, ErrMissingPrefix\n\t}\n\n\tsplitLines := strings.Split(headString, \"\\n\")\n\tif len(splitLines) < 3 {\n\t\treturn p, ErrInvalidStructure\n\t}\n\n\toid := strings.TrimPrefix(splitLines[1], MetaFileOidPrefix)\n\tif len(oid) != 64 || !oidPattern.MatchString(oid) {\n\t\treturn p, ErrInvalidOIDFormat\n\t}\n\tsize, err := strconv.ParseInt(strings.TrimPrefix(splitLines[2], \"size \"), 10, 64)\n\tif err != nil {\n\t\treturn p, err\n\t}\n\n\tp.Oid = oid\n\tp.Size = size\n\n\treturn p, nil\n}", "func printHexPtr(w io.Writer, p uintptr, isPointer bool) {\n\t// Null pointer.\n\tnum := uint64(p)\n\tif num == 0 {\n\t\tif isPointer {\n\t\t\tw.Write(nilBytes)\n\t\t} else {\n\t\t\tw.Write(zeroBytes)\n\t\t}\n\t\treturn\n\t}\n\n\t// Max uint64 is 16 bytes in hex + 2 bytes for '0x' prefix\n\tbuf := make([]byte, 18)\n\n\t// It's simpler to construct the hex string right to left.\n\tbase := uint64(16)\n\ti := len(buf) - 1\n\tfor num >= base {\n\t\tbuf[i] = hexDigits[num%base]\n\t\tnum /= base\n\t\ti--\n\t}\n\tbuf[i] = hexDigits[num]\n\n\t// Add '0x' prefix.\n\ti--\n\tbuf[i] = 'x'\n\ti--\n\tbuf[i] = '0'\n\n\t// Strip unused leading bytes.\n\tbuf = buf[i:]\n\tw.Write(buf)\n}", "func (mp *Mempool) Ptr() unsafe.Pointer {\n\treturn unsafe.Pointer(mp)\n}", "func fun_with_pointers() {\n var x int;\n var ptr *int;\n x = 7\n ptr = &x\n access_pointer(ptr)\n access_pointer(ptr)\n fmt.Printf(\"Now the value of x is %d\\n\", x)\n}", "func IsPointer(a cc.Type) bool { return a.RawDeclarator().PointerOpt != nil }", "func GetPointerv(pname uint32, params *unsafe.Pointer) {\n\tsyscall.Syscall(gpGetPointerv, 2, uintptr(pname), uintptr(unsafe.Pointer(params)), 0)\n}", "func IndexPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tC.glowIndexPointer(gpIndexPointer, (C.GLenum)(xtype), (C.GLsizei)(stride), pointer)\n}", "func (m GlobalSettingsOrderByInput) Pointer() *GlobalSettingsOrderByInput {\n\treturn &m\n}", "func (native *OpenGL) PtrOffset(offset int) unsafe.Pointer {\n\treturn gl.PtrOffset(offset)\n}", "func (recv *Value) GetPointer() uintptr {\n\tretC := C.g_value_get_pointer((*C.GValue)(recv.native))\n\tretGo := (uintptr)(unsafe.Pointer(retC))\n\n\treturn retGo\n}", "func AddressGenerator(t *rapid.T) *rapid.Generator[sdk.AccAddress] {\n\treturn rapid.Custom(func(t *rapid.T) sdk.AccAddress {\n\t\tpkBz := rapid.SliceOfN(rapid.Byte(), 20, 20).Draw(t, \"hex\")\n\t\treturn sdk.AccAddress(pkBz)\n\t})\n}", "func (k *key) FilePointer() string {\n\treturn k.GID.String()\n}", "func (s *ScaleInfo) GetPointer() int {\n\treturn s.pointer\n}", "func (p *Pointer) Format() error {\n\treturn nil\n}", "func GetPointerv(pname uint32, params *unsafe.Pointer) {\n\tC.glowGetPointerv(gpGetPointerv, (C.GLenum)(pname), params)\n}", "func GetPointerv(pname uint32, params *unsafe.Pointer) {\n\tC.glowGetPointerv(gpGetPointerv, (C.GLenum)(pname), params)\n}", "func InvokePointer() {\n\ti := 1\n\tfmt.Println(\"initial\", i)\n\n\tzeroval(i)\n\tfmt.Println(\"zeroval\", i)\n\n\tzeroptr(&i)\n\tfmt.Println(\"zeroptr\", i)\n\n\tfmt.Println(\"pointer\", &i)\n}", "func NormalPointer(xtype uint32, stride int32, pointer unsafe.Pointer) {\n\tsyscall.Syscall(gpNormalPointer, 3, uintptr(xtype), uintptr(stride), uintptr(pointer))\n}", "func Int64Pointer(i int64) *int64 {\n\treturn &i\n}", "func Int64Pointer(i int64) *int64 {\n\treturn &i\n}" ]
[ "0.61750895", "0.60294455", "0.6029336", "0.6004933", "0.5995309", "0.5967862", "0.59285", "0.58407485", "0.5783777", "0.5644884", "0.5644425", "0.5633311", "0.56280696", "0.56253093", "0.5614537", "0.55359113", "0.5499952", "0.54650974", "0.5436212", "0.54254496", "0.5418415", "0.5400243", "0.5398698", "0.53781486", "0.53740406", "0.5366316", "0.53324884", "0.5304847", "0.5292536", "0.52680767", "0.5261931", "0.5260046", "0.52380943", "0.52234715", "0.5216309", "0.52108014", "0.52097535", "0.52093357", "0.520444", "0.51842725", "0.5183319", "0.5170408", "0.5161835", "0.5160817", "0.5147184", "0.5144457", "0.5142941", "0.5136697", "0.5136256", "0.51257646", "0.51257086", "0.512033", "0.5101667", "0.50883347", "0.5086012", "0.5080601", "0.50774515", "0.5073447", "0.5072786", "0.507199", "0.5031967", "0.5025432", "0.5019924", "0.5019596", "0.5013432", "0.50071055", "0.5006841", "0.500021", "0.49961722", "0.49949858", "0.49929142", "0.49867284", "0.49825215", "0.49789035", "0.4976976", "0.49733603", "0.4972359", "0.49688703", "0.49580446", "0.4955996", "0.49409223", "0.49314928", "0.491967", "0.49183148", "0.49126187", "0.4911914", "0.4911607", "0.4911047", "0.4902047", "0.48874974", "0.4885518", "0.4884038", "0.4883145", "0.48813802", "0.48808563", "0.48808563", "0.48793367", "0.48696578", "0.48635936", "0.48635936" ]
0.8024975
0
Prepare is used by the Fake Service to prepare the placeholders for the input data.
func (fs fakeService) Prepare(ctx *context.T, method string, numArgs int) (argptrs []interface{}, tags []*vdl.Value, _ error) { inargs := make([]*vom.RawBytes, numArgs) inptrs := make([]interface{}, len(inargs)) for i := range inargs { inptrs[i] = &inargs[i] } return inptrs, nil, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *PetsEquipmentsetEntries) Prepare() {\n}", "func (c *Command) Prepare() {}", "func (s *Store) Prepare() {\n\ts.Name = strings.TrimSpace(s.Name)\n\ts.Type = strings.TrimSpace(s.Type)\n}", "func (d *DataBuckets) Prepare() {\n}", "func (c *jsiiProxy_CfnInput) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (e *engineImpl) Prepare(chain engine.ChainReader, header *block.Header) error {\n\t// TODO: implement prepare method\n\treturn nil\n}", "func (c *Controller) Prepare() {}", "func (s *SpellBuckets) Prepare() {\n}", "func (r *Response) Prepare(g *Gym) {\n\tr.InjectServiceContainer(g)\n}", "func (dbi *DB) Prepare(sqls string, arg ...string) {\r\n\tif dbi.status == false {\r\n\t\treturn\r\n\t}\r\n\r\n\tdbi.createOperation(\"DB_PREPARE\")\r\n\t// bind variables\r\n\tdbi.data.reqSql = sqls\r\n\tdbi.data.inVar = arg\r\n\t// data\r\n\tdbi.data.commPrepare()\r\n\t// communicate\r\n\tif dbi.data.comm() == false {\r\n\t\tdbi.Close()\r\n\t}\r\n\t// parse\r\n\tdbi.data.commParse()\r\n}", "func (o *ObjectContents) Prepare() {\n}", "func (p *Petitions) Prepare() {\n}", "func Prepare() error {\n\n\t// log.Println(\"Preparing work...\")\n\n\t// commands := [][]string{\n\t// \t[]string{\"yum\", \"update\", \"-y\"},\n\t// \t[]string{\"yum\", \"install\", \"-y\", \"docker\"},\n\t// \t[]string{\"service\", \"docker\", \"start\"},\n\t// \t[]string{\"docker\", \"pull\", \"tnolet/scraper:0.1.0\"},\n\t// }\n\n\t// for _, command := range commands {\n\t// \tout, err := exec.Command(command).Output()\n\n\t// \tif err != nil {\n\t// \t\tlog.Printf(\"Prepare command unsuccessful: %v, %v\", err.Error(), out)\n\t// \t\treturn err\n\t// \t}\n\n\t// \tlog.Printf(\"Succesfully executed preparation: %v\", out)\n\t// }\n\treturn nil\n\n}", "func (c *Controller) Prepare() {\n\n}", "func (s ArticleService) Prepare(ctx context.Context) {\n\tstat := `CREATE TABLE articles (id BIGSERIAL NOT NULL PRIMARY KEY, title TEXT, description TEXT, content TEXT);`\n\tif s.DB == nil {\n\t\tpanic(\"no existing database\")\n\t}\n\tif _, err := s.DB.ExecContext(ctx, stat); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (p *Project) Prepare() {\n\tp.ProjectName = p.GetName()\n\tp.Status = p.GetStatus()\n}", "func (t *Trader) Prepare() {\n}", "func (c *CreateConfig) Prepare() []error {\n\tvar errs []error\n\n\tif c.VMName == \"\" {\n\t\terrs = append(errs, fmt.Errorf(\"Target VM name is required\"))\n\t}\n\n\treturn errs\n}", "func (user *User) Prepare(step string) error {\n\tif err := user.validate(step); err != nil {\n\t\treturn err\n\t}\n\n\tif err := user.format(step); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (t *TRIAGEZIP) Prepare() {\n}", "func (bs *BaseController) Prepare() {\n\tauthorization := bs.Ctx.Request.Header.Get(\"authorization\")\n\tif authorization == \"\" {\n\t\tglog.Error(\"need authorization in header\")\n\t\treturn\n\t}\n\tif authorization != bs.tokenFactory() {\n\t\tglog.Error(\"no authorized\")\n\t\treturn\n\t}\n\tbs.UserName = bs.ConfUserName()\n\tbs.Token = bs.ConfToken()\n\treturn\n}", "func (r *jsiiProxy_Repository) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tr,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (kf *HybridKF) Prepare(Φ, Htilde *mat64.Dense) {\n\tkf.Φ = Φ\n\tkf.Htilde = Htilde\n\tkf.locked = false\n}", "func (b *BaseBuild) Prepare() error {\n\treturn nil\n}", "func (l *Lfguild) Prepare() {\n}", "func (f *FactionValues) Prepare() {\n}", "func (b *BOOKINGTRIAGETRACKER) Prepare() {\n}", "func (l *Loottable) Prepare() {\n}", "func (c *SeaterController) Prepare() {\n\tmodel, err := models.NewModel()\n\tif err != nil {\n\t\tc.TraceServerError(errors.Annotatef(err, \"failed to init model\"))\n\t}\n\tif err = model.Begin(); err != nil {\n\t\tc.TraceServerError(errors.Annotatef(err, \"failed to begin database transaction\"))\n\t}\n\tc.model = model\n\tc.orm = model.Orm()\n\tc.pagingResult = models.NewQueryParams()\n}", "func (c *BigQueryClient) Prepare(ctx context.Context, fields []FieldSchema) error {\n\tlog.From(ctx).Debug(\"creating dataset\")\n\tif err := c.CreateDataset(ctx); err != nil {\n\t\tlog.From(ctx).Error(\"creating dataset\", zap.Error(err))\n\t\treturn err\n\t}\n\n\tlog.From(ctx).Debug(\"creating table\")\n\tif err := c.CreateTable(ctx, BigQuerySchema(fields)); err != nil {\n\t\tlog.From(ctx).Error(\"creating table\", zap.Error(err))\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *jsiiProxy_CfnDetector) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func Prepare(in io.Reader, out io.Writer) error {\n\n\treader := bufio.NewReader(in)\n\n\tsep := byte('\\n') // note that double-quotes will _freak_ Go out...\n\n\tline_number := 0\n\n\tfor {\n\n\t\tb, err := reader.ReadBytes(sep)\n\n\t\tif err != nil {\n\n\t\t\tif err == io.EOF {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\treturn err\n\t\t}\n\n\t\tline_number += 1\n\n\t\tok, err := HasRequiredProperties(b)\n\n\t\tif !ok {\n\n\t\t\t// log.Printf(\"%s at line number %d\\n\", err, line_number)\n\n\t\t\tif IsMissingStreet(err) {\n\n\t\t\t\tb, err = EnsureStreet(b)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\t// log.Printf(\"%s at line number %d\\n\", err, line_number)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif IsMissingHouseNumber(err) {\n\n\t\t\t\tb, err = EnsureHouseNumber(b)\n\n\t\t\t\tif err != nil {\n\t\t\t\t\t// log.Printf(\"%s at line number %d\\n\", err, line_number)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif err != nil {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\tb2, err := EnstringifyProperties(b)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tout.Write(b2)\n\t}\n\n\treturn nil\n}", "func (h *Healthz) Prepare() error {\n\tif err := h.createBucketIfNotExist(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\n\tif err := h.createObject(); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\n\treturn nil\n}", "func (r *jsiiProxy_RepositoryBase) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tr,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (h *HTTPTester) Prepare(stmt string, bSz int, tx bool) error {\n\ts := make([]string, bSz)\n\tfor i := 0; i < len(s); i++ {\n\t\ts[i] = stmt\n\t}\n\n\tb, err := json.Marshal(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\th.br = bytes.NewReader(b)\n\n\tif tx {\n\t\th.url = h.url + \"?transaction\"\n\t}\n\n\treturn nil\n}", "func (s *StateDB) Prepare(thash, bhash corecrypto.HashType, ti int) {\n\ts.thash = thash\n\ts.bhash = bhash\n\ts.txIndex = ti\n}", "func Prepare(p Preparer, query string) (st *sql.Stmt, err error) {\n\tst, err = p.Prepare(query)\n\terr = interpretError(err)\n\treturn\n}", "func (u *User) Prepare() {\n\tu.Email = strings.TrimSpace(u.Email)\n\tu.FirstName = strings.TrimSpace(u.LastName)\n}", "func (p PreparerFunc) Prepare(r *http.Request) (*http.Request, error) {\n\treturn p(r)\n}", "func (c *jsiiProxy_CfnStudio) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (sa *ScanAPI) Prepare() {\n\t// Call super prepare method\n\tsa.BaseController.Prepare()\n\n\t// Parse parameters\n\trepoName := sa.GetString(\":splat\")\n\ttag := sa.GetString(\":tag\")\n\tprojectName, _ := utils.ParseRepository(repoName)\n\n\tpro, err := sa.ProjectMgr.Get(projectName)\n\tif err != nil {\n\t\tsa.SendInternalServerError(errors.Wrap(err, \"scan API: prepare\"))\n\t\treturn\n\t}\n\tif pro == nil {\n\t\tsa.SendNotFoundError(errors.Errorf(\"project %s not found\", projectName))\n\t\treturn\n\t}\n\tsa.pro = pro\n\n\t// Check authentication\n\tif !sa.RequireAuthenticated() {\n\t\treturn\n\t}\n\n\t// Assemble artifact object\n\tdigest, err := digestFunc(repoName, tag, sa.SecurityCtx.GetUsername())\n\tif err != nil {\n\t\tsa.SendInternalServerError(errors.Wrap(err, \"scan API: prepare\"))\n\t\treturn\n\t}\n\n\tsa.artifact = &v1.Artifact{\n\t\tNamespaceID: pro.ProjectID,\n\t\tRepository: repoName,\n\t\tTag: tag,\n\t\tDigest: digest,\n\t\tMimeType: v1.MimeTypeDockerArtifact,\n\t}\n\n\tlogger.Debugf(\"Scan API receives artifact: %#v\", sa.artifact)\n}", "func (c *CharacterBind) Prepare() {\n}", "func (c *jsiiProxy_CfnRepository) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnEnvironment) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (px *Paxos) Prepare(args *PrepareArgs, reply *PrepareReply) error {\n\tseq, n, from, done := args.Seq, args.N, args.FromPeer, args.Done\n\n\tpx.mu.Lock()\n\tdefer px.mu.Unlock()\n\n\tif px.instances[seq] == nil {\n\t\tpx.CreateInstance(seq)\n\t}\n\tif px.instances[seq].decided {\n\t\treply.Err = ErrDecided\n\t\treply.Na = px.instances[seq].na\n\t\treply.Va = px.instances[seq].va\n\t} else if n > px.instances[seq].np {\n\t\tpx.instances[seq].np = n\n\t\treply.Err = OK\n\t\treply.N = n\n\t\treply.Na = px.instances[seq].na\n\t\treply.Va = px.instances[seq].va\n\t} else {\n\t\treply.Err = ErrPrepareRejected\n\t\t//used to help proposer find new highest np\n\t\treply.Np = px.instances[seq].np\n\t}\n\n\tpx.UpdatePeersDone(from, done)\n\treply.Done = px.peersDone[px.me]\n\tpx.ForgetDone()\n\treturn nil\n}", "func (consensus *Consensus) Prepare(chain ChainReader, header *types.Header) error {\n\t// TODO: implement prepare method\n\treturn nil\n}", "func (c *jsiiProxy_CfnDetectorModel) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnLayer) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnFilter) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnDataflowEndpointGroup) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (test *PrepareAtDifferentRound) Prepare(t *testing.T) {\n\ttest.lambda = []byte{1, 2, 3, 4}\n\ttest.prevLambda = []byte{0, 0, 0, 0}\n\ttest.inputValue = spectesting.TestInputValue()\n\n\ttest.instance = spectesting.TestIBFTInstance(t, test.lambda, test.prevLambda)\n\ttest.instance.State.Round = 1\n\n\t// load messages to queue\n\tfor _, msg := range test.MessagesSequence(t) {\n\t\ttest.instance.MsgQueue.AddMessage(&network.Message{\n\t\t\tLambda: test.lambda,\n\t\t\tMsg: msg,\n\t\t\tType: network.IBFTBroadcastingType,\n\t\t})\n\t}\n}", "func (i *Inventory) Prepare() {\n}", "func (c *jsiiProxy_CfnPreset) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnConfig) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (f Factory) Prepare(clientCtx client.Context) (Factory, error) {\n\tif clientCtx.Offline {\n\t\treturn f, nil\n\t}\n\n\tfc := f\n\tfrom := clientCtx.GetFromAddress()\n\n\tif err := fc.accountRetriever.EnsureExists(clientCtx, from); err != nil {\n\t\treturn fc, err\n\t}\n\n\tinitNum, initSeq := fc.accountNumber, fc.sequence\n\tif initNum == 0 || initSeq == 0 {\n\t\tnum, seq, err := fc.accountRetriever.GetAccountNumberSequence(clientCtx, from)\n\t\tif err != nil {\n\t\t\treturn fc, err\n\t\t}\n\n\t\tif initNum == 0 {\n\t\t\tfc = fc.WithAccountNumber(num)\n\t\t}\n\n\t\tif initSeq == 0 {\n\t\t\tfc = fc.WithSequence(seq)\n\t\t}\n\t}\n\n\treturn fc, nil\n}", "func (d *Download) Prepare(logger *log.Entry, group, secretType, vaultPath string, data SecretData) error {\n\tvar keyName string\n\tvar payload []byte\n\tvar err error\n\n\tvaultPath = d.vault.composePath(data, vaultPath)\n\tlogger = logger.WithField(\"vaultPath\", vaultPath)\n\n\tif data.Key != \"\" {\n\t\tkeyName = data.Key\n\t} else {\n\t\tkeyName = data.Name\n\t}\n\n\tlogger.Infof(\"Reading data from Vault, key '%s'\", keyName)\n\tif payload, err = d.vault.Read(vaultPath, keyName); err != nil {\n\t\treturn err\n\t}\n\n\tlogger.Info(\"Creating a file instance\")\n\tfile := NewFile(group, secretType, &data, payload)\n\tif data.Zip {\n\t\tif err = file.Unzip(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\td.Files = append(d.Files, file)\n\treturn nil\n}", "func (c *jsiiProxy_CfnInstance) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (vc *VehicleContainer) Prepare(apiEntries []*APIVehicleEntry) error {\n\ttimezone, err := time.LoadLocation(\"Europe/Warsaw\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvc.Vehicles = make(map[string]*Vehicle, len(apiEntries))\n\n\tfor _, ae := range apiEntries {\n\t\tv, err := NewVehicle(ae, timezone)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tvc.Vehicles[v.ID] = v\n\t}\n\n\treturn nil\n}", "func (sb *spdkBackend) Prepare(req storage.BdevPrepareRequest) (*storage.BdevPrepareResponse, error) {\n\tsb.log.Debugf(\"spdk backend prepare (script call): %+v\", req)\n\treturn sb.prepare(req, DetectVMD, cleanHugepages)\n}", "func (px *Paxos) Prepare(args PrepareArgs, reply *PrepareReply) error {\n\tpx.mu.Lock()\n\tdefer px.mu.Unlock()\n\n\tvar pi PaxosInstance\n\tif pii, ok := px.instances[args.Seq]; ok {\n\t\tpi = pii\n\t} else {\n\t\tpi = PaxosInstance{-1, -1, nil, false}\n\t}\n\n\t// acceptor's prepare(n) handler:\n\t// if n > n_p\n\tif args.N > pi.N_p {\n\t\t// n_p = n\n\t\tinstance := PaxosInstance{args.N, pi.N_a, pi.V_a, pi.Decided}\n\t\tpx.updatePaxos(args.Seq, instance)\n\n\t\t// reply prepare_ok(n_a, v_a)\n\t\treply.N = pi.N_a\n\t\treply.V = pi.V_a\n\t} else {\n\t\t// else reply prepare_reject\n\t\treply.Reject = true\n\t}\n\n\treturn nil\n}", "func (c *Message) Prepare() {\n\tc.ID = 0\n\tc.Body = html.EscapeString(strings.TrimSpace(c.Body))\n\tc.CreatedAt = time.Now()\n\tc.UpdatedAt = time.Now()\n}", "func Preparedata(s string) string {\n\treturn \"Preparedata for \" + s\n}", "func (m *ManagerClient) Prepare(method, url string, body io.Reader) (*http.Request, error) {\n\tr, err := http.NewRequest(method, m.buildURI(url), body)\n\n\tif err == nil {\n\t\tr.Header.Add(\"User-Agent\", UserAgent)\n\t\tr.Header.Add(\"Api-Key\", m.key)\n\t}\n\treturn r, err\n}", "func (g *Grid) Prepare() {\n}", "func (controller *Controller) Prepare() {\n\tcontroller.Controller.Prepare()\n\tcontroller.SetTemplatePath(\"admin/developers/data\")\n\tcontroller.loadTable()\n}", "func (b *Builder) Prepare(args ...interface{}) ([]string, []string, error) {\n\tvar (\n\t\terrs *packer.MultiError\n\t\twarnings []string\n\t)\n\n\tif err := config.Decode(&b.config, &config.DecodeOpts{\n\t\tInterpolate: true,\n\t\tInterpolateFilter: &interpolate.RenderFilter{},\n\t}, args...); err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfileWarns, fileErrs := b.InitConfig(&b.config.ctx)\n\twarnings = append(fileWarns, fileWarns...)\n\terrs = packer.MultiErrorAppend(errs, fileErrs...)\n\n\tif errs != nil && len(errs.Errors) > 0 {\n\t\treturn nil, warnings, errs\n\t}\n\n\treturn nil, warnings, nil\n}", "func (c *jsiiProxy_CfnIPSet) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (db *TestDB) Prepare(query string) (gorpish.IStmt, error) {\n\targs := db.Called(query)\n\treturn args.Get(0).(gorpish.IStmt), args.Error(1)\n}", "func (t *transaction) prepare() {\n\tm := t.Mock\n\tr := t.Request\n\n\tuc, uo, err := mockURL(m, r)\n\tif err != nil {\n\t\tm.Testing.Fatal(err)\n\t}\n\tt.url = uo\n\tr.URL = uc\n\n\tif r.Body != nil {\n\t\tb, err := copyBody(r.Body)\n\t\tif err != nil {\n\t\t\tm.Testing.Fatal(err)\n\t\t}\n\n\t\tt.body = b\n\t\tr.Body = ioutil.NopCloser(b)\n\t}\n}", "func (a *basicAuthentication) Prepare(ctx context.Context, conn driver.Connection) error {\n\t// No need to do anything here\n\treturn nil\n}", "func (in *InBuffer) Prepare(size int) {\n\tif cap(in.Data) < size {\n\t\tin.Data = make([]byte, size)\n\t} else {\n\t\tin.Data = in.Data[0:size]\n\t}\n}", "func (action *ActionExportCreate) Prepare() *ActionExportCreateInvocation {\n\treturn &ActionExportCreateInvocation{\n\t\tAction: action,\n\t\tPath: \"/v6.0/exports\",\n\t}\n}", "func (i *Installer) Prepare(d Device) error {\n\t// Sanity check inputs.\n\tif i.config == nil {\n\t\treturn fmt.Errorf(\"installer missing config: %w\", errConfig)\n\t}\n\tif i.config.ImageFile() == \"\" {\n\t\treturn fmt.Errorf(\"missing image: %w\", errInput)\n\t}\n\text := regExFileExt.FindString(i.config.ImageFile())\n\tif ext == \"\" {\n\t\treturn fmt.Errorf(\"could not find extension for %q: %w\", i.config.ImageFile(), errFile)\n\t}\n\tf, err := os.Stat(filepath.Join(i.cache, i.config.ImageFile()))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%v: %w\", err, errPath)\n\t}\n\t// Compensate for very small image files that can cause the wrong partition\n\t// to be selected.\n\tsize := uint64(f.Size())\n\tif size < oneGB {\n\t\tsize = oneGB\n\t}\n\t// Prepare the devices for provisioning.\n\tswitch {\n\tcase ext == \".iso\" && i.config.UpdateOnly():\n\t\treturn i.prepareForISOWithoutElevation(d, size)\n\tcase ext == \".iso\":\n\t\treturn i.prepareForISOWithElevation(d, size)\n\tcase ext == \".img\":\n\t\treturn i.prepareForRaw(d)\n\t}\n\treturn fmt.Errorf(\"%q is not a supported image type: %w\", ext, errProvision)\n}", "func (v ValidHTML) Prepare() error {\n\t_, err := ignoreMask(v.Ignore)\n\treturn err\n}", "func (i *Image) Prepare(ctx context.Context, c *cache) error {\n\tif i.u == nil {\n\t\treturn nil\n\t}\n\terr := downloadData(ctx, i.u, i.decomp, c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ti.p = c.Path(i.u.String())\n\treturn nil\n}", "func (s *Msg) Prepare(queryPrepared string) error {\n\ts.QueryPrepared = OrString(queryPrepared, s.QueryPrepared)\n\treturn s.CallDtm(&s.MsgData, \"prepare\")\n}", "func (ngram *nGram) Prepare(db *Database) error {\n\treturn nil\n}", "func (action *ActionLocationNetworkCreate) Prepare() *ActionLocationNetworkCreateInvocation {\n\treturn &ActionLocationNetworkCreateInvocation{\n\t\tAction: action,\n\t\tPath: \"/v6.0/location_networks\",\n\t}\n}", "func (p *Player) Prepare(outcomeAt time.Time, config *Config) error {\n\tfor outcomeAt.Sub(p.fixedAt) > config.RatingPeriod {\n\t\tif err := p.estimated.Fix(config.Tau); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tp.fixedAt = p.fixedAt.Add(config.RatingPeriod)\n\t}\n\treturn nil\n}", "func (b *bnkd) Prepare(p backend.Payload) error {\n\tlog.Printf(\"prepare payload bytes: %s\\n\", string(p.Bytes()))\n\n\trepos, err := parseRepos(p.Type(), p.Bytes())\n\tif err != nil {\n\t\treturn errors.New(\"error unmarshalling installation event: \" + err.Error())\n\t}\n\n\tlog.Printf(\"repositories: %+v\\n\", repos)\n\tfor _, repo := range repos {\n\t\tlog.Printf(\"repository: %s\\n\", *repo.FullName)\n\t\tfullName := strings.Split(*repo.FullName, \"/\")\n\t\tpullRequests, err := b.help.pullRequests(b.client, fullName[0], fullName[1])\n\t\tif err != nil {\n\t\t\treturn errors.New(\"error getting pull requests: \" + err.Error())\n\t\t}\n\n\t\tlog.Printf(\"pull requests: %+v\\n\", pullRequests)\n\t\tfor _, pr := range pullRequests {\n\t\t\tlog.Printf(\"pull request: %+v\\n\", pr)\n\t\t\tclosed := pr.ClosedAt\n\t\t\tmerged := *pr.Merged\n\t\t\tlog.Printf(\"closed: %s, merged: %t\\n\", closed, merged)\n\t\t\tif closed != nil && merged {\n\t\t\t\tif err := b.help.comment(b.client, fullName[0], fullName[1], pr); err != nil {\n\t\t\t\t\treturn errors.New(\"error posting comment: \" + err.Error())\n\t\t\t\t}\n\t\t\t\tlog.Println(\"posting pull request comment\")\n\t\t\t}\n\t\t}\n\t}\n\n\tlog.Println(\"successful prepare invocation\")\n\treturn nil\n}", "func (c *jsiiProxy_CfnMaster) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func Prepare(f func(Initiator)) {\n\tprepares = append(prepares, f)\n}", "func (c *jsiiProxy_CfnPublicRepository) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (p AzurePKEClusterCreationParamsPreparer) Prepare(ctx context.Context, params *AzurePKEClusterCreationParams) error {\n\tif params.Name == \"\" {\n\t\treturn validationErrorf(\"Name cannot be empty\")\n\t}\n\tif params.OrganizationID == 0 {\n\t\treturn validationErrorf(\"OrganizationID cannot be 0\")\n\t}\n\t// TODO check org exists\n\t// TODO check creator user exists if present\n\tif params.SecretID == \"\" {\n\t\treturn validationErrorf(\"SecretID cannot be empty\")\n\t}\n\t// TODO validate secret ID\n\t// TODO validate SSH secret ID if present\n\n\tif params.ResourceGroup == \"\" {\n\t\tparams.ResourceGroup = fmt.Sprintf(\"%s-rg\", params.Name)\n\t\tp.logger.Debugf(\"ResourceGroup not specified, defaulting to [%s]\", params.ResourceGroup)\n\t}\n\n\tif err := p.k8sPreparer.Prepare(&params.Kubernetes); err != nil {\n\t\treturn emperror.Wrap(err, \"failed to prepare k8s network\")\n\t}\n\n\tsir, err := secret.Store.Get(params.OrganizationID, params.SecretID)\n\tif err != nil {\n\t\treturn emperror.Wrap(err, \"failed to fetch secret from store\")\n\t}\n\tcc, err := azure.NewCloudConnection(&autoazure.PublicCloud, azure.NewCredentials(sir.Values))\n\tif err != nil {\n\t\treturn emperror.Wrap(err, \"failed to create Azure cloud connection\")\n\t}\n\tif err := p.getVNetPreparer(cc, params.Name, params.ResourceGroup).Prepare(ctx, &params.Network); err != nil {\n\t\treturn emperror.Wrap(err, \"failed to prepare cluster network\")\n\t}\n\n\tif err := p.nodePoolsPreparer.Prepare(params.NodePools); err != nil {\n\t\treturn emperror.Wrap(err, \"failed to prepare node pools\")\n\t}\n\n\treturn nil\n}", "func (p *Preparer) Prepare(ctx context.Context, render resource.Renderer) (resource.Task, error) {\n\tif strings.TrimSpace(p.Source) == \"\" {\n\t\treturn nil, errors.New(\"\\\"source\\\" must contain a value\")\n\t}\n\n\t_, err := url.Parse(p.Source)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to parse \\\"source\\\"\")\n\t}\n\n\tif strings.TrimSpace(p.Destination) == \"\" {\n\t\treturn nil, errors.New(\"\\\"destination\\\" must contain a value\")\n\t}\n\n\tif p.HashType != nil && p.Hash == nil {\n\t\treturn nil, errors.New(\"\\\"hash\\\" required with use of \\\"hash_type\\\"\")\n\t} else if p.HashType == nil && p.Hash != nil {\n\t\treturn nil, errors.New(\"\\\"hash_type\\\" required with use of \\\"hash\\\"\")\n\t}\n\n\tif p.HashType != nil {\n\t\tif !isValidHashType(*p.HashType) {\n\t\t\treturn nil, fmt.Errorf(\"\\\"hash_type\\\" must be one of \\\"%s,%s,%s,%s\\\"\", string(HashMD5), string(HashSHA1), string(HashSHA256), string(HashSHA512))\n\t\t}\n\t}\n\n\tif p.Hash != nil {\n\t\tif strings.TrimSpace(*p.Hash) == \"\" {\n\t\t\treturn nil, errors.New(\"\\\"hash\\\" must contain a value\")\n\t\t}\n\n\t\tif !isValidHash(*p.HashType, *p.Hash) {\n\t\t\treturn nil, fmt.Errorf(\"\\\"hash\\\" is invalid length for %s\", *p.HashType)\n\t\t}\n\t}\n\n\tunarchive := &Unarchive{\n\t\tSource: p.Source,\n\t\tDestination: p.Destination,\n\t\tForce: p.Force,\n\t}\n\n\tif p.HashType != nil {\n\t\tunarchive.HashType = *p.HashType\n\t}\n\n\tif p.Hash != nil {\n\t\tunarchive.Hash = *p.Hash\n\t}\n\n\tunarchive.fetch = fetch.Fetch{\n\t\tSource: unarchive.Source,\n\t\tDestination: os.TempDir(),\n\t\tHashType: unarchive.HashType,\n\t\tHash: unarchive.Hash,\n\t\tUnarchive: true,\n\t}\n\n\treturn unarchive, nil\n}", "func (f *jsiiProxy_FileSystemBase) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tf,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (p VspherePKEClusterCreationParamsPreparer) Prepare(ctx context.Context, params *VspherePKEClusterCreationParams) error {\n\tif params.Name == \"\" {\n\t\treturn validationErrorf(\"Name cannot be empty\")\n\t}\n\tif params.OrganizationID == 0 {\n\t\treturn validationErrorf(\"OrganizationID cannot be 0\")\n\t}\n\n\t_, err := auth.GetOrganizationById(params.OrganizationID)\n\tif err != nil {\n\t\treturn validationErrorf(\"OrganizationID cannot be found %s\", err.Error())\n\t}\n\n\t// validate secretID\n\tif params.SecretID == \"\" {\n\t\treturn validationErrorf(\"SecretID cannot be empty\")\n\t}\n\tif err := p.verifySecretIsOfType(params.OrganizationID, params.SecretID, secrettype.Vsphere); err != nil {\n\t\treturn err\n\t}\n\n\t// validate storageSecretID if present\n\tif err := p.verifySecretIsOfType(params.OrganizationID, params.StorageSecretID, secrettype.Vsphere); err != nil {\n\t\treturn err\n\t}\n\n\t// validate SSH secret ID if present\n\tif err := p.verifySecretIsOfType(params.OrganizationID, params.SSHSecretID, secrettype.SSHSecretType); err != nil {\n\t\treturn err\n\t}\n\n\tif err := p.k8sPreparer.Prepare(&params.Kubernetes); err != nil {\n\t\treturn errors.WrapIf(err, \"failed to prepare k8s network\")\n\t}\n\n\tif err := p.getNodePoolsPreparer(clusterCreatorNodePoolPreparerDataProvider{}).Prepare(ctx, params.NodePools); err != nil {\n\t\treturn errors.WrapIf(err, \"failed to prepare node pools\")\n\t}\n\n\treturn nil\n}", "func (action *ActionUserRequestRegistrationCreate) Prepare() *ActionUserRequestRegistrationCreateInvocation {\n\treturn &ActionUserRequestRegistrationCreateInvocation{\n\t\tAction: action,\n\t\tPath: \"/v6.0/user_request/registrations\",\n\t}\n}", "func (c *jsiiProxy_CfnStack) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (c *jsiiProxy_CfnStudioComponent) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (action *ActionVpsConfigCreate) Prepare() *ActionVpsConfigCreateInvocation {\n\treturn &ActionVpsConfigCreateInvocation{\n\t\tAction: action,\n\t\tPath: \"/v6.0/vps_configs\",\n\t}\n}", "func (a *RedisAction) Prepare() {\n\tredis := redis.NewClient(\n\t\t&redis.Options{\n\t\t\tAddr: a.RedisAddr,\n\t\t\tPassword: \"\",\n\t\t\tDB: 0,\n\t\t},\n\t)\n\tfmt.Printf(\"redisAddr: %s\\n\", a.RedisAddr)\n\tgo a.setAddressToRedis(redis)\n}", "func (e *TknRunner) Prepare(t *testing.T) func(args ...string) icmd.Cmd {\n\trun := func(args ...string) icmd.Cmd {\n\t\treturn icmd.Cmd{Command: append([]string{e.Path}, args...), Timeout: 10 * time.Minute}\n\t}\n\treturn run\n}", "func (c *Config) Prepare(raws ...interface{}) ([]string, error) {\n\n\tvar md mapstructure.Metadata\n\terr := config.Decode(c, &config.DecodeOpts{\n\t\tMetadata: &md,\n\t\tInterpolate: true,\n\t\tInterpolateContext: &c.ctx,\n\t\tInterpolateFilter: &interpolate.RenderFilter{\n\t\t\tExclude: []string{\n\t\t\t\t\"run_command\",\n\t\t\t},\n\t\t},\n\t}, raws...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Defaults\n\tif c.APIToken == \"\" {\n\t\t// Default to environment variable for api_token, if it exists\n\t\tc.APIToken = os.Getenv(\"CIVO_TOKEN\")\n\t}\n\tif c.SnapshotName == \"\" {\n\t\tdef, err := interpolate.Render(\"civo-packer-{{timestamp}}\", nil)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\t// Default to civo-packer-{{ unix timestamp (utc) }}\n\t\tc.SnapshotName = def\n\t}\n\n\tif c.InstanceName == \"\" {\n\t\t// Default to packer-[time-ordered-uuid]\n\t\tc.InstanceName = fmt.Sprintf(\"packer-%s\", uuid.TimeOrderedUUID())\n\t}\n\n\tif c.StateTimeout == 0 {\n\t\t// Default to 6 minute timeouts waiting for\n\t\t// desired state. i.e waiting for instance to become active\n\t\tc.StateTimeout = 6 * time.Minute\n\t}\n\n\tif c.SnapshotTimeout == 0 {\n\t\t// Default to 60 minutes timeout, waiting for snapshot action to finish\n\t\tc.SnapshotTimeout = 60 * time.Minute\n\t}\n\n\tif c.PublicNetworking == \"\" {\n\t\tc.PublicNetworking = \"true\"\n\t}\n\n\tvar errs *packer.MultiError\n\n\tif es := c.Comm.Prepare(&c.ctx); len(es) > 0 {\n\t\terrs = packer.MultiErrorAppend(errs, es...)\n\t}\n\tif c.APIToken == \"\" {\n\t\t// Required configurations that will display errors if not set\n\t\terrs = packer.MultiErrorAppend(\n\t\t\terrs, errors.New(\"api_token for auth must be specified\"))\n\t}\n\n\tif c.Region == \"\" {\n\t\terrs = packer.MultiErrorAppend(\n\t\t\terrs, errors.New(\"region is required\"))\n\t}\n\n\tif c.Size == \"\" {\n\t\terrs = packer.MultiErrorAppend(\n\t\t\terrs, errors.New(\"size is required\"))\n\t}\n\n\tif c.Template == \"\" {\n\t\terrs = packer.MultiErrorAppend(\n\t\t\terrs, errors.New(\"template is required\"))\n\t}\n\n\tif errs != nil && len(errs.Errors) > 0 {\n\t\treturn nil, errs\n\t}\n\n\tpacker.LogSecretFilter.Set(c.APIToken)\n\treturn nil, nil\n}", "func (adapter *Shell) Prepare() {\n\tfmt.Fprint(adapter.output, \"shbot> \")\n}", "func (c *jsiiProxy_CfnDeploymentStrategy) Prepare() {\n\t_jsii_.InvokeVoid(\n\t\tc,\n\t\t\"prepare\",\n\t\tnil, // no parameters\n\t)\n}", "func (px *Paxos) Prepare(args *PrepareArgs, reply *PrepareReply) error {\n\tpx.mu.Lock()\n\tpx.maxSeq = max(px.maxSeq, args.Seq)\n\treply.Seq = px.doneSeq[px.me]\n\tpx.mu.Unlock()\n\n\tpx.acceptorMgr.mu.Lock()\n\tacceptor, ok := px.acceptorMgr.acceptors[args.Seq]\n\tif !ok {\n\t\tacceptor = &Acceptor{mu: sync.Mutex{}, nP: 0, nA: 0, vA: nil}\n\t\tpx.acceptorMgr.acceptors[args.Seq] = acceptor\n\t}\n\tpx.acceptorMgr.mu.Unlock()\n\n\tacceptor.mu.Lock()\n\tdefer acceptor.mu.Unlock()\n\tif args.N > acceptor.nP {\n\t\tacceptor.nP = args.N\n\t\treply.N = args.N\n\t\treply.Na = acceptor.nA\n\t\treply.Va = acceptor.vA\n\t\treply.Fail = false\n\t} else {\n\t\treply.Fail = true\n\t\treply.N = acceptor.nP\n\t}\n\treturn nil\n}", "func (p *Preparer) Prepare(fullInfo chan resource.Resource, done chan bool, mapWg *sync.WaitGroup) {\n\tmapWg.Wait()\n\n\tvar wg sync.WaitGroup\n\n\tidentifierSend := false\n\n\tfor data := range fullInfo {\n\t\tif !identifierSend {\n\t\t\terr := p.prep.SendIdentifier()\n\t\t\tif err != nil {\n\t\t\t\tlog.WithFields(log.Fields{\"error\": err}).Fatal(\"error send identifier\")\n\t\t\t}\n\t\t\tidentifierSend = true\n\t\t}\n\t\twg.Add(1)\n\t\tgo p.prep.Preparation(data, &wg)\n\t}\n\n\twg.Wait()\n\n\t// If the identifier was not sent, there is no resource to prepare and send,\n\t// a gRPC connection was not open and no finishing and closing of a connection are needed.\n\tif identifierSend {\n\t\tp.prep.Finish()\n\t}\n\n\tdone <- true\n}", "func (px *Paxos) Prepare(args *PrepareArgs, reply *PrepareReply) error {\n\tif args.PNum > px.APp[args.Seq] {\n\t\t// prepare request with higher Proposal Number\n\t\tpx.APp[args.Seq] = args.PNum\n\t\treply.Err = OK\n\t\treply.Proposal = px.APa[args.Seq]\n\t} else {\n\t\t// Already promised to Proposal with a higher Proposal Number\n\t\treply.Err = Reject\n\t\treply.Proposal = Proposal{px.APp[args.Seq], nil}\n\t}\n\treturn nil\n}", "func (this *baseController) Prepare() {\n\t// Reset language option.\n\tthis.Lang = \"\" // This field is from i18n.Locale.\n\n\t// 1. Get language information from 'Accept-Language'.\n\tal := this.Ctx.Request.Header.Get(\"Accept-Language\")\n\tif len(al) > 4 {\n\t\tal = al[:5] // Only compare first 5 letters.\n\t\tif i18n.IsExist(al) {\n\t\t\tthis.Lang = al\n\t\t}\n\t}\n\n\t// 2. Default language is English.\n\tif len(this.Lang) == 0 {\n\t\tthis.Lang = \"en-US\"\n\t}\n\n\t// Set template level language option.\n\tthis.Data[\"Lang\"] = this.Lang\n}" ]
[ "0.68137974", "0.67878276", "0.6714405", "0.6667867", "0.6620684", "0.6522466", "0.65172446", "0.64604", "0.64363766", "0.63994175", "0.638997", "0.63875437", "0.6365766", "0.63546795", "0.63523304", "0.63433295", "0.62986594", "0.62902296", "0.62783366", "0.62242854", "0.62224984", "0.62213206", "0.6221056", "0.6220164", "0.6214212", "0.6212622", "0.6190359", "0.6184286", "0.6180637", "0.6174719", "0.6170782", "0.6166481", "0.6161116", "0.6155834", "0.6146977", "0.6139233", "0.6135079", "0.6119832", "0.61151123", "0.6112825", "0.6108935", "0.6105239", "0.6088099", "0.60862267", "0.6081418", "0.60771024", "0.6062591", "0.60587704", "0.6035801", "0.60292447", "0.60219276", "0.6012095", "0.5996275", "0.5980548", "0.5976448", "0.59739655", "0.59479785", "0.59446335", "0.5929507", "0.5926459", "0.591912", "0.59019744", "0.5900229", "0.58933467", "0.5882686", "0.5869928", "0.58617496", "0.5846174", "0.58407915", "0.5816281", "0.5816185", "0.58089167", "0.5808288", "0.580094", "0.5799446", "0.57877123", "0.5759978", "0.57561195", "0.57555777", "0.57484967", "0.57476074", "0.5746686", "0.5745908", "0.57431227", "0.5742092", "0.57387525", "0.5736567", "0.57332134", "0.5724444", "0.5716224", "0.57158417", "0.57154095", "0.5714797", "0.5710185", "0.57095844", "0.5708363", "0.57047176", "0.5701997", "0.5701635", "0.5687425" ]
0.69741684
0
Invoke calls the mojom service based on the suffix and converts the mojom results (a struct) to Vanadium results (a slice of vom.RawBytes). Note: The argptrs from Prepare are reused here. The vom bytes should have been decoded into these argptrs, so there are actual values inside now.
func (fs fakeService) Invoke(ctx *context.T, call rpc.StreamServerCall, method string, argptrs []interface{}) (results []interface{}, _ error) { // fs.suffix consists of the mojo url and the application/interface name. // The last part should be the name; everything else is the url. parts := strings.Split(fs.suffix, "/") mojourl := strings.Join(parts[:len(parts)-1], "/") // e.g., mojo:go_remote_echo_server. May be defined in a BUILD.gn file. mojoname := parts[len(parts)-1] // e.g., mojo::examples::RemoteEcho. Defined from the interface + module. // Create the generic message pipe. r is a bindings.InterfaceRequest, and // p is a bindings.InterfacePointer. r, p := bindings.CreateMessagePipeForMojoInterface() v := v23ServiceRequest{ request: r, name: mojoname, } // v is an application.ServiceRequest with mojoname // Connect to the mojourl. fs.appctx.ConnectToApplication(mojourl).ConnectToService(&v) // Then assign a new router the FakeService. // This will never conflict because each FakeService is only invoked once. fs.router = bindings.NewRouter(p.PassMessagePipe(), bindings.GetAsyncWaiter()) defer fs.Close_Proxy() ctx.Infof("Fake Service Invoke (Remote Signature: %q -- %q)", mojourl, mojoname) // Vanadium relies on type information, so we will retrieve that first. mojomInterface, desc, err := fs.callRemoteSignature(mojourl, mojoname) if err != nil { return nil, err } ctx.Infof("Fake Service Invoke Signature %v", mojomInterface) ctx.Infof("Fake Service Invoke (Remote Method: %v)", method) // With the type information, we can make the method call to the remote interface. methodResults, err := fs.callRemoteMethod(ctx, method, mojomInterface, desc, argptrs) if err != nil { ctx.Errorf("Method called failed: %v", err) return nil, err } ctx.Infof("Fake Service Invoke Results %v", methodResults) // Convert methodResult to results. results = make([]interface{}, len(methodResults)) for i := range methodResults { results[i] = &methodResults[i] } return results, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s XMLVTService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.XMLVTService.GenerateEntity:\n\t\tvar args = struct {\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t\tEntity string `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"namespace\", \"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GenerateEntity(args.Namespace, args.Entity))\n\n\tcase RPC.XMLVTService.LoadEntity:\n\t\tvar args = struct {\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t\tEntity string `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"namespace\", \"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.LoadEntity(args.Namespace, args.Entity))\n\n\tcase RPC.XMLVTService.UpdateEntity:\n\t\tvar args = struct {\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t\tEntity *mfd.VTEntity `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"namespace\", \"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.UpdateEntity(args.Namespace, args.Entity))\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (c *MainChannelCC) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n funcName, args := stub.GetFunctionAndParameters()\n\n switch funcName {\n // 任务上传\n case \"requestUpload\":\n return requestUpload(stub, args)\n // 查询任务\n case \"requestQuery\":\n return requestQuery(stub, args)\n // 查询全部任务\n case \"requestQueryArr\":\n return requestQueryArr(stub, args)\n // 难度值上传\n case \"difficultyUpload\":\n return difficultyUpload(stub, args)\n // 难度值查询\n case \"difficultyQuery\":\n return difficultyQuery(stub, args)\n // 难度值统一查询\n case \"difficultyQueryArr\":\n return difficultyQueryArr(stub, args)\n // 判断胜利者\n case \"winnerUpload\":\n return winnerUpload(stub, args)\n // 查询胜利者\n case \"winnerQuery\":\n return winnerQuery(stub, args)\n // 查询全部胜利者\n case \"winnerQueryArr\":\n return winnerQueryArr(stub, args)\n // 子channel上传\n case \"subChannelUpload\":\n return subChannelUpload(stub, args)\n // 子channel查询\n case \"subChannelQuery\":\n return subChannelQuery(stub, args)\n // 数据上传\n case \"dataUpload\":\n return dataUpload(stub, args)\n // 查询数据\n case \"dataQuery\":\n return dataQuery(stub, args)\n // 数据统一查询\n case \"dataQueryArr\":\n return dataQueryArr(stub, args)\n // 奖励发放\n case \"rewardsUpload\":\n return rewardsUpload(stub, args)\n // 奖励获取\n case \"rewardsReceive\":\n return rewardsReceive(stub, args)\n }\n\n\treturn shim.Success(nil)\n}", "func (self *stubHandler) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\n\targs := stub.GetArgs()\n\n\tvar params []byte\n\n\tif len(args) > 1 {\n\t\tparams = args[1]\n\t}\n\n\tdispatcher, index, err := self.decodeFunction(string(args[0]))\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn dispatcher.Dispatch(stub, index, params)\n}", "func (cc *Chaincode) Invoke(stub shim.ChaincodeStubInterface) sc.Response {\n\tfcn, params := stub.GetFunctionAndParameters()\n\n\tswitch fcn {\n\tcase \"TransferOwnership\":\n\t\treturn cc.TransferOwnership(stub, params)\n\tcase \"GetHistoryForAsset\":\n\t\treturn cc.GetHistoryForAsset(stub, params)\n\tcase \"UploadMolecule\":\n\t\treturn cc.UploadMolecule(stub, params)\n\tcase \"CreateUser\":\n\t\treturn cc.CreateUser(stub, params)\n\tcase \"UpdateUser\":\n\t\treturn cc.UpdateUser(stub, params)\n\tcase \"QueryMolecules\":\n\t\treturn cc.QueryMolecules(stub, params)\n\tdefault:\n\t\treturn shim.Error(\"No match in function name\")\n\t}\n}", "func (stub *MockStub) MockInvoke(uuid string, args [][]byte) pb.Response {\n\tstub.args = args\n\tstub.MockTransactionStart(uuid)\n\tres := stub.cc.Invoke(stub)\n\tstub.MockTransactionEnd(uuid)\n\treturn res\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n if function == \"createAsset\" {\n return t.createAsset(stub, args)\n } else if function == \"updateAsset\" {\n return t.updateAsset(stub, args)\n } else if function == \"deleteAsset\" {\n return t.deleteAsset(stub, args)\n } else if function == \"deleteAllAssets\" {\n return t.deleteAllAssets(stub, args)\n } else if function == \"deletePropertiesFromAsset\" {\n return t.deletePropertiesFromAsset(stub, args)\n } else if function == \"setLoggingLevel\" {\n return nil, t.setLoggingLevel(stub, args)\n } else if function == \"setCreateOnUpdate\" {\n return nil, t.setCreateOnUpdate(stub, args)\n }\n err := fmt.Errorf(\"Invoke received unknown invocation: %s\", function)\n log.Warning(err)\n return nil, err\n}", "func (t *AssetManagementChaincode) Invoke(stub *shim.ChaincodeStub, function string, args []string) ([]byte, error) {\n\n if function == \"create\" {\n // create asset\n return t.create(stub, args)\n } else if function == \"update\" {\n // update asset (transfer ownership etc)\n return t.update(stub, args)\n }\n\n return nil, errors.New(\"Received unknown function invocation\")\n}", "func invoke(any interface{}, name string, args ...interface{}) []reflect.Value {\n\tinputs := make([]reflect.Value, len(args))\n\tfor i, _ := range args {\n\t\tinputs[i] = reflect.ValueOf(args[i])\n\t}\n\tv := reflect.ValueOf(any)\n\t//log.Println(v)\n\tm := v.MethodByName(name)\n\t//log.Println(m)\n\n\treturn m.Call(inputs)\n}", "func (t *MedChain) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\t\tfunction, args := stub.GetFunctionAndParameters()\n\t\tfmt.Println(\"\\n\\n------------------\\n\")\n\t\tfmt.Println(\"invoke is running -> \"+ function)\n\n\t\t// Handle different functions\n\t\t\n\t\tif function == \"queryAsset\" { \n\t\t\treturn t.queryAsset(stub, args)\n\t\t} else if function == \"getAll\" { \n\t\t\treturn t.getAll(stub, args)\n\t\t} else if function == \"getHistoryForRecord\" { \n\t\t\treturn t.getHistoryForRecord(stub, args)\n\t\t} else if function == \"addHospital\" { \n\t\t\treturn t.addHospital(stub, args)\n\t\t} else if function == \"updateHospital\" { \n\t\t\treturn t.updateHospital(stub, args)\n\t\t} else if function == \"getAllHospital\" { \n\t\t\treturn t.getAllHospital(stub, args)\n\t\t} else if function == \"addHospitalToPatient\" { \n\t\t\treturn t.addHospitalToPatient(stub, args)\n\t\t} else if function == \"updateHospitalToPatient\" { \n\t\t\treturn t.updateHospitalToPatient(stub, args)\n\t\t} else if function == \"getAllHospitalToPatient\" { \n\t\t\treturn t.getAllHospitalToPatient(stub, args)\n\t\t}\n\n\t\t//error\n\t\tfmt.Println(\"invoke did not find func: \" + function) \n\t\treturn shim.Error(\"Received unknown function invocation\")\n\t}", "func (t *evidence_management) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"function is ==> :\" + function)\n\taction := args[0]\n\tfmt.Println(\" action is ==> :\" + action)\n\tfmt.Println(args)\n\n\tif action == \"queryAsset\" {\n\t\treturn t.queryAsset(stub, args)\n\t} else if action == \"queryAllAsset\" {\n\t\treturn t.queryAllAsset(stub, args)\n\t} else if action == \"getHistoryForRecord\" {\n\t\treturn t.getHistoryForRecord(stub, args)\n\t} else if action == \"createCase\" {\n\t\treturn t.createCase(stub, args)\n\t} else if action == \"updateCase\" {\n\t\treturn t.updateCase(stub, args)\n\t} else if action == \"updateCaseStatus\" {\n\t\treturn t.updateCaseStatus(stub, args)\n\t} else if action == \"createFIR\" {\n\t\treturn t.createFIR(stub, args)\n\t} else if action == \"createDoc\" {\n\t\treturn t.createDoc(stub, args)\n\t} else if action == \"putPrivateData\" {\n\t\treturn t.putPrivateData(stub, args)\n\t} else if action == \"getPrivateData\" {\n\t\treturn t.getPrivateData(stub, args)\n\t} else if action == \"addAccused\" {\n\t\treturn t.addAccused(stub, args)\n\t} else if action == \"addSuspect\" {\n\t\treturn t.addSuspect(stub, args)\n\t} else if action == \"addVictim\" {\n\t\treturn t.addVictim(stub, args)\n\t}\n\n\tfmt.Println(\"invoke did not find func: \" + action) //error\n\n\treturn shim.Error(\"Received unknown function\")\n}", "func (cc *Chaincode) Invoke(stub shim.ChaincodeStubInterface) sc.Response {\n\tfcn, params := stub.GetFunctionAndParameters()\n\tvar result []byte\n\tvar err error\n\tif fcn == \"GetAllPatients\" {\n\t\tresult, err = cc.GetAllPatients(stub)\n\t} else if fcn == \"GetPatient\" {\n\t\tresult, err = cc.GetPatient(stub, params)\n\t} else if fcn == \"CreatePatient\" {\n\t\tresult, err = cc.CreatePatient(stub, params)\n\t} else if fcn == \"AddRecordToPatient\" {\n\t\tresult, err = cc.AddRecordToPatient(stub, params)\n\t}\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn shim.Success(result)\n}", "func (t *HeroesServiceChaincode) invoke(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tfmt.Println(\"########### VegetableOwnership invoke ###########\")\n\n\tif len(args) < 2 {\n\t\treturn shim.Error(\"The number of arguments is insufficient.\")\n\t}\n\n\t// Changing Ownership of a Vegetable item by Accepting Key and Value\n\tif args[1] == \"changeOwner\" && len(args) == 4 {\n\n\t\tvegAsBytes, _ := stub.GetState(args[2])\n\t\tveg := Veg{}\n\n\t\tjson.Unmarshal(vegAsBytes, &veg)\n\t\tveg.Owner = args[3]\n\n\t\tvegAsBytes, _ = json.Marshal(veg)\n\t\tstub.PutState(args[2], vegAsBytes)\n\n\t\t// Notify listeners that an event \"eventInvoke\" have been executed (check line 19 in the file invoke.go)\n\t\terr := stub.SetEvent(\"eventChangeVegOwner\", []byte{})\n\t\tif err != nil {\n\t\t\treturn shim.Error(err.Error())\n\t\t}\n\n\t\treturn shim.Success(nil)\n\t}\n\n\t/*\n\t\t Updating all fields of record\n\t*/\n\tif args[1] == \"updateRecord\" && len(args) == 4 {\n\t\tfmt.Println(\"Update All\")\n\t\tvar newVeg Veg\n\t\tjson.Unmarshal([]byte(args[3]), &newVeg)\n\t\tvar veg = Veg{ Name: newVeg.Name, Id: newVeg.Id, Quality: newVeg.Quality, Owner: newVeg.Owner}\n\t\tvegAsBytes, _ := json.Marshal(veg)\n\n\t\t// Updating Record\n\n\t\tstub.PutState(args[2], vegAsBytes)\n\n\t\t// Notify listeners that an event \"eventInvoke\" have been executed (check line 19 in the file invoke.go)\n\t\t\n\t\terr := stub.SetEvent(\"eventUpdateRecords\", []byte{})\n\t\tif err != nil {\n\t\t\treturn shim.Error(err.Error())\n\t\t}\n\n\t\treturn shim.Success(nil)\n\t}\n\n\t// If the arguments given don’t match any function, we return an error\n\n\treturn shim.Error(\"Unknown invoke action, check the second argument.\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" { //initialize the chaincode state, used as reset\n\t\treturn t.Init(stub)\n\t} else if function == \"read\" { //generic read ledger\n\t\treturn read(stub, args)\n\t} else if function == \"write\" { //generic writes to ledger\n\t\treturn write(stub, args)\n\t} else if function == \"delete_marble\" { //deletes a marble from state\n\t\treturn delete_marble(stub, args)\n\t} else if function == \"init_marble\" { //create a new marble\n\t\treturn init_marble(stub, args)\n\t} else if function == \"set_owner\" { //change owner of a marble\n\t\treturn set_owner(stub, args)\n\t} else if function == \"init_owner\"{ //create a new marble owner\n\t\treturn init_owner(stub, args)\n\t} else if function == \"read_everything\"{ //read everything, (owners + marbles + companies)\n\t\treturn read_everything(stub)\n\t} else if function == \"getHistory\"{ //read history of a marble (audit)\n\t\treturn getHistory(stub, args)\n\t} else if function == \"getMarblesByRange\"{ //read a bunch of marbles by start and stop id\n\t\treturn getMarblesByRange(stub, args)\n\t} else if function == \"disable_owner\"{ //disable a marble owner from appearing on the UI\n\t\treturn disable_owner(stub, args)\n\t}\n\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"uploadArtwork\" { //create a new Artwork\n\t\treturn t.uploadArtwork(stub, args)\n\t} else if function == \"transferArtwork\" { //change owner of a specific Artwork\n\t\treturn t.transferArtwork(stub, args)\n\t} else if function == \"deleteArtwork\" { //delete a Artwork\n\t\treturn t.deleteArtwork(stub, args)\n\t} else if function == \"readArtwork\" { //read a Artwork\n\t\treturn t.readArtwork(stub, args)\n\t} else if function == \"getHistoryForArtwork\" { //get history of values for a Artwork\n\t\treturn t.getHistoryForArtwork(stub, args)\n\t} else if function == \"queryAll\" { //get history of values for a Artwork\n\t\treturn t.queryAll(stub, args)\n\t}\n\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (s PublicService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\n\tswitch method {\n\tcase RPC.PublicService.GoPGVersions:\n\t\tresp.Set(s.GoPGVersions())\n\n\tcase RPC.PublicService.Modes:\n\t\tresp.Set(s.Modes())\n\n\tcase RPC.PublicService.SearchTypes:\n\t\tresp.Set(s.SearchTypes())\n\n\tcase RPC.PublicService.Types:\n\t\tresp.Set(s.Types())\n\n\tcase RPC.PublicService.DBTypes:\n\t\tresp.Set(s.DBTypes())\n\n\tcase RPC.PublicService.HTMLTypes:\n\t\tresp.Set(s.HTMLTypes())\n\n\tcase RPC.PublicService.Ping:\n\t\tresp.Set(s.Ping())\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (s ProjectService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.ProjectService.Open:\n\t\tvar args = struct {\n\t\t\tFilePath string `json:\"filePath\"`\n\t\t\tConnection string `json:\"connection\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"filePath\", \"connection\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.Open(args.FilePath, args.Connection))\n\n\tcase RPC.ProjectService.Current:\n\t\tresp.Set(s.Current())\n\n\tcase RPC.ProjectService.Update:\n\t\tvar args = struct {\n\t\t\tProject mfd.Project `json:\"project\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"project\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.Update(args.Project))\n\n\tcase RPC.ProjectService.Save:\n\t\tresp.Set(s.Save())\n\n\tcase RPC.ProjectService.Tables:\n\t\tresp.Set(s.Tables())\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\n\tswitch function {\n\tcase \"GetSegment\":\n\t\treturn s.GetSegment(APIstub, args)\n\tcase \"FindSegments\":\n\t\treturn s.FindSegments(APIstub, args)\n\tcase \"GetMapIDs\":\n\t\treturn s.GetMapIDs(APIstub, args)\n\tcase \"SaveSegment\":\n\t\treturn s.SaveSegment(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n}", "func (o *FakeObject) Invoke(args ...interface{}) Object {\n\ta2 := make([]reflect.Value, len(args))\n\tfor i, a := range args {\n\t\ta2[i] = reflect.ValueOf(a)\n\t}\n\treturn MakeFakeObject(reflect.ValueOf(o.Value).CallSlice(a2))\n}", "func (c *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfun, args := stub.GetFunctionAndParameters()\n\n\tfmt.Println(\"Executing => \"+fun)\n\n\tswitch fun{\n\tcase \"AddCpu\":\n\t\treturn c.AddCpu(stub,args)\n\tcase \"GetUsage\":\n\t\treturn c.GetUsage(stub,args)\n\tdefault:\n\t\treturn shim.Error(\"Not a vaild function\")\t\n\t}\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tswitch function {\n\tcase \"initFileTransfer\":\n\t\t//create a new file transfer\n\t\treturn t.initFileTransfer(stub, args)\n\tcase \"readFileTransfer\":\n\t\t//read a file transfer\n\t\treturn t.readFileTransfer(stub, args)\n\tcase \"readFileTransferPrivateDetails\":\n\t\t//read a file transfer private details\n\t\treturn t.readFileTransferPrivateDetails(stub, args)\n\t/*case \"transferMarble\":\n\t//change owner of a specific marble\n\treturn t.transferMarble(stub, args)*/\n\tcase \"delete\":\n\t\t//delete a file transfer\n\t\treturn t.delete(stub, args)\n\tcase \"queryFileTransferByOriginator\":\n\t\t//find transfer for owner X using rich query\n\t\treturn t.queryFileTransferByOriginator(stub, args)\n\tcase \"queryTransfers\":\n\t\t//find transfers based on an ad hoc rich query\n\t\treturn t.queryTransfers(stub, args)\n\t/*case \"getMarblesByRange\":\n\t//get marbles based on range query\n\treturn t.getMarblesByRange(stub, args)*/\n\tcase \"accessFile\":\n\t\t// get the file and mark is as having been accessed by the recipient\n\t\treturn t.accessFile(stub, args)\n\tdefault:\n\t\t//error\n\t\tfmt.Println(\"invoke did not find func: \" + function)\n\t\treturn shim.Error(\"Received unknown function invocation\")\n\t}\n}", "func (conversion *ConversionChaincode) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\r\n\tfuncName, args := stub.GetFunctionAndParameters()\r\n\tif funcName == \"state\" {\r\n\t\treturn SetState(stub)\r\n\t} else if funcName == \"district\" {\r\n\t\treturn SetDistrict(stub)\r\n\t} else if funcName == \"subdistrict\" {\r\n\t\treturn SetSubDistrict(stub)\r\n\t} else if funcName == \"village\" {\r\n\t\treturn SetVillage(stub)\r\n\t} else if funcName == \"get\" {\r\n\t\treturn GetUID(stub, args[0])\r\n\t}\r\n\treturn shim.Error(\"Fuction name '\" + funcName + \"' is wrong or \" + funcName + \" function does not exist!!!\")\r\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n fmt.Println(\"Calling Invoke method.\")\n\n // Retrieve the requested Smart Contract function and arguments\n function, args := APIstub.GetFunctionAndParameters()\n fmt.Println(\"Function name: \" + function)\n\n // Route to the appropriate handler function to interact with the ledger appropriately\n if function == \"queryAsset\" {\n return s.queryAsset(APIstub, args)\n } else if function == \"makeAsset\" {\n return s.makeAsset(APIstub, args)\n } else if function == \"changeAsset\" {\n return s.changeAsset(APIstub, args)\n } else if function == \"deleteAsset\" {\n return s.deleteAsset(APIstub, args)\n } else if function == \"listHistory\" {\n return s.listHistory(APIstub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\tcase \"queryProduct\":\n\t\treturn s.queryProduct(APIstub, args)\n\tcase \"createProduct\":\n\t\treturn s.createProduct(APIstub, args)\n\tcase \"queryAllProduct\":\n\t\treturn s.queryAllProduct(APIstub)\n\tcase \"changeProductStatus\":\n\t\treturn s.changeProductStatus(APIstub, args)\n\tcase \"getHistoryForProduct\":\n\t\treturn s.getHistoryForProduct(APIstub, args)\n\tcase \"queryProductByStatus\":\n\t\treturn s.queryProductByStatus(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n\n\t// return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (m *MegaCorp) OnInvoke(md Metadata, function string, params []string) ([]byte, error) {\n\tswitch function {\n\n\t// create-account function creates a new account\n\tcase \"create-account\":\n\n\t\tif len(params) < 2 {\n\t\t\treturn nil, fmt.Errorf(\"first name and last name are required\")\n\t\t}\n\n\t\tnewAcct, err := m.createAccount(params[0], params[1])\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to create account\")\n\t\t}\n\t\treturn ToJSON(newAcct), nil\n\n\t// create-employee function creates a new employee.\n\t// requires account id and position passed in params.\n\tcase \"create-employee\":\n\t\tif len(params) < 2 {\n\t\t\treturn nil, fmt.Errorf(\"account id and position are required\")\n\t\t}\n\t\tnewEmp, err := m.createEmployee(params[0], params[1])\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to create employee: %s\", err)\n\t\t}\n\t\treturn ToJSON(newEmp), nil\n\n\t// get-all-employees fetches all employees\n\tcase \"get-all-employees\":\n\t\treturn ToJSON(m.getAllEmployees()), nil\n\n\t// get-account fetches an account\n\tcase \"get-account\":\n\t\tif len(params) < 1 {\n\t\t\treturn nil, fmt.Errorf(\"account is required\")\n\t\t}\n\t\tacct, err := m.getAccount(params[0])\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to get account: %s\", err)\n\t\t}\n\t\treturn ToJSON(acct), nil\n\n\tcase \"get-all-accounts\":\n\t\treturn ToJSON(m.getAllAccounts()), nil\n\n\t// get-total-supply fetches the total supply of megacoin\n\tcase \"get-total-supply\":\n\t\treturn []byte(m.getRemainingCoinSupply()), nil\n\n\t// pay-salaries initiates salary payment. This also called by cron()\n\tcase \"pay-salaries\":\n\t\tm.paySalaries()\n\t\treturn []byte(\"done\"), nil\n\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported function\")\n\t}\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\n\tif function == \"queryLaptop\" {\n\t\treturn s.queryLaptop(APIstub, args)\n\t} else if function == \"initLedger\" {\n\t\treturn s.initLedger(APIstub)\n\t} else if function == \"createLaptop\" {\n\t\treturn s.createLaptop(APIstub, args)\n\t} else if function == \"deleteLaptop\" {\n\t\treturn s.deleteLaptop(APIstub, args)\n\t} else if function == \"queryAllLaptops\" {\n\t\treturn s.queryAllLaptops(APIstub)\n\t} else if function == \"cambiarPropietarioLaptop\" {\n\t\treturn s.cambiarPropietarioLaptop(APIstub, args)\n\t}\n\n\treturn shim.Error(\"Nombre de funcion del SmartContract invalido o inexistente.\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" { //initialize the chaincode state, used as reset\n\t\treturn t.Init(stub)\n\t} else if function == \"read\" { //generic read ledger\n\t\treturn read(stub, args)\n\t} else if function == \"write\" { //generic writes to ledger\n\t\treturn write(stub, args)\n\t} else if function == \"update_source\" { //updates a listing from state\n\t\treturn update_source(stub, args)\n\t} else if function == \"init_listing\" { //create a new listing\n\t\treturn init_listing(stub, args)\n\t} else if function == \"set_state\" { //change state of a listing\n\t\treturn set_state(stub, args)\n\t} else if function == \"init_state\" { //create a new listing state\n\t\treturn init_state(stub, args)\n\t} else if function == \"read_everything\" { //read everything, (states + CC Demo + companies)\n\t\treturn read_everything(stub)\n\t} else if function == \"getHistory\" { //read history of a listing (audit)\n\t\treturn getHistory(stub, args)\n\t} else if function == \"getListingsByRange\" { //read a bunch of CC Demo by start and stop id\n\t\treturn getListingsByRange(stub, args)\n\t} else if function == \"disable_state\" { //disable a listing state from appearing on the UI\n\t\treturn disable_state(stub, args)\n\t} else if function == \"query\" { //query data\n\t\treturn getQueryResultForQueryString(stub, args)\n\t}\n\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (s XMLService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.XMLService.GenerateEntity:\n\t\tvar args = struct {\n\t\t\tTable string `json:\"table\"`\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"table\", \"namespace\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GenerateEntity(args.Table, args.Namespace))\n\n\tcase RPC.XMLService.LoadEntity:\n\t\tvar args = struct {\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t\tEntity string `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"namespace\", \"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.LoadEntity(args.Namespace, args.Entity))\n\n\tcase RPC.XMLService.UpdateEntity:\n\t\tvar args = struct {\n\t\t\tEntity *mfd.Entity `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.UpdateEntity(args.Entity))\n\n\tcase RPC.XMLService.GenerateModelCode:\n\t\tvar args = struct {\n\t\t\tEntity mfd.Entity `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GenerateModelCode(args.Entity))\n\n\tcase RPC.XMLService.GenerateSearchModelCode:\n\t\tvar args = struct {\n\t\t\tEntity mfd.Entity `json:\"entity\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"entity\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GenerateSearchModelCode(args.Entity))\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (q *QueryServiceTestHelper) Invoke(_ gocontext.Context, method string, args, reply interface{}, _ ...grpc.CallOption) error {\n\tquerier := q.Route(method)\n\tif querier == nil {\n\t\treturn fmt.Errorf(\"handler not found for %s\", method)\n\t}\n\treqBz, err := q.cdc.Marshal(args)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tres, err := querier(q.Ctx, &abci.RequestQuery{Data: reqBz})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = q.cdc.Unmarshal(res.Value, reply)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\t\tcase \"createDEC\":\n\t\t\treturn s.createDEC(APIstub, args)\n\t\tcase \"updateDEC\":\n\t\t\treturn s.updateDEC(APIstub, args)\n\t\tcase \"traceDEC\":\n\t\t\treturn s.traceDEC(APIstub, args)\n\t\tcase \"getDEC\":\n\t\t\treturn s.getDEC(APIstub, args)\n\t}\n\t\n\treturn shim.Error(\"Invoke Function Not Success.\")\n\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub)\n\t} else if function == \"query\" {\n\t\treturn t.query(stub, args)\n\t} else if function == \"addNode\" {\n\t\treturn addNode(stub, args)\n\t} else if function == \"distribute\" {\n\t\treturn distribute(stub, args)\n\t}\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (s Service) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.Service.Getinfo:\n\t\tresp.Set(s.Getinfo())\n\n\tcase RPC.Service.GetNewAddress:\n\t\tresp.Set(s.GetNewAddress())\n\n\tcase RPC.Service.ValidateAddress:\n\t\tvar args = struct {\n\t\t\tAddr string `json:\"addr\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.ValidateAddress(args.Addr))\n\n\tcase RPC.Service.ListTransactions:\n\t\tvar args = struct {\n\t\t\tAddr *string `json:\"addr\"`\n\t\t\tCount *int `json:\"count\"`\n\t\t\tSkip *int `json:\"skip\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\", \"count\", \"skip\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\t//zenrpc:addr=\"*\"\n\t\tif args.Addr == nil {\n\t\t\tvar v string = \"*\"\n\t\t\targs.Addr = &v\n\t\t}\n\n\t\t//zenrpc:count=300\n\t\tif args.Count == nil {\n\t\t\tvar v int = 300\n\t\t\targs.Count = &v\n\t\t}\n\n\t\t//zenrpc:skip=0\n\t\tif args.Skip == nil {\n\t\t\tvar v int = 0\n\t\t\targs.Skip = &v\n\t\t}\n\n\t\tresp.Set(s.ListTransactions(*args.Addr, *args.Count, *args.Skip))\n\n\tcase RPC.Service.SendToAddress:\n\t\tvar args = struct {\n\t\t\tAddr string `json:\"addr\"`\n\t\t\tAmount json.Number `json:\"amount\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\", \"amount\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.SendToAddress(args.Addr, args.Amount))\n\n\tcase RPC.Service.GetRecords:\n\t\tvar args = struct {\n\t\t\tSTime int64 `json:\"sTime\"`\n\t\t\tETime int64 `json:\"eTime\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"sTime\", \"eTime\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GetRecords(args.STime, args.ETime))\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (ri *RPCInvoker) Invoke(ctx context.Context, microServiceName, schemaID, operationID string, arg interface{}, reply interface{}, options ...InvocationOption) error {\n\topts := getOpts(options...)\n\tif opts.Protocol == \"\" {\n\t\topts.Protocol = common.ProtocolHighway\n\t}\n\n\ti := invocation.New(ctx)\n\ti.MicroServiceName = microServiceName\n\twrapInvocationWithOpts(i, opts)\n\ti.SchemaID = schemaID\n\ti.OperationID = operationID\n\ti.Args = arg\n\ti.Reply = reply\n\terr := ri.invoke(i)\n\tif err == nil {\n\t\tsetCookieToCache(*i, getNamespaceFromMetadata(opts.Metadata))\n\t}\n\treturn err\n}", "func run(cmd string, v interface{}, extraArgs ...string) error {\n\t// lvmlock can be nil, as it is a global variable that is intended to be\n\t// initialized from calling code outside this package. We have no way of\n\t// knowing whether the caller performed that initialization and must\n\t// defensively check. In the future, we may decide to simply panic with a\n\t// nil pointer dereference.\n\tif lvmlock != nil {\n\t\t// We use Lock instead of TryLock as we have no alternative way of\n\t\t// making progress. We expect lvm2 command-line utilities invoked by\n\t\t// this package to return within a reasonable amount of time.\n\t\tif lerr := lvmlock.Lock(); lerr != nil {\n\t\t\treturn fmt.Errorf(\"lvm: acquire lock failed: %v\", lerr)\n\t\t}\n\t\tdefer func() {\n\t\t\tif lerr := lvmlock.Unlock(); lerr != nil {\n\t\t\t\tpanic(fmt.Sprintf(\"lvm: release lock failed: %v\", lerr))\n\t\t\t}\n\t\t}()\n\t}\n\tvar args []string\n\tif v != nil {\n\t\targs = append(args, \"--reportformat=json\")\n\t\targs = append(args, \"--units=b\")\n\t\targs = append(args, \"--nosuffix\")\n\t}\n\targs = append(args, extraArgs...)\n\tc := exec.Command(cmd, args...)\n\tlog.Printf(\"Executing: %v\", c)\n\tstdout, stderr := new(bytes.Buffer), new(bytes.Buffer)\n\tc.Stdout = stdout\n\tc.Stderr = stderr\n\tif err := c.Run(); err != nil {\n\t\terrstr := ignoreWarnings(stderr.String())\n\t\tlog.Print(\"stdout: \" + stdout.String())\n\t\tlog.Print(\"stderr: \" + errstr)\n\t\treturn errors.New(errstr)\n\t}\n\tstdoutbuf := stdout.Bytes()\n\tstderrbuf := stderr.Bytes()\n\terrstr := ignoreWarnings(string(stderrbuf))\n\tlog.Printf(\"stdout: \" + string(stdoutbuf))\n\tlog.Printf(\"stderr: \" + errstr)\n\tif v != nil {\n\t\tif err := json.Unmarshal(stdoutbuf, v); err != nil {\n\t\t\treturn fmt.Errorf(\"%v: [%v]\", err, string(stdoutbuf))\n\t\t}\n\t}\n\treturn nil\n}", "func (t *TnT) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\tfmt.Printf(\"Invoke called, determining function\")\n\t\n\t// Handle different functions\n\tif function == \"init\" {\n\t\tfmt.Printf(\"Function is init\")\n\t\treturn t.Init(stub, function, args)\n\t} else if function == \"createAssembly\" {\n\t\tfmt.Printf(\"Function is createAssembly\")\n\t\treturn t.createAssembly(stub, args)\n\t} else if function == \"updateAssemblyByID\" {\n\t\tfmt.Printf(\"Function is updateAssemblyByID\")\n\t\treturn t.updateAssemblyByID(stub, args)\n\t} else if function == \"createPackage\" {\n\t\tfmt.Printf(\"Function is createPackage\")\n\t\treturn t.createPackage(stub, args)\n\t} else if function == \"updatePackage\" {\n\t\tfmt.Printf(\"Function is updatePackage\")\n\t\treturn t.updatePackage(stub, args)\n\t} else if function == \"updateAssemblyInfo2ByID\" {\n\t\tfmt.Printf(\"Function is updateAssemblyInfo2ByID\")\n\t\treturn t.updateAssemblyInfo2ByID(stub, args)\n\t} else if function == \"updatePackageInfo2ById\" {\n\t\tfmt.Printf(\"Function is updatePackageInfo2ById\")\n\t\treturn t.updatePackageInfo2ById(stub, args)\n\t} \n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\tfmt.Println(args)\n\n\tif function == \"query\"{\n\t\treturn t.query(stub, args)\n\t} else if function == \"enroll_donor\"{\n\t\treturn t.enroll_donor(stub, args)\n\t} else if function == \"enroll_npo\" {\n\t\treturn t.enroll_npo(stub, args)\n\t} else if function == \"enroll_needs\" {\n\t\treturn t.enroll_needs(stub, args)\n\t} else if function == \"propose_asset\" {\n\t\treturn t.propose_asset(stub, args)\n\t} else if function == \"approve_asset\" {\n\t\treturn t.approve_asset(stub, args)\n\t} else if function == \"delete_asset\" {\n\t\treturn t.delete_asset(stub, args)\n\t} else if function == \"enroll_recipient\" {\n\t\treturn t.enroll_recipient(stub, args)\n\t} else if function == \"borrow_asset\" {\n\t\treturn t.borrow_asset(stub, args)\n\t} else if function == \"give_asset\" {\n\t\treturn t.give_asset(stub, args)\n\t} else if function == \"get_back_asset\" {\n\t\treturn t.get_back_asset(stub, args)\n\t} else if function == \"read_everything\" {\n\t\treturn t.read_everything(stub)\n\t} else if function == \"get_history\" {\n\t\treturn t.get_history(stub, args)\n\t} else if function == \"enroll_initial_needs\"{\n\t\treturn t.enroll_initial_needs(stub)\n\t}\n\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\tif function == \"registProducts\" {\n\t\treturn t.registProducts(stub, args)\n\t} else if function == \"getProductList\" {\n\t\treturn t.getProductList(stub, args)\n\t} else if function == \"getProduct\" {\n\t\treturn t.getProduct(stub, args)\n\t} else if function == \"transferOwner\" {\n\t\treturn t.transferOwner(stub, args)\n\t} else if function == \"registCategory\" {\n\t\treturn t.registCategory(stub, args)\n\t} else if function == \"getCategories\" {\n\t\treturn t.getCategories(stub, args)\n\t} else if function == \"purchase\" {\n\t\treturn t.purchase(stub, args)\n\t}\n\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\tcase \"queryById\":\n\t\treturn s.queryById(APIstub, args)\n\tcase \"createRecord\":\n\t\treturn s.createRecord(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n}", "func (i *Invoker) Invoke(\n\tctx context.Context,\n\tusername string,\n\tactName string,\n\targ string,\n) (rst string, e error) {\n\tvar (\n\t\tres *invokerV1API.InvokeResponse\n\t)\n\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\n\tgo func() {\n\t\tdefer wg.Done()\n\n\t\tif res, e = i.Client.Invoke(ctx, &invokerV1API.InvokeRequest{\n\t\t\tUsername: username,\n\t\t\tActName: actName,\n\t\t\tArg: arg,\n\t\t}); e != nil {\n\t\t\treturn\n\t\t}\n\t\trst = res.GetResult()\n\t}()\n\n\twg.Wait()\n\n\treturn\n}", "func (t *SmartContract) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n // Extract the function and args from the transaction proposal\n fn, args := stub.GetFunctionAndParameters()\n\n var result string\n var err error\n if fn == \"set\" {\n result, err = set(stub, args)\n\n } else if fn == \"get\"{ // assume 'get' even if fn is nil\n result, err = get(stub, args)\n } else if fn == \"getHistory\"{\n \treturn getHistory(stub, args)\n } else if fn == \"getall\"{\n \treturn getall(stub)\n }\n if err != nil {\n return shim.Error(err.Error())\n }\n return shim.Success([]byte(result))\n\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\t// tMap, _ := stub.GetTransient()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tswitch function {\n\tcase \"create_project\":\n\t\treturn create_project(stub, args)\n\tcase \"get_project_by_id\":\n\t\treturn get_project_by_id(stub, args)\n\tcase \"query_all_projects\":\n\t\treturn query_all_projects(stub, args)\n\tcase \"query_paging_projects\":\n\t\treturn query_paging_projects(stub, args)\n\tcase \"remove_project\":\n\t\treturn remove_project(stub, args)\n\tcase \"modify_project\":\n\t\treturn modify_project(stub, args)\n\tcase \"create_linear_workflow\":\n\t\treturn create_linear_workflow(stub, args)\n\tcase \"get_workflow_by_id\":\n\t\treturn get_workflow_by_id(stub, args)\n\tcase \"query_all_workflows\":\n\t\treturn query_all_workflows(stub, args)\n\tcase \"enable_or_disable_workflow\":\n\t\treturn enable_or_disable_workflow(stub, args)\n\tcase \"modify_workflow_def\":\n\t\treturn modify_workflow_def(stub, args)\n\tcase \"query_accessable_workflows\":\n\t\treturn query_accessable_workflows(stub, args)\n\tcase \"start_process\":\n\t\treturn start_process(stub, args)\n\tcase \"get_process_by_id\":\n\t\treturn get_process_by_id(stub, args)\n\tcase \"query_logs_by_process_id\":\n\t\treturn query_logs_by_process_id(stub, args)\n\tcase \"transfer_process\":\n\t\treturn transfer_process(stub, args)\n\tcase \"return_process\":\n\t\treturn return_process(stub, args)\n\tcase \"withdraw_process\":\n\t\treturn withdraw_process(stub, args)\n\tcase \"cancel_process\":\n\t\treturn cancel_process(stub, args)\n\tcase \"query_todo_process\":\n\t\treturn query_todo_process(stub, args)\n\tcase \"query_done_process\":\n\t\treturn query_done_process(stub, args)\n\tcase \"save_org_public_key\":\n\t\treturn save_org_public_key(stub, args)\n\tcase \"encrypt_data\":\n\t\treturn encrypt_data(stub, args)\n\tcase \"decrypt_data\":\n\t\treturn decrypt_data(stub, args)\n\tdefault:\n\t\t// error out\n\t\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\t\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n\t}\n}", "func (w *deviceWrapper) InvokeCommand(cmdName enums.Command, param map[string]interface{}) {\n\tmethod, ok := w.commands[cmdName]\n\tif !ok {\n\t\tw.Ctor.Logger.Warn(\"Device doesn't support this command\",\n\t\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\t\treturn\n\t}\n\n\tw.Ctor.Logger.Debug(\"Invoking device command\",\n\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\n\tvar results []reflect.Value\n\n\tif method.Type().NumIn() > 0 {\n\t\tobj, err := json.Marshal(param)\n\t\tif err != nil {\n\t\t\tw.Ctor.Logger.Error(\"Got error while marshalling data for device command\", err,\n\t\t\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\t\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\t\t\treturn\n\t\t}\n\n\t\tobjNew := reflect.New(method.Type().In(0)).Interface()\n\t\tval := reflect.ValueOf(objNew)\n\n\t\terr = json.Unmarshal(obj, &objNew)\n\t\tif err != nil {\n\t\t\tw.Ctor.Logger.Error(\"Got error while preparing data for device command\", err,\n\t\t\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\t\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\t\t\treturn\n\t\t}\n\n\t\tif !w.Ctor.Validator.Validate(objNew) {\n\t\t\tw.Ctor.Logger.Warn(\"Received incorrect command params\",\n\t\t\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\t\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\t\t\treturn\n\t\t}\n\t\tif reflect.ValueOf(objNew).Kind() != method.Type().In(0).Kind() {\n\t\t\tval = val.Elem()\n\t\t}\n\n\t\tresults = method.Call([]reflect.Value{val})\n\t} else {\n\t\tresults = method.Call(nil)\n\t}\n\n\tif len(results) > 0 && results[0].Interface() != nil {\n\t\tw.Ctor.Logger.Error(\"Got error while invoking device command\", results[0].Interface().(error),\n\t\t\tcommon.LogDeviceTypeToken, w.Ctor.DeviceType.String(), common.LogDeviceNameToken, w.ID,\n\t\t\tcommon.LogDeviceCommandToken, cmdName.String())\n\n\t\treturn\n\t}\n\tif w.Spec.PostCommandDeferUpdate > 0 {\n\t\ttime.Sleep(w.Spec.PostCommandDeferUpdate)\n\t}\n\n\tw.pullUpdate()\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\tif function == \"create\" {\n\t\treturn t.create(stub, args)\n\t} else if function == \"update\" {\n\t\treturn t.update(stub, args)\n\t} else if function == \"get\" {\n\t\treturn t.get(stub, args)\n\t}\n\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (t *AnswerChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tif function == \"submitAnswer\" { //create a new marble\n\t\treturn submitAnswer(stub, args)\n\t} else if function == \"thumbsUpToAnswer\" { //update_answer\n\t\treturn thumbsUpToAnswer(stub, args)\n\t} else if function == \"queryAnswersByThumsUpCount\" { //queryAnswersByStatus\n\t\treturn queryAnswersByThumsUpCount(stub, args)\n\t} else if function == \"queryAnswerByAnswerHashId\" { //queryAnswerStatusByHash\n\t\treturn queryAnswerByAnswerHashId(stub, args)\n\t}\n\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (t *SimpleChaincode) Invoke(stub *shim.ChaincodeStub, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t} else if function == \"addLoc\" {\n\t\tfmt.Println(\"**** First argument in addLoc:****\" + args[0])\n\t\treturn t.addLoc(stub, args)\n\t}\n\t\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (t *VersityChaincode) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\t// Handle different functions\n\tif function == \"initRecord\" { //create a new record\n\t\treturn t.initRecord(stub, args)\n\t} else if function == \"readRecord\" { //read a record\n\t\treturn t.readRecord(stub, args)\n\t} else if function == \"validateRecord\" { //grant permission for an employer to view a record(s)\n\t\treturn t.validateRecord(stub, args)\n\t} else if function == \"addViewerToRecords\" { //grant permission for an employer to view a record(s)\n\t\treturn t.addViewerToRecords(stub, args)\n\t} else if function == \"queryRecordsByOwner\" { //find records for owner X using rich query\n\t\treturn t.queryRecordsByOwner(stub, args)\n\t} else if function == \"queryRecords\" { //find records based on an ad hoc rich query\n\t\treturn t.queryRecords(stub, args)\n\t} else if function == \"getHistoryForRecord\" { //get history of values for a record\n\t\treturn t.getHistoryForRecord(stub, args)\n\t}\n\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (t *BenchmarkerChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\trt := reflect.ValueOf(t)\n\ttheFunc, ok := rt.Type().MethodByName(function)\n\tif !ok {\n\t\tvar methods []string\n\t\tfor i := 0; i < rt.NumMethod(); i++ {\n\t\t\tmethodName := rt.Type().Method(i).Name\n\t\t\tif methodName != \"Invoke\" && methodName != \"Init\" {\n\t\t\t\tmethods = append(methods, methodName)\n\t\t\t}\n\t\t}\n\t\treturn shim.Error(fmt.Sprintf(\"Invalid method name. Supported methods: %+v (%d) methods\", methods, rt.NumMethod()-2))\n\t}\n\tif theFunc.Type.NumIn() != len(args)+2 {\n\t\treturn shim.Error(fmt.Sprintf(\"Expected %d arguments. Got %d.\", theFunc.Type.NumIn()-2, len(args)))\n\t}\n\tin := make([]reflect.Value, theFunc.Type.NumIn())\n\tin[0] = reflect.ValueOf(t)\n\tin[1] = reflect.ValueOf(stub)\n\n\tfor i := 2; i < theFunc.Type.NumIn(); i++ {\n\t\tt := theFunc.Type.In(i)\n\t\targ := args[i-2]\n\t\tif t.Kind() == reflect.Int {\n\t\t\tx, err := strconv.Atoi(arg)\n\t\t\tif err != nil {\n\t\t\t\treturn shim.Error(fmt.Sprintf(\"Expected argument#%d to be convertable to Int. Got %s.\", i-2, arg))\n\t\t\t}\n\t\t\tin[i] = reflect.ValueOf(x)\n\t\t} else if t.Kind() == reflect.Bool {\n\t\t\tx, err := strconv.ParseBool(arg)\n\t\t\tif err != nil {\n\t\t\t\treturn shim.Error(fmt.Sprintf(\"Expected argument#%d to be convertable to Bool. Got %s.\", i-2, arg))\n\t\t\t}\n\t\t\tin[i] = reflect.ValueOf(x)\n\t\t} else if t.Kind() == reflect.Float64 {\n\t\t\tx, err := strconv.ParseFloat(arg, 64)\n\t\t\tif err != nil {\n\t\t\t\treturn shim.Error(fmt.Sprintf(\"Expected argument#%d to be convertable to Float64. Got %s.\", i-2, arg))\n\t\t\t}\n\t\t\tin[i] = reflect.ValueOf(x)\n\t\t} else if t.Kind() == reflect.String {\n\t\t\tin[i] = reflect.ValueOf(arg)\n\t\t} else {\n\t\t\treturn shim.Error(fmt.Sprintf(\"Unsupported type %s in chaincode.\", t.Kind()))\n\t\t}\n\t}\n\n\treturn theFunc.Func.Call(in)[0].Interface().(pb.Response)\n}", "func (client *BaseClient) Invoke(\n\tname string,\n\targs []reflect.Value,\n\tsettings *InvokeSettings) (results []reflect.Value, err error) {\n\tcontext := client.GetClientContext(settings)\n\tresults, err = client.handlerManager.invokeHandler(name, args, context)\n\tif results == nil && len(context.ResultTypes) > 0 {\n\t\tn := len(context.ResultTypes)\n\t\tresults = make([]reflect.Value, n)\n\t\tfor i := 0; i < n; i++ {\n\t\t\tresults[i] = reflect.New(context.ResultTypes[i]).Elem()\n\t\t}\n\t}\n\tclient.contextPool.Put(context)\n\treturn\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"createTaskMatching\" { //create a new taskmatching\n\t\treturn t.createTaskMatching(stub, args)\n\t} else if function == \"readTaskMatching\" { //reads a taskmatching\n\t\treturn t.readTaskMatching(stub, args)\n\t} else if function == \"Initialize\" { //initialize the network\n\t\treturn t.Initialize(stub)\n\t} else if function == \"calculateTaskMatching\" { //calculate a taskmatching\n\t\tt.calculateTaskMatching(stub, args)\n\n\t\tif t.allPeersDone(stub) {\n\t\t\treturn t.setBestSol(stub)\n\t\t} else {\n\t\t\treturn shim.Success(nil)\n\t\t}\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (fs fakeService) callRemoteMethod(ctx *context.T, method string, mi mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, argptrs []interface{}) ([]*vom.RawBytes, error) {\n\t// We need to parse the signature result to get the method relevant info out.\n\tfound := false\n\tvar ordinal uint32\n\tfor ord, mm := range mi.Methods {\n\t\tif *mm.DeclData.ShortName == method {\n\t\t\tordinal = ord\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"callRemoteMethod: method %s does not exist\", method)\n\t}\n\n\tmm := mi.Methods[ordinal]\n\n\t// A void function must have request id of 0, whereas one with response params\n\t// should have a unique request id.\n\theader := bindings.MessageHeader{\n\t\tType: ordinal,\n\t\tFlags: bindings.MessageExpectsResponseFlag,\n\t\tRequestId: fs.ids.Count(),\n\t}\n\n\t// Now produce the *bindings.Message that we will send to the other side.\n\tinType, err := transcoder.MojomStructToVDLType(mm.Parameters, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmessage, err := encodeMessageFromVom(header, argptrs, inType)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Otherwise, make a generic call with the message.\n\toutMessage, err := fs.callRemoteWithResponse(ctx, message)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Decode the *vom.RawBytes from the mojom bytes and mojom type.\n\toutType, err := transcoder.MojomStructToVDLType(*mm.ResponseParams, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttarget := util.StructSplitTarget()\n\tif err := transcoder.FromMojo(target, outMessage.Payload, outType); err != nil {\n\t\treturn nil, fmt.Errorf(\"transcoder.FromMojo failed: %v\", err)\n\t}\n\treturn target.Fields(), nil\n}", "func (b *BackendTransportService) Invoke(service flux.BackendService, ctx flux.Context) (interface{}, *flux.ServeError) {\n\ttypes, values, err := b.ArgumentsAssembleFunc(service.Arguments, ctx)\n\tif nil != err {\n\t\treturn nil, &flux.ServeError{\n\t\t\tStatusCode: flux.StatusServerError,\n\t\t\tErrorCode: flux.ErrorCodeGatewayInternal,\n\t\t\tMessage: flux.ErrorMessageDubboAssembleFailed,\n\t\t\tInternal: err,\n\t\t}\n\t} else {\n\t\treturn b.ExecuteWith(types, values, service, ctx)\n\t}\n}", "func (xmlmc *XmlmcInstStruct) Invoke(servicename string, methodname string) (string, error) {\n\n\t//-- Add Api Tracing\n\ttracename := \"\"\n\tif xmlmc.trace != \"\" {\n\t\ttracename = \"/\" + tracename\n\t}\n\n\txmlmclocal := \"<methodCall service=\\\"\" + servicename + \"\\\" method=\\\"\" + methodname + \"\\\" trace=\\\"goApi\" + tracename + \"\\\">\"\n\tif len(xmlmc.paramsxml) == 0 {\n\t\txmlmclocal = xmlmclocal + \"</methodCall>\"\n\t} else {\n\t\txmlmclocal = xmlmclocal + \"<params>\" + xmlmc.paramsxml\n\t\txmlmclocal = xmlmclocal + \"</params>\" + \"</methodCall>\"\n\t}\n\n\tstrURL := xmlmc.server + \"/\" + servicename + \"/?method=\" + methodname\n\n\tvar xmlmcstr = []byte(xmlmclocal)\n\n\treq, err := http.NewRequest(\"POST\", strURL, bytes.NewBuffer(xmlmcstr))\n\txmlmc.count++\n\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Unable to create http request in esp_xmlmc.go\")\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"text/xmlmc\")\n\tif xmlmc.apiKey != \"\" {\n\t\treq.Header.Add(\"Authorization\", \"ESP-APIKEY \"+xmlmc.apiKey)\n\t}\n\treq.Header.Set(\"User-Agent\", xmlmc.userAgent)\n\treq.Header.Add(\"Cookie\", xmlmc.sessionID)\n\tif xmlmc.jsonresp == true {\n\t\treq.Header.Add(\"Accept\", \"text/json\")\n\t}\n\tduration := time.Second * time.Duration(xmlmc.timeout)\n\tclient := &http.Client{Transport: xmlmc.transport, Timeout: duration}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\txmlmc.statuscode = resp.StatusCode\n\n\tdefer resp.Body.Close()\n\n\t//-- Check for HTTP Response\n\tif resp.StatusCode != 200 {\n\t\terrorString := fmt.Sprintf(\"Invalid HTTP Response: %d\", resp.StatusCode)\n\t\terr = errors.New(errorString)\n\t\t//Drain the body so we can reuse the connection\n\t\tio.Copy(ioutil.Discard, resp.Body)\n\t\treturn \"\", err\n\t}\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Cant read the body of the response\")\n\t}\n\t// If we have a new EspSessionId set it\n\tSessionIds := strings.Split(resp.Header.Get(\"Set-Cookie\"), \";\")\n\tif SessionIds[0] != \"\" {\n\t\txmlmc.sessionID = SessionIds[0]\n\t}\n\n\txmlmc.paramsxml = \"\"\n\treturn string(body), nil\n}", "func (w *worker) Invoke(args interface{}) error { return ErrNotImplement }", "func (conversion *OwnershipChaincode) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\tfuncName, args := stub.GetFunctionAndParameters()\n\tif funcName == \"set\" {\n\t\treturn SetOwnership(stub, args[0])\n\t} else if funcName == \"get\" {\n\t\treturn GetOwnership(stub, args[0])\n\t} else if funcName == \"list\" {\n\t\treturn GetSurveyNoList(stub, args[0])\n\t}\n\treturn shim.Error(\"Fuction name '\" + funcName + \"' is wrong or \" + funcName + \" function does not exist!!!\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\t\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" { //initialize the chaincode state, used as reset\n\t\treturn t.Init(stub)\n\t} else if function == \"getAllAbattoirReceived\" {\n\t\treturn getAllAbattoirReceived(stub, args[0], args[1])\t\n\t} else if function == \"saveAbattoirReceived\" {\n\t\treturn saveAbattoirReceived(stub, args)\n\t} else if function == \"getAllAbattoirDispatch\" {\n\t\treturn getAllAbattoirDispatch(stub, args[0], args[1])\n\t} else if function == \"saveAbattoirDispatch\" {\n\t\treturn saveAbattoirDispatch(stub, args)\n\t} else if function == \"getAllLogisticTransactions\" {\n\t\treturn getAllLogisticTransactions(stub, args[0], args[1])\n\t} else if function == \"saveLogisticTransaction\" {\n\t\treturn saveLogisticTransaction(stub, args)\n\t} else if function == \"updateLogisticTransactionStatus\" {\n\t\treturn updateLogisticTransactionStatus(stub, args)\n\t} else if function == \"pushIotDetailsToLogisticTransaction\" {\n\t\treturn pushIotDetailsToLogisticTransaction(stub, args)\n\t} else if function == \"getUniqueId\" {\n\t\treturn getUniqueId(stub, args[0], args[1])\n\t} else if function == \"getAllProcessorPOs\" {\n\t\treturn getAllProcessorPOs(stub, args[0], args[1])\n\t}\n\t\n\t// error out\n\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n}", "func (cc *Chaincode) Invoke(stub shim.ChaincodeStubInterface) sc.Response {\n\tfcn, params := stub.GetFunctionAndParameters()\n\tswitch fcn {\n\tcase \"register\":\n\t\treturn userRegister(stub,params)\n\tcase \"change\":\n\t\treturn changeUserStatus(stub,params)\n\tcase \"query\":\n\t\treturn queryUserInfo(stub,params)\n\tcase \"queryAll\":\n\t\treturn queryAllUserInfo(stub)\n\tdefault:\n\t\treturn shim.Error(fmt.Sprintf(\"unsupported function: %s\",fcn))\n\t}\n}", "func (s *SmartContract) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n \n fn, args := stub.GetFunctionAndParameters()\n\n if fn == \"AddComponent\" {\n return s.AddComponent(stub, args)\n } else if fn == \"InitLedger\" {\n return s.InitLedger(stub)\n } else if fn == \"QueryCar\" {\n return s.QueryCar(stub, args)\n } else if fn == \"QueryComponent\" {\n return s.QueryComponent(stub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name.\")\n \n}", "func (t *SimpleChaincode) Invoke(stub *shim.ChaincodeStub, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (t *SimpleChaincode) Invoke(stub *shim.ChaincodeStub, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (t *Subrogationcode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\n\t// Handle different functions\n\tif function == \"init\" { //initialize the chaincode state, used as reset\n\t\tres, err := t.Init(stub, \"init\", args)\n\t\treturn res, err\n\t} else if function == \"reg_claim\" { //create a new klaim\n\t\tres, err := t.reg_claim(stub, args)\n\t\treturn res, err\n\t}\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (t *IPDCChaincode) invoke_bulk(stub shim.ChaincodeStubInterface, args []string) pb.Response {\r\n\r\n\tfmt.Println(\"***********Entering invoke_bulk***********\")\r\n\r\n\tif len(args) > (PROCESSING_LIMIT + 10) {\r\n\r\n\t\tfmt.Println(\"Error: Too many invoke calls in bulk invoke\")\r\n\r\n\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\treturn shim.Error(\"Error: Too many invoke calls in bulk invoke\")\r\n\r\n\t}\r\n\r\n\tsuccess_string := \"[\"\r\n\r\n\tfor i_number, individual_invoke_args := range args {\r\n\r\n\t\ti := fmt.Sprint(i_number)\r\n\r\n\t\tvar list_args_interface []interface{}\r\n\r\n\t\terr_json := json.Unmarshal([]byte(individual_invoke_args), &list_args_interface)\r\n\r\n\t\tif err_json != nil {\r\n\r\n\t\t\tfmt.Println(\"Error: Unable to read the arguments for Invoke no. \" + i + err_json.Error())\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Unable to read the arguments for Invoke no. \" + i + err_json.Error())\r\n\t\t}\r\n\r\n\t\tif len(list_args_interface) < 1 {\r\n\r\n\t\t\tfmt.Println(\"Error: empty payload for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: empty payload for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tvar list_args []string\r\n\r\n\t\t//list_args = make([]string, len(list_args_interface))\r\n\r\n\t\tfor _, value_interface := range list_args_interface {\r\n\r\n\t\t\tvalue_string, ok_value_string := value_interface.(string)\r\n\r\n\t\t\tif !ok_value_string {\r\n\r\n\t\t\t\tfmt.Println(\"Error: Invalid format of payload for Invoke no. \" + i)\r\n\r\n\t\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\t\treturn shim.Error(\"Error: Invalid format of payload for Invoke no. \" + i)\r\n\t\t\t}\r\n\r\n\t\t\tlist_args = append(list_args, value_string)\r\n\t\t}\r\n\r\n\t\tfunction_name := list_args[0]\r\n\r\n\t\targs_to_pass := list_args[1:]\r\n\r\n\t\tkey_for_func := \"FunctionName*\" + function_name\r\n\r\n\t\tvalAsBytes, err := stub.GetState(key_for_func)\r\n\r\n\t\tif err != nil {\r\n\r\n\t\t\tfmt.Println(fmt.Sprintf(\"Error: Failed to get state: \" + err.Error() + \" for Invoke no. \" + i))\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Failed to get state: \" + err.Error() + \" for Invoke no. \" + i)\r\n\r\n\t\t} else if valAsBytes == nil {\r\n\r\n\t\t\tfmt.Println(\"Error: No value for key : \" + key_for_func + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: No value for key : \" + key_for_func + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tvar json_specification interface{}\r\n\r\n\t\terr = json.Unmarshal(valAsBytes, &json_specification)\r\n\r\n\t\tif err != nil {\r\n\r\n\t\t\tfmt.Println(\"Error in decoding Specification JSON\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error in decoding Specification JSON\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tmap_specification, ok1 := json_specification.(map[string]interface{})\r\n\r\n\t\tif !ok1 {\r\n\t\t\tfmt.Println(\"Error Parsing map_specification\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error Parsing map_specification\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\toperation, ok2 := map_specification[\"operation\"]\r\n\r\n\t\tif !ok2 {\r\n\t\t\tfmt.Println(\"Error Parsing operation\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error Parsing operation\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tprimitive_list, ok3 := operation.(map[string]interface{})\r\n\r\n\t\tif !ok3 {\r\n\t\t\tfmt.Println(\"Error Parsing primitive list\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error Parsing primitive list\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tif _, ok3 = primitive_list[\"primitive\"]; !ok3 {\r\n\r\n\t\t\tfmt.Println(\"Error: no primitive operation\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: no primitive operation\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tvar primitive_operation string\r\n\r\n\t\tprimitive_operation, ok3 = primitive_list[\"primitive\"].(string)\r\n\r\n\t\tif !ok3 {\r\n\r\n\t\t\tfmt.Println(\"Error: Invalid primitive operation\" + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Invalid primitive operation\" + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tfmt.Println(\"Primitive operation for Invoke no. \" + i + \" : \" + primitive_operation)\r\n\r\n\t\tvar invoke_response pb.Response\r\n\r\n\t\tif primitive_operation == \"invoke_insert_update\" {\r\n\r\n\t\t\tinvoke_response = t.invoke_insert_update(stub, args_to_pass, map_specification)\r\n\r\n\t\t} else if primitive_operation == \"invoke_update_status\" {\r\n\r\n\t\t\tinvoke_response = t.invoke_update_status(stub, args_to_pass, map_specification)\r\n\r\n\t\t} else if primitive_operation == \"invoke_update_status_with_modification_check\" {\r\n\r\n\t\t\tinvoke_response = t.invoke_update_status_with_modification_check(stub, args_to_pass, map_specification)\r\n\r\n\t\t} else if primitive_operation == \"invoke_delete_record\" {\r\n\r\n\t\t\tinvoke_response = t.invoke_delete_record(stub, args_to_pass, map_specification)\r\n\r\n\t\t} else if primitive_operation == \"query_primary_key\" || primitive_operation == \"query_primary_key_history\" || primitive_operation == \"query_update_status\" || primitive_operation == \"query_customer_invoice_disbursed\" || primitive_operation == \"query_customer_invoice_asn_disbursed\" {\r\n\r\n\t\t\tfmt.Println(\"Error: Query function received as Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Query function received as Invoke no. \" + i)\r\n\r\n\t\t} else if primitive_operation == \"invoke_delete_all_records\" {\r\n\r\n\t\t\tfmt.Println(\"Error: Delete all invoke call not allowed but received as Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Delete all invoke call not allowed but received as Invoke no. \" + i)\r\n\r\n\t\t} else {\r\n\r\n\t\t\tfmt.Println(\"Error: Invalid function \" + function_name + \" for Invoke no. \" + i)\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error: Invalid function \" + function_name + \" for Invoke no. \" + i)\r\n\t\t}\r\n\r\n\t\tif invoke_response.Status != shim.OK {\r\n\r\n\t\t\tfmt.Println(\"Error in executing Invoke no. \" + i + \" : \" + string(invoke_response.Message))\r\n\r\n\t\t\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\t\t\treturn shim.Error(\"Error in executing Invoke no. \" + i + \" : \" + string(invoke_response.Message))\r\n\r\n\t\t} else if i_number == 0 {\r\n\r\n\t\t\tsuccess_string = success_string + string(invoke_response.Message)\r\n\t\t} else {\r\n\r\n\t\t\tsuccess_string = success_string + \",\" + string(invoke_response.Message)\r\n\t\t}\r\n\r\n\t\tfmt.Println(\"Response of executing Invoke no. \" + i + \" : \" + string(invoke_response.Message))\r\n\r\n\t}\r\n\r\n\tsuccess_string = success_string + \"]\"\r\n\r\n\tfmt.Println(\"***********Exiting invoke_bulk***********\")\r\n\r\n\treturn shim.Success([]byte(success_string))\r\n\r\n}", "func (s *UserClient) RpcInvoke(req []byte) ([]byte, error) {\n\t// rpc.send\n\terr := s.rpc.Send(req)\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\t// rpc.receive\n\treturn s.rpc.Receive()\n}", "func (c *HTTPClient) Invoke(name string, params map[string]interface{}, body io.Reader) (Response, error) {\n\tcmd, err := NewCommand(name, params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcmd.SetBody(body)\n\treturn c.Query(cmd)\n}", "func (c *VsmStatsCommand) Run(args []string) int {\n\n\tvar (\n\t\terr, err1, err2 error\n\t\tstatus Status\n\t\tstats1, stats2 VolumeStats\n\t\tstatusArray []string\n\t)\n\n\tflags := c.Meta.FlagSet(\"vsm-stats\", FlagSetClient)\n\tflags.Usage = func() { c.Ui.Output(c.Help()) }\n\tflags.StringVar(&c.Json, \"json\", \"\", \"\")\n\n\tif err := flags.Parse(args); err != nil {\n\t\treturn 1\n\t}\n\targs = flags.Args()\n\tif len(args) < 1 {\n\t\tc.Ui.Error(c.Help())\n\t\treturn 1\n\t}\n\n\tannotations, err := GetVolAnnotations(args[0])\n\tif err != nil || annotations == nil {\n\t\treturn -1\n\t}\n\tif annotations.ControllerStatus != \"Running\" {\n\t\tfmt.Println(\"Volume not reachable\")\n\t\treturn -1\n\t}\n\n\treplicas := strings.Split(annotations.Replicas, \",\")\n\tfor _, replica := range replicas {\n\t\terr, errCode1 := GetStatus(replica+\":9502\", &status)\n\t\tif err != nil {\n\t\t\tif errCode1 == 500 || strings.Contains(err.Error(), \"EOF\") {\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", replica))\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", \"Waiting\"))\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", \"Unknown\"))\n\n\t\t\t} else {\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", replica))\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", \"Offline\"))\n\t\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", \"Unknown\"))\n\t\t\t}\n\t\t} else {\n\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", replica))\n\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%s\", \"Online\"))\n\t\t\tstatusArray = append(statusArray, fmt.Sprintf(\"%d\", status.RevisionCounter))\n\n\t\t}\n\t}\n\n\t//GetVolumeStats gets volume stats\n\terr1, _ = GetVolumeStats(annotations.ClusterIP+\":9501\", &stats1)\n\ttime.Sleep(1 * time.Second)\n\terr2, _ = GetVolumeStats(annotations.ClusterIP+\":9501\", &stats2)\n\n\tif (err1 != nil) || (err2 != nil) {\n\t\tfmt.Println(\"Volume not reachable\")\n\t}\n\n\t//StatsOutput displays output\n\terr = StatsOutput(c, annotations, args, statusArray, stats1, stats2)\n\n\treturn 0\n}", "func (t *RBCApproval) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\targs := stub.GetArgs()\n\tif len(args) < 3 {\n\t\treturn shim.Error(fmt.Sprintf(\"Incorrect number of arguments%v. Expecting 1\", args))\n\t}\n\n\tf := string(args[1])\n\n\tswitch f {\n\tcase \"queryRBCCName\":\n\t\treturn shim.Success([]byte(\"rbcapproval\"))\n\tcase \"queryStateHistory\":\n\t\treturn shim.Success(stub.QueryStateHistory(string(args[2])))\n\tcase \"newApproval\":\n\t\treturn t.newApproval(stub, string(args[2]))\n\tcase \"approval\":\n\t\treturn t.approval(stub, string(args[2]))\n\tcase \"queryOneApproval\":\n\t\treturn t.queryOneApproval(stub, string(args[2]))\n\tcase \"queryAllApproval\":\n\t\treturn t.queryAllApproval(stub, string(args[2]))\n\tdefault:\n\t\tjsonResp := fmt.Sprintf(\"function %s is not found\", f)\n\t\treturn shim.Error(jsonResp)\n\t}\n}", "func (v *VsctlMock) Exec(args ...string) ([]byte, error) {\n\tif debugMocks {\n\t\tfmt.Printf(\"MOCK [Vsctl received: ovs-vsctl %s]\\n\", args)\n\t}\n\tv.ReceivedArgs = append(v.ReceivedArgs, args)\n\n\tif len(v.VsctlResults) == 0 {\n\t\treturn nil, errors.New(\"VsctlMock - results not set\")\n\t}\n\n\tout, err := v.VsctlResults[0].ResultOutcome, v.VsctlResults[0].ResultError\n\tv.VsctlResults = v.VsctlResults[1:]\n\tif debugMocks {\n\t\tfmt.Printf(\"MOCK [Vsctl response: %s]\\n\", out)\n\t}\n\treturn []byte(out), err\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n fmt.Println(\"########### oow18_example Invoke ###########\")\n\tfunction, args := stub.GetFunctionAndParameters()\n\t\n\tif function != \"invoke\" {\n return shim.Error(\"Unknown function call\")\n\t}\n\n\tif len(args) < 2 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting at least 2\")\n\t}\n\n\tif args[0] == \"delete\" {\n\t\t// Deletes an entity from its state\n\t\t// TODO: remove/modify\n\t\treturn t.delete(stub, args)\n\t}\n\n\tif args[0] == \"query\" {\n\t\t// queries an entity state\n\t\treturn t.query(stub, args)\n\t}\n\tif args[0] == \"move\" {\n\t\t// Moves an asset between entities\n\t\t// TODO: remove\n\t\treturn t.move(stub, args)\n\t}\n\tif args[0] == \"create\" {\n\t\t// creates an entity with asset\n\t\t// TODO: remove\n\t\treturn t.create(stub, args)\n\t}\n\tif args[0] == \"createTreeOrder\" {\n\t\t// Create a tree order from username and money\n\t\treturn t.createTreeOrder(stub, args)\n\t}\n\treturn shim.Error(\"Unknown action, check the first argument, must be one of 'delete', 'query', create, or 'move'\")\n}", "func (s *Feature) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\r\n\r\n\t\r\n\r\n\t// Retrieve the requested Smart Contract function and arguments\r\n\tfunction, args := APIstub.GetFunctionAndParameters()\r\n\r\n\t// Route to the appropriate handler function to interact with the ledger appropriately\r\n\tif function == \"initData\" {\r\n\t\treturn initData(APIstub)\r\n\t} else if function == \"getByQuery\" {\r\n\t\treturn s.getByQuery(APIstub, args)\r\n\t} else if function == \"createFeature\" {\r\n\t\treturn createFeature(APIstub, args)\r\n\t} else if function == \"deleteFeature\" {\r\n\t\treturn deleteFeature(APIstub, args)\r\n\t}\r\n\r\n\treturn shim.Error(\"Invalid Smart Contract function name: \" + function)\r\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n //args := APIstub.GetFunctionAndParameters()\n args := APIstub.GetStringArgs()\n function := args[0]\n if function == \"addRecord\" {\n return s.addRecord(APIstub, args)\n } else if function == \"getRecord\" {\n return s.getRecord(APIstub, args)\n } else if function == \"encRecord\" {\n return s.encRecord(APIstub, args)\n } else if function == \"decRecord\" {\n return s.decRecord(APIstub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name \"+function)\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tlogger.Info(\"########### example_cc0 Invoke ###########\")\n\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\tif function == \"delete\" {\n\t\t// Deletes an entity from its state\n\t\treturn t.delete(stub, args)\n\t}\n\n\tif function == \"query\" {\n\t\t// queries an entity state\n\t\treturn t.query(stub, args)\n\t}\n\tif function == \"move\" {\n\t\t// Deletes an entity from its state\n\t\tlogger.Infof(\"MOVE\")\n\t\treturn t.move(stub, args)\n\t}\n\t\t\n\tif function == \"read_everything\"{ //read everything, \n\t\treturn read_everything(stub)\n\t}\n\t\n\tif function == \"getHistory\"{ //getHistory , \n\tlogger.Infof(\"getHistory\")\n\t\treturn getHistory(stub,args)\n\t}\n\tif function == \"queryByCP\"{ //queryByCP, \n\tlogger.Infof(\"queryByCP\")\n\t\treturn queryByCP(stub,args)\n\t}\n\tif function == \"bulkInsert\"{ //bulkInsert, \n\tlogger.Infof(\"bulkInsert\")\n\t\treturn bulkInsert(stub,args)\n\t}\t\n\t\n\tif function == \"update\"{ //bulkInsert, \n\tlogger.Infof(\"update\")\n\t\treturn update(stub,args)\n\t}\t\n\t\n\t\n\t\n\n\tlogger.Errorf(\"Unknown action, check the first argument, must be one of 'delete', 'query', read_everything, getHistory, queryByCP, update or 'move'. But got: %v\", args[0])\n\treturn shim.Error(fmt.Sprintf(\"Unknown action, check the first argument, must be one of 'delete', 'query', read_everything, getHistory, queryByCP, update or 'move'. But got: %v\", args[0]))\n}", "func (t *StoreCC) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tlogger.Info(\"########### example_cc0 Invoke ###########\")\n\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\tif function == \"load\" {\n\t\t// queries an entity state\n\t\treturn t.load(stub, args)\n\t} else if function == \"save\" {\n\t\treturn t.save(stub, args)\n\t}\n\n\tlogger.Errorf(\"Unknown action, check the first argument, must be one of 'delete', 'query', or 'move'. But got: %v\", args[0])\n\treturn shim.Error(fmt.Sprintf(\"Unknown action, check the first argument, must be one of 'delete', 'query', or 'move'. But got: %v\", args[0]))\n}", "func (t *CertificateChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"issue\" { //create a new marble\n\t\treturn t.issue(stub, args)\n\t} else if function == \"assign\" { //change owner of a specific marble\n\t\treturn t.assign(stub, args)\n\t} else if function == \"append\" { //transfer all marbles of a certain color\n\t\treturn t.update(stub, args)\n\t} else if function == \"query\" { //transfer all marbles of a certain color\n\t\treturn t.query(stub, args)\n\t}\n\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\t// Route to the appropriate handler function to interact with the ledger appropriately\n\tif function == \"addVehicle\" {\n\t\treturn s.addVehicle(APIstub, args)\n\t} else if function == \"getVehicle\" {\n\t\treturn s.getVehicle(APIstub,args)\n\t} else if function == \"transferVehicle\" {\n\t\treturn s.transferVehicle(APIstub, args)\n\t} else if function == \"getVehicleHistory\" {\n\t\treturn s.getVehicleHistory(APIstub, args)\n\t}\n\t/*\n\telse if function == \"addInsurace\" {\n\t\treturn s.addInsurace(APIstub, args)\n\t} else if function == \"addRTANumber\" { \n\t\treturn s.addRTANumber(APIstub, args)\n\t} else if function == \"changeOwnership\" {\n\t\treturn s.changeOwnership(APIstub, args)\n\t} else if function == \"addService\" {\n\t\treturn s.addService(APIstub, args)\n\t} else if function == \"addnewClaim\" {\n\t\treturn s.addnewClaim(APIstub, args)\n\t}\n\t*/\n\n\treturn shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tfmt.Println(\"function\", function)\n\t// Route to the appropriate handler function to interact with the ledger appropriately\n\tif function == \"initLedger\" {\n\t\treturn s.initLedger(APIstub)\n\t} else if function == \"queryAllVehicles\" {\n\t\treturn s.queryAllVehicles(APIstub)\n\t} else if function == \"updateVehicleVelocity\" {\n\t\treturn s.updateVehicleVelocity(APIstub, args)\n\t} else if function == \"queryVehicle\" {\n\t return s.queryVehicle(APIstub, args)\n\t} else if function == \"createVehicle\" {\n\t return s.createVehicle(APIstub, args)\n\t}\n\n\treturn shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tlogger.Info(\"########### example_cc0 Invoke ###########\")\n\n\tfunction, args := stub.GetFunctionAndParameters()\n\t\n\tif function == \"saveCustomer\" {\n\t\t// save an entity in its state\n\t\treturn t.saveCustomer(stub, args)\n\t}\n\tif function == \"addAttachment\" {\n\t\t// save an entity in its state\n\t\treturn t.addAttachment(stub, args)\n\t}\n\tif function == \"changeStatus\" {\n\t\t// queries an entity state\n\t\treturn t.changeStatus(stub, args)\n\t}\n\tif function == \"query\" {\n\t\t// queries an entity state\n\t\treturn t.query(stub, args)\n\t}\n\n\tlogger.Errorf(\"Unknown action, check the first argument, must be one of 'delete', 'query', or 'move'. But got: %v\", args[0])\n\treturn shim.Error(fmt.Sprintf(\"Unknown action, check the first argument, must be one of 'delete', 'query', or 'move'. But got: %v\", args[0]))\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t} else if function == \"addemployer\" {\n\t\treturn t.addEmployer(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation: \" + function)\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"initDegree\" { //create a newdegree\n\t\treturn t.initDegree(stub, args)\n\t} else if function == \"readDegree\" { //read a degree\n\t\treturn t.readDegree(stub, args)\n\t} else if function == \"queryDegreeByOwner\" { //find marbles for owner X using rich query\n\t\treturn t.queryDegreeByOwner(stub, args)\n\t} else if function == \"queryDegrees\" { //find marbles based on an ad hoc rich query\n\t\treturn t.queryDegrees(stub, args)\n\t} else if function == \"getHistoryForDegree\" { //get history of values for a marble\n\t\treturn t.getHistoryForDegree(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\treturn shim.Error(\"Received unknown function invocation\")\n}", "func (e *ExecutableInvoker) Invoke(ctx context.Context, m *Manifest, cfg *InvokerConfig) error {\n\texecPath := path.Join(e.PluginDir, path.Join(m.Command...), m.Exec)\n\tcmd := execCommandContext(ctx, execPath, cfg.Args...)\n\tcmd.Env = append(cmd.Env, cfg.Env...)\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\n\treturn cmd.Run()\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t} else if function == \"startShipment\" {\n\t\treturn t.addNewShipment(stub, args)\n\t}else if function == \"transferOwner\" {\n\t\treturn t.transferOwner(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation: \" + function)\n}", "func (s *server) OnInvoke(ctx context.Context, in *commonv1pb.InvokeRequest) (*commonv1pb.InvokeResponse, error) {\n\tvar response string\n\n\tswitch in.Method {\n\tcase \"EchoMethod\":\n\t\tresponse = s.EchoMethod()\n\t}\n\n\treturn &commonv1pb.InvokeResponse{\n\t\tContentType: \"text/plain; charset=UTF-8\",\n\t\tData: &any.Any{Value: []byte(response)},\n\t}, nil\n}", "func (t *ManageMerchant) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\r\n\tfmt.Println(\"invoke is running \" + function)\r\n\r\n\t// Handle different functions\r\n\tif function == \"init\" {\t\t\t\t\t\t\t\t\t\t\t\t\t//initialize the chaincode state, used as reset\r\n\t\treturn t.Init(stub, \"init\", args)\r\n\t}else if function == \"createMerchant\" {\t\t\t\t\t\t\t\t\t\t\t//create a new Merchant\r\n\t\treturn t.createMerchant(stub, args)\r\n\t}else if function == \"deleteMerchant\" {\t\t\t\t\t\t\t\t\t// delete a Merchant\r\n\t\treturn t.deleteMerchant(stub, args)\r\n\t}else if function == \"updateMerchant\" {\t\t\t\t\t\t\t\t\t//update a Merchant\r\n\t\treturn t.updateMerchant(stub, args)\r\n\t}\r\n\tfmt.Println(\"invoke did not find func: \" + function)\r\n\terrMsg := \"{ \\\"message\\\" : \\\"Received unknown function invocation\\\", \\\"code\\\" : \\\"503\\\"}\"\r\n\terr := stub.SetEvent(\"errEvent\", []byte(errMsg))\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t} \r\n\treturn nil, nil\t\t\t//error\r\n}", "func (wk *Worker) InvokeService(args serverless.RPCArgs, _ *struct{}) error {\n\t// TODO: implement me\n\t// Hint: You should locate the interested service registered from serviceMap.\n\t// and call `service.interf.DoService` to make the call to the plugin service.\n\t// TODO TODO TODO\n\t//\n\treturn nil\n}", "func (t *SimpleChaincode) invoke(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tfmt.Printf(\"================ Invoke on chaincode 05\\n\")\n\tif len(args) != 3 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 3\")\n\t}\n\n\tchaincodeURL := args[0]\n\tval1 := args[1]\n\tval2 := args[1]\n channelName := args[2]\n\tfmt.Printf(\"Channel Name is %s\\n\", channelName)\n\n\t// Invoke on chaincode_example02\n\tf := \"put\"\n\tinvokeArgs := util.ToChaincodeArgs(f, \"a\", val1)\n\tresponse := stub.InvokeChaincode(chaincodeURL, invokeArgs, channelName)\n\tif response.Status != shim.OK {\n\t\terrStr := fmt.Sprintf(\"Failed to query chaincode. Got error: %s\", response.Payload)\n\t\tfmt.Printf(errStr)\n\t\treturn shim.Error(errStr)\n\t}\n\tfmt.Printf(\"Invoke response %s\\n\", string(response.Payload))\n\n\tinvokeArgs = util.ToChaincodeArgs(f, \"b\", val2)\n\tresponse = stub.InvokeChaincode(chaincodeURL, invokeArgs, channelName)\n\tif response.Status != shim.OK {\n\t\terrStr := fmt.Sprintf(\"Failed to query chaincode. Got error: %s\", response.Payload)\n\t\tfmt.Printf(errStr)\n\t\treturn shim.Error(errStr)\n\t}\n\tfmt.Printf(\"Invoke response %s\\n\", string(response.Payload))\n\treturn shim.Success([]byte(\"OK\"))\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"write\" {\n\t\treturn t.write(stub, args)\n\t}else if function == \"putcontract\"{\n\t\treturn t.putcontract(stub, args)\n\t}else if function == \"statequery\"{\n\t\treturn t.statequery(stub, args)\n\t}\n\n\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation: \" + function)\n}", "func (t *ProductChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tlogger.Debug(\"Invoke\")\n\n\tfunction, args := stub.GetFunctionAndParameters()\n\tlogger.Debug(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"initProduct\" { //create a new product\n\t\treturn t.initProduct(stub, args)\n\t} else if function == \"updateProduct\" { //update an existing product\n\t\treturn t.updateProduct(stub, args)\n\t} else if function == \"updateOwner\" { //update an owner of an existing product\n\t\treturn t.updateOwner(stub, args)\n\t} else if function == \"readProduct\" { //read a product\n\t\treturn t.readProduct(stub, args)\n\t} else if function == \"queryProductsByOwner\" { //find products for the owner X using rich query\n\t\treturn t.queryProductsByOwner(stub, args)\n\t} else if function == \"queryProducts\" { //find products based on an ad hoc rich query\n\t\treturn t.queryProducts(stub, args)\n\t} else if function == \"getHistoryForProduct\" { //get history of values for a product\n\t\treturn t.getHistoryForProduct(stub, args)\n\t} else if function == \"getProductsByLabel\" { // Get product by label\n\t\treturn t.queryProductsByLabel(stub, args)\n\t} else if function == \"increaseQuantity\" { // Increase Quantity\n\t\treturn t.increaseQuantity(stub, args)\n\t} else if function == \"createUser\" {\n\t\treturn t.createUser(stub, args)\n\t} else if function == \"queryUser\" {\n\t\treturn t.queryUser(stub, args)\n\t} else if function == \"queryManufacturerByRole\" {\n\t\treturn t.queryManufacturerByRole(stub, args)\n\t} else if function == \"queryDealerByRole\" {\n\t\treturn t.queryDealerByRole(stub, args)\n\t} else if function == \"queryCustomerByRole\" {\n\t\treturn t.queryCustomerByRole(stub, args)\n\t} else if function == \"queryProductsByName\" {\n\t\treturn t.queryProductsByName(stub, args)\n\t}\n\n\tlogger.Debug(\"invoke did not find func: \" + function) //error\n\treturn pb.Response{Status: 403, Message: \"Invalid invoke function name.\"}\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\t// Route to the appropriate handler function to interact with the ledger appropriately\n\tswitch function {\n\tcase \"initVote\":\n\t\treturn s.initVote(APIstub)\n\tcase \"beginSignUp\":\n\t\treturn s.beginSignUp(APIstub, args)\n\tcase \"finishRegistrationPhase\":\n\t\treturn s.finishRegistrationPhase(APIstub)\n\tcase \"submitVote\":\n\t\treturn s.submitVote(APIstub, args)\n\tcase \"setEligible\":\n\t\treturn s.setEligible(APIstub, args)\n\tcase \"register\":\n\t\treturn s.register(APIstub, args)\n\tcase \"computeTally\":\n\t\treturn s.computeTally(APIstub)\n\tcase \"question\":\n\t\treturn s.question(APIstub)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name: \" + function)\n\t}\n}", "func (d *DefaultActorContainerContext) Invoke(ctx context.Context, methodName string, param []byte) ([]reflect.Value, actorErr.ActorErr) {\n\tmethodType, ok := d.methodType[methodName]\n\tif !ok {\n\t\treturn nil, actorErr.ErrActorMethodNoFound\n\t}\n\targsValues := make([]reflect.Value, 0)\n\targsValues = append(argsValues, reflect.ValueOf(d.actor), reflect.ValueOf(ctx))\n\tif len(methodType.argsType) > 0 {\n\t\ttyp := methodType.argsType[0]\n\t\tparamValue := reflect.New(typ)\n\t\tparamInterface := paramValue.Interface()\n\t\tif err := d.serializer.Unmarshal(param, paramInterface); err != nil {\n\t\t\treturn nil, actorErr.ErrActorMethodSerializeFailed\n\t\t}\n\t\targsValues = append(argsValues, reflect.ValueOf(paramInterface).Elem())\n\t}\n\treturnValue := methodType.method.Func.Call(argsValues)\n\treturn returnValue, actorErr.Success\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\n\tvar r Request\n\n\tif function == \"create_request\" {\n return t.create_request(stub, args)\n\t} else {\n\t\t\t// If the function is not a create then there must be a request so we need to retrieve it.\n\t\t\tbytes, err := stub.GetState(args[0])\n\n\t\t\tif err != nil {\tfmt.Printf(\"INVOKE: reqeust can't be found : %s\", err); return nil, errors.New(\"INVOKE: reqeust can't be found \")\t}\n\n\t\t\terr = json.Unmarshal(bytes, &r);\n\t\t\tif err != nil {\tfmt.Printf(\"INVOKE: request corrupted : %s\", err); return nil, errors.New(\"INVOKE: reqeust corrupted \"+string(bytes))\t}\n\n\t\t\treturn t.review_request(stub, r)\n\t\t}\n\n\treturn nil, errors.New(\"Function of the name \"+ function +\" doesn't exist.\")\n\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\tfmt.Println(\"invoke is running \" + function)\n\n\t// Handle different functions\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t} else if function == \"initAssset\" {\n\t\treturn t.initAssset(stub, args)\n\t} else if function == \"ownerUpdation\" {\n\t\treturn t.updateOwner(stub, args)\n\t} else if function == \"initContract\" {\n\t\treturn t.initContract(stub, args)\n\t} else if function == \"contractUpdation\" {\n\t\treturn t.updateContract(stub, args)\n\t} else if function == \"readyForShipment\" {\n\t\treturn t.toReadyForShipment(stub, args)\n\t} else if function == \"inTransit\" {\n\t\treturn t.toInTransit(stub, args)\n\t} else if function == \"shipmentReached\" {\n\t\treturn t.toShipmentReached(stub, args)\n\t} else if function == \"shipmentDelivered\" {\n\t\treturn t.toShipmentDelivered(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function) //error\n\n\treturn nil, errors.New(\"Received unknown function invocation: \" + function)\n}", "func (c *connector) Invoke(ctx radio.Context) error {\n\tif ctx == nil {\n\t\treturn errors.New(\"no context given\")\n\t}\n\n\t// Obtaining exclusive lock on command\n\tc.lock.Lock()\n\tdefer c.lock.Unlock()\n\n\t// Sending indicator, that command in progress\n\tctx.CommandInProgress()\n\n\t// Running with os.Std replacements\n\tvar err error\n\tsout, serr := std.BindStrings(func() {\n\t\tc.cmd.SetArgs(ctx.GetArgs())\n\t\terr = c.cmd.Execute()\n\t})\n\n\t// Packing response\n\tif len(serr) > 0 {\n\t\tctx.SendMessage(serr)\n\t}\n\tif len(sout) > 0 {\n\t\tctx.SendMessage(sout)\n\t}\n\tif err != nil {\n\t\tctx.SendMessage(err)\n\t}\n\n\treturn err\n}", "func (t *SimpleChaincode) invoke(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tvar id, name string // Entities\n\tvar info string\n\tvar err error\n\n\tif len(args) != 3 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 3\")\n\t}\n\n\tid = args[0]\n\tname = args[1]\n\n\t// Get the state from the ledger\n\t// TODO: will be nice to have a GetAllState call to ledger\n\tidValbytes, err := stub.GetState(id)\n\tnameValbytes, err := stub.GetState(name)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get state\")\n\t}\n\tif idValbytes == nil {\n\t\treturn shim.Error(\"Entity not found\")\n\t}\n\n\t//info = string(idValbytes)\n\n\t// Perform the execution\n\tinfo = args[2]\n\n\t// Write the state back to the ledger\n\terr = stub.PutState(id, []byte(info))\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\t// modify data to couchdb\n\tdb := client.Use(\"test\")\n\tdoc := &Document{\n Id: id,\n Name : name,\n Data : info,\n }\n result, err := db.Put(doc)\n if err != nil {\n panic(err)\n }\n\n\treturn shim.Success(nil)\n}", "func (t *SimpleChaincode) Invoke(APIstub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\t// Handle different functions\n\n\tif function == \"addStudent\" { //add a Student\n\t\treturn t.addStudent(APIstub, args)\n\t} else if function == \"readStudent\" { //read a Student\n\t\treturn t.readStudent(APIstub, args)\n\t} else if function == \"addCert\" { //add a Certificate\n\t\treturn t.addCert(APIstub, args)\n\t} else if function == \"readCert\" { //read a Certificate\n\t\treturn t.readCert(APIstub, args)\n\t} else if function == \"transferCert\" { //transfer a Certificate\n\t\treturn t.transferCert(APIstub, args)\n\t} else if function == \"initLedger\" {\n\t\treturn t.initLedger(APIstub, args)\n\t} else if function == \"queryAllCert\" {\n\t\treturn t.queryAllCert(APIstub, args)\n\t} else if function == \"login\" {\n\t\treturn t.login(APIstub, args)\n\t} else if function == \"uniCredentials\" {\n\t\treturn t.uniCredentials(APIstub, args)\n\t} else if function == \"creatorCredentials\" {\n\t\treturn t.creatorCredentials(APIstub, args)\n\t}\n\treturn shim.Error(\"Received unknown function invocation\")\n\n}", "func (t *AssetManagementChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\n\t// Handle different functions\n\tif function == \"createInvoice\" {\n\t\t// Assign ownership\n\t\treturn t.createInvoice(stub, args)\n\t} else if function == \"approveInvoice\" {\n\t\t// Transfer ownership\n\t\treturn t.approveInvoice(stub, args)\n\t} else if function == \"createPaymentRequest\" {\n\t\t// Transfer ownership\n\t\treturn t.createPaymentRequest(stub, args)\n\t} else if function == \"assignPaymentRequest\" {\n\t\t// Transfer ownership\n\t\treturn t.assignPaymentRequest(stub, args)\n\t}\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\t// Route to the appropriate handler function to interact with the ledger appropriately\n\tif function == \"initLedger\" {\n\t\treturn s.initLedger(APIstub)\n\t} else if function == \"queryAll\" { //return all the assets on the ledger\n\t\treturn s.queryAll(APIstub, args)\n\t} else if function == \"query\" { //single bank or customer or forexPair\n\t\treturn s.query(APIstub, args)\n\t} else if function == \"pay\" { //execute a payment between two currencies\n\t\treturn s.pay(APIstub, args)\n\t} else if function == \"createBank\" {\n\t\treturn s.createBank(APIstub, args)\n\t} else if function == \"createCustomer\" {\n\t\treturn s.createCustomer(APIstub, args)\n\t} else if function == \"createForex\" {\n\t\treturn s.createForex(APIstub, args)\n\t}\n\n\treturn shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (cc *Chaincode) Invoke(stub shim.ChaincodeStubInterface) sc.Response {\n\tfcn, params := stub.GetFunctionAndParameters()\n\tfmt.Println(\"Invoke()\", fcn, params)\n\n\tif fcn == \"createNewChargeSheet\" {\n\t\treturn cc.createNewChargeSheet(stub, params)\n\t} else if fcn == \"readChargeSheet\" {\n\t\treturn cc.readChargeSheet(stub, params)\n\t} else if fcn == \"addFIRIDs\" {\n\t\treturn cc.addFIRIDs(stub, params)\n\t} else if fcn == \"addSectionOfLaw\" {\n\t\treturn cc.addSectionOfLaw(stub, params)\n\t} else if fcn == \"addInvestigatingOfficer\" {\n\t\treturn cc.addInvestigatingOfficer(stub, params)\n\t} else if fcn == \"addInvestigationID\" {\n\t\treturn cc.addInvestigationID(stub, params)\n\t} else if fcn == \"addAccusedPerson\" {\n\t\treturn cc.addAccusedPerson(stub, params)\n\t} else if fcn == \"addBriefReport\" {\n\t\treturn cc.addBriefReport(stub, params)\n\t} else if fcn == \"addChargedPerson\" {\n\t\treturn cc.addChargedPerson(stub, params)\n\t} else if fcn == \"addJudgementID\" {\n\t\treturn cc.addJudgementID(stub, params)\n\t} else if fcn == \"setComplete\" {\n\t\treturn cc.setComplete(stub, params)\n\t} else {\n\t\tfmt.Println(\"Invoke() did not find func: \" + fcn)\n\t\treturn shim.Error(\"Received unknown function invocation!\")\n\t}\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n\n\t// Handle different functions\n\tfmt.Println(\"invoke is running \" + function)\n\tif function == \"init\" {\n\t\treturn t.Init(stub, \"init\", args)\n\t\t//\t} else if function == \"write\" {\n\t\t//\t\treturn t.write(stub, args)\n\t} else if function == \"transfer\" {\n\t\treturn t.transfer(stub, args)\n\t} else if function == \"addAccount\" {\n\t\treturn t.addAccount(stub, args)\n\t}\n\tfmt.Println(\"invoke did not find func: \" + function)\n\n\treturn nil, errors.New(\"Received unknown function invocation\")\n}", "func (s XMLLangService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.XMLLangService.LoadTranslation:\n\t\tvar args = struct {\n\t\t\tLanguage string `json:\"language\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"language\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.LoadTranslation(args.Language))\n\n\tcase RPC.XMLLangService.TranslateEntity:\n\t\tvar args = struct {\n\t\t\tNamespace string `json:\"namespace\"`\n\t\t\tEntity string `json:\"entity\"`\n\t\t\tLanguage string `json:\"language\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"namespace\", \"entity\", \"language\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.TranslateEntity(args.Namespace, args.Entity, args.Language))\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (c *Client) Invoke(ctx context.Context, method string, args, res interface{}, opts ...grpc.CallOption) error {\n\tconn, err := c.Conn()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn conn.Invoke(ctx, method, args, res, opts...)\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\t// Route to the appropriate handler function to interact with the ledger appropriately\n\tif function == \"queryID\" {\n\t\treturn s.queryID(APIstub,args)\n\t}else if function == \"createBirth\" {\n\t\treturn s.createBirth(APIstub, args)\n\t}else if function == \"createHuman\" {\n\t\treturn s.createHuman(APIstub, args)\n\t}else if function == \"marry\" {\n\t\treturn s.marry(APIstub, args)\n\t}else if function == \"marryCheck\" {\n\t\treturn s.marryCheck(APIstub, args)\n\t}else if function == \"divorceCheck\" {\n\t\treturn s.divorceCheck(APIstub, args)\n\t}else if function == \"divorce\" {\n\t\treturn s.divorce(APIstub, args)\n\t}else if function == \"createCheck\" {\n\t\treturn s.createCheck(APIstub, args)\n\t}else if function == \"queryMarryCheck\" {\n\t\treturn s.queryMarryCheck(APIstub,args)\n\t}else if function == \"queryCreatCheck\" {\n\t\treturn s.queryCreatCheck(APIstub,args)\n\t}\n\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t\n}", "func (t *Task) invoke(ctx context.Context, resch chan TaskResult) {\n\tresult := &TaskResult{\n\t\tindex: t.index,\n\t\tName: t.name,\n\t\tDone: true,\n\t}\n\tparams := []reflect.Value{}\n\n\tgo func() {\n\t\tres := t.valueOf.Call(params)\n\t\tswitch t.returnType {\n\t\tcase TaskReturnErrorOnly:\n\t\t\tif v := res[0].Interface(); v != nil {\n\t\t\t\terr, ok := v.(error)\n\t\t\t\tif !ok {\n\t\t\t\t\tresult.Error = fmt.Errorf(\"value is not an error\")\n\t\t\t\t}\n\t\t\t\tresult.Error = err\n\t\t\t}\n\t\tcase TaskReturnResultOnly:\n\t\t\tresult.Result = res[0].Interface()\n\t\tcase TaskReturnResultAndError:\n\t\t\tresult.Result = res[0].Interface()\n\t\t\tif v := res[1].Interface(); v != nil {\n\t\t\t\terr, ok := v.(error)\n\t\t\t\tif !ok {\n\t\t\t\t\tresult.Error = fmt.Errorf(\"value is not an error\")\n\t\t\t\t}\n\t\t\t\tresult.Error = err\n\t\t\t}\n\t\tdefault:\n\t\t\tresult.Error = fmt.Errorf(\"invalid function return type (%d)\", t.returnType)\n\t\t}\n\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\tcase resch <- *result:\n\t\t}\n\t}()\n}", "func (cc *Game) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\n\tfunction, args := stub.GetFunctionAndParameters()\n\n\tvar result []byte\n\tvar err error\n\n\tif function == SETUPGAME {\n\t\tresult, err = setupGame(stub, args)\n\t} else if function == PLAY {\n\t\tresult, err = play(stub, args)\n\t} else if function == GETPLAYERPOINTS {\n\t\tresult, err = getPlayerPoints(stub, args)\n\t} else if function == GETPLAYERTURN {\n\t\tresult, err = getPlayerTurn(stub, args)\n\t} else if function == \"\" {\n\t\terr = errors.New(\"Chaincode invoke function name should not be empty\")\n\t} else {\n\t\terr = errors.New(\"Invalid chaincode invoke function name\")\n\t}\n\n\tif err != nil {\n\t\tfmt.Println(\"Error occured on chaincode invoke: - \", err.Error())\n\t\treturn shim.Error(err.Error())\n\t}\n\n\treturn shim.Success(result)\n}" ]
[ "0.56265765", "0.5580322", "0.55538565", "0.53817976", "0.5344088", "0.5284873", "0.52635103", "0.52544475", "0.52204365", "0.5205676", "0.5204454", "0.51735866", "0.51717585", "0.5162761", "0.51490843", "0.51134276", "0.50971574", "0.50762224", "0.50643224", "0.50294834", "0.5024659", "0.5001411", "0.4999155", "0.49783912", "0.497471", "0.49703318", "0.49668232", "0.49598077", "0.49581715", "0.49556375", "0.49399358", "0.49372873", "0.49357915", "0.49313384", "0.49265522", "0.49215698", "0.4915033", "0.4911383", "0.4910838", "0.49108174", "0.49004602", "0.48881143", "0.48740914", "0.48731393", "0.487126", "0.48706427", "0.48690036", "0.48540926", "0.48499689", "0.48473", "0.48454726", "0.48429453", "0.4841707", "0.48331213", "0.4832685", "0.48317033", "0.48300394", "0.48300394", "0.4826869", "0.4812641", "0.48124456", "0.47945628", "0.47891477", "0.47883868", "0.47756335", "0.4775019", "0.47719452", "0.476807", "0.47677702", "0.4754203", "0.47534242", "0.47416556", "0.4729402", "0.472527", "0.47205046", "0.47194734", "0.47193295", "0.47105664", "0.47086003", "0.47005838", "0.46966407", "0.46949995", "0.4694622", "0.46946064", "0.46945193", "0.46924385", "0.46846813", "0.46826854", "0.46812564", "0.46805218", "0.46783692", "0.46716583", "0.46715096", "0.4666076", "0.46657515", "0.46629605", "0.46614805", "0.465915", "0.46588737", "0.4645198" ]
0.6023194
0
callRemoteSignature obtains type and header information from the remote mojo service. Remote mojo interfaces all define a signature method.
func (fs fakeService) callRemoteSignature(mojourl string, mojoname string) (mojomInterface mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, err error) { // TODO(afandria): The service_describer mojom file defines the constant, but // it is not actually present in the generated code: // https://github.com/domokit/mojo/issues/469 // serviceDescriberInterfaceName := "_ServiceDescriber" r, p := service_describer.CreateMessagePipeForServiceDescriber() fs.appctx.ConnectToApplication(mojourl).ConnectToService(&r) sDescriber := service_describer.NewServiceDescriberProxy(p, bindings.GetAsyncWaiter()) defer sDescriber.Close_Proxy() r2, p2 := service_describer.CreateMessagePipeForServiceDescription() err = sDescriber.DescribeService(mojoname, r2) if err != nil { return } sDescription := service_describer.NewServiceDescriptionProxy(p2, bindings.GetAsyncWaiter()) defer sDescription.Close_Proxy() mojomInterface, err = sDescription.GetTopLevelInterface() if err != nil { return } descPtr, err := sDescription.GetAllTypeDefinitions() if err != nil { return } return mojomInterface, *descPtr, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (ri reflectInvoker) Signature(ctx *context.T, call ServerCall) ([]signature.Interface, error) {\n\treturn signature.CopyInterfaces(ri.sig), nil\n}", "func (fs fakeService) Signature(ctx *context.T, call rpc.ServerCall) ([]signature.Interface, error) {\n\tctx.Infof(\"Fake Service Signature???\")\n\treturn nil, nil\n}", "func (fs fakeService) callRemoteMethod(ctx *context.T, method string, mi mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, argptrs []interface{}) ([]*vom.RawBytes, error) {\n\t// We need to parse the signature result to get the method relevant info out.\n\tfound := false\n\tvar ordinal uint32\n\tfor ord, mm := range mi.Methods {\n\t\tif *mm.DeclData.ShortName == method {\n\t\t\tordinal = ord\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"callRemoteMethod: method %s does not exist\", method)\n\t}\n\n\tmm := mi.Methods[ordinal]\n\n\t// A void function must have request id of 0, whereas one with response params\n\t// should have a unique request id.\n\theader := bindings.MessageHeader{\n\t\tType: ordinal,\n\t\tFlags: bindings.MessageExpectsResponseFlag,\n\t\tRequestId: fs.ids.Count(),\n\t}\n\n\t// Now produce the *bindings.Message that we will send to the other side.\n\tinType, err := transcoder.MojomStructToVDLType(mm.Parameters, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmessage, err := encodeMessageFromVom(header, argptrs, inType)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Otherwise, make a generic call with the message.\n\toutMessage, err := fs.callRemoteWithResponse(ctx, message)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Decode the *vom.RawBytes from the mojom bytes and mojom type.\n\toutType, err := transcoder.MojomStructToVDLType(*mm.ResponseParams, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttarget := util.StructSplitTarget()\n\tif err := transcoder.FromMojo(target, outMessage.Payload, outType); err != nil {\n\t\treturn nil, fmt.Errorf(\"transcoder.FromMojo failed: %v\", err)\n\t}\n\treturn target.Fields(), nil\n}", "func (ri reflectInvoker) MethodSignature(ctx *context.T, call ServerCall, method string) (signature.Method, error) {\n\t// Return the first method in any interface with the given method name.\n\tfor _, iface := range ri.sig {\n\t\tif msig, ok := iface.FindMethod(method); ok {\n\t\t\treturn signature.CopyMethod(msig), nil\n\t\t}\n\t}\n\treturn signature.Method{}, verror.ErrUnknownMethod.Errorf(ctx, \"method does not exist: %v\", method)\n}", "func (fs fakeService) MethodSignature(ctx *context.T, call rpc.ServerCall, method string) (signature.Method, error) {\n\tctx.Infof(\"Fake Service Method Signature???\")\n\treturn signature.Method{}, nil\n}", "func (fs fakeService) Invoke(ctx *context.T, call rpc.StreamServerCall, method string, argptrs []interface{}) (results []interface{}, _ error) {\n\t// fs.suffix consists of the mojo url and the application/interface name.\n\t// The last part should be the name; everything else is the url.\n\tparts := strings.Split(fs.suffix, \"/\")\n\tmojourl := strings.Join(parts[:len(parts)-1], \"/\") // e.g., mojo:go_remote_echo_server. May be defined in a BUILD.gn file.\n\tmojoname := parts[len(parts)-1] // e.g., mojo::examples::RemoteEcho. Defined from the interface + module.\n\n\t// Create the generic message pipe. r is a bindings.InterfaceRequest, and\n\t// p is a bindings.InterfacePointer.\n\tr, p := bindings.CreateMessagePipeForMojoInterface()\n\tv := v23ServiceRequest{\n\t\trequest: r,\n\t\tname: mojoname,\n\t} // v is an application.ServiceRequest with mojoname\n\n\t// Connect to the mojourl.\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&v)\n\n\t// Then assign a new router the FakeService.\n\t// This will never conflict because each FakeService is only invoked once.\n\tfs.router = bindings.NewRouter(p.PassMessagePipe(), bindings.GetAsyncWaiter())\n\tdefer fs.Close_Proxy()\n\n\tctx.Infof(\"Fake Service Invoke (Remote Signature: %q -- %q)\", mojourl, mojoname)\n\n\t// Vanadium relies on type information, so we will retrieve that first.\n\tmojomInterface, desc, err := fs.callRemoteSignature(mojourl, mojoname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Signature %v\", mojomInterface)\n\tctx.Infof(\"Fake Service Invoke (Remote Method: %v)\", method)\n\n\t// With the type information, we can make the method call to the remote interface.\n\tmethodResults, err := fs.callRemoteMethod(ctx, method, mojomInterface, desc, argptrs)\n\tif err != nil {\n\t\tctx.Errorf(\"Method called failed: %v\", err)\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Results %v\", methodResults)\n\n\t// Convert methodResult to results.\n\tresults = make([]interface{}, len(methodResults))\n\tfor i := range methodResults {\n\t\tresults[i] = &methodResults[i]\n\t}\n\treturn results, nil\n}", "func (c *CallCommon) Signature() *types.Signature {\n\tif c.Method != nil {\n\t\treturn c.Method.Type().(*types.Signature)\n\t}\n\treturn typeparams.CoreType(c.Value.Type()).(*types.Signature)\n}", "func Signature(method string) []byte {\n\t// hash method\n\thasher := sha3.NewLegacyKeccak256()\n\thasher.Write([]byte(method))\n\tb := hasher.Sum(nil)\n\treturn b[:4]\n}", "func GetInvokeTransactionSignature(stub *shim.ChaincodeStub, args []string) ([]byte, error) {\n\n\tif(isRegisteredInvoker(stub)){\n\t\t//Retrieve the invoker initiating the transaction\n\t\tinvoker,error := getInvoker(stub)\n\t\tif error != nil {\n\t\t\treturn nil, errors.New(\"Error retrieving the Invoker\")\n\t\t}\n\n\t\t//Get invoker's Enrollment Certificate\n\t\teCert, error := getInvokerEcert(invoker)\n\t\tif error != nil {\n\t\t\treturn nil, errors.New(\"Error retrieving the Invoker Enrollment Certificate\")\n\t\t}\n\n\t\t//Check if the Invoker has the correct affiliation\n\t\taffiliation, error := getInvokerAffiliation(eCert)\n\t\tif error != nil {\n\t\t\treturn nil, errors.New(\"Error retrieving the Invoker Enrollment Certificate\")\n\t\t}\n\n\t\tif ((strings.Compare(affiliation,strings.ToLower(args[0]))) != 0){\n\t\t\treturn nil, errors.New(\"Invoker does not have the necessary affiliation\")\n\t\t}\n\t}\n\n\trecipientAdministrator := getAdminUser(args[1])\n\tinvokerAdministrator := getAdminUser(args[0])\n\n\ttSigma, tCertificate, tBinding, _ := generateTransactionSigma(recipientAdministrator, args[2])\n\tinvokerAdminSigma,_ := generateInvokerSigma(invokerAdministrator, args[2])\n\n\tsecMetadata := SecurityMetaData{tSigma, tCertificate, tBinding, invokerAdminSigma}\n\tsecMetadataRaw, err := asn1.Marshal(secMetadata)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn secMetadataRaw, nil\n}", "func (w *Waiter) InterfaceSignature() string {\n\tvar buf bytes.Buffer\n\tif err := waiterTmpls.ExecuteTemplate(&buf, \"waiter interface\", w); err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn strings.TrimSpace(buf.String())\n}", "func (s *service) Signature(ctx *shared.Context, kind contract.ResourceKind, filter map[string]interface{}) (core.Record, error) {\n\tresult, err := s.Signatures(ctx, kind, filter)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(result) == 1 {\n\t\treturn result[0], nil\n\t} else if len(result) > 1 {\n\t\treturn nil, fmt.Errorf(\"expected one record, but had: %v\", len(result))\n\t}\n\treturn nil, nil\n}", "func (g *grpc) generateServerSignature(servName string, method *pb.MethodDescriptorProto) string {\n\torigMethName := method.GetName()\n\tmethName := generator.CamelCase(origMethName)\n\tif reservedClientName[methName] {\n\t\tmethName += \"_\"\n\t}\n\n\tvar reqArgs []string\n\tret := \"error\"\n\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, contextPkg+\".Context\")\n\t\tret = \"(*\" + g.typeName(method.GetOutputType()) + \", error)\"\n\t}\n\tif !method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, \"*\"+g.typeName(method.GetInputType()))\n\t}\n\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, servName+\"_\"+generator.CamelCase(origMethName)+\"Server\")\n\t}\n\n\treturn methName + \"(\" + strings.Join(reqArgs, \", \") + \") \" + ret\n}", "func (fs fakeService) callRemoteWithResponse(ctx *context.T, message *bindings.Message) (outMessage *bindings.Message, err error) {\n\tctx.Infof(\"callRemoteGeneric: Send message along the router\")\n\n\treadResult := <-fs.router.AcceptWithResponse(message)\n\tif err = readResult.Error; err != nil {\n\t\treturn\n\t}\n\n\tctx.Infof(\"callRemoteGeneric: Audit response message header flag\")\n\t// The message flag we receive back must be a bindings.MessageIsResponseFlag\n\tif readResult.Message.Header.Flags != bindings.MessageIsResponseFlag {\n\t\terr = &bindings.ValidationError{bindings.MessageHeaderInvalidFlags,\n\t\t\tfmt.Sprintf(\"invalid message header flag: %v\", readResult.Message.Header.Flags),\n\t\t}\n\t\treturn\n\t}\n\n\tctx.Infof(\"callRemoteGeneric: Audit response message header type\")\n\t// While the mojo service we called into will return a header whose\n\t// type must match our outgoing one.\n\tif got, want := readResult.Message.Header.Type, message.Header.Type; got != want {\n\t\terr = &bindings.ValidationError{bindings.MessageHeaderUnknownMethod,\n\t\t\tfmt.Sprintf(\"invalid method in response: expected %v, got %v\", want, got),\n\t\t}\n\t\treturn\n\t}\n\n\treturn readResult.Message, nil\n}", "func (g *grpc) generateServerSignatureWithParamNames(servName string, method *pb.MethodDescriptorProto) string {\n\torigMethName := method.GetName()\n\tmethName := generator.CamelCase(origMethName)\n\tif reservedClientName[methName] {\n\t\tmethName += \"_\"\n\t}\n\n\tvar reqArgs []string\n\tret := \"error\"\n\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, \"ctx \"+contextPkg+\".Context\")\n\t\tret = \"(*\" + g.typeName(method.GetOutputType()) + \", error)\"\n\t}\n\tif !method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, \"req *\"+g.typeName(method.GetInputType()))\n\t}\n\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\treqArgs = append(reqArgs, \"srv \"+servName+\"_\"+generator.CamelCase(origMethName)+\"Server\")\n\t}\n\n\treturn methName + \"(\" + strings.Join(reqArgs, \", \") + \") \" + ret\n}", "func ComputeSignature(authToken, uri string, params map[string]string) string {\n\toriginalString := fmt.Sprintf(\"%s%s\", uri, headersWithSep(params, \"\", \"\", false))\n\tmac := hmac.New(sha1.New, []byte(authToken))\n\tmac.Write([]byte(originalString))\n\treturn base64.StdEncoding.EncodeToString(mac.Sum(nil))\n}", "func (_PBridge *PBridgeCaller) SignatureLength(opts *bind.CallOpts) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _PBridge.contract.Call(opts, &out, \"signatureLength\")\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func (s *SharedMemory) Signature() string {\n\treturn util.DecodeCharPtr(unsafe.Pointer(&s.shmem.dwSignature), C.sizeof_DWORD)\n}", "func Signature(arg, ret Type) string {\n\targs := make([]string, arg.NumOut())\n\tfor i := range args {\n\t\targs[i] = arg.Out(i).String()\n\t}\n\trets := make([]string, ret.NumOut())\n\tfor i := range rets {\n\t\trets[i] = ret.Out(i).String()\n\t}\n\tvar b strings.Builder\n\tb.WriteString(\"func(\")\n\tb.WriteString(strings.Join(args, \", \"))\n\tb.WriteString(\")\")\n\tswitch len(rets) {\n\tcase 0:\n\tcase 1:\n\t\tb.WriteString(\" \")\n\t\tb.WriteString(rets[0])\n\tdefault:\n\t\tb.WriteString(\" (\")\n\t\tb.WriteString(strings.Join(rets, \", \"))\n\t\tb.WriteString(\")\")\n\t}\n\treturn b.String()\n}", "func GetSignature(key string, method string, req map[string]string, fcResource string) string {\n\theader := &headers{}\n\tlowerKeyHeaders := map[string]string{}\n\tfor k, v := range req {\n\t\tlowerKey := strings.ToLower(k)\n\t\tif strings.HasPrefix(lowerKey, HTTPHeaderPrefix) {\n\t\t\theader.Keys = append(header.Keys, lowerKey)\n\t\t\theader.Vals = append(header.Vals, v)\n\t\t}\n\t\tlowerKeyHeaders[lowerKey] = v\n\t}\n\tsort.Sort(header)\n\n\tfcHeaders := \"\"\n\tfor i := range header.Keys {\n\t\tfcHeaders += header.Keys[i] + \":\" + header.Vals[i] + \"\\n\"\n\t}\n\n\tdate := req[HTTPHeaderDate]\n\tif expires, ok := getExpiresFromURLQueries(fcResource); ok {\n\t\tdate = expires\n\t}\n\n\tsignStr := method + \"\\n\" + lowerKeyHeaders[strings.ToLower(HTTPHeaderContentMD5)] + \"\\n\" + lowerKeyHeaders[strings.ToLower(HTTPHeaderContentType)] + \"\\n\" + date + \"\\n\" + fcHeaders + fcResource\n\n\th := hmac.New(func() hash.Hash { return sha256.New() }, []byte(key))\n\tio.WriteString(h, signStr)\n\tsignedStr := base64.StdEncoding.EncodeToString(h.Sum(nil))\n\n\treturn signedStr\n}", "func signature(req *http.Request, awsSecretAccessKey string) string {\n\treturn signWithKey(stringToSign(req), awsSecretAccessKey)\n}", "func (g *grpc) generateClientSignature(servName string, method *pb.MethodDescriptorProto) string {\n\torigMethName := method.GetName()\n\tmethName := generator.CamelCase(origMethName)\n\tif reservedClientName[methName] {\n\t\tmethName += \"_\"\n\t}\n\treqArg := \", in *\" + g.typeName(method.GetInputType())\n\tif method.GetClientStreaming() {\n\t\treqArg = \"\"\n\t}\n\trespName := \"*\" + g.typeName(method.GetOutputType())\n\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\trespName = servName + \"_\" + generator.CamelCase(origMethName) + \"Client\"\n\t}\n\treturn fmt.Sprintf(\"%s(ctx %s.Context%s, opts ...%s.CallOption) (%s, error)\", methName, contextPkg, reqArg, grpcPkg, respName)\n}", "func (client *BaseClient) GetSignature(request *tea.Request, secret string) string {\n\tstringToSign := buildRpcStringToSign(request)\n\tsignature := client.Sign(stringToSign, secret, \"&\")\n\treturn signature\n}", "func (a *Authorization) Signature() ([]byte, error) {\n\treturn a.signature, nil\n}", "func (me *XsdGoPkgHasElem_SignatureMethod) Walk() (err error) {\n\tif fn := WalkHandlers.XsdGoPkgHasElem_SignatureMethod; me != nil {\n\t\tif fn != nil {\n\t\t\tif err = fn(me, true); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif err = me.SignatureMethod.Walk(); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\treturn\n\t\t}\n\t\tif fn != nil {\n\t\t\tif err = fn(me, false); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (b Binding) Signature() (ret string) {\n\tret = b.ValidationString()\n\tif receivesBinaryContent(b) {\n\t\tret = \":\" + ret\n\t}\n\treturn\n}", "func (k *RSAPrivKey) Signature(payload []byte) (string, error) {\n\tif k.key == nil {\n\t\treturn \"\", ErrorKeyUninitialized\n\t}\n\n\tsha256 := crypto.SHA256.New()\n\t_, err := sha256.Write(payload)\n\tif err != nil {\n\t\treturn \"\", errors.AddStack(err)\n\t}\n\n\thashed := sha256.Sum(nil)\n\n\tsig, err := rsa.SignPSS(rand.Reader, k.key, crypto.SHA256, hashed, nil)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn base64.StdEncoding.EncodeToString(sig), nil\n}", "func (s *service) ChunkSignature(ctx *shared.Context, kind contract.ResourceKind, offset, limit int, filter map[string]interface{}) (*core.Signature, error) {\n\tdbResource := s.dbResource(kind)\n\treturn s.chunkSignature(ctx, dbResource, offset, limit, filter)\n}", "func signature(accessKeySecret, method, uri string,\n\theaders map[string]string) (digest string, err error) {\n\tvar contentMD5, contentType, date, canoHeaders, canoResource string\n\tvar slsHeaderKeys sort.StringSlice\n\n\tif val, ok := headers[\"Content-MD5\"]; ok {\n\t\tcontentMD5 = val\n\t}\n\n\tif val, ok := headers[\"Content-Type\"]; ok {\n\t\tcontentType = val\n\t}\n\n\tdate, ok := headers[\"Date\"]\n\tif !ok {\n\t\terr = fmt.Errorf(\"Can't find 'Date' header\")\n\t\treturn\n\t}\n\n\t// Calc CanonicalizedSLSHeaders\n\tslsHeaders := make(map[string]string, len(headers))\n\tfor k, v := range headers {\n\t\tl := strings.TrimSpace(strings.ToLower(k))\n\t\tif strings.HasPrefix(l, \"x-log-\") || strings.HasPrefix(l, \"x-acs-\") {\n\t\t\tslsHeaders[l] = strings.TrimSpace(v)\n\t\t\tslsHeaderKeys = append(slsHeaderKeys, l)\n\t\t}\n\t}\n\n\tsort.Sort(slsHeaderKeys)\n\tfor i, k := range slsHeaderKeys {\n\t\tcanoHeaders += k + \":\" + slsHeaders[k]\n\t\tif i+1 < len(slsHeaderKeys) {\n\t\t\tcanoHeaders += \"\\n\"\n\t\t}\n\t}\n\n\t// Calc CanonicalizedResource\n\tu, err := url.Parse(uri)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tcanoResource += u.EscapedPath()\n\tif u.RawQuery != \"\" {\n\t\tvar keys sort.StringSlice\n\n\t\tvals := u.Query()\n\t\tfor k := range vals {\n\t\t\tkeys = append(keys, k)\n\t\t}\n\n\t\tsort.Sort(keys)\n\t\tcanoResource += \"?\"\n\t\tfor i, k := range keys {\n\t\t\tif i > 0 {\n\t\t\t\tcanoResource += \"&\"\n\t\t\t}\n\n\t\t\tfor _, v := range vals[k] {\n\t\t\t\tcanoResource += k + \"=\" + v\n\t\t\t}\n\t\t}\n\t}\n\n\tsignStr := method + \"\\n\" +\n\t\tcontentMD5 + \"\\n\" +\n\t\tcontentType + \"\\n\" +\n\t\tdate + \"\\n\" +\n\t\tcanoHeaders + \"\\n\" +\n\t\tcanoResource\n\n\t// Signature = base64(hmac-sha1(UTF8-Encoding-Of(SignString),AccessKeySecret))\n\tmac := hmac.New(sha1.New, []byte(accessKeySecret))\n\t_, err = mac.Write([]byte(signStr))\n\tif err != nil {\n\t\treturn\n\t}\n\tdigest = base64.StdEncoding.EncodeToString(mac.Sum(nil))\n\treturn\n}", "func (t Trailer) GetSignature() (v string, err quickfix.MessageRejectError) {\n\tvar f field.SignatureField\n\tif err = t.Get(&f); err == nil {\n\t\tv = f.Value()\n\t}\n\treturn\n}", "func (exe *HTTPRemote) Execute(param map[string]interface{}) (map[string]interface{}, error) {\n\texeID, _ := util.GetStringParam(param, \"id\")\n\n\tparamJSON, err := json.Marshal(param)\n\tif err != nil {\n\t\texe.Errorf(\"Generate json param error: %s\", err)\n\t\treturn nil, errors.New(\"Generate json param error\")\n\t}\n\n\tsignatureBytes := util.CalculateMAC(paramJSON, []byte(exe.Secret))\n\tsignature := hex.EncodeToString(signatureBytes)\n\n\treq, err := http.NewRequest(\"POST\", exe.Host, bytes.NewBuffer(paramJSON))\n\tif err != nil {\n\t\texe.Errorf(\"Create request failed: %s\", err)\n\t\treturn nil, errors.New(\"Create request failed\")\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"X-Herald-Signature\", signature)\n\n\texe.Infof(\"Start to connect to: %s\", exe.Host)\n\n\tclient := &http.Client{\n\t\tTimeout: exe.Timeout,\n\t}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\texe.Errorf(\"Remote execution request failed: %s\", err)\n\t\treturn nil, errors.New(\"Remote execution request failed\")\n\t}\n\tdefer resp.Body.Close()\n\n\tcontentType := resp.Header.Get(\"Content-Type\")\n\n\texe.Debugf(\"Response status: %s\", resp.Status)\n\texe.Debugf(\"Response content type: %s\", contentType)\n\n\tif resp.StatusCode != http.StatusOK {\n\t\texe.Errorf(\"Http status not OK: %s\", resp.Status)\n\t\tbody, _ := ioutil.ReadAll(resp.Body)\n\t\texe.Errorf(\"Remote error: %s\", string(body))\n\t\treturn nil, fmt.Errorf(`Http status %d: \"%s\"`, resp.StatusCode, string(body))\n\t}\n\n\tmediaType, mtParams, err := mime.ParseMediaType(contentType)\n\tif err != nil {\n\t\texe.Errorf(\"Parse media type error: %s\", err)\n\t\treturn nil, errors.New(\"Parse media type error\")\n\t}\n\n\tresult := make(map[string]interface{})\n\n\texe.Debugf(\"Parsed context type: %s\", mediaType)\n\tresult[\"context_type\"] = mediaType\n\n\tif mediaType == \"application/json\" {\n\t\texe.processJSONPart(result, resp.Body)\n\t} else if strings.HasPrefix(mediaType, \"multipart/\") {\n\t\texe.processMultiPart(result, resp.Body, mtParams[\"boundary\"], exeID)\n\t} else {\n\t\texe.Errorf(\"Unknown media type: %s\", mediaType)\n\t\tbody, _ := ioutil.ReadAll(resp.Body)\n\t\tresult[\"response\"] = string(body)\n\t\treturn result, errors.New(\"Unknown media type\")\n\t}\n\n\texitCodeFloat, err := util.GetFloatParam(result, \"exit_code\")\n\texitCode := int(exitCodeFloat)\n\tif exitCode != 0 {\n\t\treturn result, fmt.Errorf(\"Command failed with code %d\", exitCode)\n\t}\n\n\treturn result, nil\n}", "func calculateSignature(handler hndlr, insts interface{}) signature {\n\tif reflect.TypeOf(insts).Kind() != reflect.Slice {\n\t\treturn zeroSignature\n\t}\n\n\tinstances := reflect.ValueOf(insts)\n\n\t// sort the instances by name\n\tinstanceMap := make(map[string]inst)\n\tinstanceNames := make([]string, instances.Len())\n\tfor i := 0; i < instances.Len(); i++ {\n\t\tii := instances.Index(i).Interface()\n\t\tinstance := ii.(inst)\n\t\tinstanceMap[instance.GetName()] = instance\n\t\tinstanceNames[i] = instance.GetName()\n\t}\n\tsort.Strings(instanceNames)\n\n\tbuf := pool.GetBuffer()\n\tencoded := true\n\n\tencoded = encoded && encode(buf, handler.AdapterName())\n\tif handler.AdapterParams() != nil &&\n\t\t(reflect.ValueOf(handler.AdapterParams()).Kind() != reflect.Ptr || !reflect.ValueOf(handler.AdapterParams()).IsNil()) {\n\t\tencoded = encoded && encode(buf, handler.AdapterParams())\n\t}\n\tfor _, name := range instanceNames {\n\t\tinstance := instanceMap[name]\n\t\tencoded = encoded && encode(buf, instance.TemplateName())\n\t\tencoded = encoded && encode(buf, instance.TemplateParams())\n\t}\n\n\tif encoded {\n\t\tsha := sha1.Sum(buf.Bytes())\n\t\tpool.PutBuffer(buf)\n\t\treturn sha\n\t}\n\n\tpool.PutBuffer(buf)\n\treturn zeroSignature\n}", "func (r *GetResponseBody) SetSignature(v *refs.Signature) {\n\tif r != nil {\n\t\tr.sig = v\n\t}\n}", "func (s *BasevhdlListener) ExitSignature(ctx *SignatureContext) {}", "func (c *Client) GetSignature(req *http.Request) string {\n\t// Sort fcHeaders.\n\theaders := &fcHeaders{}\n\tfor k := range req.Header {\n\t\tif strings.HasPrefix(strings.ToLower(k), \"x-fc-\") {\n\t\t\theaders.Keys = append(headers.Keys, strings.ToLower(k))\n\t\t\theaders.Values = append(headers.Values, req.Header.Get(k))\n\t\t}\n\t}\n\tsort.Sort(headers)\n\tfcHeaders := \"\"\n\tfor i := range headers.Keys {\n\t\tfcHeaders += headers.Keys[i] + \":\" + headers.Values[i] + \"\\n\"\n\t}\n\n\thttpMethod := req.Method\n\tcontentMd5 := req.Header.Get(\"Content-MD5\")\n\tcontentType := req.Header.Get(\"Content-Type\")\n\tdate := req.Header.Get(\"Date\")\n\tfcResource := req.URL.Path\n\n\tsignStr := httpMethod + \"\\n\" + contentMd5 + \"\\n\" + contentType + \"\\n\" + date + \"\\n\" + fcHeaders + fcResource\n\n\th := hmac.New(func() hash.Hash { return sha256.New() }, []byte(c.accessKeySecret))\n\t_, _ = io.WriteString(h, signStr)\n\treturn base64.StdEncoding.EncodeToString(h.Sum(nil))\n}", "func ExtractSignature(b []byte) ([]byte, Signature, error) {\n\t// BUG(cryptix): this expects signature on the root of the object.\n\t// some functions (like createHistoryStream with keys:true) nest the message on level deeper and this fails\n\tmatches := signatureRegexp.FindSubmatch(b)\n\tif n := len(matches); n != 2 {\n\t\treturn nil, \"\", fmt.Errorf(\"message Encode: expected signature in formatted bytes. Only %d matches\", n)\n\t}\n\tsig := Signature(matches[1])\n\tout := signatureRegexp.ReplaceAll(b, []byte{})\n\treturn out, sig, nil\n}", "func extractSignature(in metav1.Object) ([]byte, error) {\n\t// signature should be contained in annotation\n\tsig, ok := in.GetAnnotations()[SignatureAnnotation]\n\tif !ok {\n\t\treturn nil, nil\n\t}\n\t// extract signature\n\tsignature, err := base64.StdEncoding.DecodeString(sig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn signature, nil\n}", "func (c *Client) FetchImageSignatures(ctx context.Context, targetRepository string) ([]oci.Signature, error) {\n\timage, err := c.pullTargetImage(ctx, targetRepository)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmanifest, err := getManifest(ctx, image)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsignatures := make([]oci.Signature, 0, len(manifest.Layers))\n\tfor _, layer := range manifest.Layers {\n\t\tblob, err := content.ReadBlob(ctx, image.ContentStore(), layer)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsig := &cosign.Sig{\n\t\t\tLayer: layer,\n\t\t\tBlob: blob,\n\t\t}\n\t\tsignatures = append(signatures, sig)\n\t}\n\treturn signatures, nil\n}", "func (client ClientImpl) generateAuthSignature(message []byte) (string, error) {\n\treturn generateSignatureFromKey(message, client.privateKey)\n}", "func GetSign(srcdata interface{}, bizkey string) string {\n\tmd5ctx := md5.New()\n\n\tswitch v := reflect.ValueOf(srcdata); v.Kind() {\n\tcase reflect.String:\n\t\tmd5ctx.Write([]byte(v.String() + bizkey))\n\t\treturn hex.EncodeToString(md5ctx.Sum(nil))\n\tcase reflect.Map:\n\t\torderStr := orderParam(v.Interface(), bizkey)\n\t\tmd5ctx.Write([]byte(orderStr))\n\t\treturn hex.EncodeToString(md5ctx.Sum(nil))\n\tcase reflect.Struct:\n\t\torderStr := Struct2map(v.Interface(), bizkey)\n\t\tmd5ctx.Write([]byte(orderStr))\n\t\treturn hex.EncodeToString(md5ctx.Sum(nil))\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func (me *XsdGoPkgHasElems_SignatureMethod) Walk() (err error) {\n\tif fn := WalkHandlers.XsdGoPkgHasElems_SignatureMethod; me != nil {\n\t\tif fn != nil {\n\t\t\tif err = fn(me, true); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tfor _, x := range me.SignatureMethods {\n\t\t\tif err = x.Walk(); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif fn != nil {\n\t\t\tif err = fn(me, false); xsdt.OnWalkError(&err, &WalkErrors, WalkContinueOnError, WalkOnError) {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (c MockClient) VerifySignature(signature string, body []byte) error {\n\tif c.VerifySignatureError {\n\t\treturn ErrSignatureMismatch\n\t}\n\n\treturn nil\n}", "func (_PBridge *PBridgeCallerSession) SignatureLength() (*big.Int, error) {\n\treturn _PBridge.Contract.SignatureLength(&_PBridge.CallOpts)\n}", "func (fnGet) Sig() (paramTypes []data.Type, isVariadic bool) {\n\treturn []data.Type{data.TypeObject, data.TypeString}, false\n}", "func signV2( r *rest.Rest, accessKeyID, secretAccessKey string ) (string,error) {\n\n\t// Calculate HMAC for secretAccessKey.\n\tstringToSign := stringToSignV2( r )\n\n\thm := hmac.New(sha1.New, []byte(secretAccessKey))\n\thm.Write([]byte(stringToSign))\n\n\t// Prepare auth header.\n\tauthHeader := new(bytes.Buffer)\n\tauthHeader.WriteString(fmt.Sprintf(\"%s %s:\", signV2Algorithm, accessKeyID))\n\n\tencoder := base64.NewEncoder(base64.StdEncoding, authHeader)\n\tencoder.Write(hm.Sum(nil))\n\tencoder.Close()\n\n\t// Authorization header.\n return authHeader.String(), nil\n}", "func (*SignatureResponse) Descriptor() ([]byte, []int) {\n\treturn file_p2p_p2p_proto_rawDescGZIP(), []int{3}\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\tcase \"queryById\":\n\t\treturn s.queryById(APIstub, args)\n\tcase \"createRecord\":\n\t\treturn s.createRecord(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n}", "func getTokenSignature(header *Header, body *Body, secret *pbauth.Secret) (string, error) {\n\tif err := ValidateHeader(header); err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := ValidateBody(body); err != nil {\n\t\treturn \"\", err\n\t}\n\tif err := ValidateSecret(secret); err != nil {\n\t\treturn \"\", err\n\t}\n\tif body.Permission == Admin && header.Alg != Hs512 {\n\t\treturn \"\", consts.ErrInvalidPermission\n\t}\n\tif header.TokenTyp != Jwt && header.TokenTyp != Jet {\n\t\treturn \"\", consts.ErrUnknownTokenType\n\t}\n\t// Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>\n\t// 1. Encode the header\n\tencodedHeader, err := base64Encode(header)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// 2. Encode the body\n\tencodedBody, err := base64Encode(body)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// 3. Build <encoded header>.<encoded body>\n\t// 4. Build <hashed(<encoded header>.<encoded body>)>\n\t// 5. Build Token Signature = <encoded header>.<encoded body>.<hashed(<encoded header>.<encoded body>)>\n\treturn buildTokenSignature(encodedHeader, encodedBody, header.Alg, secret)\n}", "func (*SignatureHelpClientCapabilities_SignatureInformation) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{114, 0}\n}", "func (client ArtifactsClient) getContainerImageSignature(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/container/imageSignatures/{imageSignatureId}\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response GetContainerImageSignatureResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/registry/20160918/ContainerImageSignature/GetContainerImageSignature\"\n\t\terr = common.PostProcessServiceError(err, \"Artifacts\", \"GetContainerImageSignature\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (r *GetResponseBody) GetSignature() *refs.Signature {\n\tif r != nil {\n\t\treturn r.sig\n\t}\n\n\treturn nil\n}", "func (pi *PackageInfo) Signature(obj types.Object) string {\n\tif obj == nil {\n\t\treturn \"\"\n\t} else if pi.owner == nil {\n\t\tpi.owner = make(map[types.Object]types.Object)\n\t\townerByPos := make(map[token.Position]types.Object)\n\t\tunownedByPos := make(map[token.Position][]types.Object)\n\t\tpi.addOwners(pi.Package, ownerByPos, unownedByPos)\n\t\tfor _, pkg := range pi.Dependencies {\n\t\t\tpi.addOwners(pkg, ownerByPos, unownedByPos)\n\t\t}\n\t}\n\tif sig, ok := pi.sigs[obj]; ok {\n\t\treturn sig\n\t}\n\ttag, base := pi.newSignature(obj)\n\tsig := base\n\tif tag != \"\" {\n\t\tsig = tag + \" \" + base\n\t}\n\tpi.sigs[obj] = sig\n\treturn sig\n}", "func (s *BasevhdlListener) EnterSignature(ctx *SignatureContext) {}", "func (s *storageImageSource) GetSignatures(ctx context.Context, instanceDigest *digest.Digest) (signatures [][]byte, err error) {\n\tif instanceDigest != nil {\n\t\treturn nil, ErrNoManifestLists\n\t}\n\tvar offset int\n\tsigslice := [][]byte{}\n\tsignature := []byte{}\n\tif len(s.SignatureSizes) > 0 {\n\t\tsignatureBlob, err := s.imageRef.transport.store.ImageBigData(s.image.ID, \"signatures\")\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"error looking up signatures data for image %q\", s.image.ID)\n\t\t}\n\t\tsignature = signatureBlob\n\t}\n\tfor _, length := range s.SignatureSizes {\n\t\tsigslice = append(sigslice, signature[offset:offset+length])\n\t\toffset += length\n\t}\n\tif offset != len(signature) {\n\t\treturn nil, errors.Errorf(\"signatures data contained %d extra bytes\", len(signatures)-offset)\n\t}\n\treturn sigslice, nil\n}", "func SignatureV2(req *http.Request, Auth interface{}) (err error) {\n\tauth, _ := Auth.(map[string]string)\n\tqueryVals := req.URL.Query()\n\tqueryVals.Set(\"AWSAccessKeyId\", auth[\"AccessKey\"])\n\tqueryVals.Set(\"SignatureVersion\", \"2\")\n\tqueryVals.Set(\"SignatureMethod\", \"HmacSHA256\")\n\n\tqueryStr, err := canonicalQueryString(queryVals)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tpath := req.URL.Path\n\tif path == \"\" {\n\t\tpath = \"/\"\n\t}\n\n\tpayload := new(bytes.Buffer)\n\n\tpayloadstring := checkrequestMethod(req.Method) + \"\\n\" + req.Host + \"\\n\" + path + \"\\n\" + queryStr\n\n\tfmt.Fprintf(payload, \"%s\", payloadstring)\n\n\thash := hmac.New(sha256.New, []byte(auth[\"SecretKey\"]))\n\n\thash.Write(payload.Bytes())\n\n\tsignature := make([]byte, base64.StdEncoding.EncodedLen(hash.Size()))\n\n\tbase64.StdEncoding.Encode(signature, hash.Sum(nil))\n\n\tqueryVals.Set(\"Signature\", string(signature))\n\n\treq.URL.RawQuery = queryVals.Encode()\n\n\treturn nil\n}", "func (p *BlsCosi) generateSignature(responses ResponseMap) (kyber.Point, *cosi.Mask, error) {\n\tpublics := p.Publics()\n\n\t//generate personal mask\n\tpersonalMask, err := cosi.NewMask(p.suite, publics, p.Public())\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// generate personal signature and append to other sigs\n\tpersonalSig, err := bls.Sign(p.suite, p.Private(), p.Msg)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\t// fill the map with the Root signature\n\tresponses[p.Public().String()] = &Response{\n\t\tMask: personalMask.Mask(),\n\t\tSignature: personalSig,\n\t}\n\n\t// Aggregate all signatures\n\tresponse, err := makeAggregateResponse(p.suite, publics, responses)\n\tif err != nil {\n\t\tlog.Lvlf3(\"%v failed to create aggregate signature\", p.ServerIdentity())\n\t\treturn nil, nil, err\n\t}\n\n\t//create final aggregated mask\n\tfinalMask, err := cosi.NewMask(p.suite, publics, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\terr = finalMask.SetMask(response.Mask)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tfinalSignature, err := response.Signature.Point(p.suite)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tlog.Lvlf3(\"%v is done aggregating signatures with total of %d signatures\", p.ServerIdentity(), finalMask.CountEnabled())\n\n\treturn finalSignature, finalMask, err\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n //args := APIstub.GetFunctionAndParameters()\n args := APIstub.GetStringArgs()\n function := args[0]\n if function == \"addRecord\" {\n return s.addRecord(APIstub, args)\n } else if function == \"getRecord\" {\n return s.getRecord(APIstub, args)\n } else if function == \"encRecord\" {\n return s.encRecord(APIstub, args)\n } else if function == \"decRecord\" {\n return s.decRecord(APIstub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name \"+function)\n}", "func (_PBridge *PBridgeSession) SignatureLength() (*big.Int, error) {\n\treturn _PBridge.Contract.SignatureLength(&_PBridge.CallOpts)\n}", "func verifySignature(n string, t reflect.Type) {\n\tif t.Kind() != reflect.Func {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"handler for %q is not a function but a %v\", n, t.Kind(),\n\t\t))\n\t}\n\n\tif t.NumIn() != 2 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"expected handler for %q to have 2 arguments but it had %v\",\n\t\t\tn, t.NumIn(),\n\t\t))\n\t}\n\n\tif t.NumOut() != 1 {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"expected handler for %q to have 1 results but it had %v\",\n\t\t\tn, t.NumOut(),\n\t\t))\n\t}\n\n\tif t.Out(0) != _errorType {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"the last resultsof the handler for %q must be of type error, \"+\n\t\t\t\t\"and not: %v\",\n\t\t\tn, t.Out(0),\n\t\t))\n\t}\n\n\treqBodyType := t.In(1)\n\n\tif !isValidReqResType(reqBodyType) {\n\t\tpanic(fmt.Sprintf(\n\t\t\t\"the second argument of the handler for %q must be \"+\n\t\t\t\t\"a struct pointer, or interface{}, and not: %v\",\n\t\t\tn, reqBodyType,\n\t\t))\n\t}\n}", "func (*SignatureInformation) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{120}\n}", "func (v vector) signature() string {\n\tbuf := &bytes.Buffer{}\n\tfor _, v := range v {\n\t\tbinary.Write(buf, binary.LittleEndian, v)\n\t}\n\n\treturn base64.URLEncoding.EncodeToString(buf.Bytes())\n}", "func (v *Plugin_Handshake_Args) MethodName() string {\n\treturn \"handshake\"\n}", "func (p *SASQueryParameters) Signature() string {\n\treturn p.signature\n}", "func GetSignedImageFlavor(flavorString string, rsaPrivateKeyLocation string) (string, error) {\n\tlog.Trace(\"flavor/image_flavor:GetSignedImageFlavor() Entering\")\n\tdefer log.Trace(\"flavor/image_flavor:GetSignedImageFlavor() Leaving\")\n\tvar privateKey *rsa.PrivateKey\n\tvar flavorInterface ImageFlavor\n\tif rsaPrivateKeyLocation == \"\" {\n\t\tlog.Error(\"No RSA Key file path provided\")\n\t\treturn \"\", errors.New(\"No RSA Key file path provided\")\n\t}\n\n\tpriv, err := ioutil.ReadFile(rsaPrivateKeyLocation)\n\tif err != nil {\n\t\tlog.Error(\"No RSA private key found\")\n\t\treturn \"\", err\n\t}\n\n\tprivPem, _ := pem.Decode(priv)\n\tparsedKey, err := x509.ParsePKCS8PrivateKey(privPem.Bytes)\n\tif err != nil {\n\t\tlog.Error(\"Cannot parse RSA private key from file\")\n\t\treturn \"\", err\n\t}\n\n\tprivateKey, ok := parsedKey.(*rsa.PrivateKey)\n\tif !ok {\n\t\tlog.Error(\"Unable to parse RSA private key\")\n\t\treturn \"\", err\n\t}\n\thashEntity := sha512.New384()\n\thashEntity.Write([]byte(flavorString))\n\tsignature, err := rsa.SignPKCS1v15(rand.Reader, privateKey, crypto.SHA384, hashEntity.Sum(nil))\n\tsignatureString := base64.StdEncoding.EncodeToString(signature)\n\n\tjson.Unmarshal([]byte(flavorString), &flavorInterface)\n\n\tsignedFlavor := SignedImageFlavor{\n\t\tImageFlavor: flavorInterface.Image,\n\t\tSignature: signatureString,\n\t}\n\n\tsignedFlavorJSON, err := json.Marshal(signedFlavor)\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Error while marshalling signed image flavor: \" + err.Error())\n\t}\n\n\treturn string(signedFlavorJSON), nil\n}", "func GetSignedFlavor(flavorString string, rsaPrivateKeyLocation string) (string, error) {\n\tvar privateKey *rsa.PrivateKey\n\tvar flavorInterface flavor.ImageFlavor\n\tif rsaPrivateKeyLocation == \"\" {\n\t\tlog.Error(\"No RSA Key file path provided\")\n\t\treturn \"\", errors.New(\"No RSA Key file path provided\")\n\t}\n\n\tpriv, err := ioutil.ReadFile(rsaPrivateKeyLocation)\n\tif err != nil {\n\t\tlog.Error(\"No RSA private key found\")\n\t\treturn \"\", err\n\t}\n\n\tprivPem, _ := pem.Decode(priv)\n\tparsedKey, err := x509.ParsePKCS8PrivateKey(privPem.Bytes)\n\tif err != nil {\n\t\tlog.Error(\"Cannot parse RSA private key from file\")\n\t\treturn \"\", err\n\t}\n\n\tprivateKey, ok := parsedKey.(*rsa.PrivateKey)\n\tif !ok {\n\t\tlog.Error(\"Unable to parse RSA private key\")\n\t\treturn \"\", err\n\t}\n\thashEntity := sha512.New384()\n\t_, err = hashEntity.Write([]byte(flavorString))\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Error while writing flavor hash: \" + err.Error())\n\t}\n\tsignature, err := rsa.SignPKCS1v15(rand.Reader, privateKey, crypto.SHA384, hashEntity.Sum(nil))\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Error while signing flavor: \" + err.Error())\n\t}\n\tsignatureString := base64.StdEncoding.EncodeToString(signature)\n\n\terr = json.Unmarshal([]byte(flavorString), &flavorInterface)\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Error while unmarshalling signed image flavor: \" + err.Error())\n\t}\n\n\tsignedFlavor := &flavor.SignedImageFlavor{\n\t\tImageFlavor: flavorInterface.Image,\n\t\tSignature: signatureString,\n\t}\n\n\tsignedFlavorJSON, err := json.Marshal(signedFlavor)\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Error while marshalling signed image flavor: \" + err.Error())\n\t}\n\n\treturn string(signedFlavorJSON), nil\n}", "func extractTagsForMethod(desc []InterfaceDesc, name string) ([]*vdl.Value, error) {\n\tseenFirst := false\n\tvar first []*vdl.Value\n\tfor _, descIface := range desc {\n\t\tfor _, descMethod := range descIface.Methods {\n\t\t\tif name == descMethod.Name {\n\t\t\t\tswitch tags := descMethod.Tags; {\n\t\t\t\tcase !seenFirst:\n\t\t\t\t\tseenFirst = true\n\t\t\t\t\tfirst = tags\n\t\t\t\tcase !equalTags(first, tags):\n\t\t\t\t\treturn nil, abortedf(fmt.Errorf(\"different tags %v and %v\", first, tags))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn first, nil\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n fmt.Println(\"Calling Invoke method.\")\n\n // Retrieve the requested Smart Contract function and arguments\n function, args := APIstub.GetFunctionAndParameters()\n fmt.Println(\"Function name: \" + function)\n\n // Route to the appropriate handler function to interact with the ledger appropriately\n if function == \"queryAsset\" {\n return s.queryAsset(APIstub, args)\n } else if function == \"makeAsset\" {\n return s.makeAsset(APIstub, args)\n } else if function == \"changeAsset\" {\n return s.changeAsset(APIstub, args)\n } else if function == \"deleteAsset\" {\n return s.deleteAsset(APIstub, args)\n } else if function == \"listHistory\" {\n return s.listHistory(APIstub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func (check *checker) funcType(recv *ast.FieldList, ftyp *ast.FuncType, def *Named) *Signature {\n\tsig := new(Signature)\n\tif def != nil {\n\t\tdef.underlying = sig\n\t}\n\n\tscope := NewScope(check.topScope)\n\tcheck.recordScope(ftyp, scope)\n\n\trecv_, _ := check.collectParams(scope, recv, false)\n\tparams, isVariadic := check.collectParams(scope, ftyp.Params, true)\n\tresults, _ := check.collectParams(scope, ftyp.Results, false)\n\n\tif len(recv_) > 0 {\n\t\t// There must be exactly one receiver.\n\t\tif len(recv_) > 1 {\n\t\t\tcheck.invalidAST(recv_[1].Pos(), \"method must have exactly one receiver\")\n\t\t\t// ok to continue\n\t\t}\n\t\trecv := recv_[0]\n\t\t// spec: \"The receiver type must be of the form T or *T where T is a type name.\"\n\t\t// (ignore invalid types - error was reported before)\n\t\tif t, _ := deref(recv.typ); t != Typ[Invalid] {\n\t\t\tvar err string\n\t\t\tif T, _ := t.(*Named); T != nil {\n\t\t\t\t// spec: \"The type denoted by T is called the receiver base type; it must not\n\t\t\t\t// be a pointer or interface type and it must be declared in the same package\n\t\t\t\t// as the method.\"\n\t\t\t\tif T.obj.pkg != check.pkg {\n\t\t\t\t\terr = \"type not defined in this package\"\n\t\t\t\t} else {\n\t\t\t\t\tswitch u := T.underlying.(type) {\n\t\t\t\t\tcase *Basic:\n\t\t\t\t\t\t// unsafe.Pointer is treated like a regular pointer\n\t\t\t\t\t\tif u.kind == UnsafePointer {\n\t\t\t\t\t\t\terr = \"unsafe.Pointer\"\n\t\t\t\t\t\t}\n\t\t\t\t\tcase *Pointer, *Interface:\n\t\t\t\t\t\terr = \"pointer or interface type\"\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\terr = \"basic or unnamed type\"\n\t\t\t}\n\t\t\tif err != \"\" {\n\t\t\t\tcheck.errorf(recv.pos, \"invalid receiver %s (%s)\", recv.typ, err)\n\t\t\t\t// ok to continue\n\t\t\t}\n\t\t}\n\t\tsig.recv = recv\n\t}\n\n\tsig.scope = scope\n\tsig.params = NewTuple(params...)\n\tsig.results = NewTuple(results...)\n\tsig.isVariadic = isVariadic\n\n\treturn sig\n}", "func (sigTest) Describe__() []rpc.InterfaceDesc {\n\treturn []rpc.InterfaceDesc{\n\t\t{\n\t\t\tName: \"Iface1\",\n\t\t\tPkgPath: \"a/b/c\",\n\t\t\tDoc: \"Doc Iface1\",\n\t\t\tEmbeds: []rpc.EmbedDesc{\n\t\t\t\t{Name: \"Iface1Embed1\", PkgPath: \"x/y\", Doc: \"Doc embed1\"},\n\t\t\t},\n\t\t\tMethods: []rpc.MethodDesc{\n\t\t\t\t{\n\t\t\t\t\tName: \"Sig3\",\n\t\t\t\t\tDoc: \"Doc Sig3\",\n\t\t\t\t\tInArgs: []rpc.ArgDesc{{Name: \"i0_3\", Doc: \"Doc i0_3\"}},\n\t\t\t\t\tOutArgs: []rpc.ArgDesc{{Name: \"o0_3\", Doc: \"Doc o0_3\"}},\n\t\t\t\t\tInStream: rpc.ArgDesc{Name: \"is_3\", Doc: \"Doc is_3\"},\n\t\t\t\t\tOutStream: rpc.ArgDesc{Name: \"os_3\", Doc: \"Doc os_3\"},\n\t\t\t\t\tTags: []*vdl.Value{tagAlpha, tagBeta},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tName: \"Iface2\",\n\t\t\tPkgPath: \"d/e/f\",\n\t\t\tDoc: \"Doc Iface2\",\n\t\t\tMethods: []rpc.MethodDesc{\n\t\t\t\t{\n\t\t\t\t\t// The same Sig3 method is described here in a different interface.\n\t\t\t\t\tName: \"Sig3\",\n\t\t\t\t\tDoc: \"Doc Sig3x\",\n\t\t\t\t\tInArgs: []rpc.ArgDesc{{Name: \"i0_3x\", Doc: \"Doc i0_3x\"}},\n\t\t\t\t\tOutArgs: []rpc.ArgDesc{{Name: \"o0_3x\", Doc: \"Doc o0_3x\"}},\n\t\t\t\t\tInStream: rpc.ArgDesc{Name: \"is_3x\", Doc: \"Doc is_3x\"},\n\t\t\t\t\tOutStream: rpc.ArgDesc{Name: \"os_3x\", Doc: \"Doc os_3x\"},\n\t\t\t\t\t// Must have the same tags as every other definition of this method.\n\t\t\t\t\tTags: []*vdl.Value{tagAlpha, tagBeta},\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tName: \"Sig4\",\n\t\t\t\t\tDoc: \"Doc Sig4\",\n\t\t\t\t\tInArgs: []rpc.ArgDesc{\n\t\t\t\t\t\t{Name: \"i0_4\", Doc: \"Doc i0_4\"}, {Name: \"i1_4\", Doc: \"Doc i1_4\"}},\n\t\t\t\t\tOutArgs: []rpc.ArgDesc{\n\t\t\t\t\t\t{Name: \"o0_4\", Doc: \"Doc o0_4\"}, {Name: \"o1_4\", Doc: \"Doc o1_4\"}},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (s UserSet) Signature() string {\n\tres, _ := s.RecordCollection.Get(models.NewFieldName(\"Signature\", \"signature\")).(string)\n\treturn res\n}", "func execWriteSignature(_ int, p *gop.Context) {\n\targs := p.GetArgs(3)\n\ttypes.WriteSignature(args[0].(*bytes.Buffer), args[1].(*types.Signature), args[2].(types.Qualifier))\n}", "func execNewSignature(_ int, p *gop.Context) {\n\targs := p.GetArgs(4)\n\tret := types.NewSignature(args[0].(*types.Var), args[1].(*types.Tuple), args[2].(*types.Tuple), args[3].(bool))\n\tp.Ret(4, ret)\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\r\n\r\n\t//mspid := \"ORG1\"\r\n\t//cid := new ClientIdentity(APIstub);\r\n\t//mspid, err := cid.GetMSPID(APIstub)\r\n\t//if err != nil {\r\n\t//\tfmt.Println(\"Error - GetMSPID()\")\r\n\t//}\r\n\t//fmt.Print(\"MSPID: \")\r\n\t//fmt.Print(mspid)\r\n\t//fmt.Println(\"\\n\\n -\")\r\n\r\n\t// Get necessary creator info for calling the chaincode\r\n\t//id, err := cid.GetID(APIstub)\r\n\t//if err != nil {\r\n\t//\tfmt.Println(\"Error - cide.GetID()\")\r\n\t//}\r\n\t//fmt.Print(\"ID: \")\r\n\t//fmt.Print(id)\r\n\t//fmt.Println(\"\\n\\n -\")\r\n\r\n\t//mspid, err := cid.GetMSPID(APIstub)\r\n\t//if err != nil {\r\n\t//\tfmt.Println(\"Error - cide.GetMSPID()\")\r\n\t//}\r\n\t//fmt.Print(\"MSPID: \")\r\n\t//fmt.Print(mspid)\r\n\t//fmt.Println(\"\\n\\n -\")\r\n\r\n\t//cert, err := cid.GetX509Certificate(APIstub)\r\n\t//if err != nil {\r\n\t//\tfmt.Println(\"Error - cide.GetX509Certificate()\")\r\n\t//}\r\n\t//fmt.Print(\"GetX509Certificate: \\n\\n\")\r\n\t//fmt.Print(\"IPAddresses: \")\r\n\t//fmt.Println(cert.IPAddresses)\r\n\r\n\t//fmt.Print(\"IsCA: \")\r\n\t//fmt.Println(cert.IsCA)\r\n\r\n\t//fmt.Print(\"Issuer: \")\r\n\t//fmt.Println(cert.Issuer)\r\n\r\n\t//fmt.Print(\"IssuingCertificateURL: \")\r\n\t//fmt.Println(cert.IssuingCertificateURL)\r\n\r\n\t//fmt.Print(\"Public Key: \")\r\n\t//fmt.Println(cert.PublicKey)\r\n\r\n\t//fmt.Println(\"\\n\\n -\")\r\n\r\n\t// Retrieve the requested Smart Contract function and arguments\r\n\tfunction, args := APIstub.GetFunctionAndParameters()\r\n\t// Route to the appropriate handler function to interact with the ledger appropriately\r\n\tif function == \"queryTRU\" {\r\n\t\treturn s.queryTRU(APIstub, args)\r\n\t} else if function == \"querySCA\" {\r\n\t\treturn s.querySCA(APIstub, args)\r\n\t} else if function == \"initLedger\" {\r\n\t\treturn s.initLedger(APIstub)\r\n\t} else if function == \"updateTRU\" {\r\n\t\treturn s.updateTRU(APIstub, args)\r\n\t} else if function == \"createTRU\" {\r\n\t\treturn s.createTRU(APIstub, args)\r\n\t} else if function == \"splitTRU\" {\r\n\t\treturn s.splitTRU(APIstub, args)\r\n\t} else if function == \"updateSCA\" {\r\n\t\treturn s.updateSCA(APIstub, args)\r\n\t} else if function == \"queryAllSCA\" {\r\n\t\treturn s.queryAllSCA(APIstub)\r\n\t} else if function == \"queryAllTRU\" {\r\n\t\treturn s.queryAllTRU(APIstub)\r\n\t} else if function == \"queryTRACE\" {\r\n\t\treturn s.queryTRACE(APIstub, args)\r\n\t} else if function == \"updateTRACE\" {\r\n\t\treturn s.updateTRACE(APIstub, args)\r\n\t} else if function == \"changeOWN\" {\r\n\t\treturn s.changeOWN(APIstub, args)\r\n\t} else if function == \"searchTRU\" {\r\n\t\treturn s.searchTRU(APIstub, args)\t\r\n\t} else if function == \"queryNextTRU\" {\r\n\t\treturn s.queryNextTRU(APIstub)\r\n\t}\r\n\r\n\treturn shim.Error(\"Invalid Smart Contract function name.\")\r\n}", "func (i PullAll) Signature() uint8 {\n\treturn PullAllSignature\n}", "func (a Asset) signature() string {\n\tif !a.Config.URL.SignURL {\n\t\treturn \"\"\n\t}\n\n\talgo, length := a.getSignatureAlgorithmAndLength()\n\n\ttoSign := joinUrl([]interface{}{a.Transformation, a.PublicID})\n\n\treturn signature.SignURL(toSign, a.Config.Cloud.APISecret, algo, length)\n}", "func (e *engineImpl) VerifyHeaderWithSignature(chain engine.ChainReader, header *block.Header, commitSig []byte, commitBitmap []byte, reCalculate bool) error {\n\tif chain.Config().IsStaking(header.Epoch()) {\n\t\t// Never recalculate after staking is enabled\n\t\treCalculate = false\n\t}\n\tpublicKeys, err := GetPublicKeys(chain, header, reCalculate)\n\tif err != nil {\n\t\treturn ctxerror.New(\"[VerifyHeaderWithSignature] Cannot get publickeys for block header\").WithCause(err)\n\t}\n\n\tpayload := append(commitSig[:], commitBitmap[:]...)\n\taggSig, mask, err := ReadSignatureBitmapByPublicKeys(payload, publicKeys)\n\tif err != nil {\n\t\treturn ctxerror.New(\"[VerifyHeaderWithSignature] Unable to deserialize the commitSignature and commitBitmap in Block Header\").WithCause(err)\n\t}\n\thash := header.Hash()\n\n\tif e := header.Epoch(); chain.Config().IsStaking(e) {\n\t\tslotList, err := chain.ReadShardState(e)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"cannot read shard state\")\n\t\t}\n\t\td := quorum.NewDecider(quorum.SuperMajorityStake)\n\t\td.SetShardIDProvider(func() (uint32, error) {\n\t\t\treturn header.ShardID(), nil\n\t\t})\n\t\td.SetMyPublicKeyProvider(func() (*bls.PublicKey, error) {\n\t\t\treturn nil, nil\n\t\t})\n\t\td.SetVoters(slotList.FindCommitteeByID(header.ShardID()).Slots, true)\n\t\tif !d.IsQuorumAchievedByMask(mask, true) {\n\t\t\treturn ctxerror.New(\n\t\t\t\t\"[VerifySeal] Not enough voting power in commitSignature from Block Header\",\n\t\t\t)\n\t\t}\n\t} else {\n\t\tquorumCount, err := QuorumForBlock(chain, header, reCalculate)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err,\n\t\t\t\t\"cannot calculate quorum for block %s\", header.Number())\n\t\t}\n\t\tif count := utils.CountOneBits(mask.Bitmap); count < int64(quorumCount) {\n\t\t\treturn ctxerror.New(\"[VerifyHeaderWithSignature] Not enough signature in commitSignature from Block Header\",\n\t\t\t\t\"need\", quorumCount, \"got\", count)\n\t\t}\n\t}\n\tblockNumHash := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(blockNumHash, header.Number().Uint64())\n\tcommitPayload := append(blockNumHash, hash[:]...)\n\n\tif !aggSig.VerifyHash(mask.AggregatePublic, commitPayload) {\n\t\treturn ctxerror.New(\"[VerifySeal] Unable to verify aggregated signature for block\", \"blockNum\", header.Number().Uint64()-1, \"blockHash\", hash)\n\t}\n\treturn nil\n}", "func (s *SmartContract) Invoke(stub shim.ChaincodeStubInterface) sc.Response {\n\tfunctionname, args := stub.GetFunctionAndParameters()\n\tif functionname == \"queryChainKey\" {\n\t\treturn s.queryChainKey(stub, args)\n\t} else if functionname == \"listChainKeys\" {\n\t\treturn s.listChainKeys(stub)\n\t} else if functionname == \"recordChainKeys\" {\n\t\tfmt.Println(\"Recording hash...\")\n\t\treturn s.recordChainKey(stub, args)\n\t} else {\n\t\tfmt.Println(\"Invalid function name\")\n\t}\n\treturn shim.Success(nil)\n}", "func GetSignatureHeader(bytes []byte) (*common.SignatureHeader, error) {\n\tsh := &common.SignatureHeader{}\n\terr := proto.Unmarshal(bytes, sh)\n\treturn sh, err\n}", "func (s *service) Signatures(ctx *shared.Context, kind contract.ResourceKind, filter map[string]interface{}) (core.Records, error) {\n\tdbResource := s.dbResource(kind)\n\treturn s.signatures(ctx, dbResource, filter)\n}", "func signatureProcessing(writer http.ResponseWriter, request *http.Request, container bool) {\n\tswitch request.Method {\n\tcase \"GET\":\n\tcase \"POST\":\n\t\tif err := request.ParseForm(); err != nil {\n\t\t\tlog.Printf(\"ParseForm() err: %#v\", err)\n\t\t\treturn\n\t\t}\n\t\tlog.Printf(\n\t\t\t\"Request from client (request.PostFrom): %#v\\n\", request.PostForm,\n\t\t)\n\t\tif container {\n\t\t\t// Process the form values and return a container signature\n\t\t\t// file.\n\t\t\tsignatureFile, fileName := processContainerForm(request.PostForm)\n\t\t\tsetHeaders(writer, len(signatureFile), fileName)\n\t\t\t// Stream the content to the client.\n\t\t\treader := strings.NewReader(signatureFile)\n\t\t\tio.Copy(writer, reader)\n\t\t\treturn\n\t\t}\n\t\t// Process the form values and return a standard signature file.\n\t\tsignatureFile, fileName := processStandardForm(request.PostForm)\n\t\tsetHeaders(writer, len(signatureFile), fileName)\n\t\t// Stream the content to the client.\n\t\treader := strings.NewReader(signatureFile)\n\t\tio.Copy(writer, reader)\n\t\treturn\n\tdefault:\n\t\tconst onlyPost = \"Sorry, only POST methods are supported.\"\n\t\tfmt.Fprintf(writer, onlyPost)\n\t}\n}", "func GetResponseSignature(response types.ProxyResponse) (string, string) {\n\t// Get Response Hash\n\tresponseHash := GetBlake2bHash(response.Response)\n\n\t// Generate json to be signed\n\tsign := new(types.ResponseSign)\n\tsign.Chainid = response.Chainid\n\tsign.Block = response.Block\n\tsign.Blocktime = response.Blocktime\n\tsign.Timestamp = response.Timestamp\n\tsign.Response = responseHash\n\tsignBytes, err := json.Marshal(sign)\n\tif err != nil {\n\t\treturn \"\", responseHash\n\t}\n\n\t// Get Signature\n\tsignature, err := config.Config.PrivKey.Sign(signBytes)\n\tif err != nil {\n\t\treturn \"\", responseHash\n\t}\n\n\treturn base64.StdEncoding.EncodeToString([]byte(signature)), responseHash\n}", "func (ips *inProcessSigner) Sign(destinationURL string, body []byte) (string, error) {\n\treq := &api.AuthenticatedConnectionSignatureRequest{\n\t\tRequestInfo: createRequestInfo(destinationURL, body),\n\t}\n\tsignatureResponse, err := ips.signatory.SignAuthenticatedConnection(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn getSignatureMessage(signatureResponse)\n}", "func ComputeSignatureV2(authToken, uri string, nonce string) string {\n\tparsedUrl, err := url.Parse(uri)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tvar originalString string = parsedUrl.Scheme + \"://\" + parsedUrl.Host + parsedUrl.Path + nonce\n\tmac := hmac.New(sha256.New, []byte(authToken))\n\tmac.Write([]byte(originalString))\n\tvar messageMAC string = base64.StdEncoding.EncodeToString(mac.Sum(nil))\n\treturn messageMAC\n}", "func (s *SmartContract) ReadSignature(ctx contractapi.TransactionContextInterface, id string) (string, error) {\n\tctcJSON, err := ctx.GetStub().GetState(id)\n\tfmt.Print(string(ctcJSON))\n\tif err != nil {\n\t\treturn string(ctcJSON), fmt.Errorf(\"Unexpected error : %q\", err)\n\t}\n\tctc := new(Contract)\n\t_ = json.Unmarshal(ctcJSON, ctc)\n\treturn ctc.Signature, nil\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\t\tcase \"createDEC\":\n\t\t\treturn s.createDEC(APIstub, args)\n\t\tcase \"updateDEC\":\n\t\t\treturn s.updateDEC(APIstub, args)\n\t\tcase \"traceDEC\":\n\t\t\treturn s.traceDEC(APIstub, args)\n\t\tcase \"getDEC\":\n\t\t\treturn s.getDEC(APIstub, args)\n\t}\n\t\n\treturn shim.Error(\"Invoke Function Not Success.\")\n\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\tcase \"queryProduct\":\n\t\treturn s.queryProduct(APIstub, args)\n\tcase \"createProduct\":\n\t\treturn s.createProduct(APIstub, args)\n\tcase \"queryAllProduct\":\n\t\treturn s.queryAllProduct(APIstub)\n\tcase \"changeProductStatus\":\n\t\treturn s.changeProductStatus(APIstub, args)\n\tcase \"getHistoryForProduct\":\n\t\treturn s.getHistoryForProduct(APIstub, args)\n\tcase \"queryProductByStatus\":\n\t\treturn s.queryProductByStatus(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n\n\t// return shim.Error(\"Invalid Smart Contract function name.\")\n}", "func signatureBase(r io.ReadSeeker, s *Signature) (sig *Signature, msg, dkimheader []byte, err error) {\n\theaders := make(map[string][]Header)\n\tfor raw, conv, err := ReadSMTPHeaderRelaxed(r); err == nil; raw, conv, err = ReadSMTPHeaderRelaxed(r) {\n\t\tsplit := bytes.SplitN(conv, []byte{':'}, 2)\n\t\tname := string(split[0])\n\t\t// headers acts as an upside-down stack. We add the oldest ones\n\t\t// to the start, and consume from the front in dkimMessageBase.\n\t\theaders[name] = append([]Header{Header{raw, conv}}, headers[name]...)\n\t\tif name == \"dkim-signature\" && s == nil {\n\t\t\tsig = ParseSignature(raw)\n\t\t}\n\t}\n\tcbody, err := ReadSMTPBodyRelaxed(r)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\tsha := sha256.Sum256(cbody[:])\n\tencoded := string(base64.StdEncoding.EncodeToString(sha[:]))\n\tif s != nil {\n\t\tsig = s\n\t\ts.BodyHash = encoded\n\t\traw := []byte(s.String())\n\t\trelax := relaxHeader(raw)\n\t\theaders[\"dkim-signature\"] = append([]Header{{raw, relax}}, headers[\"dkim-signature\"]...)\n\t}\n\tif sig == nil {\n\t\treturn nil, nil, nil, fmt.Errorf(\"Permanent failure: no DKIM signature\")\n\t}\n\tif encoded != sig.BodyHash {\n\t\treturn nil, nil, nil, fmt.Errorf(\"Permanent failure: body hash does not match\")\n\t}\n\n\tvar tohash []byte\n\tfor _, h := range sig.Headers {\n\t\tvar hval Header\n\t\tlh := strings.ToLower(h)\n\t\tif header, ok := headers[lh]; ok && len(header) > 0 {\n\t\t\t// If there is a header, consume it so that if a header\n\t\t\t// is included in sig.Headers multiple times the next\n\t\t\t// one is correct.\n\t\t\thval = header[0]\n\t\t\theaders[lh] = header[1:]\n\t\t}\n\t\tswitch sig.HeaderCanonicalization {\n\t\tcase \"simple\":\n\t\t\ttohash = append(tohash, hval.Raw...)\n\t\tcase \"relaxed\":\n\t\t\ttohash = append(tohash, hval.Relaxed...)\n\t\t}\n\t}\n\tvar sighead []byte\n\trawsig, ok := headers[\"dkim-signature\"]\n\tif !ok {\n\t\treturn nil, nil, nil, fmt.Errorf(\"Permanent failure: No DKIM-Signature\")\n\t}\n\tif sig.HeaderCanonicalization == \"relaxed\" {\n\t\tsighead = bytes.TrimRight(rawsig[0].Relaxed, \"\\r\\n\")\n\t} else {\n\t\tsighead = bytes.TrimRight(rawsig[0].Raw, \"\\r\\n\")\n\t}\n\treturn sig, tohash, sighead, nil\n\n}", "func (t Type) CallSignature() *CallSignature {\n\tcallImpl, isCallable := t.impl.(typeCallable)\n\tif !isCallable {\n\t\treturn nil\n\t}\n\n\treturn callImpl.CallSignature()\n}", "func (*SignatureDescriptor) Descriptor() ([]byte, []int) {\n\treturn file_x_authn_v1alpha1_signing_proto_rawDescGZIP(), []int{1}\n}", "func (s PublicService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\n\tswitch method {\n\tcase RPC.PublicService.GoPGVersions:\n\t\tresp.Set(s.GoPGVersions())\n\n\tcase RPC.PublicService.Modes:\n\t\tresp.Set(s.Modes())\n\n\tcase RPC.PublicService.SearchTypes:\n\t\tresp.Set(s.SearchTypes())\n\n\tcase RPC.PublicService.Types:\n\t\tresp.Set(s.Types())\n\n\tcase RPC.PublicService.DBTypes:\n\t\tresp.Set(s.DBTypes())\n\n\tcase RPC.PublicService.HTMLTypes:\n\t\tresp.Set(s.HTMLTypes())\n\n\tcase RPC.PublicService.Ping:\n\t\tresp.Set(s.Ping())\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func CreateSignature(algo x509.SignatureAlgorithm, authMethod protocol.AuthMethod, signed []byte, private crypto.Signer, log log.Logger) ([]byte, error) {\n\t// if using a plain old signature, this is all we need\n\tswitch authMethod {\n\tcase protocol.AUTH_RSA_DIGITAL_SIGNATURE, protocol.AUTH_DSS_DIGITAL_SIGNATURE:\n\t\treturn signData(x509.SHA1WithRSA, signed, private, log)\n\tcase protocol.AUTH_DIGITAL_SIGNATURE:\n\tdefault:\n\t\treturn nil, errors.Errorf(\"Authentication Method is not supported: %s\", authMethod)\n\t}\n\tsignature, err := signData(algo, signed, private, log)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// encode rfc7427 signature\n\tsigAuth := &protocol.SignatureAuth{\n\t\tAsn1Data: certAuthToAsn[algo],\n\t\tSignature: signature,\n\t}\n\treturn sigAuth.Encode(), nil\n}", "func (ss StdSignature) GetSignature() []byte {\n\treturn ss.Signature\n}", "func (y Yubikey) Signature() (*Slot, error) {\n\treturn y.Slot(Signature)\n}", "func (o *OperationService) decodeSignature(sig string) (string, error) {\n\tdecBytes, err := crypto.Decode(sig)\n\tif err != nil {\n\t\treturn \"\", errors.Wrap(err, \"could not decode signature\")\n\t}\n\treturn hex.EncodeToString(decBytes), nil\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\t// Retrieve the requested Smart Contract function and arguments\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\n\tswitch function {\n\tcase \"GetSegment\":\n\t\treturn s.GetSegment(APIstub, args)\n\tcase \"FindSegments\":\n\t\treturn s.FindSegments(APIstub, args)\n\tcase \"GetMapIDs\":\n\t\treturn s.GetMapIDs(APIstub, args)\n\tcase \"SaveSegment\":\n\t\treturn s.SaveSegment(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n}", "func (repo repository) GetUserSignatures(ctx context.Context, params signatures.GetUserSignaturesParams, pageSize int64) (*models.Signatures, error) {\n\tf := logrus.Fields{\n\t\t\"functionName\": \"GetUserSignatures\",\n\t\tutils.XREQUESTID: ctx.Value(utils.XREQUESTID),\n\t}\n\t// This is the keys we want to match\n\tcondition := expression.Key(\"signature_reference_id\").Equal(expression.Value(params.UserID))\n\n\t// Use the nice builder to create the expression\n\texpr, err := expression.NewBuilder().WithKeyCondition(condition).WithProjection(buildProjection()).Build()\n\tif err != nil {\n\t\tlog.WithFields(f).Warnf(\"error building expression for user signature query, userID: %s, error: %v\",\n\t\t\tparams.UserID, err)\n\t\treturn nil, err\n\t}\n\n\t// Assemble the query input parameters\n\tqueryInput := &dynamodb.QueryInput{\n\t\tExpressionAttributeNames: expr.Names(),\n\t\tExpressionAttributeValues: expr.Values(),\n\t\tKeyConditionExpression: expr.KeyCondition(),\n\t\tFilterExpression: expr.Filter(),\n\t\tProjectionExpression: expr.Projection(),\n\t\tTableName: aws.String(repo.signatureTableName),\n\t\tIndexName: aws.String(SignatureReferenceIndex), // Name of a secondary index to scan\n\t\tLimit: aws.Int64(pageSize), // The maximum number of items to evaluate (not necessarily the number of matching items)\n\t}\n\n\t// If we have the next key, set the exclusive start key value\n\tif params.NextKey != nil {\n\t\tlog.WithFields(f).Debugf(\"Received a nextKey, value: %s\", *params.NextKey)\n\t\t// The primary key of the first item that this operation will evaluate.\n\t\t// and the query key (if not the same)\n\t\tqueryInput.ExclusiveStartKey = map[string]*dynamodb.AttributeValue{\n\t\t\t\"signature_id\": {\n\t\t\t\tS: params.NextKey,\n\t\t\t},\n\t\t\t\"signature_reference_id\": {\n\t\t\t\tS: &params.UserID,\n\t\t\t},\n\t\t}\n\t}\n\n\tsigs := make([]*models.Signature, 0)\n\tvar lastEvaluatedKey string\n\n\t// Loop until we have all the records\n\tfor ok := true; ok; ok = lastEvaluatedKey != \"\" {\n\t\t// Make the DynamoDB Query API call\n\t\tresults, errQuery := repo.dynamoDBClient.Query(queryInput)\n\t\tif errQuery != nil {\n\t\t\tlog.WithFields(f).Warnf(\"error retrieving user signatures for user: %s/%s, error: %v\",\n\t\t\t\tparams.UserID, *params.UserName, errQuery)\n\t\t\treturn nil, errQuery\n\t\t}\n\n\t\t// Convert the list of DB models to a list of response models\n\t\tsignatureList, modelErr := repo.buildProjectSignatureModels(ctx, results, \"\", LoadACLDetails)\n\t\tif modelErr != nil {\n\t\t\tlog.WithFields(f).Warnf(\"error converting DB model to response model for signatures for user %s/%s, error: %v\",\n\t\t\t\tparams.UserID, *params.UserName, modelErr)\n\t\t\treturn nil, modelErr\n\t\t}\n\n\t\t// Add to the signatures response model to the list\n\t\tsigs = append(sigs, signatureList...)\n\n\t\t// log.WithFields(f).Debugf(\"LastEvaluatedKey: %+v\", results.LastEvaluatedKey[\"signature_id\"])\n\t\tif results.LastEvaluatedKey[\"signature_id\"] != nil {\n\t\t\tlastEvaluatedKey = *results.LastEvaluatedKey[\"signature_id\"].S\n\t\t\tqueryInput.ExclusiveStartKey = results.LastEvaluatedKey\n\t\t} else {\n\t\t\tlastEvaluatedKey = \"\"\n\t\t}\n\n\t\tif int64(len(sigs)) >= pageSize {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// How many total records do we have - may not be up-to-date as this value is updated only periodically\n\tdescribeTableInput := &dynamodb.DescribeTableInput{\n\t\tTableName: &repo.signatureTableName,\n\t}\n\tdescribeTableResult, err := repo.dynamoDBClient.DescribeTable(describeTableInput)\n\tif err != nil {\n\t\tlog.WithFields(f).Warnf(\"error retrieving total record count for user: %s/%s, error: %v\",\n\t\t\tparams.UserID, *params.UserName, err)\n\t\treturn nil, err\n\t}\n\n\t// Meta-data for the response\n\ttotalCount := *describeTableResult.Table.ItemCount\n\n\treturn &models.Signatures{\n\t\tProjectID: \"\",\n\t\tResultCount: int64(len(sigs)),\n\t\tTotalCount: totalCount,\n\t\tLastKeyScanned: lastEvaluatedKey,\n\t\tSignatures: sigs,\n\t}, nil\n}", "func (obj *request) Signature() signature.Signature {\n\treturn obj.sig\n}", "func (this *Select) Signature() value.Value {\n\treturn this.subresult.Signature()\n}", "func (client ArtifactsClient) listContainerImageSignatures(ctx context.Context, request common.OCIRequest, binaryReqBody *common.OCIReadSeekCloser, extraHeaders map[string]string) (common.OCIResponse, error) {\n\n\thttpRequest, err := request.HTTPRequest(http.MethodGet, \"/container/imageSignatures\", binaryReqBody, extraHeaders)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar response ListContainerImageSignaturesResponse\n\tvar httpResponse *http.Response\n\thttpResponse, err = client.Call(ctx, &httpRequest)\n\tdefer common.CloseBodyIfValid(httpResponse)\n\tresponse.RawResponse = httpResponse\n\tif err != nil {\n\t\tapiReferenceLink := \"https://docs.oracle.com/iaas/api/#/en/registry/20160918/ContainerImageSignatureSummary/ListContainerImageSignatures\"\n\t\terr = common.PostProcessServiceError(err, \"Artifacts\", \"ListContainerImageSignatures\", apiReferenceLink)\n\t\treturn response, err\n\t}\n\n\terr = common.UnmarshalResponse(httpResponse, &response)\n\treturn response, err\n}", "func (d *Deck) GetSignature() string {\n\tsig := \"\"\n\tfor _, card := range d.cards {\n\t\tsig += card.GetSignature()\n\t}\n\treturn sig\n}", "func (s *SmartContract) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n \n fn, args := stub.GetFunctionAndParameters()\n\n if fn == \"AddComponent\" {\n return s.AddComponent(stub, args)\n } else if fn == \"InitLedger\" {\n return s.InitLedger(stub)\n } else if fn == \"QueryCar\" {\n return s.QueryCar(stub, args)\n } else if fn == \"QueryComponent\" {\n return s.QueryComponent(stub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name.\")\n \n}" ]
[ "0.66295636", "0.607431", "0.5982824", "0.5577489", "0.551738", "0.5453798", "0.53808355", "0.5306812", "0.5105116", "0.5049495", "0.49796417", "0.4950732", "0.4820985", "0.48139128", "0.48001745", "0.47923473", "0.4780425", "0.47619063", "0.4744137", "0.47363123", "0.4726382", "0.47251844", "0.46585235", "0.46386397", "0.46285927", "0.4627002", "0.460637", "0.46042982", "0.45928943", "0.4591743", "0.45702475", "0.4522126", "0.45204952", "0.45119816", "0.45095053", "0.4495009", "0.447554", "0.44689327", "0.44596177", "0.4453525", "0.4429842", "0.44054243", "0.43978614", "0.43972746", "0.43877354", "0.43813595", "0.43741757", "0.43707857", "0.4364599", "0.43611434", "0.43552646", "0.43498194", "0.4342677", "0.4301917", "0.42979088", "0.4295054", "0.42903018", "0.42791528", "0.42659953", "0.42641458", "0.4259489", "0.42565843", "0.42430568", "0.423893", "0.4231397", "0.4219246", "0.4213835", "0.4212028", "0.42101374", "0.4209618", "0.4209018", "0.42072958", "0.4195134", "0.4191269", "0.41792855", "0.41772708", "0.41770285", "0.41689184", "0.41683766", "0.41655964", "0.41631442", "0.41562772", "0.41557562", "0.41497403", "0.41489983", "0.41455448", "0.4143452", "0.41414392", "0.41405195", "0.41382086", "0.413503", "0.41313782", "0.41300687", "0.41283253", "0.41227612", "0.41179547", "0.41174763", "0.41118777", "0.41059804", "0.41044694" ]
0.77472115
0
A helper function that sends a remote message that expects a response.
func (fs fakeService) callRemoteWithResponse(ctx *context.T, message *bindings.Message) (outMessage *bindings.Message, err error) { ctx.Infof("callRemoteGeneric: Send message along the router") readResult := <-fs.router.AcceptWithResponse(message) if err = readResult.Error; err != nil { return } ctx.Infof("callRemoteGeneric: Audit response message header flag") // The message flag we receive back must be a bindings.MessageIsResponseFlag if readResult.Message.Header.Flags != bindings.MessageIsResponseFlag { err = &bindings.ValidationError{bindings.MessageHeaderInvalidFlags, fmt.Sprintf("invalid message header flag: %v", readResult.Message.Header.Flags), } return } ctx.Infof("callRemoteGeneric: Audit response message header type") // While the mojo service we called into will return a header whose // type must match our outgoing one. if got, want := readResult.Message.Header.Type, message.Header.Type; got != want { err = &bindings.ValidationError{bindings.MessageHeaderUnknownMethod, fmt.Sprintf("invalid method in response: expected %v, got %v", want, got), } return } return readResult.Message, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (client *Client) SendResponse(message string) {\n\tlog.Printf(\"=> %v\", message)\n\tconnection := client.connection\n\t_, err := connection.Cmd(message)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (m Matcher) sendResponse(requestMessage telegram.RequestMessage) error {\n\tresponseMessage := telegram.Message{\n\t\tText: \"pong\",\n\t\tReplyToMessageID: requestMessage.ID,\n\t}\n\n\treturn telegram.SendMessage(requestMessage, responseMessage)\n}", "func (p *player) sendResponse(messageType uint, data ...interface{}) {\n\tp.sendBytes(createResponse(messageType, data...))\n}", "func sendSimpleReply(conn net.Conn, rep byte) error {\n\tp := socks5.NewReply(rep, socks5.ATYPIPv4, []byte{0x00, 0x00, 0x00, 0x00}, []byte{0x00, 0x00})\n\t_, err := p.WriteTo(conn)\n\treturn err\n}", "func TestServerResponseOk(t *testing.T) {\n\ttestServer := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.WriteHeader(http.StatusOK)\n\t\tw.Write([]byte(`{}`))\n\t}))\n\tresult, err := sendMessage(testServer.URL, \"[email protected]\", \"test\")\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\tif result != false {\n\t\tt.Fail()\n\t}\n}", "func sendResponseRaw(w http.ResponseWriter, r *http.Request, resp []byte, err error) {\n\n\tif err != nil {\n\t\tlog.Printf(err.Error())\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tmsg := fmt.Sprintf(`{\"success\" : true, \"data\" : \"%x\"}`, resp)\n\tfmt.Fprintln(w, msg)\n\n}", "func sendTextResponse(w http.ResponseWriter, msg string) {\n\tw.Write([]byte(msg))\n}", "func (r *Receiver) sendResponse(packet *ptp.SyncDelayReq, sourceIP string, rawPacket gopacket.Packet) error {\n\tdst, err := net.ResolveIPAddr(\"ip6:ipv6-icmp\", sourceIP)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to resolve sender address: %w\", err)\n\t}\n\tconn, err := net.ListenPacket(\"ip6:ipv6-icmp\", \"\")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to establish connection: %w\", err)\n\t}\n\tdefer conn.Close()\n\n\tmess := icmp.Message{\n\t\tType: ipv6.ICMPTypeTimeExceeded, Code: 0,\n\t\tBody: &icmp.RawBody{\n\t\t\tData: rawPacket.Data()[PTPUnusedSize:],\n\t\t},\n\t}\n\tbuf, err := mess.Marshal(nil)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to marshal the icmp packet: %w\", err)\n\t}\n\tif _, err := conn.WriteTo(buf, dst); err != nil {\n\t\treturn fmt.Errorf(\"unable to write to connection: %w\", err)\n\t}\n\treturn nil\n}", "func send_response() {\r\n\r\n\tfmt.Printf(\"<RESPONSE>\\n<STATUS>\\n\")\r\n\tfmt.Printf(\"<status_code>%d</status_code>\\n\",status_info.status_code)\r\n\tfmt.Printf(\"<error_message>%s</error_message>\\n\",status_info.error_message)\r\n\tfmt.Printf(\"<error_details>%s</error_details>\\n\",status_info.error_details)\r\n\tfmt.Printf(\"</STATUS>\\n\")\r\n\tif response_data != \"\" {\r\n\t\tfmt.Printf(\"%s\",response_data)\r\n\t}\r\n\tfmt.Printf(\"</RESPONSE>\")\r\n\tos.Exit(0)\r\n}", "func sendResponse(message spark.Message) error {\n\t// create a new message\n\tm, err := s.CreateMessage(message)\n\tif err != nil {\n\t\tlog.Printf(\"Unable to create message.\\nM: %v\\n\", m)\n\t}\n\treturn err\n}", "func sendAndGetExpect(c *Conn, cmd, expectResult string) error {\n\n\tresp, err := send(c, cmd)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif resp != expectResult {\n\t\treturn parseError(resp)\n\t}\n\treturn nil\n}", "func SendResponse(output string){\n\n url := \"https://127.0.0.1/\" + url.PathEscape(output)\n _, err := http.Get(url)\n if err != nil {\n\t\t//log.Fatalln(err)\n\t\treturn\n }\n\n}", "func send(context string, req *client.HTTPRequest, resBody interface{}, expectedRespCode int) error {\n\terr := client.Invoke(context, req, resBody, expectedRespCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func sendReply(conn *Conn, p *Packet) {\n\tactualRequest := string([]byte(p.Msg))\n\tif actualRequest == expectedRequest {\n\t\tconn.UnicastTo([]byte(expectedReply), p.Addr)\n\t} else {\n\t\tconn.UnicastTo([]byte(\"Go away!\"), p.Addr)\n\t}\n}", "func TestReqResp(t *testing.T) {\n\t// Connect to NATS\n\tm := NewMessenger(testConfig)\n\tdefer m.Close()\n\n\t// Use a WaitGroup to wait for the message to arrive\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\n\t// Subscribe to the source subject with the message processing function\n\ttestSubject := \"test_subject\"\n\ttestMsgContent := []byte(\"Some text to send...\")\n\ttestRespContent := []byte(\"Some text to send back as response...\")\n\tm.Response(testSubject, func(content []byte) ([]byte, error) {\n\t\tdefer wg.Done()\n\t\trequire.EqualValues(t, content, testMsgContent)\n\t\treturn testRespContent, nil\n\t})\n\n\t// Send a message\n\tresp, err := m.Request(testSubject, testMsgContent, 50*time.Millisecond)\n\tassert.Nil(t, err)\n\trequire.EqualValues(t, resp, testRespContent)\n\n\t// Wait for the message to come in\n\twg.Wait()\n}", "func sendResponse(w http.ResponseWriter, r *http.Request, resp api.SmartCardResponse, err error) {\n\n\tif err != nil {\n\t\tlog.Printf(err.Error())\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tb, err := json.Marshal(resp)\n\tif err != nil {\n\t\tlog.Printf(err.Error())\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tmsg := fmt.Sprintf(`{\"success\" : true, \"data\" : %s}`, string(b))\n\tfmt.Fprintln(w, msg)\n\n}", "func send(c *Conn, cmd string) (string, error) {\n\t_, err := sendFull(c, []byte(cmd))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t//wait for response\n\tresp, err := c.bufReader.ReadString('\\n')\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn resp, nil\n}", "func (server *Server) sendReturn(message []string) error {\n\tserver.Lock()\n\tdefer server.Unlock()\n\tvar buf bytes.Buffer\n\tenc := gob.NewEncoder(&buf)\n\tstatus := datums.ClientResult{Name: clientHost, Results: message}\n\terr := enc.Encode(&status)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = server.Channel.Write(buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (_m *WebSocketServer) SendReply(message interface{}) {\n\t_m.Called(message)\n}", "func (r *serverRequest) sendResponse(reply interface{}, errmsg string) error {\n\tif DEBUG_PRINT { //debug\n\t\tfmt.Println(\"Send Response\")\n\t}\n\tresp := new(serverResponse)\n\n\tif errmsg != \"\" {\n\t\tresp.Error = errmsg\n\t\treply = nil //invalidRequest\n\t}\n\n\tresp.Id = r.Id\n\n\tif errmsg == \"\" {\n\t\tresp.Error = nil\n\t\tresp.Result = reply\n\t} else {\n\t\tresp.Error = errmsg\n\t}\n\treturn r.parent.enc.Encode(resp)\n\n}", "func sendResponse(sim *backends.SimulatedBackend, contract *contracts.Verifier, hash string, responseNodeVer *big.Int) {\n\tvar auth *bind.TransactOpts\n\tfor key := range initAddresses {\n\t\tauth = key\n\t\tdelete(initAddresses, auth)\n\t\tbreak\n\t}\n\n\t_, err := contract.Verify(auth, hash, responseNodeVer)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\tsim.Commit()\n}", "func (req *clientRequest) sendResponse(ch *clientHandle, err error) {\n\treq.response <- clientResponse{clientHandle: ch, Err: err}\n}", "func echo(conn net.Conn){\n defer conn.Close()\n bData, _ := recvDataB(conn)\n sendDataB(conn, bData, OK_CODE)\n}", "func sendClientResponse(resp *messages.ClientPollResponse, w http.ResponseWriter) {\n\tdata, err := resp.EncodePollResponse()\n\tif err != nil {\n\t\tlog.Printf(\"error encoding answer\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t} else {\n\t\tif _, err := w.Write([]byte(data)); err != nil {\n\t\t\tlog.Printf(\"unable to write answer with error: %v\", err)\n\t\t}\n\t}\n}", "func sendResponse(clientAddr *net.UDPAddr, msgID []byte, respPay pb.KVResponse) {\n\tif self.Addr.String() == clientAddr.String() {\n\t\tfmt.Println(\"Stop. No need to send response to self. \", clientAddr.Port)\n\t\tfmt.Println(respPay)\n\t\treturn\n\t}\n\n\trespPayBytes, err := proto.Marshal(&respPay)\n\tif err != nil {\n\t\tfmt.Println(\"Could not marshall the payload\")\n\t\treturn\n\t}\n\n\tcheckSum := getChecksum(msgID, respPayBytes)\n\n\trespMsg := pb.Msg{\n\t\tMessageID: msgID,\n\t\tPayload: respPayBytes,\n\t\tCheckSum: checkSum,\n\t\tType: 1, // response\n\t}\n\n\trespMsgBytes, err := proto.Marshal(&respMsg)\n\tif err != nil {\n\t\tfmt.Println(\"Could not marshall the message\")\n\t\treturn\n\t}\n\n\tfmt.Println(self.Addr.String(), \"🤖ABOUT TO SENT\", respMsgBytes, clientAddr.String())\n\t//Cache the response if it's not a server overload\n\tif respPay.ErrCode != SYS_OVERLOAD_ERR {\n\t\t// If there is no space to add to the cache, send a server\n\t\t// overload response instead\n\t\tif !responseCache.Add(msgID, getNetAddress(clientAddr), respMsgBytes) {\n\t\t\trespPay.ErrCode = SYS_OVERLOAD_ERR\n\t\t\trespPay.OverloadWaitTime = &overloadWaitTimeMs\n\t\t\tsendResponse(clientAddr, msgID, respPay)\n\t\t}\n\t}\n\n\tfmt.Println(\"sending \", respMsgBytes, \" to \", clientAddr.Port)\n\t// Send the message back to the client\n\t_, err = conn.WriteToUDP(respMsgBytes, clientAddr)\n\tif err != nil {\n\t\tfmt.Println(\"sendResponse WriteToUDP\", err)\n\t}\n\n\tif _, ok := incomingCache.Delete(msgID); !ok {\n\t\tfmt.Println(\"Error: No Request Found in RequestsCache.\", hex.EncodeToString(msgID))\n\t}\n}", "func (a *agent) SendResponse(ctx context.Context, resp *api.Response) (*api.Empty, error) {\n\ta.output[resp.GetAgentID()] <- resp\n\n\t// if there was an error handling the command, the error message will be set\n\t// in the Command's Out field. The server should log any errors that the\n\t// agent encounters. This function will only return an error if there is an\n\t// issue handling/forwarding the response to a client.\n\tif resp.GetStatus() == api.Status_ERROR {\n\t\tlog.Error(string(resp.GetData()))\n\t}\n\treturn api.EmptyMessage, nil\n}", "func (m *Message) send() *http.Response {\n\tvar client = &http.Client{\n\t\tTimeout: time.Second * 10,\n\t}\n\n\treq, err := http.NewRequest(m.method, m.url, bytes.NewReader(m.body))\n\tif err != nil {\n\t\tlog.Printf(\"%v\\n\", err)\n\t}\n\treq.Header = m.headers\n\n\t// Prvent remote server from keeping connection alive\n\treq.Close = true\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlog.Printf(\"%v\\n\", err)\n\t\treturn resp\n\t}\n\n\tif m.debug {\n\t\tlog.Printf(\"%T: %v\\n\", resp, resp)\n\t}\n\n\t// explicitly close body to avoid leaks\n\tresp.Body.Close()\n\treturn resp\n}", "func TestSendError(t *testing.T) {\n\terr := SendError(\"Send Error\", \"https://status.btfs.io\", \"my peer id\", \"my HValue\")\n\tif err != nil {\n\t\tt.Errorf(\"Send error message to status server failed, reason: %v\", err)\n\t} else {\n\t\tt.Log(\"Send error message to status server successfully!\")\n\t}\n}", "func (net *netService) response(session *session.Session, data []byte) error {\n\t// current message is notify message, can not response\n\tif session.LastID <= 0 {\n\t\treturn ErrSessionOnNotify\n\t}\n\tm, err := message.Encode(&message.Message{\n\t\tType: message.MessageType(message.Response),\n\t\tID: session.LastID,\n\t\tData: data,\n\t})\n\tif err != nil {\n\t\tlog.Errorf(err.Error())\n\t\treturn err\n\t}\n\n\tp := packet.Packet{\n\t\tType: packet.Data,\n\t\tLength: len(m),\n\t\tData: m,\n\t}\n\tep, err := p.Pack()\n\tif err != nil {\n\t\tlog.Errorf(err.Error())\n\t\treturn err\n\t}\n\n\tnet.send(session, ep)\n\treturn nil\n}", "func (r *response) Reply(message string) {\n\t_, _ = r.client.SendMessage(r.channel, message)\n}", "func reply(ch *amqp.Channel, out *output) {\n\tvar body []byte\n\tvar err error\n\tvar responseMsg amqp.Publishing\n\n\tresponseMsg.MessageId = out.ID\n\tif out.Err != nil {\n\t\tlog.Printf(\"Failed HTTP protocol to Maestro (will deliver error 599): %v, msg ID: %v\\n\",\n\t\t\tout.Err, out.ID)\n\t\tresponseMsg.Body = []byte(out.Err.Error())\n\t\tresponseMsg.Type = strconv.Itoa(599)\n\t} else {\n\t\tbody, err = ioutil.ReadAll(out.Response.Body)\n\t\tLogOnError(err, \"Read Response.Body: \"+out.ID)\n\t\tout.Response.Body.Close()\n\t\tresponseMsg.Body = body\n\t\tresponseMsg.Type = strconv.Itoa(out.Response.StatusCode)\n\t}\n\t// We use an ephemeral callback queue instead of an HTTP callback\n\t// see https://www.rabbitmq.com/direct-reply-to.html\n\terr = ch.Publish(\n\t\t\"\",\n\t\tout.ReplyTo,\n\t\tfalse, // mandatory\n\t\tfalse, // immediate\n\t\tresponseMsg,\n\t)\n\tif err != nil {\n\t\t// This failure case (if possible) can lead to deadlock,\n\t\t// the client will be stuck waiting for this out message\n\t\t// so we won't ack the original message\n\t\t// FIXME (msf): should we logfatal and exit? only when client exits will msg be released\n\t\tlog.Printf(\"Failed rabbitmq.Publish() to: %v, err: %v, will NOT ACK %v\\n\",\n\t\t\tout.ReplyTo,\n\t\t\terr,\n\t\t\tout.ID)\n\t} else {\n\t\tlog.Printf(\"Delivered result for message: %v, len: %v, replyTo: %v\\n\",\n\t\t\tout.ID, len(responseMsg.Body), out.ReplyTo,\n\t\t)\n\t\terr = ch.Ack(out.DeliveryTag, false)\n\t\tLogOnError(err, fmt.Sprintf(\"ERROR, failed to ack message: %v, deliveryTag: %v\\n\",\n\t\t\tout.ID, out.DeliveryTag))\n\t}\n}", "func waitForResponse(originalResp *Response, channel chan *rpc.Call) {\n\tresp := <-channel\n\tvar test *Response\n\ttest = resp.Reply.(*Response)\n\ttest.client.Close()\n\toriginalResp.Reply = test.Reply\n\treturn\n}", "func (c *Client) SendResponse(requestId string, response ocpp.Response) error {\n\tcallResult, err := c.CreateCallResult(response, requestId)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonMessage, err := callResult.MarshalJSON()\n\tif err != nil {\n\t\treturn ocpp.NewError(GenericError, err.Error(), requestId)\n\t}\n\tif err = c.client.Write(jsonMessage); err != nil {\n\t\tlog.Errorf(\"error sending response [%s]: %v\", callResult.GetUniqueId(), err)\n\t\treturn ocpp.NewError(GenericError, err.Error(), requestId)\n\t}\n\tlog.Debugf(\"sent CALL RESULT [%s]\", callResult.GetUniqueId())\n\tlog.Debugf(\"sent JSON message to server: %s\", string(jsonMessage))\n\treturn nil\n}", "func respondToRaw(sock mangos.Socket) {\n\n\tmsg, err := sock.RecvMsg()\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\tlog.Println(string(msg.Body))\n\n}", "func sendRequest(conn net.Conn, text string) {\n message := text;\n \n if _,err := conn.Write([]byte(message + \"\\n\")); err != nil {\n log.Fatal(err)\n }\n}", "func sendMessage(url string, email, text string) (bool, error) {\n\treqBody := messageRequest{email, text}\n\tmarshallBody, err := json.Marshal(reqBody)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\treq, err := http.NewRequest(\"POST\", url, bytes.NewReader(marshallBody))\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tclient := &http.Client{}\n\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode >= http.StatusBadRequest {\n\t\treturn false, fmt.Errorf(errBadCodeTpl, resp.StatusCode)\n\t}\n\tvar respBody response\n\tif err := json.NewDecoder(resp.Body).Decode(&respBody); err != nil {\n\t\treturn false, err\n\t}\n\n\tif respBody.Code != 0 && respBody.Code != http.StatusOK {\n\t\treturn false, fmt.Errorf(errBadCodeTpl, respBody.Code)\n\t}\n\treturn respBody.Paid, nil\n}", "func wait_for_response(w dns.ResponseWriter, conn *dns.Conn, request *dns.Msg) (response *dns.Msg) {\n\tfor {\n\t\tresponse, err := conn.ReadMsg()\n\t\t// some sort of error reading reply\n\t\tif err != nil {\n\t\t\t_D(\"%s QID:%d error reading message: %s\", w.RemoteAddr(), request.Id, err)\n\t\t\tSRVFAIL(w, request)\n\t\t\treturn nil\n\t\t}\n\t\t// got a response, life is good\n\t\tif response.Id == request.Id {\n\t\t\t_D(\"%s QID:%d got reply\", w.RemoteAddr(), request.Id)\n\t\t\treturn response\n\t\t}\n\t\t// got a response, but it was for a different QID... ignore\n\t\t_D(\"%s QID:%d ignoring reply to wrong QID:%d\", w.RemoteAddr(), request.Id, response.Id)\n\t}\n}", "func sendReply(w io.Writer, resp uint8, addr *AddrSpec) error {\n\t// Format the address\n\tvar addrType uint8\n\tvar addrBody []byte\n\tvar addrPort uint16\n\tswitch {\n\tcase addr == nil:\n\t\taddrType = ipv4Address\n\t\taddrBody = []byte{0, 0, 0, 0}\n\t\taddrPort = 0\n\n\tcase addr.FQDN != \"\":\n\t\taddrType = fqdnAddress\n\t\taddrBody = append([]byte{byte(len(addr.FQDN))}, addr.FQDN...)\n\t\taddrPort = uint16(addr.Port)\n\n\tcase addr.IP.To4() != nil:\n\t\taddrType = ipv4Address\n\t\taddrBody = []byte(addr.IP.To4())\n\t\taddrPort = uint16(addr.Port)\n\n\tcase addr.IP.To16() != nil:\n\t\taddrType = ipv6Address\n\t\taddrBody = []byte(addr.IP.To16())\n\t\taddrPort = uint16(addr.Port)\n\n\tdefault:\n\t\treturn fmt.Errorf(\"Failed to format address: %v\", addr)\n\t}\n\n\t// Format the message\n\tmsg := make([]byte, 6+len(addrBody))\n\tmsg[0] = socks5Version\n\tmsg[1] = resp\n\tmsg[2] = 0 // Reserved\n\tmsg[3] = addrType\n\tcopy(msg[4:], addrBody)\n\tmsg[4+len(addrBody)] = byte(addrPort >> 8)\n\tmsg[4+len(addrBody)+1] = byte(addrPort & 0xff)\n\n\t// Send the message\n\t_, err := w.Write(msg)\n\treturn err\n}", "func TestServerReturnBadCode(t *testing.T) {\n\ttestServer := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tw.Write([]byte(`{}`))\n\t}))\n\t_, err := sendMessage(testServer.URL, \"[email protected]\", \"test\")\n\tif err == nil {\n\t\tt.Fail()\n\t}\n}", "func (this *SmtpWriter) SendResponse(response string) error {\n\tvar err error\n\n\tif err = this.Connection.SetWriteDeadline(time.Now().Add(time.Second * 2)); err != nil {\n\t\tlog.Printf(\"Error setting write deadline: %s\", err.Error())\n\t}\n\n\t_, err = this.Connection.Write([]byte(string(response + smtpconstants.SMTP_CRLF)))\n\treturn err\n}", "func (s *Sender) SendResponse(connID string, id string, result interface{}, err *ResError) error {\n\treturn s.sendMsg(connID, Response{ID: id, Result: result, Error: err})\n}", "func SendReply(w http.ResponseWriter, rc int, s string) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tm := UResp{Status: s, ReplyCode: rc, Timestamp: time.Now().Format(time.RFC822)}\n\tstr, err := json.Marshal(m)\n\tif nil != err {\n\t\tfmt.Fprintf(w, \"{\\n\\\"Status\\\": \\\"%s\\\"\\n\\\"Timestamp:\\\": \\\"%s\\\"\\n}\\n\",\n\t\t\t\"encoding error\", time.Now().Format(time.RFC822))\n\t} else {\n\t\tfmt.Fprintf(w, string(str))\n\t}\n}", "func (socket *MockSocket) SendCommand(command socket.Commander) chan *socket.Response {\n\treturn command.Response()\n}", "func respond(code int, msg string) (events.APIGatewayProxyResponse, error) {\n\treturn events.APIGatewayProxyResponse{\n\t\tBody: msg,\n\t\tStatusCode: code,\n\t}, nil\n}", "func sendAndExit(response *plugins.Response) {\n\tresponseBytes, _ := proto.Marshal(response)\n\tos.Stdout.Write(responseBytes)\n\tos.Exit(0)\n}", "func sendMessageToDaemon(message string) string {\n var response string\n var err error\n\n if response, err = service.Send(settings.Socket, message); err != nil {\n StdErr.Write([]byte(err.Error()))\n os.Exit(ERR_GENERAL)\n }\n\n return response\n}", "func SendMessage(message string) {\n\tif message == \"\" {\n\t\tfmt.Println(\"Message cannot be empty\")\n\t\tos.Exit(1)\n\t}\n\n\ttype Status struct {\n\t\tSuccess bool\n\t\tMessage interface{}\n\t}\n\n\ttype Result struct {\n\t\tNewMessageID string `json:\"new_message_id\"`\n\t}\n\ttype Response struct {\n\t\tStatus Status `json:\"status\"`\n\t\tResult interface{}\n\t}\n\n\tbaseURL, err := url.Parse(apiURL)\n\n\tif err != nil {\n\t\tfmt.Println(\"Malformed URL: \", err.Error())\n\t\treturn\n\t}\n\n\troomID := viper.GetString(\"roomID\")\n\tparams := url.Values{}\n\tparams.Add(\"room_id\", roomID)\n\n\tbaseURL.RawQuery = params.Encode()\n\n\tvar replacedString = strings.ReplaceAll(message, \"\\\\n\", \"\\n\")\n\t// re := regexp.MustCompile(`\\r?\\n`)\n\t// replacedString := re.ReplaceAllString(message, `\\n`)\n\t// fmt.Printf(\"fd;fhkhfkhf %v\", replacedString)\n\tjsonValues, err := json.Marshal(map[string]string{\n\t\t\"text\": replacedString,\n\t\t\"_t\": viper.GetString(\"token\"),\n\t})\n\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\n\tformData := url.Values{\n\t\t\"pdata\": {string(jsonValues)},\n\t}\n\n\treq, err := http.NewRequest(\n\t\thttp.MethodPost,\n\t\tbaseURL.String(),\n\t\tstrings.NewReader(formData.Encode()),\n\t)\n\tif err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\treq.Header.Add(\"Content-Type\", \"application/x-www-form-urlencoded;charset=UTF-8\")\n\treq.Header.Add(\"cookie\", fmt.Sprintf(\"%v=%v\", cookieName, viper.GetString(\"cookie\")))\n\n\tresponse, err := http.DefaultClient.Do(req)\n\n\tif err != nil {\n\t\tlog.Fatalf(\"Error making send-message request, %v\\n\", err)\n\t}\n\n\tdefer response.Body.Close()\n\tbody, err := ioutil.ReadAll(response.Body)\n\n\tif err != nil {\n\t\tfmt.Printf(\"Error parsing response %v\\n\", err)\n\t\treturn\n\t}\n\n\tvar parsedBody Response\n\terr = json.Unmarshal(body, &parsedBody)\n\tif err != nil {\n\t\tlog.Fatalln(\"Error parsing response\", err)\n\t}\n\n\tif parsedBody.Status.Success != true {\n\t\tvar errMsg string\n\t\tswitch v := parsedBody.Status.Message.(type) {\n\t\tcase string:\n\t\t\terrMsg = v\n\t\tcase []interface{}:\n\t\t\t// for _, val := range v {\n\t\t\t// \tfmt.Println(val)\n\t\t\t// }\n\t\t\terrMsg = v[0].(string)\n\t\t}\n\t\tfmt.Printf(\"❌ Send report failed: %v\\n\", errMsg)\n\n\t} else {\n\t\tfmt.Println(\"✅ Send report successfully\")\n\t}\n\t// fmt.Printf(\"%s\\n\", string(body))\n\n}", "func reply(c *sknet.Context, r interface{}) {\n\tc.Set(\"response\", r)\n}", "func (r *Runner) SendMessageIgnoringResponse(channel model.Snowflake, message string) {\n\tr.T.Helper()\n\n\tsendMessage(r.DiscordSession, r.Handler, channel, message)\n\tr.DiscordMessagesCount++\n\tr.AssertState()\n}", "func (p *Postback) Respond(value string) {\n\tvar (\n\t\tparams map[string]string\n\t\tclient *http.Client = &http.Client{}\n\t\tres *http.Response\n\t\tbody []byte\n\t\treq *http.Request\n\t\terr error\n\t)\n\n\t// convert the json string into string map\n\tif err = json.Unmarshal([]byte(value), &params); err != nil {\n\t\tlog.Print(err)\n\t\treturn\n\t}\n\n\t// create a new request using the endpoint method/url and data params\n\tif req, err = http.NewRequest(p.Method, p.Fill(params), nil); err != nil {\n\t\tlog.Print(err)\n\t\treturn\n\t}\n\n\tdelivered := time.Now()\n\n\t// execute the http request\n\tif res, err = client.Do(req); err != nil {\n\t\tlog.Print(err)\n\t\treturn\n\t}\n\n\t// read the body\n\tif body, err = ioutil.ReadAll(res.Body); err != nil {\n\t\tlog.Print(err)\n\t\treturn\n\t}\n\n\tlog.Printf(\n\t\t\"received: %s\\n\\tdelivery time: \\t%v\\n\\tresponse code: \\t%s\\n\\tresponse time: \\t%v\\n\\tresponse body: \\t'%s'\\n\",\n\t\treq.URL.String(),\n\t\tdelivered,\n\t\tres.Status,\n\t\ttime.Now(),\n\t\tstring(body),\n\t)\n}", "func sendMessage(daprURL string) *http.Response {\n\tbody := fmt.Sprintf(format, time.Now().String())\n\tresp, err := http.Post(daprURL, contentType, strings.NewReader(body))\n\tif err != nil {\n\t\tfmt.Println(\"failed to send request to dapr -\", err)\n\t\treturn nil\n\t}\n\tif resp == nil {\n\t\tfmt.Println(\"nil response\")\n\t\treturn nil\n\t}\n\tfmt.Println(\"sent message\", body)\n\treturn resp\n}", "func sendResponse(rw http.ResponseWriter, response responseJSON.Response) {\n\tfmt.Fprint(rw, response)\n}", "func callAndVerify(msg string, client pb.GreeterClient, shouldFail bool) error {\n\tctx, cancel := context.WithTimeout(context.Background(), time.Second)\n\tdefer cancel()\n\t_, err := client.SayHello(ctx, &pb.HelloRequest{Name: msg})\n\tif want, got := shouldFail == true, err != nil; got != want {\n\t\treturn fmt.Errorf(\"want and got mismatch, want shouldFail=%v, got fail=%v, rpc error: %v\", want, got, err)\n\t}\n\treturn nil\n}", "func sendResponse(w http.ResponseWriter, statusCode int, message string) {\n\tw.WriteHeader(statusCode)\n\tjson.NewEncoder(w).Encode(appResponse{Message: message})\n}", "func sendResponse(w http.ResponseWriter, code int, message string, data interface{}) {\n\t// Standard marshalled envelope for success.\n\ta := Response{\n\t\tData: data,\n\t\tMessage: message,\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(code)\n\terr := json.NewEncoder(w).Encode(a)\n\tif err != nil {\n\t\tsendResponse(w, http.StatusInternalServerError, \"Something unexpected happened\", nil)\n\t\treturn\n\t}\n}", "func (y *YeeLight) sendCommand(cmd *command) (*Answer, error) {\n\tif y.tcpSocket == nil {\n\t\ty.releaseAnswerChan(cmd.ID, nil)\n\t\treturn nil, errors.WithStack(ErrConnNotInitialized)\n\t}\n\ty.idMutex.RLock()\n\trespChan, ok := y.pendingCmds[cmd.ID]\n\ty.idMutex.RUnlock()\n\tif !ok {\n\t\treturn nil, errors.WithStack(ErrFailedCmd)\n\t}\n\ty.tcpSocket.Write(cmd.json())\n\tselect {\n\tcase a := <-respChan:\n\t\treturn &a, nil\n\tcase <-time.After(commandTimeout):\n\t\ty.releaseAnswerChan(cmd.ID, nil)\n\t\treturn nil, errors.Wrapf(ErrTimedOut, \"failed command %v\", cmd)\n\t}\n}", "func (q *query) sendResponse(data []byte, err error) {\n\tq.response <- queryResponse{data: data, err: err}\n}", "func (c *client) send(payload string) http.Response {\n\tbuf := new(bytes.Buffer)\n\tif _, err := buf.WriteString(payload); err != nil {\n\t\tpanic(err)\n\t}\n\trequest, err := http.NewRequest(\"POST\", fmt.Sprintf(\"http://%s/packets\", c.adapter), buf)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\tresp, err := c.Do(request)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn *resp\n}", "func HomeTest(w http.ResponseWriter, r *http.Request) {\n\n fmt.Println(\"Home test request from: \" + r.RemoteAddr)\n\n // Set response headers.\n w.Header().Set(\"Content-Type\", \"application/json\")\n\n // Create the json response.\n var resp response\n resp.Message = \"Hello!\"\n resp.Error = 0\n\n // Send response\n err := json.NewEncoder(w).Encode(resp)\n fmt.Println(err)\n}", "func (p PlatformMessage) ExpectsResponse() bool {\n\treturn p.ResponseHandle != 0\n}", "func response(code int, text string, next telnet.State) *smtpResponse {\n\n\treturn &smtpResponse{\n\t\ttelnet.NewLineResponse(text, next),\n\t\tcode,\n\t}\n}", "func sendReply(ch *api.Channel, reply *api.VppReply) {\n\tselect {\n\tcase ch.ReplyChan <- reply:\n\t\t// reply sent successfully\n\tcase <-time.After(time.Millisecond * 100):\n\t\t// receiver still not ready\n\t\tlog.WithFields(logger.Fields{\n\t\t\t\"channel\": ch,\n\t\t\t\"msg_id\": reply.MessageID,\n\t\t}).Warn(\"Unable to send the reply, reciever end not ready.\")\n\t}\n}", "func (m *MockLibvirt) reply(buf []byte) []byte {\n\tatomic.AddUint32(&m.serial, 1)\n\tbinary.BigEndian.PutUint32(buf[20:24], m.serial)\n\n\treturn buf\n}", "func (h *middlewareHarness) sendAccept(msgID uint64,\n\tresponseReplacement proto.Message) {\n\n\tvar replacementBytes []byte\n\tif responseReplacement != nil {\n\t\tvar err error\n\t\treplacementBytes, err = proto.Marshal(responseReplacement)\n\t\trequire.NoError(h.t, err)\n\t}\n\n\terr := h.stream.Send(&lnrpc.RPCMiddlewareResponse{\n\t\tMiddlewareMessage: &lnrpc.RPCMiddlewareResponse_Feedback{\n\t\t\tFeedback: &lnrpc.InterceptFeedback{\n\t\t\t\tReplaceResponse: len(replacementBytes) > 0,\n\t\t\t\tReplacementSerialized: replacementBytes,\n\t\t\t},\n\t\t},\n\t\tRefMsgId: msgID,\n\t})\n\trequire.NoError(h.t, err)\n}", "func sendToTServer(addr string, protocol string, msg string) error {\n\tlog.Printf(\"Sending '%s' to %s\", msg, addr)\n\tconn, err := net.Dial(protocol, addr)\n\tif err != nil {\n\t\tlog.Printf(\"Request failed. Couldn't Connect to server %s...\", addr)\n\t\treturn err\n\t}\n\tfmt.Fprint(conn, msg+\"\\n\")\n\treturn nil\n}", "func Send(tsx TsxServer, cmd string, cmdName string) string {\n\n\t// connect to this socket\n\n\tsrv := fmt.Sprintf(\"%s:%d\", tsx.Addr, tsx.Port)\n\tconn, err := net.Dial(\"tcp\", srv)\n\tif err != nil {\n\t\tfmt.Printf(\"Try tcp to server: %s\\n\", srv)\n\t\tfmt.Printf(\"error: %s\\n\", err)\n\t\tpanic(\"We have no connection !\")\n\t}\n\tdefer conn.Close()\n\n\t// send to socket\n\tfmt.Printf(\"Get status from TheSkyX: %s\\n\", cmdName)\n\tconn.Write([]byte(cmd))\n\n\t// listen for reply\n\tret := listenReply(conn)\n\n\treturn string(ret)\n}", "func TestReqRespTimeoutErr(t *testing.T) {\n\t// Connect to NATS\n\tm := NewMessenger(testConfig)\n\tdefer m.Close()\n\n\t// Use a WaitGroup to wait for the message to arrive\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\n\t// Subscribe to the source subject with the message processing function\n\ttestSubject := \"test_subject\"\n\ttestMsgContent := []byte(\"Some text to send...\")\n\tm.Response(testSubject, func(content []byte) ([]byte, error) {\n\t\tdefer wg.Done()\n\t\trequire.EqualValues(t, content, testMsgContent)\n\t\ttime.Sleep(100 * time.Millisecond)\n\t\treturn []byte(``), nil\n\t})\n\n\t// Send a message\n\t_, err := m.Request(testSubject, testMsgContent, 50*time.Millisecond)\n\tassert.NotNil(t, err)\n\tassert.Equal(t, err, errors.New(\"nats: timeout\"), \"should be equal\")\n\n\t// Wait for the message to come in\n\twg.Wait()\n}", "func SendDelayResponse(ip net.Addr, id uint8, tM time.Time) {\n\t// Build message, encode and send\n\tmessage := DelayResponseMessage{\n\t\tMessageCode: DelayResponse,\n\t\tId: id,\n\t\tTime: tM.UnixNano() / int64(time.Microsecond),\n\t}\n\n\tencoded := encode(message)\n\tsendUnicast(ip, UnicastSlavePort, encoded)\n}", "func (transport *MockOutboundTransport) Send(data []byte, destination string) (string, error) {\n\treturn transport.ExpectedResponse, transport.SendErr\n}", "func (ctx *rabbitMQContext) SendResponseTo(response CommandMessage, key string) error {\n\tmessageBytes, err := json.MarshalIndent(response, \"\", \" \")\n\tif err != nil {\n\t\treturn &Error{\"Unable to marshal response to json!\", err}\n\t}\n\n\terr = ctx.channel.Publish(commandExchange, key, true, false, amqp.Publishing{\n\t\tContentType: response.ContentType(),\n\t\tBody: messageBytes,\n\t})\n\tif err != nil {\n\t\treturn &Error{\"Failed to publish a command to \" + key + \"!\", err}\n\t}\n\n\treturn nil\n}", "func SendResponse(output string) (int){\n\n // Flag to tell output to be directed to the Pastebin intergration\n const pb_Flag bool = false \n\n if pb_Flag{\n SendtoPB(output)\n }else{\n url := \"http://127.0.0.1:8080/\" + url.PathEscape(output)\n _, err := http.Get(url)\n if err != nil {\n //log.Fatalln(err)\n return 0\n }\n }\n\n return 0\n\n}", "func SendMessage(address string, method string, request *Request, reply *Reply) error {\n\tclient, err := rpc.Dial(\"tcp\", address)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = client.Call(method, request, reply)\n\tif err != nil {\n\t\tclient.Close()\n\t\treturn err\n\t}\n\tclient.Close()\n\treturn nil\n}", "func (request *request) send() (*response, error) {\n\thttpClient, err := request.channel.socksProxy.GetHTTPClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\thttpResponse, err := httpClient.Post(\n\t\tstring(request.server),\n\t\t\"application/x-www-form-urlencoded\",\n\t\tstrings.NewReader(request.encode()),\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif httpResponse.StatusCode != http.StatusOK {\n\t\treturn nil, errp.New(\"Proxy Server did not respond with OK http status code, it is probably offline\")\n\t}\n\tdefer func() { _ = httpResponse.Body.Close() }()\n\tbody, err := io.ReadAll(httpResponse.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar serverResponse response\n\terr = json.Unmarshal(body, &serverResponse)\n\treturn &serverResponse, err\n}", "func (client *TcpBridgeClient) sendReceive(operation byte,\n statusRequire int,\n meta interface{},\n bodyLength int64,\n bodyWriterHandler func(manager *ConnectionManager, frame *Frame) error,\n ) (*Frame, error) {\n client.connManager.RequireStatus(statusRequire)\n frame := &Frame{}\n frame.SetOperation(operation)\n frame.SetMeta(meta)\n frame.SetMetaBodyLength(bodyLength)\n frame.BodyWriterHandler = bodyWriterHandler\n if err := client.connManager.Send(frame); err != nil {\n return nil, err\n }\n response, e := client.connManager.Receive()\n if e != nil {\n return nil, e\n }\n if response != nil {\n return response, nil\n } else {\n return nil, errors.New(\"receive empty response from server\")\n }\n}", "func (s *Sender) Send(msg *Message) (*Response, error) {\n\tif err := checkSender(s); err != nil {\n\t\treturn nil, err\n\t} else if err := checkMessage(msg); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Send the message for the first time.\n\tresp, err := s.SendNoRetry(msg)\n\tif err != nil {\n\t\treturn nil, err\n\t} else if resp.Failure == 0 {\n\t\t//no errors\n\t\treturn resp, nil\n\t}\n\t\n\tif msg.RegistrationIDs == nil{\n\t\tif len(resp.Results) == 1{\n\t\t\tresp.Results[0].OldRegistrationID = msg.To\n\t\t}else{\n\t\t\treturn resp, errors.New(\"Incorrect result\")\n\t\t}\t\n\t}else{\n\t\tif len(resp.Results) == len(msg.RegistrationIDs){\n\t\t\tl := len(resp.Results)\n\t\t\tfor i := 0; i < l; i++{\n\t\t\t\tresp.Results[i].OldRegistrationID = msg.RegistrationIDs[i]\n\t\t\t}\n\t\t}else{\n\t\t\treturn resp, errors.New(\"Incorrect result\")\n\t\t}\n\t}\n\n\treturn resp, nil\n}", "func (w *worker) sendResponse(ctx *URLContext, visited bool, harvested interface{}, idleDeath bool) {\n\t// Push harvested urls back to crawler, even if empty (uses the channel communication\n\t// to decrement reference count of pending URLs)\n\tif ctx == nil || !isRobotsURL(ctx.url) {\n\t\t// If a stop signal has been received, ignore the response, since the push\n\t\t// channel may be full and could block indefinitely.\n\t\tselect {\n\t\tcase <-w.stop:\n\t\t\tw.logFunc(LogInfo, \"ignoring send response, will stop.\")\n\t\t\treturn\n\t\tdefault:\n\t\t\t// Nothing, just continue...\n\t\t}\n\n\t\t// No stop signal, send the response\n\t\tres := &workerResponse{\n\t\t\tctx,\n\t\t\tvisited,\n\t\t\tharvested,\n\t\t\tw.host,\n\t\t\tidleDeath,\n\t\t}\n\t\tw.push <- res\n\t}\n}", "func (suite *TransportTestSuite) TestSendReceive() {\n\ttrans := suite.Transport\n\tinboundChan := make(chan message.Request, 1)\n\ttrans.Listen(testService, inboundChan)\n\n\tgo func() {\n\t\tselect {\n\t\tcase req := <-inboundChan:\n\t\t\tsuite.Assert().NotNil(req)\n\t\t\tsuite.Assert().Equal(\"ping\", string(req.Payload()))\n\t\t\tsuite.Assert().Equal(\"Shut up and take my money!\", req.Headers()[\"X-Fry\"])\n\t\t\tsuite.Assert().Equal(testService, req.Service())\n\t\t\tsuite.Assert().Equal(\"/foo\", req.Endpoint())\n\n\t\t\trsp := message.NewResponse()\n\t\t\trsp.SetId(req.Id())\n\t\t\trsp.SetPayload([]byte(\"pong\"))\n\t\t\tsuite.Assert().NoError(trans.Respond(req, rsp))\n\n\t\tcase <-trans.Tomb().Dying():\n\t\t}\n\t}()\n\n\treq := message.NewRequest()\n\treq.SetService(testService)\n\treq.SetEndpoint(\"/foo\")\n\treq.SetPayload([]byte(\"ping\"))\n\treq.SetId(\"1\")\n\treq.SetHeader(\"X-Fry\", \"Shut up and take my money!\")\n\trsp, err := trans.Send(req, time.Second)\n\tsuite.Assert().NoError(err)\n\tsuite.Assert().NotNil(rsp)\n\tsuite.Assert().Equal(req.Id(), rsp.Id())\n\tsuite.Assert().Equal(\"pong\", string(rsp.Payload()))\n}", "func (tv TV) send(sharpCommand string, sharpParameter string) (string, error) {\n\tcmdString := fmt.Sprintf(\"%4s%-4s\\r\", sharpCommand, sharpParameter)\n\n\tconnectString := fmt.Sprintf(\"%s:%s\", tv.IP, tv.Port)\n\tconn, err := net.DialTimeout(\"tcp\", connectString, time.Duration(100*time.Millisecond))\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"Error connecting to TV: %s\", err)\n\t}\n\tdefer conn.Close()\n\n\tfmt.Fprintf(conn, cmdString)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"Error sending command to TV: %s, err\")\n\t}\n\n\tapiResult := make([]byte, 32)\n\tbytesRead, err := conn.Read(apiResult)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"Error reading response from TV: Only read in %d bytes:\",\n\t\t\tbytesRead)\n\t} else {\n\t\tresultString := parseResult(apiResult)\n\t\tif resultString == \"ERR\" {\n\t\t\treturn resultString, errors.New(\"Error(ERR) returned by TV in response to command.\")\n\t\t}\n\t\treturn resultString, nil\n\t}\n\n\t// Can we even get here? Hmm...\n\treturn \"\", errors.New(\"BUG: Send() in utils.go fell through to the end. That's not supposed to happen.\")\n}", "func (instance *NDiscovery) sendCommand(conn *lygo_n_net.NConn, command string, params map[string]interface{}) *lygo_n_commons.Response {\n\treturn conn.Send(command, params)\n}", "func TestSendBundleResponse(t *testing.T) {\n\t// should be tx hash\n\treq_sendRawTransaction := server.NewJsonRpcRequest(1, \"eth_sendRawTransaction\", []interface{}{\"0xf8ac8201018527d064ee00830197f594269616d549d7e8eaa82dfb17028d0b212d11232a80b844a9059cbb000000000000000000000000c5daad04f42f923ed03a4e1e192e9ca9f46a14d50000000000000000000000000000000000000000000000000e92596fd629000025a013838b4bc34c2c3bf77f635cfa8d910e19092f38a8d7326077dbcc05f1f3fab1a06740cde8bdd8c27df60b5dd260f671b2f560e5387a83618a18d0793e17a17e02\"})\n\trpcResult := sendRpcAndParseResponseOrFailNowString(t, req_sendRawTransaction)\n\trequire.Equal(t, \"0xfc211edc6cfe4de65c8aa654d2bf5fec366486729b5b0867d4a7595f0bb5b6d5\", rpcResult)\n}", "func (h *facts) simpleResponse(message slack.Msg, text string) {\n\tif text == \"\" {\n\t\treturn\n\t}\n\tr := new(plugin.SlackResponse)\n\tr.Channel = message.Channel\n\tr.Options = append(r.Options, slack.MsgOptionText(text, false))\n\th.sink <- r\n}", "func (cmd *Command) Respond(response *Response) {\n\tcmd.response <- response\n}", "func (p *DiscoveryProtocol) createSendResponse(data *api.DiscoveryRequest) bool {\n\t// Get the init node ID\n\tinitPeerID, _ := peer.IDB58Decode(data.DiscoveryMsgData.InitNodeID)\n\n\tresp := &api.DiscoveryResponse{DiscoveryMsgData: NewDiscoveryMsgData(data.DiscoveryMsgData.MessageData.Id, false, p.p2pHost),\n\t\tMessage: api.DiscoveryMessage_DiscoveryRes}\n\n\tresp.DiscoveryMsgData.InitHash = data.DiscoveryMsgData.InitHash\n\t// sign the data\n\tkey := p.p2pHost.Peerstore().PrivKey(p.p2pHost.ID())\n\tresp.DiscoveryMsgData.MessageData.Sign = signProtoMsg(resp, key)\n\n\tlog.Printf(\"%s: Discovery response to: %s was sent. Message Id: %s, Message: %s\",\n\t\tp.p2pHost.ID(), initPeerID, resp.DiscoveryMsgData.MessageData.Id, resp.Message)\n\n\t// send the response\n\treturn sendMsg(p.p2pHost, initPeerID, resp, protocol.ID(discoveryResponse))\n}", "func (x *pwClientState) sendHMACReceiveOK(c net.Conn, client *PWClient) error {\n\th := sha256.New()\n\th.Write(x.salt)\n\th.Write([]byte(client.password))\n\tsum := new(big.Int).SetBytes(h.Sum([]byte{}))\n\n\tsecret := new(big.Int).Add(client.x, sum)\n\tsecret.Exp(x.serverPub, secret, client.p)\n\n\th.Reset()\n\th.Write(secret.Bytes())\n\tk := h.Sum([]byte{})\n\th = hmac.New(sha256.New, x.salt)\n\th.Write(k)\n\n\tfmt.Fprintf(c, \"hmac: %x\\n\", h.Sum([]byte{}))\n\n\tvar s string\n\tif _, err := fmt.Fscanln(c, &s); err != nil {\n\t\treturn err\n\t} else if s != \"ok\" {\n\t\treturn errors.New(\"ReceiveOK: invalid response\")\n\t}\n\treturn nil\n}", "func (flu *FlutterEngine) SendPlatformMessageResponse(\n\tresponseTo PlatformMessageResponseHandle,\n\tencodedMessage []byte,\n) Result {\n\tres := C.FlutterEngineSendPlatformMessageResponse(\n\t\tflu.Engine,\n\t\t(*C.FlutterPlatformMessageResponseHandle)(unsafe.Pointer(responseTo)),\n\t\t// TODO: who is responsible for free-ing this C alloc? And can they be\n\t\t// freed when this call returns? Or are they stil used at that time?\n\t\t(*C.uint8_t)(C.CBytes(encodedMessage)),\n\t\t(C.size_t)(len(encodedMessage)),\n\t)\n\n\treturn (Result)(res)\n}", "func TestReqRespServerErr(t *testing.T) {\n\t// Connect to NATS\n\tm := NewMessenger(testConfig)\n\tdefer m.Close()\n\n\t// Use a WaitGroup to wait for the message to arrive\n\twg := sync.WaitGroup{}\n\twg.Add(1)\n\n\t// Subscribe to the source subject with the message processing function\n\ttestSubject := \"test_subject\"\n\ttestMsgContent := []byte(\"Some text to send...\")\n\ttestRespErr := errors.New(\"Server error\")\n\tm.Response(testSubject, func(content []byte) ([]byte, error) {\n\t\tdefer wg.Done()\n\t\trequire.EqualValues(t, content, testMsgContent)\n\t\treturn nil, testRespErr\n\t})\n\n\t// Send a message\n\tresp, err := m.Request(testSubject, testMsgContent, 50*time.Millisecond)\n\tassert.Nil(t, err)\n\trequire.EqualValues(t, resp, testRespErr.Error())\n\n\t// Wait for the message to come in\n\twg.Wait()\n}", "func makeRemoteCall(remoteNode *NodeAddr, method string, req interface{}, rsp interface{}) error {\n\t// Dial the server if we don't already have a connection to it\n\tremoteNodeAddrStr := remoteNode.Addr\n\tvar err error\n\tclient, ok := connMap[remoteNodeAddrStr]\n\tif !ok {\n\t\tclient, err = rpc.Dial(\"tcp\", remoteNode.Addr)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tconnMap[remoteNodeAddrStr] = client\n\t}\n\n\t// Make the request\n\tuniqueMethodName := fmt.Sprintf(\"%v.%v\", remoteNodeAddrStr, method)\n\t// fmt.Println(uniqueMethodName)\n\terr = client.Call(uniqueMethodName, req, rsp)\n\tif err != nil {\n\t\tclient.Close()\n\t\tdelete(connMap, remoteNodeAddrStr)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func SendOKReply(s *StatusReq) {\n\tSendReply(s.w, RespOK, \"OK\")\n}", "func SendEphemeralCommandResponse(message string) (*model.CommandResponse, *model.AppError) {\n\treturn &model.CommandResponse{\n\t\tUsername: config.BotDisplayName,\n\t\tIconURL: config.BotIconURL,\n\t\tType: model.COMMAND_RESPONSE_TYPE_EPHEMERAL,\n\t\tText: message,\n\t}, nil\n}", "func waitForGuestbookResponse(ctx context.Context, c clientset.Interface, cmd, arg, expectedResponse string, timeout time.Duration, ns string) bool {\n\tfor start := time.Now(); time.Since(start) < timeout && ctx.Err() == nil; time.Sleep(5 * time.Second) {\n\t\tres, err := makeRequestToGuestbook(ctx, c, cmd, arg, ns)\n\t\tif err == nil && res == expectedResponse {\n\t\t\treturn true\n\t\t}\n\t\tframework.Logf(\"Failed to get response from guestbook. err: %v, response: %s\", err, res)\n\t}\n\treturn false\n}", "func sendMessage(w http.ResponseWriter, r *http.Request) {\n\t// Store the message\n\tvar message Message\n\t_ = json.NewDecoder(r.Body).Decode(&message)\n\tmessage.Username = getRandUsername()\n\tstoreMessage(message)\n\n\t// Broadcast the message\n\tbroadcast <- message\n\n\tvar msgs []Message\n\tmsgs = append(msgs, message)\n\tgetMessageResponse(w, msgs, nil)\n}", "func sendVerack(address string) {\n\tcmd := commandToBytes(\"verack\")\n\tSendCmd(address, cmd)\n}", "func HandleRemoteOutput(message models.Message, bot *models.Bot) {\n\tc := &Client{\n\t\tCredentials: bot.GoogleChatCredentials,\n\t\tProjectID: bot.GoogleChatProjectID,\n\t\tSubscriptionID: bot.GoogleChatSubscriptionID,\n\t}\n\n\t// Send messages to Google Chat\n\tgo c.Send(message, bot)\n}", "func (s *Server) unicastResponse(resp *dns.Msg, from net.Addr) error {\n\tbuf, err := resp.Pack()\n\tif err != nil {\n\t\treturn err\n\t}\n\taddr := from.(*net.UDPAddr)\n\tif addr.IP.To4() != nil {\n\t\t_, err = s.ipv4conn.WriteToUDP(buf, addr)\n\t\treturn err\n\t} else {\n\t\t_, err = s.ipv6conn.WriteToUDP(buf, addr)\n\t\treturn err\n\t}\n}", "func (m *mockStream) PullSend() (*Response, error) {\n\tselect {\n\tcase r := <-m.sendChan:\n\t\treturn r, nil\n\tcase <-time.After(timeOut):\n\t\treturn nil, errors.New(\"Timeout on PullSend\")\n\t}\n}", "func testSendCommand(t *testing.T, addr string, args []string, status int,\n\tout, err string, outbuf, errbuf *bytes.Buffer) {\n\tioCmp(t, \"sendCommand\", sendCommand(addr, args...), status, out, err,\n\t\toutbuf, errbuf)\n}", "func (h Response) Tell(message string) {\n\th.emit(\":tell\", strings.Replace(strings.Replace(message, \"\\n\", \" \", -1), \"\\t\", \"\", -1))\n}", "func (rf *Raft) sendRequestVote(server int, args *RequestVoteArgs, reply *RequestVoteReply) bool {\n ok := rf.peers[server].Call(\"Raft.RequestVote\", args, reply)\n return ok\n}", "func (mb *client) send(request *ProtocolDataUnit) (response *ProtocolDataUnit, err error) {\n\taduRequest, err := mb.packager.Encode(request)\n\tif err != nil {\n\t\treturn\n\t}\n\taduResponse, err := mb.transporter.Send(aduRequest)\n\tif err != nil {\n\t\treturn\n\t}\n\tif err = mb.packager.Verify(aduRequest, aduResponse); err != nil {\n\t\treturn\n\t}\n\tresponse, err = mb.packager.Decode(aduResponse)\n\tif err != nil {\n\t\treturn\n\t}\n\t// Check correct function code returned (exception)\n\tif response.FunctionCode != request.FunctionCode {\n\t\terr = responseError(response)\n\t\treturn\n\t}\n\tif response.Data == nil || len(response.Data) == 0 {\n\t\t// Empty response\n\t\terr = fmt.Errorf(\"modbus: response data is empty\")\n\t\treturn\n\t}\n\treturn\n}", "func sendReq(req *http.Request, result interface{}) (resp *http.Response, err error) {\n\n\tswitch {\n\t// TODO: This wil dump the authorization token. Which it probably shouldn't.\n\tcase debug():\n\t\treqDump, dumpErr := httputil.DumpRequestOut(req, true)\n\t\treqStr := string(reqDump)\n\t\tif dumpErr != nil {\n\t\t\tfmt.Printf(\"Error dumping request (display as generic object): %v\\n\", dumpErr)\n\t\t\treqStr = fmt.Sprintf(\"%v\", req)\n\t\t}\n\t\tfmt.Printf(\"%s %s\\n\", t.Title(\"Request\"), t.Text(reqStr))\n\t\tfmt.Println()\n\tcase verbose():\n\t\tfmt.Printf(\"%s %s\\n\", t.Title(\"Request:\"), t.Text(\"%s %s\", req.Method, req.URL))\n\t\tfmt.Println()\n\t}\n\n\tresp, err = hubClient.Do(req)\n\tif err == nil {\n\n\t\tif debug() {\n\t\t\trespDump, dumpErr := httputil.DumpResponse(resp, true)\n\t\t\trespStr := string(respDump)\n\t\t\tif dumpErr != nil {\n\t\t\t\tfmt.Printf(\"Error dumping response (display as generic object): %v\\n\", dumpErr)\n\t\t\t\trespStr = fmt.Sprintf(\"%v\", resp)\n\t\t\t}\n\t\t\tfmt.Printf(\"%s\\n%s\\n\", t.Title(\"Respose:\"), t.Text(respStr))\n\t\t\tfmt.Println()\n\t\t}\n\n\t\t// Do this after the Dump, the dump reads out the response for reprting and\n\t\t// replaces the reader with anotherone that has the data.\n\t\terr = checkReturnCode(*resp)\n\t\tif result != nil {\n\t\t\tif err == nil {\n\t\t\t\terr = unmarshal(resp, result)\n\t\t\t}\n\t\t}\n\n\t}\n\treturn resp, err\n}" ]
[ "0.64340484", "0.62959814", "0.578626", "0.5778683", "0.5730046", "0.56900716", "0.5680978", "0.56303036", "0.56154716", "0.56125027", "0.55935466", "0.5575292", "0.5554475", "0.55145407", "0.54929286", "0.549079", "0.54845786", "0.54530555", "0.5369651", "0.5367394", "0.5358621", "0.5327492", "0.5293739", "0.5291876", "0.528732", "0.52847254", "0.5269139", "0.5246922", "0.5236567", "0.52315664", "0.52266276", "0.52261764", "0.52197194", "0.5211618", "0.5195309", "0.51937413", "0.51900065", "0.51889896", "0.5182133", "0.51438415", "0.5132556", "0.51256394", "0.51177883", "0.5108474", "0.51069474", "0.51048374", "0.5099192", "0.5080489", "0.5078435", "0.50746435", "0.5071833", "0.50617844", "0.5042494", "0.5035525", "0.5023041", "0.50183016", "0.501119", "0.5003634", "0.49883598", "0.4983157", "0.4974952", "0.49745953", "0.49691668", "0.49651477", "0.49621218", "0.49620774", "0.49591568", "0.4958283", "0.49578303", "0.49528652", "0.49509856", "0.49459273", "0.49446827", "0.49414867", "0.49227703", "0.49206525", "0.49085936", "0.4899209", "0.48956394", "0.4895001", "0.48893604", "0.48888808", "0.48864666", "0.4884814", "0.48844597", "0.48766017", "0.48711", "0.48657233", "0.4865059", "0.48616552", "0.48577744", "0.4856373", "0.48524833", "0.48479098", "0.48477894", "0.4844552", "0.48415554", "0.48199922", "0.48182285", "0.48149052" ]
0.6446415
0
callRemoteMethod calls the method remotely in a generic way. Produces []vom.RawBytes at the end for the invoker to return.
func (fs fakeService) callRemoteMethod(ctx *context.T, method string, mi mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, argptrs []interface{}) ([]*vom.RawBytes, error) { // We need to parse the signature result to get the method relevant info out. found := false var ordinal uint32 for ord, mm := range mi.Methods { if *mm.DeclData.ShortName == method { ordinal = ord found = true break } } if !found { return nil, fmt.Errorf("callRemoteMethod: method %s does not exist", method) } mm := mi.Methods[ordinal] // A void function must have request id of 0, whereas one with response params // should have a unique request id. header := bindings.MessageHeader{ Type: ordinal, Flags: bindings.MessageExpectsResponseFlag, RequestId: fs.ids.Count(), } // Now produce the *bindings.Message that we will send to the other side. inType, err := transcoder.MojomStructToVDLType(mm.Parameters, desc) if err != nil { return nil, err } message, err := encodeMessageFromVom(header, argptrs, inType) if err != nil { return nil, err } // Otherwise, make a generic call with the message. outMessage, err := fs.callRemoteWithResponse(ctx, message) if err != nil { return nil, err } // Decode the *vom.RawBytes from the mojom bytes and mojom type. outType, err := transcoder.MojomStructToVDLType(*mm.ResponseParams, desc) if err != nil { return nil, err } target := util.StructSplitTarget() if err := transcoder.FromMojo(target, outMessage.Payload, outType); err != nil { return nil, fmt.Errorf("transcoder.FromMojo failed: %v", err) } return target.Fields(), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *ObjectRef) CallNonvirtualMethod(env *Env, className string, methodName string, dest interface{}, args ...interface{}) error {\n\trType, rClassName, err := typeOfReturnValue(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tretVal, err := o.genericCallNonvirtualMethod(env, className, methodName, rType, rClassName, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif v, ok := dest.(ToGoConverter); ok && (rType&Object == Object || rType&Array == Array) {\n\t\treturn v.ConvertToGo(retVal.(*ObjectRef))\n\t} else if rType.isArray() && rType != Object|Array {\n\t\t// If return type is an array of convertable java to go types, do the conversion\n\t\tconverted, err := env.toGoArray(retVal.(*ObjectRef).jobject, rType)\n\t\tdeleteLocalRef(env.jniEnv, retVal.(*ObjectRef).jobject)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn assignDest(converted, dest)\n\t} else {\n\t\treturn assignDest(retVal, dest)\n\t}\n}", "func (ri reflectInvoker) Invoke(ctx *context.T, call StreamServerCall, method string, argptrs []interface{}) ([]interface{}, error) {\n\tinfo, ok := ri.methods[method]\n\tif !ok {\n\t\treturn nil, verror.ErrUnknownMethod.Errorf(ctx, \"method does not exist: %v\", method)\n\t}\n\t// Create the reflect.Value args for the invocation. The receiver of the\n\t// method is always first, followed by the required ctx and call args.\n\trvArgs := make([]reflect.Value, len(argptrs)+3)\n\trvArgs[0] = ri.rcvr\n\trvArgs[1] = reflect.ValueOf(ctx)\n\tif info.rtStreamCall == nil {\n\t\t// There isn't a typesafe streaming call, just use the call.\n\t\trvArgs[2] = reflect.ValueOf(call)\n\t} else {\n\t\t// There is a typesafe streaming call with type rtStreamCall. We perform\n\t\t// the equivalent of the following:\n\t\t// ctx := new(rtStreamCall)\n\t\t// ctx.Init(call)\n\t\trvStreamCall := reflect.New(info.rtStreamCall)\n\t\tinfo.rvStreamCallInit.Call([]reflect.Value{rvStreamCall, reflect.ValueOf(call)})\n\t\trvArgs[2] = rvStreamCall\n\t}\n\t// Positional user args follow.\n\tfor ix, argptr := range argptrs {\n\t\trvArgs[ix+3] = reflect.ValueOf(argptr).Elem()\n\t}\n\t// Invoke the method, and handle the final error out-arg.\n\trvResults := info.rvFunc.Call(rvArgs)\n\tif len(rvResults) == 0 {\n\t\treturn nil, abortedf(errNoFinalErrorOutArg)\n\t}\n\trvErr := rvResults[len(rvResults)-1]\n\trvResults = rvResults[:len(rvResults)-1]\n\tif rvErr.Type() != rtError {\n\t\treturn nil, abortedf(errNoFinalErrorOutArg)\n\t}\n\tif iErr := rvErr.Interface(); iErr != nil {\n\t\treturn nil, iErr.(error)\n\t}\n\t// Convert the rest of the results into interface{}.\n\tif len(rvResults) == 0 {\n\t\treturn nil, nil\n\t}\n\tresults := make([]interface{}, len(rvResults))\n\tfor ix, r := range rvResults {\n\t\tresults[ix] = r.Interface()\n\t}\n\treturn results, nil\n}", "func (_Univ2 *Univ2Raw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Univ2.Contract.Univ2Caller.contract.Call(opts, result, method, params...)\n}", "func makeRemoteCall(remoteNode *NodeAddr, method string, req interface{}, rsp interface{}) error {\n\t// Dial the server if we don't already have a connection to it\n\tremoteNodeAddrStr := remoteNode.Addr\n\tvar err error\n\tclient, ok := connMap[remoteNodeAddrStr]\n\tif !ok {\n\t\tclient, err = rpc.Dial(\"tcp\", remoteNode.Addr)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tconnMap[remoteNodeAddrStr] = client\n\t}\n\n\t// Make the request\n\tuniqueMethodName := fmt.Sprintf(\"%v.%v\", remoteNodeAddrStr, method)\n\t// fmt.Println(uniqueMethodName)\n\terr = client.Call(uniqueMethodName, req, rsp)\n\tif err != nil {\n\t\tclient.Close()\n\t\tdelete(connMap, remoteNodeAddrStr)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (o *ObjectRef) CallMethod(env *Env, methodName string, dest interface{}, args ...interface{}) error {\n\trType, rClassName, err := typeOfReturnValue(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tretVal, err := o.genericCallMethod(env, methodName, rType, rClassName, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif v, ok := dest.(ToGoConverter); ok && (rType&Object == Object || rType&Array == Array) {\n\t\treturn v.ConvertToGo(retVal.(*ObjectRef))\n\t} else if rType.isArray() && rType != Object|Array {\n\t\t// If return type is an array of convertable java to go types, do the conversion\n\t\tconverted, err := env.toGoArray(retVal.(*ObjectRef).jobject, rType)\n\t\tdeleteLocalRef(env.jniEnv, retVal.(*ObjectRef).jobject)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn assignDest(converted, dest)\n\t} else {\n\t\treturn assignDest(retVal, dest)\n\t}\n\n}", "func makeRemoteCall(address string, structtype string, method string, req interface{}, rsp interface{}) error {\n\t// Dial the server\n\tclient, err := rpc.Dial(\"tcp\", address)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Make the request\n\tfqm := fmt.Sprintf(\"%v.%v\", structtype, method)\n\terr = client.Call(fqm, req, rsp)\n\n\tclient.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (_BytesLib *BytesLibRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _BytesLib.Contract.BytesLibCaller.contract.Call(opts, result, method, params...)\n}", "func (_BytesLib *BytesLibRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BytesLib.Contract.BytesLibCaller.contract.Call(opts, result, method, params...)\n}", "func (_BytesLib *BytesLibCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _BytesLib.Contract.contract.Call(opts, result, method, params...)\n}", "func callRemoteService(ip string) ipInfoResult {\n\tvar obj ipInfoResult\n\n\tapi := \"/json\"\n\tif 0 == len(ip) {\n\t\tapi = \"json\"\n\t}\n\turl := \"https://ipinfo.io/\" + ip + api\n\tresp, err := http.Get(url)\n\tif err != nil {\n\t\tfmt.Println(\"error: \", err)\n\t\treturn obj\n\t}\n\tdefer resp.Body.Close()\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(\"error: \", err)\n\t\treturn obj\n\t}\n\n\tif strings.Contains(string(body), \"Rate limit exceeded\") {\n\t\tfmt.Println(\"\\nError for:\", url)\n\t\tfmt.Println(string(body))\n\t\tos.Exit(1)\n\t}\n\n\tjson.Unmarshal(body, &obj)\n\treturn obj\n}", "func (_Univ2 *Univ2CallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Univ2.Contract.contract.Call(opts, result, method, params...)\n}", "func (c *YHClient) MCallRaw(method string, params ...interface{}) ([]byte, error) {\n\treturn c.mcall(method, params)\n}", "func (fs fakeService) Invoke(ctx *context.T, call rpc.StreamServerCall, method string, argptrs []interface{}) (results []interface{}, _ error) {\n\t// fs.suffix consists of the mojo url and the application/interface name.\n\t// The last part should be the name; everything else is the url.\n\tparts := strings.Split(fs.suffix, \"/\")\n\tmojourl := strings.Join(parts[:len(parts)-1], \"/\") // e.g., mojo:go_remote_echo_server. May be defined in a BUILD.gn file.\n\tmojoname := parts[len(parts)-1] // e.g., mojo::examples::RemoteEcho. Defined from the interface + module.\n\n\t// Create the generic message pipe. r is a bindings.InterfaceRequest, and\n\t// p is a bindings.InterfacePointer.\n\tr, p := bindings.CreateMessagePipeForMojoInterface()\n\tv := v23ServiceRequest{\n\t\trequest: r,\n\t\tname: mojoname,\n\t} // v is an application.ServiceRequest with mojoname\n\n\t// Connect to the mojourl.\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&v)\n\n\t// Then assign a new router the FakeService.\n\t// This will never conflict because each FakeService is only invoked once.\n\tfs.router = bindings.NewRouter(p.PassMessagePipe(), bindings.GetAsyncWaiter())\n\tdefer fs.Close_Proxy()\n\n\tctx.Infof(\"Fake Service Invoke (Remote Signature: %q -- %q)\", mojourl, mojoname)\n\n\t// Vanadium relies on type information, so we will retrieve that first.\n\tmojomInterface, desc, err := fs.callRemoteSignature(mojourl, mojoname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Signature %v\", mojomInterface)\n\tctx.Infof(\"Fake Service Invoke (Remote Method: %v)\", method)\n\n\t// With the type information, we can make the method call to the remote interface.\n\tmethodResults, err := fs.callRemoteMethod(ctx, method, mojomInterface, desc, argptrs)\n\tif err != nil {\n\t\tctx.Errorf(\"Method called failed: %v\", err)\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Results %v\", methodResults)\n\n\t// Convert methodResult to results.\n\tresults = make([]interface{}, len(methodResults))\n\tfor i := range methodResults {\n\t\tresults[i] = &methodResults[i]\n\t}\n\treturn results, nil\n}", "func (_IUniswapV2Pair *IUniswapV2PairRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Pair.Contract.IUniswapV2PairCaller.contract.Call(opts, result, method, params...)\r\n}", "func (_PBridge *PBridgeRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _PBridge.Contract.PBridgeCaller.contract.Call(opts, result, method, params...)\n}", "func (_UniswapV2Library *UniswapV2LibraryRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _UniswapV2Library.Contract.UniswapV2LibraryCaller.contract.Call(opts, result, method, params...)\r\n}", "func (_BytesLib *BytesLibCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BytesLib.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Cakevault *CakevaultRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Cakevault.Contract.CakevaultCaller.contract.Call(opts, result, method, params...)\n}", "func (_IUniswapV2Pair *IUniswapV2PairCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Pair.Contract.contract.Call(opts, result, method, params...)\r\n}", "func (a *api) CallLocal(ctx context.Context, in *internalv1pb.InternalInvokeRequest) (*internalv1pb.InternalInvokeResponse, error) {\n\tif a.appChannel == nil {\n\t\treturn nil, status.Error(codes.Internal, messages.ErrChannelNotFound)\n\t}\n\n\treq, err := invokev1.InternalInvokeRequest(in)\n\tif err != nil {\n\t\treturn nil, status.Errorf(codes.InvalidArgument, messages.ErrInternalInvokeRequest, err.Error())\n\t}\n\tdefer req.Close()\n\n\terr = a.callLocalValidateACL(ctx, req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar callerAppID string\n\tcallerIDHeader, ok := req.Metadata()[invokev1.CallerIDHeader]\n\tif ok && len(callerIDHeader.Values) > 0 {\n\t\tcallerAppID = callerIDHeader.Values[0]\n\t} else {\n\t\tcallerAppID = \"unknown\"\n\t}\n\n\tdiag.DefaultMonitoring.ServiceInvocationRequestReceived(callerAppID, req.Message().Method)\n\n\tvar statusCode int32\n\tdefer func() {\n\t\tdiag.DefaultMonitoring.ServiceInvocationResponseSent(callerAppID, req.Message().Method, statusCode)\n\t}()\n\n\t// stausCode will be read by the deferred method above\n\tres, err := a.appChannel.InvokeMethod(ctx, req)\n\tif err != nil {\n\t\tstatusCode = int32(codes.Internal)\n\t\treturn nil, status.Errorf(codes.Internal, messages.ErrChannelInvoke, err)\n\t} else {\n\t\tstatusCode = res.Status().Code\n\t}\n\tdefer res.Close()\n\n\treturn res.ProtoWithData()\n}", "func (_UniswapV2Library *UniswapV2LibraryCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _UniswapV2Library.Contract.contract.Call(opts, result, method, params...)\r\n}", "func (_BREM *BREMRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BREM.Contract.BREMCaller.contract.Call(opts, result, method, params...)\n}", "func (_BREMToken *BREMTokenRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BREMToken.Contract.BREMTokenCaller.contract.Call(opts, result, method, params...)\n}", "func (_Outbox *OutboxRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Outbox.Contract.OutboxCaller.contract.Call(opts, result, method, params...)\n}", "func (_Cakevault *CakevaultCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Cakevault.Contract.contract.Call(opts, result, method, params...)\n}", "func (_BREMToken *BREMTokenCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BREMToken.Contract.contract.Call(opts, result, method, params...)\n}", "func (j *Env) CallStaticMethod(className string, methodName string, dest interface{}, args ...interface{}) error {\n\trType, rClassName, err := typeOfReturnValue(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tretVal, err := j.genericCallStaticMethod(className, methodName, rType, rClassName, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif v, ok := dest.(ToGoConverter); ok && (rType&Object == Object || rType&Array == Array) {\n\t\treturn v.ConvertToGo(retVal.(*ObjectRef))\n\t} else if rType.isArray() && rType != Object|Array {\n\t\t// If return type is an array of convertable java to go types, do the conversion\n\t\tconverted, err := j.toGoArray(retVal.(*ObjectRef).jobject, rType)\n\t\tdeleteLocalRef(j.jniEnv, retVal.(*ObjectRef).jobject)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn assignDest(converted, dest)\n\t} else {\n\t\treturn assignDest(retVal, dest)\n\t}\n}", "func (wd *remoteWD) execute(method, url string, data []byte) (json.RawMessage, error) {\n\treturn executeCommand(method, url, data)\n}", "func (_BREM *BREMCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BREM.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Vault *VaultRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Vault.Contract.VaultCaller.contract.Call(opts, result, method, params...)\n}", "func (_Mevsky *MevskyRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Mevsky.Contract.MevskyCaller.contract.Call(opts, result, method, params...)\n}", "func (service *rpcServiceMap) callService(mtype *rpcMethodType, argv,\n replyv reflect.Value) error {\n\n function := mtype.method.Func\n // Invoke the method, providing a new value for the reply.\n returnValues := function.Call([]reflect.Value{service.rcvr, argv, replyv})\n // The return value for the method is an error.\n errInter := returnValues[0].Interface()\n errmsg := \"\"\n if errInter != nil {\n errmsg = errInter.(error).Error()\n return fmt.Errorf(errmsg)\n }\n return nil\n}", "func (_Vault *VaultCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Vault.Contract.contract.Call(opts, result, method, params...)\n}", "func (server *Server) callService(conn *ConnDriver, seq uint64, service *service, methodType *methodType, argv, replyv reflect.Value) {\n\tfunction := methodType.method.Func\n\n\treturnValues := function.Call([]reflect.Value{service.rcvr, argv, replyv})\n\t// The return value for the method is an error.\n\terrInter := returnValues[0].Interface()\n\n\trespHeader := NewResponseHeader()\n\trespHeader.ReplyType = ReplyTypeData\n\trespHeader.Seq = seq\n\tif errInter != nil {\n\t\tswitch errInter.(type) {\n\t\tcase *Error:\n\t\t\tserver.replyCmd(conn, seq, errInter.(*Error), CmdTypeErr)\n\t\tcase Error:\n\t\t\te := errInter.(Error)\n\t\t\tserver.replyCmd(conn, seq, &e, CmdTypeErr)\n\t\tcase error:\n\t\t\tserver.replyCmd(conn, seq, &Error{500, ErrTypeLogic, errInter.(error).Error()}, CmdTypeErr)\n\t\t}\n\t\treturn\n\t}\n\tconn.Lock()\n\terr := server.SendFrame(conn, respHeader, replyv)\n\tconn.Unlock()\n\tif err != nil && !isNetError(err) {\n\t\tlog.Fatalln(\"encoding error:\" + err.Error())\n\t}\n\treturn\n}", "func (_Smartchef *SmartchefRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Smartchef.Contract.SmartchefCaller.contract.Call(opts, result, method, params...)\n}", "func (r *rpcServerService) doCall(serviceMethod string, args []byte) ([]byte,\n error) {\n\n glog.V(3).Infof(\"rpc: doCall to %s\", serviceMethod)\n glog.V(4).Infof(\"rpc: doCall to %s with %v\", serviceMethod, args)\n\n dot := strings.LastIndex(serviceMethod, \".\")\n if dot < 0 {\n err := fmt.Errorf(\"rpc: service/method ill-formed: \" + serviceMethod)\n glog.Error(err)\n return nil, err\n }\n serviceName := serviceMethod[:dot]\n methodName := serviceMethod[dot+1:]\n // Look up the request.\n serviceInf, ok := r.serviceMap.Get(serviceName)\n if !ok || serviceInf == nil {\n err := errors.New(\"rpc: can't find service \" + serviceName)\n glog.Error(err)\n return nil, err\n }\n service, okType := serviceInf.(*rpcServiceMap)\n if !okType || service == nil {\n err := errors.New(\"rpc: unexpected type error for service \" + serviceName)\n glog.Error(err)\n return nil, err\n }\n mtype := service.method[methodName]\n if mtype == nil {\n err := errors.New(\"rpc: can't find method \" + serviceMethod)\n glog.Error(err)\n return nil, err\n }\n argv := reflect.New(mtype.argType)\n errJSON := json.Unmarshal(args, argv.Interface())\n if errJSON != nil {\n glog.Error(\"error in unmarshal: \", errJSON)\n return nil, errJSON\n }\n glog.V(4).Infof(\"rpc: json unmarshalled request is: %s -> %#v\", args, argv)\n replyv := reflect.New(mtype.replyType.Elem())\n\n glog.V(3).Infof(\"rpc: calling service %v method %v with %v\",\n service, mtype, argv)\n\n errCall := service.callService(mtype, argv.Elem(), replyv)\n if errCall != nil {\n glog.V(3).Infof(\"rpc call returned error: \", errCall)\n return nil, errCall\n }\n reply, errRep := json.Marshal(replyv.Interface())\n if errRep != nil {\n glog.Error(\"rpc reply marshall error: \", errRep)\n return nil, errRep\n }\n glog.V(3).Info(\"rpc reply: \", string(reply))\n return reply, nil\n}", "func (_TellorMesosphere *TellorMesosphereRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _TellorMesosphere.Contract.TellorMesosphereCaller.contract.Call(opts, result, method, params...)\n}", "func (etherScan *EtherScan) rpcCall(params url.Values, result interface{}) error {\n\tparams.Set(\"module\", \"proxy\")\n\n\tvar wrapped struct {\n\t\tJSONRPC string `json:\"jsonrpc\"`\n\t\tID int `json:\"id\"`\n\t\tError *struct {\n\t\t\tMessage string `json:\"message\"`\n\t\t} `json:\"error\"`\n\t\tResult *json.RawMessage `json:\"result\"`\n\t}\n\tif err := etherScan.call(params, &wrapped); err != nil {\n\t\treturn err\n\t}\n\tif wrapped.Error != nil {\n\t\treturn errp.New(wrapped.Error.Message)\n\t}\n\tif result == nil {\n\t\treturn nil\n\t}\n\tif wrapped.Result == nil {\n\t\treturn errp.New(\"expected result\")\n\t}\n\tif err := json.Unmarshal(*wrapped.Result, result); err != nil {\n\t\treturn errp.WithStack(err)\n\t}\n\treturn nil\n}", "func (_PBridge *PBridgeCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _PBridge.Contract.contract.Call(opts, result, method, params...)\n}", "func (c *Client) RPC(ctx context.Context, method string, data []byte) (RPCResult, error) {\n\tif c.isClosed() {\n\t\treturn RPCResult{}, ErrClientClosed\n\t}\n\tresCh := make(chan RPCResult, 1)\n\terrCh := make(chan error, 1)\n\tc.sendRPC(ctx, method, data, func(result RPCResult, err error) {\n\t\tresCh <- result\n\t\terrCh <- err\n\t})\n\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn RPCResult{}, ctx.Err()\n\tcase res := <-resCh:\n\t\treturn res, <-errCh\n\t}\n}", "func (_Utils *UtilsRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Utils.Contract.UtilsCaller.contract.Call(opts, result, method, params...)\n}", "func (_Utils *UtilsRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Utils.Contract.UtilsCaller.contract.Call(opts, result, method, params...)\n}", "func (c *Conn) InvokeRaw(ctx context.Context, input bin.Encoder, output bin.Decoder) error {\n\treturn c.rpcContent(ctx, input, output)\n}", "func (_Mevsky *MevskyCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Mevsky.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Verifier *VerifierRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Verifier.Contract.VerifierCaller.contract.Call(opts, result, method, params...)\n}", "func call(method string, argsArray []string) {\n\tfmt.Println(\"------ start ------ \", method)\n\t// all args\n\tvar args [][]byte\n\targs = append(args, []byte(method))\n\tfmt.Printf(\"- args=[\")\n\tfmt.Printf(\"p0=%s\", method)\n\tif argsArray != nil {\n\t\tfor i := 0; i < len(argsArray); i++ {\n\t\t\targs = append(args, []byte(argsArray[i]))\n\t\t\tfmt.Printf(\",p%d=%s\", i+1, argsArray[i])\n\t\t}\n\t}\n\tfmt.Printf(\"]\")\n\tfmt.Println(\"\")\n\t// invoke\n\tresponse := stub.MockInvoke(\"uuid\", args)\n\tfmt.Printf(\"- status=\")\n\tfmt.Println(response.GetStatus())\n\tfmt.Printf(\"- error message=\")\n\tfmt.Println(response.GetMessage())\n\tfmt.Printf(\"- payload=\")\n\tfmt.Println(string(response.GetPayload()))\n\tfmt.Println(\"------ end ------ \")\n\tfmt.Println(\"\")\n}", "func (c *remotingClient) InvokeSync(ctx context.Context, addr string, request *RemotingCommand) (*RemotingCommand, error) {\n\tconn, err := c.connect(ctx, addr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp := NewResponseFuture(ctx, request.Opaque, nil)\n\tc.responseTable.Store(resp.Opaque, resp)\n\tdefer c.responseTable.Delete(request.Opaque)\n\terr = c.sendRequest(conn, request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn resp.waitResponse()\n}", "func (_TellorMesosphere *TellorMesosphereCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _TellorMesosphere.Contract.contract.Call(opts, result, method, params...)\n}", "func (_IUniswapV2Router01 *IUniswapV2Router01CallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Router01.Contract.contract.Call(opts, result, method, params...)\r\n}", "func (_CraftingI *CraftingICallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _CraftingI.Contract.contract.Call(opts, result, method, params...)\n}", "func (xmlmc *XmlmcInstStruct) Invoke(servicename string, methodname string) (string, error) {\n\n\t//-- Add Api Tracing\n\ttracename := \"\"\n\tif xmlmc.trace != \"\" {\n\t\ttracename = \"/\" + tracename\n\t}\n\n\txmlmclocal := \"<methodCall service=\\\"\" + servicename + \"\\\" method=\\\"\" + methodname + \"\\\" trace=\\\"goApi\" + tracename + \"\\\">\"\n\tif len(xmlmc.paramsxml) == 0 {\n\t\txmlmclocal = xmlmclocal + \"</methodCall>\"\n\t} else {\n\t\txmlmclocal = xmlmclocal + \"<params>\" + xmlmc.paramsxml\n\t\txmlmclocal = xmlmclocal + \"</params>\" + \"</methodCall>\"\n\t}\n\n\tstrURL := xmlmc.server + \"/\" + servicename + \"/?method=\" + methodname\n\n\tvar xmlmcstr = []byte(xmlmclocal)\n\n\treq, err := http.NewRequest(\"POST\", strURL, bytes.NewBuffer(xmlmcstr))\n\txmlmc.count++\n\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Unable to create http request in esp_xmlmc.go\")\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"text/xmlmc\")\n\tif xmlmc.apiKey != \"\" {\n\t\treq.Header.Add(\"Authorization\", \"ESP-APIKEY \"+xmlmc.apiKey)\n\t}\n\treq.Header.Set(\"User-Agent\", xmlmc.userAgent)\n\treq.Header.Add(\"Cookie\", xmlmc.sessionID)\n\tif xmlmc.jsonresp == true {\n\t\treq.Header.Add(\"Accept\", \"text/json\")\n\t}\n\tduration := time.Second * time.Duration(xmlmc.timeout)\n\tclient := &http.Client{Transport: xmlmc.transport, Timeout: duration}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\txmlmc.statuscode = resp.StatusCode\n\n\tdefer resp.Body.Close()\n\n\t//-- Check for HTTP Response\n\tif resp.StatusCode != 200 {\n\t\terrorString := fmt.Sprintf(\"Invalid HTTP Response: %d\", resp.StatusCode)\n\t\terr = errors.New(errorString)\n\t\t//Drain the body so we can reuse the connection\n\t\tio.Copy(ioutil.Discard, resp.Body)\n\t\treturn \"\", err\n\t}\n\n\tbody, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn \"\", errors.New(\"Cant read the body of the response\")\n\t}\n\t// If we have a new EspSessionId set it\n\tSessionIds := strings.Split(resp.Header.Get(\"Set-Cookie\"), \";\")\n\tif SessionIds[0] != \"\" {\n\t\txmlmc.sessionID = SessionIds[0]\n\t}\n\n\txmlmc.paramsxml = \"\"\n\treturn string(body), nil\n}", "func (_Owned *OwnedRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Owned.Contract.OwnedCaller.contract.Call(opts, result, method, params...)\n}", "func (_Lmc *LmcRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Lmc.Contract.LmcCaller.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.ContractCaller.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.ContractCaller.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.ContractCaller.contract.Call(opts, result, method, params...)\n}", "func (_Smartchef *SmartchefCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Smartchef.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Outbox *OutboxCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Outbox.Contract.contract.Call(opts, result, method, params...)\n}", "func (_IUniswapV2Router01 *IUniswapV2Router01Raw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Router01.Contract.IUniswapV2Router01Caller.contract.Call(opts, result, method, params...)\r\n}", "func (_CrToken *CrTokenCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _CrToken.Contract.contract.Call(opts, result, method, params...)\n}", "func (_BridgeRegistry *BridgeRegistryCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _BridgeRegistry.Contract.contract.Call(opts, result, method, params...)\n}", "func (_IOrakuruCore *IOrakuruCoreRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _IOrakuruCore.Contract.IOrakuruCoreCaller.contract.Call(opts, result, method, params...)\n}", "func (_Verifier *VerifierCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Verifier.Contract.contract.Call(opts, result, method, params...)\n}", "func (_CrToken *CrTokenRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _CrToken.Contract.CrTokenCaller.contract.Call(opts, result, method, params...)\n}", "func (_DevUtils *DevUtilsRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _DevUtils.Contract.DevUtilsCaller.contract.Call(opts, result, method, params...)\n}", "func (server *Server) callRPCAndFormatReply(buf []byte, ci *clientInfo, jReq *jsonRequest) (ior *ioReply) {\n\tvar (\n\t\terr error\n\t\treturnValues []reflect.Value\n\t\ttypOfReq reflect.Type\n\t\tdummyReq interface{}\n\t)\n\n\t// Setup the reply structure with common fields\n\tior = &ioReply{}\n\trid := jReq.RequestID\n\tjReply := &jsonReply{MyUniqueID: jReq.MyUniqueID, RequestID: rid}\n\n\tma := server.svrMap[jReq.Method]\n\tif ma != nil {\n\n\t\t// Another unmarshal of buf to find the parameters specific to\n\t\t// this RPC\n\t\ttypOfReq = ma.request.Elem()\n\t\tdummyReq = reflect.New(typOfReq).Interface()\n\n\t\tsReq := svrRequest{}\n\t\tsReq.Params[0] = dummyReq\n\t\terr = json.Unmarshal(buf, &sReq)\n\t\tif err != nil {\n\t\t\tserver.logger.Fatalf(\"Unmarshal sReq: %+v err: %v\", sReq, err)\n\t\t\treturn\n\t\t}\n\t\treq := reflect.ValueOf(dummyReq)\n\t\tcid := reflect.ValueOf(ci.myUniqueID)\n\n\t\t// Create the reply structure\n\t\ttypOfReply := ma.reply.Elem()\n\t\tmyReply := reflect.New(typOfReply)\n\n\t\t// Call the method\n\t\tfunction := ma.methodPtr.Func\n\t\tt := time.Now()\n\t\tif ma.passClientID {\n\t\t\treturnValues = function.Call([]reflect.Value{server.receiver, cid, req, myReply})\n\t\t} else {\n\t\t\treturnValues = function.Call([]reflect.Value{server.receiver, req, myReply})\n\t\t}\n\t\tci.setMethodStats(jReq.Method, uint64(time.Since(t).Microseconds()))\n\n\t\t// The return value for the method is an error.\n\t\terrInter := returnValues[0].Interface()\n\t\tif errInter == nil {\n\t\t\tjReply.Result = myReply.Elem().Interface()\n\t\t} else {\n\t\t\te, ok := errInter.(error)\n\t\t\tif !ok {\n\t\t\t\tserver.logger.Fatalf(\"Call returnValues invalid cast errInter: %+v\", errInter)\n\t\t\t}\n\t\t\tjReply.ErrStr = e.Error()\n\t\t}\n\t} else {\n\t\t// TODO - figure out if this is the correct error\n\n\t\t// Method does not exist\n\t\tjReply.ErrStr = fmt.Sprintf(\"errno: %d\", unix.ENOENT)\n\t}\n\n\t// Convert response into JSON for return trip\n\tior.JResult, err = json.Marshal(jReply)\n\tif err != nil {\n\t\tserver.logger.Fatalf(\"Unable to marshal jReply: %+v err: %v\", jReply, err)\n\t}\n\n\treturn\n}", "func (_IUniswapV2Factory *IUniswapV2FactoryRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Factory.Contract.IUniswapV2FactoryCaller.contract.Call(opts, result, method, params...)\r\n}", "func (c *ClientImplementation) CallRaw(method, path, contentType string, headers http.Header, body io.Reader, v interface{}) error {\n\treq, err := c.NewRequest(method, path, contentType, headers, body)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn c.Do(req, v)\n}", "func (_MiniSafe *MiniSafeRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _MiniSafe.Contract.MiniSafeCaller.contract.Call(opts, result, method, params...)\n}", "func (_TransferHelper *TransferHelperRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _TransferHelper.Contract.TransferHelperCaller.contract.Call(opts, result, method, params...)\r\n}", "func (_BREMFactory *BREMFactoryRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _BREMFactory.Contract.BREMFactoryCaller.contract.Call(opts, result, method, params...)\n}", "func makeRemoteNodeCall(remote Node, method string, req interface{}, rsp interface{}) error {\n\tDebug.Printf(\"%v(%v)\\n\", method, req)\n\treturn makeRemoteCall(remote.Address, \"TapestryRPCServer\", method, req, rsp)\n}", "func (_IUniswapV2Factory *IUniswapV2FactoryCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\r\n\treturn _IUniswapV2Factory.Contract.contract.Call(opts, result, method, params...)\r\n}", "func (g *Channel) invokeMethodV1(ctx context.Context, req *invokev1.InvokeMethodRequest) (*invokev1.InvokeMethodResponse, error) {\n\tif g.ch != nil {\n\t\tg.ch <- struct{}{}\n\t}\n\n\t// Read the request, including the data\n\tpd, err := req.ProtoWithData()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmd := invokev1.InternalMetadataToGrpcMetadata(ctx, pd.Metadata, true)\n\n\tif g.appMetadataToken != \"\" {\n\t\tmd.Set(securityConsts.APITokenHeader, g.appMetadataToken)\n\t}\n\n\t// Prepare gRPC Metadata\n\tctx = grpcMetadata.NewOutgoingContext(context.Background(), md)\n\n\tvar header, trailer grpcMetadata.MD\n\n\topts := []grpc.CallOption{\n\t\tgrpc.Header(&header),\n\t\tgrpc.Trailer(&trailer),\n\t\tgrpc.MaxCallSendMsgSize(g.maxRequestBodySizeMB << 20),\n\t\tgrpc.MaxCallRecvMsgSize(g.maxRequestBodySizeMB << 20),\n\t}\n\n\tresp, err := g.appCallbackClient.OnInvoke(ctx, pd.Message, opts...)\n\n\tif g.ch != nil {\n\t\t<-g.ch\n\t}\n\n\tvar rsp *invokev1.InvokeMethodResponse\n\tif err != nil {\n\t\t// Convert status code\n\t\trespStatus := status.Convert(err)\n\t\t// Prepare response\n\t\trsp = invokev1.NewInvokeMethodResponse(int32(respStatus.Code()), respStatus.Message(), respStatus.Proto().Details)\n\t} else {\n\t\trsp = invokev1.NewInvokeMethodResponse(int32(codes.OK), \"\", nil)\n\t}\n\n\trsp.WithHeaders(header).\n\t\tWithTrailers(trailer).\n\t\tWithMessage(resp)\n\n\t// If the data has a type_url, set protobuf as content type\n\t// This is necessary to support the HTTP->gRPC service invocation path correctly\n\ttypeURL := resp.GetData().GetTypeUrl()\n\tif typeURL != \"\" {\n\t\trsp.WithContentType(invokev1.ProtobufContentType)\n\t}\n\n\treturn rsp, nil\n}", "func (_ProofVerify *ProofVerifyRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _ProofVerify.Contract.ProofVerifyCaller.contract.Call(opts, result, method, params...)\n}", "func (client *RPCClient) Call(severName string, serverMethod string, args interface{}, reply interface{}) error {\n\taddrs, err := serverInstance.svrmgr.GetServerHosts(severName, FlagRPCHost)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// pick a random start index for round robin\n\ttotal := len(addrs)\n\tstart := client.random.Intn(total)\n\n\tfor idx := 0; idx < total; idx++ {\n\t\taddr := addrs[(start+idx)%total]\n\t\tmapkey := fmt.Sprintf(\"%s[%s]\", severName, addr)\n\t\tif client.clients[mapkey] == nil {\n\t\t\tclient.clients[mapkey], err = rpc.Dial(\"tcp\", addr)\n\t\t\tif err != nil {\n\t\t\t\tLog.Warnf(\"RPC dial error : %s\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\terr = rpcCallWithReconnect(client.clients[mapkey], addr, serverMethod, args, reply)\n\t\tif err != nil {\n\t\t\tLog.Warnf(\"RpcCallWithReconnect error : %s\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\treturn nil\n\t}\n\n\treturn errorf(err.Error())\n}", "func (_ECVerify *ECVerifyRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _ECVerify.Contract.ECVerifyCaller.contract.Call(opts, result, method, params...)\n}", "func (s *Server) Invoke(ctx context.Context, method string, args interface{}, reply interface{}) error {\n\tcallID := uuid.NewString()\n\tmsg, err := message.NewRequest(callID, method, args)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treq, err := MarshalProtoMessage(msg)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.mu.Lock()\n\twait := s.registerMethodCall(callID)\n\ts.mu.Unlock()\n\n\t// Extract the public key from context\n\tp, ok := peer.FromContext(ctx)\n\tif !ok {\n\t\treturn errors.New(\"could not extract public key\")\n\t}\n\tpubKey := p.PublicKey\n\n\terr = s.sendMsg(pubKey, req)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Wait for the response\n\tselect {\n\tcase msg := <-wait:\n\t\t// Handle error\n\t\tif msg.Error != \"\" {\n\t\t\treturn errors.New(msg.Error)\n\t\t}\n\n\t\t// Unmarshal the payload into the reply\n\t\terr := UnmarshalProtoMessage(msg.GetPayload(), reply)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase <-time.After(2 * time.Second): // TODO - Make this configurable\n\t\t// Remove the call since we have timeout\n\t\ts.mu.Lock()\n\t\ts.removeMethodCall(callID)\n\t\ts.mu.Unlock()\n\t\treturn errors.New(\"call timeout\")\n\t}\n\n\treturn nil\n}", "func (_Console *ConsoleRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Console.Contract.ConsoleCaller.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Contract *ContractCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Contract.Contract.contract.Call(opts, result, method, params...)\n}", "func (_CRLv0 *CRLv0CallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _CRLv0.Contract.contract.Call(opts, result, method, params...)\n}", "func (s *UserClient) RpcInvoke(req []byte) ([]byte, error) {\n\t// rpc.send\n\terr := s.rpc.Send(req)\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\t// rpc.receive\n\treturn s.rpc.Receive()\n}", "func (_Lmc *LmcCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Lmc.Contract.contract.Call(opts, result, method, params...)\n}", "func (_BridgeRegistry *BridgeRegistryRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _BridgeRegistry.Contract.BridgeRegistryCaller.contract.Call(opts, result, method, params...)\n}", "func (_Token *TokenCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Token.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Registry *RegistryCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Registry.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Owned *OwnedCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Owned.Contract.contract.Call(opts, result, method, params...)\n}", "func (_KNS *KNSCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _KNS.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Bridge *BridgeRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Bridge.Contract.BridgeCaller.contract.Call(opts, result, method, params...)\n}", "func (_IOrakuruCore *IOrakuruCoreCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _IOrakuruCore.Contract.contract.Call(opts, result, method, params...)\n}", "func (_EtherDelta *EtherDeltaCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _EtherDelta.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Votes *VotesCallerRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _Votes.Contract.contract.Call(opts, result, method, params...)\n}", "func (_Finalizable *FinalizableRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Finalizable.Contract.FinalizableCaller.contract.Call(opts, result, method, params...)\n}", "func (_HarmonyBridge *HarmonyBridgeRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _HarmonyBridge.Contract.HarmonyBridgeCaller.contract.Call(opts, result, method, params...)\n}", "func (_Finalizable *FinalizableCallerRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Finalizable.Contract.contract.Call(opts, result, method, params...)\n}", "func (_TokensNetwork *TokensNetworkRaw) Call(opts *bind.CallOpts, result interface{}, method string, params ...interface{}) error {\n\treturn _TokensNetwork.Contract.TokensNetworkCaller.contract.Call(opts, result, method, params...)\n}", "func (_PermInterface *PermInterfaceRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _PermInterface.Contract.PermInterfaceCaller.contract.Call(opts, result, method, params...)\n}", "func (_Weth *WethRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _Weth.Contract.WethCaller.contract.Call(opts, result, method, params...)\n}" ]
[ "0.5998847", "0.58454365", "0.58190733", "0.5798254", "0.5792092", "0.5782765", "0.57422286", "0.55782926", "0.5519295", "0.5499245", "0.54811376", "0.54350615", "0.54342645", "0.5404993", "0.53696823", "0.5339746", "0.53338146", "0.53276527", "0.5325165", "0.5311489", "0.5310276", "0.5288284", "0.52817154", "0.5267633", "0.52614766", "0.525704", "0.52385825", "0.52363163", "0.52234185", "0.5199071", "0.5182189", "0.5179205", "0.51708406", "0.51697713", "0.51672626", "0.516456", "0.5154795", "0.514618", "0.51217014", "0.51171845", "0.51092774", "0.51092774", "0.5103461", "0.50868183", "0.5080419", "0.507111", "0.50689197", "0.5058595", "0.50577587", "0.50565535", "0.5048494", "0.50427145", "0.5041307", "0.50372416", "0.50372416", "0.50372416", "0.50331837", "0.50281477", "0.5027489", "0.5019684", "0.5018875", "0.5015051", "0.5011403", "0.49911588", "0.4989358", "0.49833125", "0.49820462", "0.49740037", "0.49696103", "0.49661598", "0.49646983", "0.4964576", "0.49642342", "0.4952417", "0.4951914", "0.4950738", "0.49504733", "0.4949892", "0.494987", "0.49403414", "0.49403414", "0.49403414", "0.49376562", "0.49323654", "0.49246347", "0.49203616", "0.49201563", "0.4920047", "0.49148637", "0.4911857", "0.49100927", "0.49082404", "0.49075508", "0.49068663", "0.4906677", "0.4902579", "0.4895289", "0.48949924", "0.48905057", "0.4889528" ]
0.7403664
0
The fake service has no signature.
func (fs fakeService) Signature(ctx *context.T, call rpc.ServerCall) ([]signature.Interface, error) { ctx.Infof("Fake Service Signature???") return nil, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (fs fakeService) MethodSignature(ctx *context.T, call rpc.ServerCall, method string) (signature.Method, error) {\n\tctx.Infof(\"Fake Service Method Signature???\")\n\treturn signature.Method{}, nil\n}", "func testService() *corev1.Service {\n\treturn &corev1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: \"default\",\n\t\t\tName: \"symbols\",\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"deploy\": \"sourcegraph\",\n\t\t\t},\n\t\t},\n\t\tSpec: corev1.ServiceSpec{\n\t\t\tType: corev1.ServiceTypeClusterIP,\n\t\t\tPorts: []corev1.ServicePort{\n\t\t\t\t{\n\t\t\t\t\tName: \"http\",\n\t\t\t\t\tPort: 3184,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tStatus: corev1.ServiceStatus{},\n\t}\n}", "func (fs fakeService) callRemoteSignature(mojourl string, mojoname string) (mojomInterface mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, err error) {\n\t// TODO(afandria): The service_describer mojom file defines the constant, but\n\t// it is not actually present in the generated code:\n\t// https://github.com/domokit/mojo/issues/469\n\t// serviceDescriberInterfaceName := \"_ServiceDescriber\"\n\n\tr, p := service_describer.CreateMessagePipeForServiceDescriber()\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&r)\n\tsDescriber := service_describer.NewServiceDescriberProxy(p, bindings.GetAsyncWaiter())\n\tdefer sDescriber.Close_Proxy()\n\n\tr2, p2 := service_describer.CreateMessagePipeForServiceDescription()\n\terr = sDescriber.DescribeService(mojoname, r2)\n\tif err != nil {\n\t\treturn\n\t}\n\tsDescription := service_describer.NewServiceDescriptionProxy(p2, bindings.GetAsyncWaiter())\n\tdefer sDescription.Close_Proxy()\n\n\tmojomInterface, err = sDescription.GetTopLevelInterface()\n\tif err != nil {\n\t\treturn\n\t}\n\tdescPtr, err := sDescription.GetAllTypeDefinitions()\n\tif err != nil {\n\t\treturn\n\t}\n\treturn mojomInterface, *descPtr, nil\n}", "func (fs fakeService) Invoke(ctx *context.T, call rpc.StreamServerCall, method string, argptrs []interface{}) (results []interface{}, _ error) {\n\t// fs.suffix consists of the mojo url and the application/interface name.\n\t// The last part should be the name; everything else is the url.\n\tparts := strings.Split(fs.suffix, \"/\")\n\tmojourl := strings.Join(parts[:len(parts)-1], \"/\") // e.g., mojo:go_remote_echo_server. May be defined in a BUILD.gn file.\n\tmojoname := parts[len(parts)-1] // e.g., mojo::examples::RemoteEcho. Defined from the interface + module.\n\n\t// Create the generic message pipe. r is a bindings.InterfaceRequest, and\n\t// p is a bindings.InterfacePointer.\n\tr, p := bindings.CreateMessagePipeForMojoInterface()\n\tv := v23ServiceRequest{\n\t\trequest: r,\n\t\tname: mojoname,\n\t} // v is an application.ServiceRequest with mojoname\n\n\t// Connect to the mojourl.\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&v)\n\n\t// Then assign a new router the FakeService.\n\t// This will never conflict because each FakeService is only invoked once.\n\tfs.router = bindings.NewRouter(p.PassMessagePipe(), bindings.GetAsyncWaiter())\n\tdefer fs.Close_Proxy()\n\n\tctx.Infof(\"Fake Service Invoke (Remote Signature: %q -- %q)\", mojourl, mojoname)\n\n\t// Vanadium relies on type information, so we will retrieve that first.\n\tmojomInterface, desc, err := fs.callRemoteSignature(mojourl, mojoname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Signature %v\", mojomInterface)\n\tctx.Infof(\"Fake Service Invoke (Remote Method: %v)\", method)\n\n\t// With the type information, we can make the method call to the remote interface.\n\tmethodResults, err := fs.callRemoteMethod(ctx, method, mojomInterface, desc, argptrs)\n\tif err != nil {\n\t\tctx.Errorf(\"Method called failed: %v\", err)\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Results %v\", methodResults)\n\n\t// Convert methodResult to results.\n\tresults = make([]interface{}, len(methodResults))\n\tfor i := range methodResults {\n\t\tresults[i] = &methodResults[i]\n\t}\n\treturn results, nil\n}", "func (f *lazyCallReq) Service() []byte {\n\tl := f.Payload[_serviceLenIndex]\n\treturn f.Payload[_serviceNameIndex : _serviceNameIndex+int(l)]\n}", "func TestServiceMethodNamesCamelCase(t *testing.T) {\n\ts := httptest.NewServer(NewHaberdasherV1Server(&HaberdasherService{}, nil))\n\tdefer s.Close()\n\n\tclient := NewHaberdasherV1ProtobufClient(s.URL, http.DefaultClient)\n\n\that, err := client.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err != nil {\n\t\tt.Fatalf(\"go protobuf client err=%q\", err)\n\t}\n\tif hat.Size != 1 {\n\t\tt.Errorf(\"wrong hat size returned\")\n\t}\n}", "func (f FakeServiceGetter) Get(name string) (*v1.Service, error) {\n\treturn f.Svc, f.Err\n}", "func SophisticatedResponder(w http.ResponseWriter, httpRequest *http.Request, fakeRequest *Request) {\n\tstatusCode := fakeRequest.Response.StatusCode\n\tbody := fakeRequest.Response.BodyBuffer\n\tresponseHeader := fakeRequest.Response.Header\n\n\tif len(fakeRequest.Header) > 0 {\n\t\ts, b, err := validateHeaders(fakeRequest.Header, httpRequest.Header)\n\t\tif err != nil {\n\t\t\tstatusCode = s\n\t\t\tbody = []byte(b)\n\t\t}\n\t}\n\tif len(fakeRequest.Cookies()) > 0 {\n\t\ts, b, err := validateCookies(fakeRequest.Cookies(), httpRequest.Cookies())\n\t\tif err != nil {\n\t\t\tstatusCode = s\n\t\t\tbody = []byte(b)\n\t\t}\n\t}\n\tif (len(responseHeader)) > 0 {\n\t\tfor k := range fakeRequest.Response.Header {\n\t\t\tw.Header().Add(k, responseHeader.Get(k))\n\t\t}\n\t}\n\tif statusCode > 0 {\n\t\tw.WriteHeader(statusCode)\n\t}\n\tserviceResponses := \"\"\n\tif len(fakeRequest.ServiceEndpoints) > 0 {\n\t\tfor _, uri := range fakeRequest.ServiceEndpoints {\n\t\t\tstatus, body, err := invokeServiceEndpoint(uri, httpRequest.Header)\n\t\t\tif err == nil {\n\t\t\t\tserviceResponses += (uri + \": \")\n\t\t\t\tserviceResponses += (status + \": \")\n\t\t\t\tserviceResponses += (body)\n\t\t\t\tserviceResponses += \"<br>\"\n\t\t\t}\n\t\t}\n\t}\n\tif (len(body)) > 0 {\n\t\tb := string(body)\n\t\tif len(fakeRequest.InjectionKeys) > 0 {\n\t\t\tfor _, k := range fakeRequest.InjectionKeys {\n\t\t\t\tif k == \"path\" {\n\t\t\t\t\tbody = []byte(fmt.Sprintf(b, strings.TrimPrefix(httpRequest.URL.Path, \"/\")))\n\t\t\t\t\tb = string(body)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif len(serviceResponses) > 0 {\n\t\t\tb += serviceResponses\n\t\t}\n\n\t\tif fakeRequest.RenderHTML {\n\t\t\tb = \"<html><head><title>fakeserver</title></head><body>\" + b + \"</body></html>\"\n\t\t}\n\n\t\tw.Write([]byte(b))\n\t}\n}", "func setupFakeClient(url string) *Client {\n\treturn &Client{\n\t\tServiceEndpoint: ServiceEndpoint{\n\t\t\tRequestURL: url,\n\t\t\tDocsURL: \"some-docs-url\",\n\t\t},\n\t}\n}", "func handleService(req typhon.Request) typhon.Response {\n\tparts := reService.FindStringSubmatch(req.URL.Path)\n\tif len(parts) != 3 {\n\t\treturn typhon.Response{Error: terrors.NotFound(\"bad_endpoint\", \"Unable to determine service endpoint.\", nil)}\n\t}\n\n\treturn handle(req, \"s-\"+parts[1], parts[2])\n}", "func TestMakePublicService(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tsks *v1alpha1.ServerlessService\n\t\twant *corev1.Service\n\t}{{\n\t\tname: \"HTTP - serve\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tUID: \"1982\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolHTTP1,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"collie\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"collie\",\n\t\t\t\t\tUID: \"1982\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameHTTP1,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTPPort,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTPPort),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP - proxy\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tUID: \"1982\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tMode: v1alpha1.SKSOperationModeProxy,\n\t\t\t\tProtocolType: networking.ProtocolHTTP1,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"collie\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"collie\",\n\t\t\t\t\tUID: \"1982\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameHTTP1,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTPPort,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTPPort),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - serve\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - serve - no backends\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - proxy\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeProxy,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}}\n\n\tfor _, test := range tests {\n\t\tt.Run(test.name, func(t *testing.T) {\n\t\t\tgot := MakePublicService(test.sks)\n\t\t\tif diff := cmp.Diff(test.want, got); diff != \"\" {\n\t\t\t\tt.Errorf(\"Public K8s Service mismatch (-want, +got) = %v\", diff)\n\t\t\t}\n\t\t})\n\t}\n}", "func TestService(t *testing.T) {\n\t// Create service to test\n\ts := res.NewService(\"foo\")\n\ts.Handle(\"bar.$id\",\n\t\tres.Access(res.AccessGranted),\n\t\tres.GetModel(func(r res.ModelRequest) {\n\t\t\tr.Model(struct {\n\t\t\t\tMessage string `json:\"msg\"`\n\t\t\t}{r.PathParam(\"id\")})\n\t\t}),\n\t)\n\n\t// Create test session\n\tc := restest.NewSession(t, s)\n\tdefer c.Close()\n\n\t// Test sending get request and validate response\n\tc.Get(\"foo.bar.42\").\n\t\tResponse().\n\t\tAssertModel(map[string]string{\"msg\": \"42\"})\n}", "func newService(rcvr interface{}, guard Guard) *service {\n\ts := new(service)\n\ts.typ = reflect.TypeOf(rcvr)\n\ts.rcvr = reflect.ValueOf(rcvr)\n\ts.name = reflect.Indirect(s.rcvr).Type().Name()\n\ts.guard = guard\n\n\t// install the methods\n\ts.method = suitableMethods(s.typ, true)\n\n\treturn s\n}", "func newServiceNoPatchTest(name string, options ...controllertesting.ServiceOption) TableRow {\n\ttest := newDispatcherBasicTest(\"Existing Dispatcher Service, \" + name + \", No Patch\")\n\ttest.Objects = append(test.Objects,\n\t\tcontrollertesting.NewKafkaChannelDispatcherService(options...),\n\t\tcontrollertesting.NewKafkaChannelDispatcherDeployment())\n\treturn test\n}", "func (s *ServiceSuite) TestSingleServiceFirstParty(c *gc.C) {\n\toc := newBakery(\"bakerytest\", nil)\n\n\tprimary, err := oc.Oven.NewMacaroon(testContext, bakery.LatestVersion, ages, nil, bakery.LoginOp)\n\tc.Assert(err, gc.IsNil)\n\tc.Assert(primary.M().Location(), gc.Equals, \"bakerytest\")\n\terr = oc.Oven.AddCaveat(testContext, primary, strCaveat(\"something\"))\n\n\t_, err = oc.Checker.Auth(macaroon.Slice{primary.M()}).Allow(strContext(\"something\"), bakery.LoginOp)\n\tc.Assert(err, gc.IsNil)\n}", "func (g *grpc) generateService(file *generator.FileDescriptor, service *pb.ServiceDescriptorProto, index int) {\n\tpath := fmt.Sprintf(\"6,%d\", index) // 6 means service.\n\n\torigServName := service.GetName()\n\tfullServName := origServName\n\tif pkg := file.GetPackage(); pkg != \"\" {\n\t\tfullServName = pkg + \".\" + fullServName\n\t}\n\tservName := generator.CamelCase(origServName)\n\tdeprecated := service.GetOptions().GetDeprecated()\n\n\tg.P()\n\tg.P(fmt.Sprintf(`// %sClient is the client API for %s service.\n//\n// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.`, servName, servName))\n\n\t// Client interface.\n\tif deprecated {\n\t\tg.P(\"//\")\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"type \", servName, \"Client interface {\")\n\tfor i, method := range service.Method {\n\t\tg.gen.PrintComments(fmt.Sprintf(\"%s,2,%d\", path, i)) // 2 means method in a service.\n\t\tg.P(g.generateClientSignature(servName, method))\n\t}\n\tg.P(\"}\")\n\tg.P()\n\n\t// Client structure.\n\tg.P(\"type \", unexport(servName), \"Client struct {\")\n\tg.P(\"cc *\", grpcPkg, \".ClientConn\")\n\tg.P(\"}\")\n\tg.P()\n\n\t// NewClient factory.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"func New\", servName, \"Client (cc *\", grpcPkg, \".ClientConn) \", servName, \"Client {\")\n\tg.P(\"return &\", unexport(servName), \"Client{cc}\")\n\tg.P(\"}\")\n\tg.P()\n\n\tvar methodIndex, streamIndex int\n\tserviceDescVar := \"_\" + servName + \"_serviceDesc\"\n\t// Client method implementations.\n\tfor _, method := range service.Method {\n\t\tvar descExpr string\n\t\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\t\t// Unary RPC method\n\t\t\tdescExpr = fmt.Sprintf(\"&%s.Methods[%d]\", serviceDescVar, methodIndex)\n\t\t\tmethodIndex++\n\t\t} else {\n\t\t\t// Streaming RPC method\n\t\t\tdescExpr = fmt.Sprintf(\"&%s.Streams[%d]\", serviceDescVar, streamIndex)\n\t\t\tstreamIndex++\n\t\t}\n\t\tg.generateClientMethod(servName, fullServName, serviceDescVar, method, descExpr)\n\t}\n\n\t// Server interface.\n\tserverType := servName + \"Server\"\n\tg.P(\"// \", serverType, \" is the server API for \", servName, \" service.\")\n\tif deprecated {\n\t\tg.P(\"//\")\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"type \", serverType, \" interface {\")\n\tfor i, method := range service.Method {\n\t\tg.gen.PrintComments(fmt.Sprintf(\"%s,2,%d\", path, i)) // 2 means method in a service.\n\t\tg.P(g.generateServerSignature(servName, method))\n\t}\n\tg.P(\"}\")\n\tg.P()\n\n\t// Server Unimplemented struct for forward compatability.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.generateUnimplementedServer(servName, service)\n\n\t// Server registration.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"func Register\", servName, \"Server(s *\", grpcPkg, \".Server, srv \", serverType, \") {\")\n\tg.P(\"s.RegisterService(&\", serviceDescVar, `, srv)`)\n\tg.P(\"}\")\n\tg.P()\n\n\t// Server handler implementations.\n\tvar handlerNames []string\n\tfor _, method := range service.Method {\n\t\thname := g.generateServerMethod(servName, fullServName, method)\n\t\thandlerNames = append(handlerNames, hname)\n\t}\n\n\t// Service descriptor.\n\tg.P(\"var \", serviceDescVar, \" = \", grpcPkg, \".ServiceDesc {\")\n\tg.P(\"ServiceName: \", strconv.Quote(fullServName), \",\")\n\tg.P(\"HandlerType: (*\", serverType, \")(nil),\")\n\tg.P(\"Methods: []\", grpcPkg, \".MethodDesc{\")\n\tfor i, method := range service.Method {\n\t\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\t\tcontinue\n\t\t}\n\t\tg.P(\"{\")\n\t\tg.P(\"MethodName: \", strconv.Quote(method.GetName()), \",\")\n\t\tg.P(\"Handler: \", handlerNames[i], \",\")\n\t\tg.P(\"},\")\n\t}\n\tg.P(\"},\")\n\tg.P(\"Streams: []\", grpcPkg, \".StreamDesc{\")\n\tfor i, method := range service.Method {\n\t\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\t\tcontinue\n\t\t}\n\t\tg.P(\"{\")\n\t\tg.P(\"StreamName: \", strconv.Quote(method.GetName()), \",\")\n\t\tg.P(\"Handler: \", handlerNames[i], \",\")\n\t\tif method.GetServerStreaming() {\n\t\t\tg.P(\"ServerStreams: true,\")\n\t\t}\n\t\tif method.GetClientStreaming() {\n\t\t\tg.P(\"ClientStreams: true,\")\n\t\t}\n\t\tg.P(\"},\")\n\t}\n\tg.P(\"},\")\n\tg.P(\"Metadata: \\\"\", file.GetName(), \"\\\",\")\n\tg.P(\"}\")\n\tg.P()\n}", "func TestServiceStart(t *testing.T) {\n\trunTest(t, func(s *res.Service) {\n\t\ts.Handle(\"model\", res.GetResource(func(r res.GetRequest) { r.NotFound() }))\n\t}, nil)\n}", "func (poolMgr *Poolmgr) tapService(w http.ResponseWriter, r *http.Request) {\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\thttp.Error(w, \"Failed to read request\", 500)\n\t\treturn\n\t}\n\tsvcName := string(body)\n\tsvcHost := strings.TrimPrefix(svcName, \"http://\")\n\n\terr = poolMgr.fsCache.TouchByAddress(svcHost)\n\tif err != nil {\n\t\tlog.Printf(\"funcSvc tap error: %v\", err)\n\t\thttp.Error(w, \"Not found\", 404)\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusOK)\n}", "func (m *CloudWatchLogsServiceMock) CreateNewServiceIfUnHealthy() {\n\n}", "func svcHandler()", "func (f *FakeInstance) Start(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func Service() typhon.Service {\n\treturn Proxy\n}", "func (auth *Manager) Signed(service core.APIService) core.APIService {\n\treturn &wrapped{\n\t\tinfoService: service.InfoService,\n\t\tconfigService: func(ctx context.Context, config *v0.Config) (*v0.Response, error) {\n\t\t\tok, wrongSig, err := auth.Check(config)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tif !ok {\n\t\t\t\treturn response.MSG(v0.Code_BadRequest, fmt.Sprintf(\"incorrect signature for: %s\", wrongSig)), nil\n\t\t\t}\n\t\t\treturn service.ConfigService(ctx, config)\n\t\t},\n\t}\n}", "func (ft *FacadeUnitTest) Test_GetEvaluatedServiceSimple(c *C) {\n\tserviceID := \"0\"\n\tserviceName := \"service0\"\n\tsvc := service.Service{\n\t\tID: serviceID,\n\t\tName: serviceName,\n\t\tActions: map[string]string{\"name\": \"{{.Name}}\", \"instanceID\": \"{{.InstanceID}}\"},\n\t}\n\tft.serviceStore.On(\"GetServiceDetails\", ft.ctx, serviceID).Return(&service.ServiceDetails{ID: serviceID}, nil)\n\tft.serviceStore.On(\"Get\", ft.ctx, serviceID).Return(&svc, nil)\n\tft.configStore.On(\"GetConfigFiles\", ft.ctx, serviceID, \"/\"+serviceID).Return([]*serviceconfigfile.SvcConfigFile{}, nil)\n\n\tinstanceID := 99\n\tresult, err := ft.Facade.GetEvaluatedService(ft.ctx, serviceID, instanceID)\n\n\tc.Assert(result, Not(IsNil))\n\tc.Assert(err, IsNil)\n\n\tc.Assert(result.Actions[\"name\"], Equals, serviceName)\n\tc.Assert(result.Actions[\"instanceID\"], Equals, fmt.Sprintf(\"%d\", instanceID))\n}", "func (k *Kit) ServiceName() string { return k.name }", "func TestCallToPublicService(t *testing.T) {\n\tt.Parallel()\n\n\tclients := Setup(t)\n\n\tt.Log(\"Creating a Service for the helloworld test app.\")\n\tnames := test.ResourceNames{\n\t\tService: test.ObjectNameForTest(t),\n\t\tImage: test.HelloWorld,\n\t}\n\n\ttest.EnsureTearDown(t, clients, &names)\n\n\tresources, err := v1test.CreateServiceReady(t, clients, &names)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create initial Service: %v: %v\", names.Service, err)\n\t}\n\n\tif resources.Route.Status.URL.Host == \"\" {\n\t\tt.Fatalf(\"Route is missing .Status.URL: %#v\", resources.Route.Status)\n\t}\n\tif resources.Route.Status.Address == nil {\n\t\tt.Fatalf(\"Route is missing .Status.Address: %#v\", resources.Route.Status)\n\t}\n\n\tgatewayTestCases := []struct {\n\t\tname string\n\t\turl *url.URL\n\t\taccessibleExternally bool\n\t}{\n\t\t{\"local_address\", resources.Route.Status.Address.URL.URL(), false},\n\t\t{\"external_address\", resources.Route.Status.URL.URL(), true},\n\t}\n\n\tfor _, tc := range gatewayTestCases {\n\t\ttc := tc\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tt.Parallel()\n\t\t\tif !test.ServingFlags.DisableLogStream {\n\t\t\t\tcancel := logstream.Start(t)\n\t\t\t\tdefer cancel()\n\t\t\t}\n\t\t\ttestProxyToHelloworld(t, clients, tc.url, false /*inject*/, tc.accessibleExternally)\n\t\t})\n\t}\n}", "func (c *ApiService) CreateService(params *CreateServiceParams) (*VerifyV2Service, error) {\n\tpath := \"/v2/Services\"\n\n\tdata := url.Values{}\n\theaders := make(map[string]interface{})\n\n\tif params != nil && params.FriendlyName != nil {\n\t\tdata.Set(\"FriendlyName\", *params.FriendlyName)\n\t}\n\tif params != nil && params.CodeLength != nil {\n\t\tdata.Set(\"CodeLength\", fmt.Sprint(*params.CodeLength))\n\t}\n\tif params != nil && params.LookupEnabled != nil {\n\t\tdata.Set(\"LookupEnabled\", fmt.Sprint(*params.LookupEnabled))\n\t}\n\tif params != nil && params.SkipSmsToLandlines != nil {\n\t\tdata.Set(\"SkipSmsToLandlines\", fmt.Sprint(*params.SkipSmsToLandlines))\n\t}\n\tif params != nil && params.DtmfInputRequired != nil {\n\t\tdata.Set(\"DtmfInputRequired\", fmt.Sprint(*params.DtmfInputRequired))\n\t}\n\tif params != nil && params.TtsName != nil {\n\t\tdata.Set(\"TtsName\", *params.TtsName)\n\t}\n\tif params != nil && params.Psd2Enabled != nil {\n\t\tdata.Set(\"Psd2Enabled\", fmt.Sprint(*params.Psd2Enabled))\n\t}\n\tif params != nil && params.DoNotShareWarningEnabled != nil {\n\t\tdata.Set(\"DoNotShareWarningEnabled\", fmt.Sprint(*params.DoNotShareWarningEnabled))\n\t}\n\tif params != nil && params.CustomCodeEnabled != nil {\n\t\tdata.Set(\"CustomCodeEnabled\", fmt.Sprint(*params.CustomCodeEnabled))\n\t}\n\tif params != nil && params.PushIncludeDate != nil {\n\t\tdata.Set(\"Push.IncludeDate\", fmt.Sprint(*params.PushIncludeDate))\n\t}\n\tif params != nil && params.PushApnCredentialSid != nil {\n\t\tdata.Set(\"Push.ApnCredentialSid\", *params.PushApnCredentialSid)\n\t}\n\tif params != nil && params.PushFcmCredentialSid != nil {\n\t\tdata.Set(\"Push.FcmCredentialSid\", *params.PushFcmCredentialSid)\n\t}\n\tif params != nil && params.TotpIssuer != nil {\n\t\tdata.Set(\"Totp.Issuer\", *params.TotpIssuer)\n\t}\n\tif params != nil && params.TotpTimeStep != nil {\n\t\tdata.Set(\"Totp.TimeStep\", fmt.Sprint(*params.TotpTimeStep))\n\t}\n\tif params != nil && params.TotpCodeLength != nil {\n\t\tdata.Set(\"Totp.CodeLength\", fmt.Sprint(*params.TotpCodeLength))\n\t}\n\tif params != nil && params.TotpSkew != nil {\n\t\tdata.Set(\"Totp.Skew\", fmt.Sprint(*params.TotpSkew))\n\t}\n\tif params != nil && params.DefaultTemplateSid != nil {\n\t\tdata.Set(\"DefaultTemplateSid\", *params.DefaultTemplateSid)\n\t}\n\n\tresp, err := c.requestHandler.Post(c.baseURL+path, data, headers)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\tps := &VerifyV2Service{}\n\tif err := json.NewDecoder(resp.Body).Decode(ps); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ps, err\n}", "func fakeGcp() (*compute.Service, error) {\n\tclient := &http.Client{}\n\thttpmock.ActivateNonDefault(client)\n\treturn compute.NewService(context.Background(), option.WithoutAuthentication(), option.WithHTTPClient(client))\n}", "func newService() *service {\n\n\tctx, cancel := context.WithCancel(context.Background())\n\n\treturn &service{\n\t\tctx: ctx,\n\t\tctxCancel: cancel,\n\t}\n}", "func newFakeClient() client.Client {\n\treturn fakeclient.NewFakeClient()\n}", "func NewService(c Config) *Service {\n \n s := &Service{}\n s.instance = c.Instance\n s.hostname = c.Hostname\n s.userAgent = c.UserAgent\n s.port = c.Endpoint\n s.router = mux.NewRouter()\n s.entityHandler = c.EntityHandler\n s.readTimeout = c.ReadTimeout\n s.writeTimeout = c.WriteTimeout\n s.idleTimeout = c.IdleTimeout\n \n if c.Name == \"\" {\n s.name = \"service\"\n }else{\n s.name = c.Name\n }\n \n if c.Debug || os.Getenv(\"GOREST_DEBUG\") == \"true\" {\n s.debug = true\n }\n \n if c.TraceRegexps != nil {\n if s.traceRequests == nil {\n s.traceRequests = make(map[string]*regexp.Regexp)\n }\n for _, e := range c.TraceRegexps {\n s.traceRequests[e.String()] = e\n }\n }\n if t := os.Getenv(\"GOREST_TRACE\"); t != \"\" {\n if s.traceRequests == nil {\n s.traceRequests = make(map[string]*regexp.Regexp)\n }\n for _, e := range strings.Split(t, \";\") {\n s.traceRequests[e] = regexp.MustCompile(e)\n }\n }\n if s.debug {\n for k, _ := range s.traceRequests {\n fmt.Println(\"rest: trace:\", k)\n }\n }\n \n s.suppress = make(map[string]struct{})\n if v := os.Getenv(\"GOREST_TRACE_SUPPRESS_HEADERS\"); v != \"\" {\n if !strings.EqualFold(v, \"none\") {\n for _, e := range strings.Split(v, \",\") {\n s.suppress[strings.ToLower(e)] = struct{}{}\n }\n }\n }else{\n s.suppress[\"authorization\"] = struct{}{}\n }\n \n return s\n}", "func (g *Generator) FService() *GeneratorFile {\n\treturn g.F(FILEID_SERVICE)\n}", "func NewService(t testing.TB) *Service {\n\tmock := &Service{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func TestServiceMethodNamesUnderscores(t *testing.T) {\n\ts := httptest.NewServer(NewHaberdasherV1Server(&HaberdasherService{}, nil))\n\tdefer s.Close()\n\n\tclient := NewHaberdasherV1ProtobufClient(s.URL, compatibilityTestClient{client: http.DefaultClient},\n\t\ttwirp.WithClientLiteralURLs(true))\n\that, err := client.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err != nil {\n\t\tt.Fatalf(\"compatible protobuf client err=%q\", err)\n\t}\n\tif hat.Size != 1 {\n\t\tt.Errorf(\"wrong hat size returned\")\n\t}\n\n\tcamelCasedClient := NewHaberdasherV1ProtobufClient(s.URL, compatibilityTestClient{client: http.DefaultClient},\n\t\ttwirp.WithClientLiteralURLs(false)) // default value, send CamelCased routes\n\t_, err = camelCasedClient.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err == nil {\n\t\tt.Fatalf(\"expected error raised by the compatibilityTestClient because routes are camelcased. Got nil.\")\n\t}\n\tif err.Error() != \"twirp error internal: failed to do request: expected: /twirp/twirp.internal.twirptest.snake_case_names.Haberdasher_v1/MakeHat_v1, got: /twirp/twirp.internal.twirptest.snake_case_names.HaberdasherV1/MakeHatV1\" {\n\t\tt.Fatalf(\"expected error to be about the expected path, got err=%q\", err)\n\t}\n}", "func service() typhon.Service {\r\n\trouter := typhon.Router{}\r\n\trouter.GET(\"/oxcross\", serveResponse)\r\n\trouter.GET(\"/healthz\", serveResponse)\r\n\r\n\tsvc := router.Serve().Filter(typhon.ErrorFilter).Filter(typhon.H2cFilter)\r\n\r\n\treturn svc\r\n}", "func Test(c *gin.Context) {\n\tvar serviceTestDTO model.ServiceTest\n\n\terr := c.BindJSON(&serviceTestDTO)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\trefConf := genericServiceImpl.NewRefConf(\"dubbo-admin\", serviceTestDTO.Service, \"dubbo\")\n\ttime.Sleep(2 * time.Second)\n\tresp, err := refConf.\n\t\tGetRPCService().(*generic.GenericService).\n\t\tInvoke(\n\t\t\tc,\n\t\t\tserviceTestDTO.Method,\n\t\t\tserviceTestDTO.ParameterTypes,\n\t\t\t[]hessian.Object{\"A003\"}, // fixme\n\t\t)\n\trefConf.GetInvoker().Destroy()\n\tif err != nil {\n\t\tlogger.Error(\"Error do generic invoke for service test\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, resp)\n}", "func (f *FakeInstance) Create(_ context.Context, _ *govultr.InstanceCreateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func FakeNew() (*Client, *FakeClientset) {\n\treturn FakeNewWithIngressSupports(false, true)\n}", "func TestGetUserServicePatched (t *testing.T){\n\tuser1, err := GetUserService(user_01.SocialNumber)\n\tassert.Equal(t, 200, err.HTTPStatus)\n\tassert.Equal(t, user1.Name, new_name_user_01)\n}", "func (mt *mockTokenBuilder) SetKeyService(keyService string) {\n\t//TODO some mocking\n}", "func NewFixedService() Service {\n\treturn &fixedService{}\n}", "func newMyService(g greeter, f factory) builder {\n\treturn func(s *MyService) {\n\t\ts.greeter = g\n\t\ts.f = f\n\t}\n}", "func TestService_Handle_Inviter(t *testing.T) {\n\tmockStore := &mockstorage.MockStore{Store: make(map[string]mockstorage.DBEntry)}\n\tstoreProv := mockstorage.NewCustomMockStoreProvider(mockStore)\n\tk := newKMS(t, storeProv)\n\tprov := &protocol.MockProvider{\n\t\tStoreProvider: storeProv,\n\t\tServiceMap: map[string]interface{}{\n\t\t\tmediator.Coordination: &mockroute.MockMediatorSvc{},\n\t\t},\n\t\tCustomKMS: k,\n\t\tKeyTypeValue: kms.ED25519Type,\n\t\tKeyAgreementTypeValue: kms.X25519ECDHKWType,\n\t}\n\n\tctx := &context{\n\t\toutboundDispatcher: prov.OutboundDispatcher(),\n\t\tcrypto: &tinkcrypto.Crypto{},\n\t\tkms: k,\n\t\tkeyType: kms.ED25519Type,\n\t\tkeyAgreementType: kms.X25519ECDHKWType,\n\t}\n\n\tverPubKey, encPubKey := newSigningAndEncryptionDIDKeys(t, ctx)\n\n\tctx.vdRegistry = &mockvdr.MockVDRegistry{CreateValue: createDIDDocWithKey(verPubKey, encPubKey)}\n\n\tconnRec, err := connection.NewRecorder(prov)\n\trequire.NoError(t, err)\n\trequire.NotNil(t, connRec)\n\n\tctx.connectionRecorder = connRec\n\n\tdoc, err := ctx.vdRegistry.Create(testMethod, nil)\n\trequire.NoError(t, err)\n\n\ts, err := New(prov)\n\trequire.NoError(t, err)\n\n\tactionCh := make(chan service.DIDCommAction, 10)\n\terr = s.RegisterActionEvent(actionCh)\n\trequire.NoError(t, err)\n\n\tstatusCh := make(chan service.StateMsg, 10)\n\terr = s.RegisterMsgEvent(statusCh)\n\trequire.NoError(t, err)\n\n\tcompletedFlag := make(chan struct{})\n\trespondedFlag := make(chan struct{})\n\n\tgo msgEventListener(t, statusCh, respondedFlag, completedFlag)\n\n\tgo func() { service.AutoExecuteActionEvent(actionCh) }()\n\n\tinvitation := &Invitation{\n\t\tType: InvitationMsgType,\n\t\tID: randomString(),\n\t\tLabel: \"Bob\",\n\t\tRecipientKeys: []string{verPubKey},\n\t\tServiceEndpoint: \"http://alice.agent.example.com:8081\",\n\t}\n\n\terr = ctx.connectionRecorder.SaveInvitation(invitation.ID, invitation)\n\trequire.NoError(t, err)\n\n\tthid := randomString()\n\n\t// Invitation was previously sent by Alice to Bob.\n\t// Bob now sends a did-exchange Invitation\n\tpayloadBytes, err := json.Marshal(\n\t\t&Request{\n\t\t\tType: RequestMsgType,\n\t\t\tID: thid,\n\t\t\tLabel: \"Bob\",\n\t\t\tThread: &decorator.Thread{\n\t\t\t\tPID: invitation.ID,\n\t\t\t},\n\t\t\tDID: doc.DIDDocument.ID,\n\t\t\tDocAttach: unsignedDocAttach(t, doc.DIDDocument),\n\t\t})\n\trequire.NoError(t, err)\n\tmsg, err := service.ParseDIDCommMsgMap(payloadBytes)\n\trequire.NoError(t, err)\n\t_, err = s.HandleInbound(msg, service.NewDIDCommContext(doc.DIDDocument.ID, \"\", nil))\n\trequire.NoError(t, err)\n\n\tselect {\n\tcase <-respondedFlag:\n\tcase <-time.After(2 * time.Second):\n\t\trequire.Fail(t, \"didn't receive post event responded\")\n\t}\n\t// Alice automatically sends exchange Response to Bob\n\t// Bob replies with an ACK\n\tpayloadBytes, err = json.Marshal(\n\t\t&model.Ack{\n\t\t\tType: AckMsgType,\n\t\t\tID: randomString(),\n\t\t\tStatus: \"OK\",\n\t\t\tThread: &decorator.Thread{ID: thid},\n\t\t})\n\trequire.NoError(t, err)\n\n\tdidMsg, err := service.ParseDIDCommMsgMap(payloadBytes)\n\trequire.NoError(t, err)\n\n\t_, err = s.HandleInbound(didMsg, service.NewDIDCommContext(doc.DIDDocument.ID, \"\", nil))\n\trequire.NoError(t, err)\n\n\tselect {\n\tcase <-completedFlag:\n\tcase <-time.After(2 * time.Second):\n\t\trequire.Fail(t, \"didn't receive post event complete\")\n\t}\n\n\tvalidateState(t, s, thid, findNamespace(AckMsgType), (&completed{}).Name())\n}", "func TestService_CreateWithAutoPauseNotificationsParameters(t *testing.T) {\n\tsetup()\n\tdefer teardown()\n\n\tmux.HandleFunc(\"/services\", func(w http.ResponseWriter, r *http.Request) {\n\t\ttestMethod(t, r, \"POST\")\n\t\t_, _ = w.Write([]byte(`{\"service\": {\"id\": \"1\",\"name\":\"foo\"}}`))\n\t})\n\n\tclient := defaultTestClient(server.URL, \"foo\")\n\tinput := Service{\n\t\tName: \"foo\",\n\t\tAutoPauseNotificationsParameters: &AutoPauseNotificationsParameters{\n\t\t\tEnabled: true,\n\t\t\tTimeout: 60,\n\t\t},\n\t}\n\tres, err := client.CreateService(input)\n\n\twant := &Service{\n\t\tAPIObject: APIObject{\n\t\t\tID: \"1\",\n\t\t},\n\t\tName: \"foo\",\n\t}\n\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\ttestEqual(t, want, res)\n}", "func TestEventService(t *testing.T) {\n\tvar result EventService\n\terr := json.NewDecoder(strings.NewReader(eventServiceBody)).Decode(&result)\n\n\tif err != nil {\n\t\tt.Errorf(\"Error decoding JSON: %s\", err)\n\t}\n\n\tif result.ID != \"EventService\" {\n\t\tt.Errorf(\"Received invalid ID: %s\", result.ID)\n\t}\n\n\tif result.Name != \"Event Service\" {\n\t\tt.Errorf(\"Received invalid name: %s\", result.Name)\n\t}\n\n\tif result.DeliveryRetryAttempts != 4 {\n\t\tt.Errorf(\"Expected 4 retry attempts, got: %d\", result.DeliveryRetryAttempts)\n\t}\n\n\tif result.DeliveryRetryIntervalSeconds != 30 {\n\t\tt.Errorf(\"Expected 30 second retry interval, got: %d\", result.DeliveryRetryIntervalSeconds)\n\t}\n\n\tif result.SSEFilterPropertiesSupported.MetricReportDefinition {\n\t\tt.Error(\"MetricReportDefinition filter should be false\")\n\t}\n\n\tif !result.SSEFilterPropertiesSupported.MessageID {\n\t\tt.Error(\"Message ID filter should be true\")\n\t}\n\n\tif result.submitTestEventTarget != \"/redfish/v1/EventService/Actions/EventService.SubmitTestEvent\" {\n\t\tt.Errorf(\"Invalid SubmitTestEvent target: %s\", result.submitTestEventTarget)\n\t}\n\n\tfor _, et := range result.EventTypesForSubscription {\n\t\tif !et.IsValidEventType() {\n\t\t\tt.Errorf(\"invalid event type: %s\", et)\n\t\t}\n\t}\n\n}", "func (s *Deployment) createService() error {\n\tsvc := s.getService(s.stos.Spec.GetServiceName())\n\tsvc.Spec = corev1.ServiceSpec{\n\t\tType: corev1.ServiceType(s.stos.Spec.GetServiceType()),\n\t\tPorts: []corev1.ServicePort{\n\t\t\t{\n\t\t\t\tName: s.stos.Spec.GetServiceName(),\n\t\t\t\tProtocol: \"TCP\",\n\t\t\t\tPort: int32(s.stos.Spec.GetServiceInternalPort()),\n\t\t\t\tTargetPort: intstr.IntOrString{Type: intstr.Int, IntVal: int32(s.stos.Spec.GetServiceExternalPort())},\n\t\t\t},\n\t\t},\n\t\tSelector: map[string]string{\n\t\t\t\"app\": appName,\n\t\t\t\"kind\": daemonsetKind,\n\t\t},\n\t}\n\n\tif err := s.client.Create(context.Background(), svc); err != nil && !apierrors.IsAlreadyExists(err) {\n\t\treturn fmt.Errorf(\"failed to create %s: %v\", svc.GroupVersionKind().Kind, err)\n\t}\n\t// if err := s.createOrUpdateObject(svc); err != nil {\n\t// \treturn err\n\t// }\n\n\t// Patch storageos-api secret with above service IP in apiAddress.\n\tif !s.stos.Spec.CSI.Enable {\n\t\tsecret := &corev1.Secret{\n\t\t\tTypeMeta: metav1.TypeMeta{\n\t\t\t\tKind: \"Secret\",\n\t\t\t\tAPIVersion: \"v1\",\n\t\t\t},\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tName: s.stos.Spec.SecretRefName,\n\t\t\t\tNamespace: s.stos.Spec.SecretRefNamespace,\n\t\t\t},\n\t\t}\n\t\tnsNameSecret := types.NamespacedName{\n\t\t\tNamespace: secret.ObjectMeta.GetNamespace(),\n\t\t\tName: secret.ObjectMeta.GetName(),\n\t\t}\n\t\tif err := s.client.Get(context.Background(), nsNameSecret, secret); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tnsNameService := types.NamespacedName{\n\t\t\tNamespace: svc.ObjectMeta.GetNamespace(),\n\t\t\tName: svc.ObjectMeta.GetName(),\n\t\t}\n\t\tif err := s.client.Get(context.Background(), nsNameService, svc); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tapiAddress := fmt.Sprintf(\"tcp://%s:5705\", svc.Spec.ClusterIP)\n\t\tsecret.Data[apiAddressKey] = []byte(apiAddress)\n\n\t\tif err := s.client.Update(context.Background(), secret); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestCryptoSignerInterfaceBehavior(t *testing.T) {\n\tcs := NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.EmptyCryptoServiceInterfaceBehaviorTests(t, cs)\n\tinterfaces.CreateGetKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.CreateListKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.AddGetKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.AddListKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n}", "func newHelloService() HelloService {\n\treturn &helloServiceImpl{}\n}", "func fixPolicyFake(client *fake.Clientset) *clientset {\n\treturn &clientset{client}\n}", "func initService() {\n\tauthSvc = permit.New(conf.Conf.Auth)\n\tsvc = service.New()\n}", "func NewService(t mockConstructorTestingTNewService) *Service {\n\tmock := &Service{}\n\tmock.Mock.Test(t)\n\n\tt.Cleanup(func() { mock.AssertExpectations(t) })\n\n\treturn mock\n}", "func newSystemService(rpc RpcClient) (*SystemService, error) {\r\n\tif rpc == nil {\r\n\t\treturn nil, errors.New(\"rpc client cannot be nil\")\r\n\t}\r\n\r\n\treturn &SystemService{\r\n\t\trpc: rpc,\r\n\t}, nil\r\n}", "func (mr *MockapprunnerDescriberMockRecorder) Service() *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Service\", reflect.TypeOf((*MockapprunnerDescriber)(nil).Service))\n}", "func TestTokenSignature(t *testing.T) {\n\tdb.InitDB()\n\tvar router *gin.Engine = routes.SetupRouter()\n\n\tvar user models.UserCreate = utils.CreateUser(\"Tom\", \"qwerty1234\", t, router)\n\tuser.Token = utils.ConnectUser(\"Tom\", \"qwerty1234\", t, router)\n\n\tsplittedToken := strings.Split(user.Token, \".\")\n\tif len(splittedToken) != 3 {\n\t\tlog.Fatal(\"Bad token.\")\n\t\tt.Fail()\n\t}\n\n\theader := splittedToken[0]\n\tpayload := splittedToken[1]\n\tsignature := splittedToken[2]\n\tsignature = signature + \"modif\"\n\tmodifiedToken := header + \".\" + payload + \".\" + signature\n\n\tvar url string = \"/v1/user/\" + strconv.Itoa(user.ID)\n\tvar bearer = \"Bearer \" + modifiedToken\n\trecord := httptest.NewRecorder()\n\trequest, _ := http.NewRequest(\"GET\", url, nil)\n\trequest.Header.Add(\"Content-Type\", \"application/json\")\n\trequest.Header.Add(\"Authorization\", bearer)\n\n\trouter.ServeHTTP(record, request)\n\n\tvar message Message\n\terr := json.Unmarshal([]byte(record.Body.String()), &message)\n\tif err != nil {\n\t\tlog.Fatal(\"Bad output: \", err.Error())\n\t\tt.Fail()\n\t}\n\n\tassert.Equal(t, record.Code, 403)\n\tassert.Equal(t, message.Message, \"Bad signature\")\n\n\tuser.Token = utils.ConnectUser(\"Tom\", \"qwerty1234\", t, router)\n\tutils.CleanUser(user.ID, user.Token, t, router)\n\tdb.CloseDB()\n}", "func NewMockService(transport *http.Transport, aurl string, rurl string, surl string) Service {\n\n\treturn Service{\n\t\tclient: &http.Client{\n\t\t\tTransport: transport,\n\t\t},\n\t\tauthURL: aurl,\n\t\tregistryURL: rurl,\n\t\tserviceURL: surl,\n\t}\n}", "func TestMayaAnyK8sGenerateService(t *testing.T) {\n\ttests := map[string]struct {\n\t\tkind string\n\t\tyaml string\n\t\tisError bool\n\t}{\n\t\t\"blank service\": {kind: \"Service\", yaml: \"\", isError: true},\n\t\t\"hello service\": {kind: \"Service\", yaml: \"Hello!!\", isError: true},\n\t\t\"invalid service\": {kind: \"blah\", yaml: \"Junk!!\", isError: true},\n\t\t\"valid service\": {\n\t\t\tkind: \"Service\",\n\t\t\tisError: false,\n\t\t\tyaml: `\napiVersion: v1\nkind: Service\nmetadata:\n name: my-service\nspec:\n ports:\n - name: api\n port: 5656\n protocol: TCP\n targetPort: 5656\n selector:\n name: maya-apiserver\n sessionAffinity: None\n`},\n\t}\n\n\tfor name, test := range tests {\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\tma := &MayaAnyK8s{\n\t\t\t\tKind: test.kind,\n\t\t\t\tMayaYaml: MayaYaml{\n\t\t\t\t\tYaml: test.yaml,\n\t\t\t\t},\n\t\t\t}\n\t\t\ts, err := ma.GenerateService()\n\n\t\t\tif !test.isError && err != nil {\n\t\t\t\tt.Fatalf(\"Expected: 'no error' Actual: '%s'\", err)\n\t\t\t}\n\n\t\t\tif test.isError && s != nil {\n\t\t\t\tt.Fatalf(\"Expected: 'nil service' Actual: '%v'\", s)\n\t\t\t}\n\t\t})\n\t}\n}", "func TestServices(t *testing.T) { check.TestingT(t) }", "func TestServiceWithoutLogger(t *testing.T) {\n\ts := res.NewService(\"test\")\n\ts.SetLogger(nil)\n\ts.Handle(\"model\", res.GetResource(func(r res.GetRequest) { r.NotFound() }))\n\tsession := restest.NewSession(t, s, restest.WithKeepLogger)\n\tdefer session.Close()\n}", "func newService(namespace, name string) *v1.Service {\n\treturn &v1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: name,\n\t\t\tNamespace: namespace,\n\t\t\tLabels: labelMap(),\n\t\t},\n\t\tSpec: v1.ServiceSpec{\n\t\t\tSelector: labelMap(),\n\t\t\tPorts: []v1.ServicePort{\n\t\t\t\t{Name: \"port-1338\", Port: 1338, Protocol: \"TCP\", TargetPort: intstr.FromInt(1338)},\n\t\t\t\t{Name: \"port-1337\", Port: 1337, Protocol: \"TCP\", TargetPort: intstr.FromInt(1337)},\n\t\t\t},\n\t\t},\n\t}\n\n}", "func testDefBanner(t *testing.T, s *Service) {\n\tres, err := s.DefBanner(context.TODO())\n\tif err != nil {\n\t\tt.Logf(\"testDefBanner error(%v) \\n\", err)\n\t\treturn\n\t}\n\tt.Logf(\"testDefBanner res: %+v \\n\", res)\n}", "func Service(err error) *ErrService { return &ErrService{err, ServceUnavailable} }", "func serviceMutateFn(ctx context.Context, newService *corev1.Service, client client.Client) func() error {\n\treturn func() error {\n\t\t// TODO: probably nothing has to be done, check v1 implementation of CreateOrUpdate\n\t\t//existingService := existing.(*corev1.Service)\n\t\t//if !reflect.DeepEqual(newService.Spec, existingService.Spec) {\n\t\t//\treturn client.Update(ctx, existing)\n\t\t//}\n\t\treturn nil\n\t}\n}", "func (p *PodmanTestIntegration) StartRemoteService() {\n}", "func TestSignContractSuccess(t *testing.T) {\n\tsignatureHelper(t, false)\n}", "func (suite *AuthSuite) TestAuthUnknownServiceMember() {\n\t// Set up: Prepare the session, goth.User, callback handler, http response\n\t// and request, landing URL, and pass them into authorizeUnknownUser\n\n\thandlerConfig := suite.HandlerConfig()\n\tappnames := handlerConfig.AppNames()\n\n\t// Prepare the session and session manager\n\tfakeToken := \"some_token\"\n\tsession := auth.Session{\n\t\tApplicationName: auth.MilApp,\n\t\tIDToken: fakeToken,\n\t\tHostname: appnames.MilServername,\n\t}\n\tsessionManager := handlerConfig.SessionManagers().Mil\n\tmockSender := setUpMockNotificationSender() // We should get an email for this activity\n\n\t// Prepare the goth.User to simulate the UUID and email that login.gov would\n\t// provide\n\tfakeUUID, _ := uuid.NewV4()\n\tuser := goth.User{\n\t\tUserID: fakeUUID.String(),\n\t\tEmail: \"[email protected]\",\n\t}\n\tctx := suite.SetupSessionContext(context.Background(), &session, sessionManager)\n\n\t// Call the function under test\n\tresult := authorizeUnknownUser(ctx, suite.AppContextWithSessionForTest(&session), user,\n\t\tsessionManager, mockSender)\n\tsuite.Equal(authorizationResultAuthorized, result)\n\tmockSender.(*mocks.NotificationSender).AssertNumberOfCalls(suite.T(), \"SendNotification\", 1)\n\n\t// Look up the user and service member in the test DB\n\tfoundUser, _ := models.GetUserFromEmail(suite.DB(), user.Email)\n\tserviceMemberID := session.ServiceMemberID\n\tserviceMember, _ := models.FetchServiceMemberForUser(suite.DB(), &session, serviceMemberID)\n\t// Look up the session token in the session store (this test uses the memory store)\n\tsessionStore := sessionManager.Store()\n\t_, existsBefore, _ := sessionStore.Find(foundUser.CurrentMilSessionID)\n\n\t// Verify service member exists and its ID is populated in the session\n\tsuite.NotEmpty(session.ServiceMemberID)\n\n\t// Verify session contains UserID that points to the newly-created user\n\tsuite.Equal(foundUser.ID, session.UserID)\n\n\t// Verify user's LoginGovEmail and LoginGovUUID match the values passed in\n\tsuite.Equal(user.Email, foundUser.LoginGovEmail)\n\tsuite.Equal(user.UserID, foundUser.LoginGovUUID.String())\n\n\t// Verify that the user's CurrentMilSessionID is not empty. The value is\n\t// generated randomly, so we can't test for a specific string. Any string\n\t// except an empty string is acceptable.\n\tsuite.NotEqual(\"\", foundUser.CurrentMilSessionID)\n\n\t// Verify the session token also exists in the session store\n\tsuite.Equal(true, existsBefore)\n\n\t// Verify the service member that was created is associated with the user\n\t// that was created\n\tsuite.Equal(foundUser.ID, serviceMember.UserID)\n}", "func hello(w http.ResponseWriter, r *http.Request) {\n\tw.Write([]byte(fake.Hello()))\n}", "func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handlers, partitionID, endpoint, signingRegion, signingName, resolvedRegion string) *InputService1ProtocolTest {\n\tsvc := &InputService1ProtocolTest{\n\t\tClient: client.New(\n\t\t\tcfg,\n\t\t\tmetadata.ClientInfo{\n\t\t\t\tServiceName: \"InputService1ProtocolTest\",\n\t\t\t\tServiceID: \"InputService1ProtocolTest\",\n\t\t\t\tSigningName: signingName,\n\t\t\t\tSigningRegion: signingRegion,\n\t\t\t\tPartitionID: partitionID,\n\t\t\t\tEndpoint: endpoint,\n\t\t\t\tAPIVersion: \"\",\n\t\t\t\tResolvedRegion: resolvedRegion,\n\t\t\t\tJSONVersion: \"1.1\",\n\t\t\t\tTargetPrefix: \"com.amazonaws.foo\",\n\t\t\t},\n\t\t\thandlers,\n\t\t),\n\t}\n\n\t// Handlers\n\tsvc.Handlers.Sign.PushBackNamed(v4.SignRequestHandler)\n\tsvc.Handlers.Build.PushBackNamed(jsonrpc.BuildHandler)\n\tsvc.Handlers.Unmarshal.PushBackNamed(jsonrpc.UnmarshalHandler)\n\tsvc.Handlers.UnmarshalMeta.PushBackNamed(jsonrpc.UnmarshalMetaHandler)\n\tsvc.Handlers.UnmarshalError.PushBackNamed(jsonrpc.UnmarshalErrorHandler)\n\n\treturn svc\n}", "func Test_IndexHandler(t *testing.T) {\n\tvar (\n\t\tversionMsg Service\n\t\tresp *http.Response\n\t)\n\n\tsvc := NewService()\n\n\tts := httptest.NewServer(svc.NewRouter(\"*\"))\n\tdefer ts.Close()\n\n\treq, _ := http.NewRequest(\"GET\", ts.URL+\"/\", nil)\n\n\toutputLog := helpers.CaptureOutput(func() {\n\t\tresp, _ = http.DefaultClient.Do(req)\n\t})\n\n\tif got, want := resp.StatusCode, 200; got != want {\n\t\tt.Fatalf(\"Invalid status code, got %d but want %d\", got, want)\n\t}\n\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tt.Fatalf(\"Got an error when reading body: %s\", err.Error())\n\t}\n\n\terr = json.Unmarshal(data, &versionMsg)\n\tif err != nil {\n\t\tt.Fatalf(\"Got an error when parsing json: %s\", err.Error())\n\t}\n\tif got, want := versionMsg.Version, svc.Version; got != want {\n\t\tt.Fatalf(\"Wrong version return, got %s but want %s\", got, want)\n\t}\n\tif got, want := versionMsg.Name, svc.Name; got != want {\n\t\tt.Fatalf(\"Wrong version return, got %s but want %s\", got, want)\n\t}\n\n\tmatched, err := regexp.MatchString(`uri=/ `, outputLog)\n\tif matched != true || err != nil {\n\t\tt.Fatalf(\"request is not logged :\\n%s\", outputLog)\n\t}\n}", "func newService(repo Repository) Service {\n\n\tif repo == nil {\n\t\treturn nil\n\t}\n\treturn &service{repo}\n}", "func (_m *MockDispatchServer) mustEmbedUnimplementedDispatchServer() {\n\t_m.Called()\n}", "func MustService(connectURL string) *Service {\n\trv, err := NewService(connectURL)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn rv\n}", "func GetService(nbmaster string, httpClient *http.Client, jwt string, host string, hostUuid string, serviceName string) {\r\n fmt.Printf(\"\\nGet NetBackup service %s on %s...\\n\\n\", serviceName, host)\r\n\r\n uri := \"https://\" + nbmaster + \":\" + port + \"/netbackup/admin/hosts/\" + hostUuid + \"/services/\" + serviceName\r\n\r\n request, _ := http.NewRequest(http.MethodGet, uri, nil)\r\n request.Header.Add(\"Authorization\", jwt);\r\n request.Header.Add(\"Content-Type\", contentTypeV3);\r\n\r\n response, err := httpClient.Do(request)\r\n\r\n if err != nil {\r\n fmt.Printf(\"The HTTP request failed with error: %s\\n\", err)\r\n panic(\"Unable to get services\")\r\n } else {\r\n if response.StatusCode == 200 {\r\n resp, _ := ioutil.ReadAll(response.Body)\r\n var obj interface{}\r\n json.Unmarshal(resp, &obj)\r\n service := obj.(map[string]interface{})[\"data\"].(map[string]interface{})\r\n\r\n fmt.Printf(\"id status\\n\");\r\n fmt.Printf(\"============.=========\\n\");\r\n id := (service)[\"id\"]\r\n status := ((service)[\"attributes\"]).(map[string]interface{})[\"status\"]\r\n\r\n fmt.Printf(\"%-12s %s\\n\", id, status);\r\n } else {\r\n printErrorResponse(response)\r\n }\r\n }\r\n}", "func getServiceName(service *apiv1.Service) string {\n\treturn fmt.Sprintf(\"%s/%s\", service.Namespace, service.Name)\n}", "func TestMayaAnyK8sGenerateServiceTemplated(t *testing.T) {\n\ttests := map[string]MockMayaAnyK8s{\n\t\t\"templated service\": {\n\t\t\tkind: \"Service\",\n\t\t\tapiVersion: \"v1\",\n\t\t\towner: \"pv-123-abc\",\n\t\t\tsuffixName: \"-svc\",\n\t\t\tisError: false,\n\t\t\tyaml: `\napiVersion: {{.APIVersion}}\nkind: {{.Kind}}\nmetadata:\n name: {{.Owner}}-svc\nspec:\n ports:\n - name: api\n port: 5656\n protocol: TCP\n targetPort: 5656\n selector:\n name: maya-apiserver\n sessionAffinity: None\n`},\n\t}\n\n\tfor name, mock := range tests {\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\tma := mock.NewMayaAnyK8s()\n\t\t\ts, err := ma.GenerateService()\n\n\t\t\tif mock.isError && s != nil {\n\t\t\t\tt.Fatalf(\"Expected: 'nil service' Actual: '%v'\", s)\n\t\t\t}\n\t\t\tmock.TestObjectMeta(s.ObjectMeta, err, t)\n\t\t\tmock.TestTypeMeta(s.TypeMeta, err, t)\n\t\t})\n\t}\n}", "func newInputService1ProtocolTestClient(cfg aws.Config, handlers request.Handlers, endpoint, signingRegion string) *InputService1ProtocolTest {\n\tsvc := &InputService1ProtocolTest{\n\t\tClient: client.New(\n\t\t\tcfg,\n\t\t\tmetadata.ClientInfo{\n\t\t\t\tServiceName: \"inputservice1protocoltest\",\n\t\t\t\tSigningRegion: signingRegion,\n\t\t\t\tEndpoint: endpoint,\n\t\t\t\tAPIVersion: \"2014-01-01\",\n\t\t\t},\n\t\t\thandlers,\n\t\t),\n\t}\n\n\t// Handlers\n\tsvc.Handlers.Sign.PushBack(v4.Sign)\n\tsvc.Handlers.Build.PushBackNamed(restxml.BuildHandler)\n\tsvc.Handlers.Unmarshal.PushBackNamed(restxml.UnmarshalHandler)\n\tsvc.Handlers.UnmarshalMeta.PushBackNamed(restxml.UnmarshalMetaHandler)\n\tsvc.Handlers.UnmarshalError.PushBackNamed(restxml.UnmarshalErrorHandler)\n\n\treturn svc\n}", "func UnsecuredService(serviceName string) *goa.Service {\n\tsvc := goa.New(serviceName)\n\tsvc.Context = tokencontext.ContextWithTokenManager(svc.Context, TokenManager)\n\treturn svc\n}", "func newBackingServices(c *Client, namespace string) *backingservices {\n\treturn &backingservices{\n\t\tr: c,\n\t\tns: namespace,\n\t}\n}", "func serviceClean(svc corev1.Service) corev1.Service {\n\tsvc.Spec.ExternalIPs = []string{}\n\tsvc.Spec.LoadBalancerIP = \"\"\n\tsvc.Spec.LoadBalancerSourceRanges = []string{}\n\tsvc.Spec.ExternalName = \"\"\n\tsvc.Status.LoadBalancer = corev1.LoadBalancerStatus{}\n\treturn svc\n}", "func (t *liverpc) generateClient(file *descriptor.FileDescriptorProto, service *descriptor.ServiceDescriptorProto) {\n\tclientName := clientName(service)\n\tstructName := unexported(clientName)\n\tnewClientFunc := \"New\" + clientName\n\n\tt.P(`type `, structName, ` struct {`)\n\tt.P(` client *liverpc.Client`)\n\tt.P(`}`)\n\tt.P()\n\tt.P(`// `, newClientFunc, ` creates a client that implements the `, clientName, ` interface.`)\n\tt.P(`func `, newClientFunc, `(client *liverpc.Client) `, clientName, ` {`)\n\tt.P(` return &`, structName, `{`)\n\tt.P(` client: client,`)\n\tt.P(` }`)\n\tt.P(`}`)\n\tt.P()\n\n\tfor _, method := range service.Method {\n\t\tmethName := methodName(method)\n\t\tpkgName := pkgName(file)\n\n\t\tinputType := t.goTypeName(method.GetInputType())\n\t\toutputType := t.goTypeName(method.GetOutputType())\n\n\t\tparts := strings.Split(pkgName, \".\")\n\t\tif len(parts) < 2 {\n\t\t\tpanic(\"package name must contain at least to parts, eg: service.v1, get \" + pkgName + \"!\")\n\t\t}\n\t\tvStr := parts[len(parts)-1]\n\t\tif len(vStr) < 2 {\n\t\t\tpanic(\"package name must contain a valid version, eg: service.v1\")\n\t\t}\n\t\t_, err := strconv.Atoi(vStr[1:])\n\t\tif err != nil {\n\t\t\tpanic(\"package name must contain a valid version, eg: service.v1, get \" + vStr)\n\t\t}\n\n\t\trpcMethod := method.GetName()\n\t\trpcCtrl := service.GetName()\n\t\trpcCmd := rpcCtrl + \".\" + rpcMethod\n\n\t\tt.P(`func (c *`, structName, `) `, methName, `(ctx `, t.pkgs[\"context\"], `.Context, in *`, inputType, `, opts ...liverpc.CallOption) (*`, outputType, `, error) {`)\n\t\tt.P(` out := new(`, outputType, `)`)\n\t\tt.P(` err := doRPCRequest(ctx,c.client, `, vStr[1:], `, \"`, rpcCmd, `\", in, out, opts)`)\n\t\tt.P(` if err != nil {`)\n\t\tt.P(` return nil, err`)\n\t\tt.P(` }`)\n\t\tt.P(` return out, nil`)\n\t\tt.P(`}`)\n\t\tt.P()\n\t}\n}", "func TestSignContractFailure(t *testing.T) {\n\tsignatureHelper(t, true)\n}", "func (service *rpcServiceMap) callService(mtype *rpcMethodType, argv,\n replyv reflect.Value) error {\n\n function := mtype.method.Func\n // Invoke the method, providing a new value for the reply.\n returnValues := function.Call([]reflect.Value{service.rcvr, argv, replyv})\n // The return value for the method is an error.\n errInter := returnValues[0].Interface()\n errmsg := \"\"\n if errInter != nil {\n errmsg = errInter.(error).Error()\n return fmt.Errorf(errmsg)\n }\n return nil\n}", "func (mr *MockProviderMockRecorder) Service(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Service\", reflect.TypeOf((*MockProvider)(nil).Service), arg0)\n}", "func (p *printer) Service(service *descriptor.ServiceDescriptorProto, methodIndex int) {\n\tp.MaybeLeadingComments(service)\n\tdefer p.open(\"service %s\", service.GetName())()\n\n\tif methodIndex < 0 {\n\t\tfor i := range service.Method {\n\t\t\tp.Method(service.Method[i])\n\t\t}\n\t} else {\n\t\tp.Method(service.Method[methodIndex])\n\t\tif len(service.Method) > 1 {\n\t\t\tp.Printf(\"// other methods were omitted.\\n\")\n\t\t}\n\t}\n}", "func BenchmarkService(b *testing.B) {\n\tfp := protofile\n\n\treader, _ := os.Open(fp)\n\tdefer reader.Close()\n\n\tparser := proto.NewParser(reader)\n\tdefinition, _ := parser.Parse()\n\n\tvar (\n\t\tcurrentService *proto.Service\n\t\tcurrentTemplateMap map[string]map[string]string\n\t)\n\n\thandlerService := proto.WithService(func(s *proto.Service) {\n\t\tclearRandoms()\n\t\tif rpc != \"\" && s.Name != rpc {\n\t\t\treturn\n\t\t}\n\t\tcurrentService = s\n\t\tif m, ok := Templates[s.Name]; ok {\n\t\t\tcurrentTemplateMap = m\n\t\t}\n\t})\n\n\thandlerRPC := proto.WithRPC(func(r *proto.RPC) {\n\t\tfor _, action := range []string{\"create\", \"update\", \"delete\"} {\n\t\t\tif m, ok := currentTemplateMap[action]; ok {\n\t\t\t\tif t, ok := m[r.Name]; ok {\n\n\t\t\t\t\tb.Run(fmt.Sprintf(\"%s\", r.Name), func(b *testing.B) {\n\t\t\t\t\t\t// if ovs { // skip unvalid service: useful when various services are defined in a single proto file\n\t\t\t\t\t\t// \ttokens := strings.Split(service, \".\")\n\t\t\t\t\t\t// \tsuffix := tokens[len(tokens)-1]\n\t\t\t\t\t\t// \ttcr := strings.ToLower(currentService)\n\t\t\t\t\t\t// \tif !strings.HasPrefix(tcr, suffix) {\n\t\t\t\t\t\t// \t\treturn\n\t\t\t\t\t\t// \t}\n\t\t\t\t\t\t// }\n\n\t\t\t\t\t\t// Gather error count\n\t\t\t\t\t\tvar gotError chan bool\n\t\t\t\t\t\tvar done chan int\n\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\tgotError = make(chan bool)\n\t\t\t\t\t\t\tdone = make(chan int)\n\t\t\t\t\t\t\tgo func() {\n\t\t\t\t\t\t\t\terrNb := 0\n\t\t\t\t\t\t\t\tfor {\n\t\t\t\t\t\t\t\t\tselect {\n\t\t\t\t\t\t\t\t\tcase nb := <-done:\n\t\t\t\t\t\t\t\t\t\tresults := float32(errNb) / float32(nb) * 100\n\t\t\t\t\t\t\t\t\t\tfmt.Printf(\"Benchmarking %s - %.00f%% error rate\\n\", currentService.Name+\".\"+r.Name, results)\n\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\tcase <-gotError:\n\t\t\t\t\t\t\t\t\t\terrNb++\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}()\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t// Retrieve the test template\n\t\t\t\t\t\ttmpl, err := template.New(r.Name).Parse(t)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tfor i := 0; i < b.N; i++ {\n\t\t\t\t\t\t\tif r.StreamsRequest && !r.StreamsReturns {\n\t\t\t\t\t\t\t\terr := benchStreamClient(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else if !r.StreamsRequest && r.StreamsReturns {\n\t\t\t\t\t\t\t\terr := benchStreamServer(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\terr := benchCall(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\tdone <- b.N\n\t\t\t\t\t\t}\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t})\n\n\tproto.Walk(definition, handlerService, handlerRPC)\n}", "func createPhpService() {\n\tserviceData := &apiv1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: \"test-server\",\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"name\": \"test-server\",\n\t\t\t},\n\t\t},\n\t\tSpec: apiv1.ServiceSpec{\n\t\t\tPorts: []apiv1.ServicePort{\n\t\t\t\t{\n\t\t\t\t\tPort: 9000,\n\t\t\t\t\tProtocol: apiv1.ProtocolTCP,\n\t\t\t\t\tTargetPort: intstr.IntOrString{\n\t\t\t\t\t\tIntVal: 9000,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\tSelector: map[string]string{\n\t\t\t\t\"app\": \"test-server\",\n\t\t\t},\n\t\t},\n\t}\n\tcreateService(serviceData)\n}", "func newService(cr *argoprojv1a1.ArgoCD) *corev1.Service {\n\treturn &corev1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: cr.Name,\n\t\t\tNamespace: cr.Namespace,\n\t\t\tLabels: argoutil.LabelsForCluster(cr),\n\t\t},\n\t}\n}", "func (m *MockConsulHealth) Service(service, tag string, passingOnly bool, q *api.QueryOptions) (entries []*api.ServiceEntry, meta *api.QueryMeta, err error) {\n\n\targs := m.Called(service, tag, passingOnly, q)\n\n\tentries = nil\n\tmeta = nil\n\terr = args.Error(2)\n\n\tif e := args.Get(0); e != nil {\n\t\tentries = e.(func() []*api.ServiceEntry)()\n\t}\n\n\tif m := args.Get(1); m != nil {\n\t\tmeta = m.(*api.QueryMeta)\n\t}\n\n\treturn\n}", "func newService(c *onet.Context) (onet.Service, error) {\n\ts := &Service{\n\t\tServiceProcessor: onet.NewServiceProcessor(c),\n\t}\n\ts.RegisterProcessorFunc(cosiSendRawID, s.HandleRaw)\n\t_, err := c.ProtocolRegister(protoName, s.NewDefaultProtocol)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := s.RegisterHandlers(s.GenesisTx, s.Setup, s.TreesBLSCoSi, s.MemoryAllocated); err != nil {\n\t\treturn nil, errors.New(\"Couldn't register messages\")\n\t}\n\n\ts.propagateF, s.mypi, err = propagate.NewPropagationFunc(c, \"Propagate\", s.propagateHandler, -1)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ts.trees = make(map[onet.TreeID]*onet.Tree)\n\ts.coinToAtomic = make(map[string]int)\n\ts.atomicCoinReserved = make([]int32, 0)\n\n\tdb, bucketNameTx := s.GetAdditionalBucket([]byte(\"Tx\"))\n\t_, bucketNameLastTx := s.GetAdditionalBucket([]byte(\"LastTx\"))\n\ts.bucketNameTx = bucketNameTx\n\ts.bucketNameLastTx = bucketNameLastTx\n\ts.db = db\n\treturn s, nil\n}", "func CreateServiceStatusRequest() (request *ServiceStatusRequest) {\nrequest = &ServiceStatusRequest{\nRpcRequest: &requests.RpcRequest{},\n}\nrequest.InitWithApiInfo(\"Yundun\", \"2015-04-16\", \"ServiceStatus\", \"yundun\", \"openAPI\")\nreturn\n}", "func Test_DeviceService_Get_EmptyIP(t *testing.T) {\n\ts := DeviceService{}\n\t_, err := s.Get(\"\")\n\tassert.Error(t, err)\n}", "func (mr *MockStorageMockRecorder) GetService(srv interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"GetService\", reflect.TypeOf((*MockStorage)(nil).GetService), srv)\n}", "func (g *grpc) generateClientSignature(servName string, method *pb.MethodDescriptorProto) string {\n\torigMethName := method.GetName()\n\tmethName := generator.CamelCase(origMethName)\n\tif reservedClientName[methName] {\n\t\tmethName += \"_\"\n\t}\n\treqArg := \", in *\" + g.typeName(method.GetInputType())\n\tif method.GetClientStreaming() {\n\t\treqArg = \"\"\n\t}\n\trespName := \"*\" + g.typeName(method.GetOutputType())\n\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\trespName = servName + \"_\" + generator.CamelCase(origMethName) + \"Client\"\n\t}\n\treturn fmt.Sprintf(\"%s(ctx %s.Context%s, opts ...%s.CallOption) (%s, error)\", methName, contextPkg, reqArg, grpcPkg, respName)\n}", "func TestNewBeerService_ReturnsBeerInteractor(t *testing.T) {\n\tt.Parallel()\n\tinteractor := &mocks.BeerInteractor{}\n\tassert.NotNil(t, adapters.NewBeerService(interactor))\n}", "func (s *Service) Listen() {\n\n}", "func New() servicespec.FSService {\n\treturn &service{}\n}", "func newService(serviceName string) *Service {\n\treturn &Service{\n\t\tpluginDir: serverless.PluginDir,\n\t\tname: serviceName,\n\t\tinterf: nil,\n\t}\n}", "func TestServiceWith_WithoutMatchingPattern(t *testing.T) {\n\trunTest(t, func(s *res.Service) {\n\t\ts.Handle(\"collection\", res.GetResource(func(r res.GetRequest) { r.NotFound() }))\n\t}, func(s *restest.Session) {\n\t\terr := s.Service().With(\"test.model\", func(r res.Resource) {})\n\t\tif err == nil {\n\t\t\tt.Errorf(\"expected With to return an error, but it didn't\")\n\t\t}\n\t})\n}", "func testIATBHServiceClassCode(t testing.TB) {\n\tbh := mockIATBatchHeaderFF()\n\tbh.ServiceClassCode = 0\n\terr := bh.Validate()\n\tif !base.Match(err, ErrFieldInclusion) {\n\t\tt.Errorf(\"%T: %s\", err, err)\n\t}\n}", "func (r *apiRegister) DeleteService(svc *register.Service) error {\n\treturn fmt.Errorf(\"Not Implemented\")\n}", "func (a *FastlyIntegrationApi) CreateFastlyService(ctx _context.Context, accountId string, body FastlyServiceRequest) (FastlyServiceResponse, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarReturnValue FastlyServiceResponse\n\t)\n\n\tlocalBasePath, err := a.Client.Cfg.ServerURLWithContext(ctx, \"v2.FastlyIntegrationApi.CreateFastlyService\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, datadog.GenericOpenAPIError{ErrorMessage: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v2/integrations/fastly/accounts/{account_id}/services\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"account_id\"+\"}\", _neturl.PathEscape(datadog.ParameterToString(accountId, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\tlocalVarHeaderParams[\"Content-Type\"] = \"application/json\"\n\tlocalVarHeaderParams[\"Accept\"] = \"application/json\"\n\n\t// body params\n\tlocalVarPostBody = &body\n\tdatadog.SetAuthKeys(\n\t\tctx,\n\t\t&localVarHeaderParams,\n\t\t[2]string{\"apiKeyAuth\", \"DD-API-KEY\"},\n\t\t[2]string{\"appKeyAuth\", \"DD-APPLICATION-KEY\"},\n\t)\n\treq, err := a.Client.PrepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, nil)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.Client.CallAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := datadog.ReadBody(localVarHTTPResponse)\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 || localVarHTTPResponse.StatusCode == 403 || localVarHTTPResponse.StatusCode == 404 || localVarHTTPResponse.StatusCode == 429 {\n\t\t\tvar v APIErrorResponse\n\t\t\terr = a.Client.Decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.ErrorModel = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.Client.Decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}" ]
[ "0.6303204", "0.6056114", "0.6040024", "0.56926787", "0.56186223", "0.5607778", "0.5566932", "0.5545187", "0.54975945", "0.54929286", "0.54550153", "0.5408207", "0.53727037", "0.5369734", "0.5344562", "0.5338126", "0.53317064", "0.53117275", "0.5292198", "0.52864736", "0.5285324", "0.52824163", "0.5261824", "0.5260987", "0.52526444", "0.5248278", "0.5240406", "0.5231008", "0.52099866", "0.5201575", "0.52012426", "0.51964533", "0.5192325", "0.51847935", "0.5177962", "0.5176621", "0.517287", "0.5154695", "0.51275367", "0.5122051", "0.5120173", "0.5119281", "0.5094479", "0.5085922", "0.5076365", "0.5073792", "0.5059092", "0.5056611", "0.50555336", "0.5024775", "0.5009542", "0.50075054", "0.5005966", "0.5003163", "0.49936068", "0.49934477", "0.4987011", "0.4984404", "0.49800897", "0.4975268", "0.49744362", "0.49724674", "0.4971575", "0.49691233", "0.49666283", "0.49630454", "0.4961084", "0.49578595", "0.49543184", "0.49512342", "0.4949683", "0.4947106", "0.4946378", "0.49431077", "0.49409312", "0.49399793", "0.49365973", "0.49363104", "0.4929963", "0.49278474", "0.4927802", "0.49246264", "0.4921552", "0.492047", "0.49204326", "0.49195743", "0.49183127", "0.4918284", "0.4911952", "0.4911797", "0.49109662", "0.4910017", "0.49027", "0.48987305", "0.48844287", "0.48804983", "0.4872498", "0.48721012", "0.48683107", "0.48629624" ]
0.6982023
0
The fake service knows nothing about method signatures.
func (fs fakeService) MethodSignature(ctx *context.T, call rpc.ServerCall, method string) (signature.Method, error) { ctx.Infof("Fake Service Method Signature???") return signature.Method{}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (fs fakeService) Signature(ctx *context.T, call rpc.ServerCall) ([]signature.Interface, error) {\n\tctx.Infof(\"Fake Service Signature???\")\n\treturn nil, nil\n}", "func (fs fakeService) Invoke(ctx *context.T, call rpc.StreamServerCall, method string, argptrs []interface{}) (results []interface{}, _ error) {\n\t// fs.suffix consists of the mojo url and the application/interface name.\n\t// The last part should be the name; everything else is the url.\n\tparts := strings.Split(fs.suffix, \"/\")\n\tmojourl := strings.Join(parts[:len(parts)-1], \"/\") // e.g., mojo:go_remote_echo_server. May be defined in a BUILD.gn file.\n\tmojoname := parts[len(parts)-1] // e.g., mojo::examples::RemoteEcho. Defined from the interface + module.\n\n\t// Create the generic message pipe. r is a bindings.InterfaceRequest, and\n\t// p is a bindings.InterfacePointer.\n\tr, p := bindings.CreateMessagePipeForMojoInterface()\n\tv := v23ServiceRequest{\n\t\trequest: r,\n\t\tname: mojoname,\n\t} // v is an application.ServiceRequest with mojoname\n\n\t// Connect to the mojourl.\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&v)\n\n\t// Then assign a new router the FakeService.\n\t// This will never conflict because each FakeService is only invoked once.\n\tfs.router = bindings.NewRouter(p.PassMessagePipe(), bindings.GetAsyncWaiter())\n\tdefer fs.Close_Proxy()\n\n\tctx.Infof(\"Fake Service Invoke (Remote Signature: %q -- %q)\", mojourl, mojoname)\n\n\t// Vanadium relies on type information, so we will retrieve that first.\n\tmojomInterface, desc, err := fs.callRemoteSignature(mojourl, mojoname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Signature %v\", mojomInterface)\n\tctx.Infof(\"Fake Service Invoke (Remote Method: %v)\", method)\n\n\t// With the type information, we can make the method call to the remote interface.\n\tmethodResults, err := fs.callRemoteMethod(ctx, method, mojomInterface, desc, argptrs)\n\tif err != nil {\n\t\tctx.Errorf(\"Method called failed: %v\", err)\n\t\treturn nil, err\n\t}\n\n\tctx.Infof(\"Fake Service Invoke Results %v\", methodResults)\n\n\t// Convert methodResult to results.\n\tresults = make([]interface{}, len(methodResults))\n\tfor i := range methodResults {\n\t\tresults[i] = &methodResults[i]\n\t}\n\treturn results, nil\n}", "func (fs fakeService) callRemoteSignature(mojourl string, mojoname string) (mojomInterface mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, err error) {\n\t// TODO(afandria): The service_describer mojom file defines the constant, but\n\t// it is not actually present in the generated code:\n\t// https://github.com/domokit/mojo/issues/469\n\t// serviceDescriberInterfaceName := \"_ServiceDescriber\"\n\n\tr, p := service_describer.CreateMessagePipeForServiceDescriber()\n\tfs.appctx.ConnectToApplication(mojourl).ConnectToService(&r)\n\tsDescriber := service_describer.NewServiceDescriberProxy(p, bindings.GetAsyncWaiter())\n\tdefer sDescriber.Close_Proxy()\n\n\tr2, p2 := service_describer.CreateMessagePipeForServiceDescription()\n\terr = sDescriber.DescribeService(mojoname, r2)\n\tif err != nil {\n\t\treturn\n\t}\n\tsDescription := service_describer.NewServiceDescriptionProxy(p2, bindings.GetAsyncWaiter())\n\tdefer sDescription.Close_Proxy()\n\n\tmojomInterface, err = sDescription.GetTopLevelInterface()\n\tif err != nil {\n\t\treturn\n\t}\n\tdescPtr, err := sDescription.GetAllTypeDefinitions()\n\tif err != nil {\n\t\treturn\n\t}\n\treturn mojomInterface, *descPtr, nil\n}", "func SophisticatedResponder(w http.ResponseWriter, httpRequest *http.Request, fakeRequest *Request) {\n\tstatusCode := fakeRequest.Response.StatusCode\n\tbody := fakeRequest.Response.BodyBuffer\n\tresponseHeader := fakeRequest.Response.Header\n\n\tif len(fakeRequest.Header) > 0 {\n\t\ts, b, err := validateHeaders(fakeRequest.Header, httpRequest.Header)\n\t\tif err != nil {\n\t\t\tstatusCode = s\n\t\t\tbody = []byte(b)\n\t\t}\n\t}\n\tif len(fakeRequest.Cookies()) > 0 {\n\t\ts, b, err := validateCookies(fakeRequest.Cookies(), httpRequest.Cookies())\n\t\tif err != nil {\n\t\t\tstatusCode = s\n\t\t\tbody = []byte(b)\n\t\t}\n\t}\n\tif (len(responseHeader)) > 0 {\n\t\tfor k := range fakeRequest.Response.Header {\n\t\t\tw.Header().Add(k, responseHeader.Get(k))\n\t\t}\n\t}\n\tif statusCode > 0 {\n\t\tw.WriteHeader(statusCode)\n\t}\n\tserviceResponses := \"\"\n\tif len(fakeRequest.ServiceEndpoints) > 0 {\n\t\tfor _, uri := range fakeRequest.ServiceEndpoints {\n\t\t\tstatus, body, err := invokeServiceEndpoint(uri, httpRequest.Header)\n\t\t\tif err == nil {\n\t\t\t\tserviceResponses += (uri + \": \")\n\t\t\t\tserviceResponses += (status + \": \")\n\t\t\t\tserviceResponses += (body)\n\t\t\t\tserviceResponses += \"<br>\"\n\t\t\t}\n\t\t}\n\t}\n\tif (len(body)) > 0 {\n\t\tb := string(body)\n\t\tif len(fakeRequest.InjectionKeys) > 0 {\n\t\t\tfor _, k := range fakeRequest.InjectionKeys {\n\t\t\t\tif k == \"path\" {\n\t\t\t\t\tbody = []byte(fmt.Sprintf(b, strings.TrimPrefix(httpRequest.URL.Path, \"/\")))\n\t\t\t\t\tb = string(body)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif len(serviceResponses) > 0 {\n\t\t\tb += serviceResponses\n\t\t}\n\n\t\tif fakeRequest.RenderHTML {\n\t\t\tb = \"<html><head><title>fakeserver</title></head><body>\" + b + \"</body></html>\"\n\t\t}\n\n\t\tw.Write([]byte(b))\n\t}\n}", "func TestServiceMethodNamesCamelCase(t *testing.T) {\n\ts := httptest.NewServer(NewHaberdasherV1Server(&HaberdasherService{}, nil))\n\tdefer s.Close()\n\n\tclient := NewHaberdasherV1ProtobufClient(s.URL, http.DefaultClient)\n\n\that, err := client.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err != nil {\n\t\tt.Fatalf(\"go protobuf client err=%q\", err)\n\t}\n\tif hat.Size != 1 {\n\t\tt.Errorf(\"wrong hat size returned\")\n\t}\n}", "func (service *rpcServiceMap) callService(mtype *rpcMethodType, argv,\n replyv reflect.Value) error {\n\n function := mtype.method.Func\n // Invoke the method, providing a new value for the reply.\n returnValues := function.Call([]reflect.Value{service.rcvr, argv, replyv})\n // The return value for the method is an error.\n errInter := returnValues[0].Interface()\n errmsg := \"\"\n if errInter != nil {\n errmsg = errInter.(error).Error()\n return fmt.Errorf(errmsg)\n }\n return nil\n}", "func (r *rpcServerService) doCall(serviceMethod string, args []byte) ([]byte,\n error) {\n\n glog.V(3).Infof(\"rpc: doCall to %s\", serviceMethod)\n glog.V(4).Infof(\"rpc: doCall to %s with %v\", serviceMethod, args)\n\n dot := strings.LastIndex(serviceMethod, \".\")\n if dot < 0 {\n err := fmt.Errorf(\"rpc: service/method ill-formed: \" + serviceMethod)\n glog.Error(err)\n return nil, err\n }\n serviceName := serviceMethod[:dot]\n methodName := serviceMethod[dot+1:]\n // Look up the request.\n serviceInf, ok := r.serviceMap.Get(serviceName)\n if !ok || serviceInf == nil {\n err := errors.New(\"rpc: can't find service \" + serviceName)\n glog.Error(err)\n return nil, err\n }\n service, okType := serviceInf.(*rpcServiceMap)\n if !okType || service == nil {\n err := errors.New(\"rpc: unexpected type error for service \" + serviceName)\n glog.Error(err)\n return nil, err\n }\n mtype := service.method[methodName]\n if mtype == nil {\n err := errors.New(\"rpc: can't find method \" + serviceMethod)\n glog.Error(err)\n return nil, err\n }\n argv := reflect.New(mtype.argType)\n errJSON := json.Unmarshal(args, argv.Interface())\n if errJSON != nil {\n glog.Error(\"error in unmarshal: \", errJSON)\n return nil, errJSON\n }\n glog.V(4).Infof(\"rpc: json unmarshalled request is: %s -> %#v\", args, argv)\n replyv := reflect.New(mtype.replyType.Elem())\n\n glog.V(3).Infof(\"rpc: calling service %v method %v with %v\",\n service, mtype, argv)\n\n errCall := service.callService(mtype, argv.Elem(), replyv)\n if errCall != nil {\n glog.V(3).Infof(\"rpc call returned error: \", errCall)\n return nil, errCall\n }\n reply, errRep := json.Marshal(replyv.Interface())\n if errRep != nil {\n glog.Error(\"rpc reply marshall error: \", errRep)\n return nil, errRep\n }\n glog.V(3).Info(\"rpc reply: \", string(reply))\n return reply, nil\n}", "func (fs fakeService) callRemoteMethod(ctx *context.T, method string, mi mojom_types.MojomInterface, desc map[string]mojom_types.UserDefinedType, argptrs []interface{}) ([]*vom.RawBytes, error) {\n\t// We need to parse the signature result to get the method relevant info out.\n\tfound := false\n\tvar ordinal uint32\n\tfor ord, mm := range mi.Methods {\n\t\tif *mm.DeclData.ShortName == method {\n\t\t\tordinal = ord\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif !found {\n\t\treturn nil, fmt.Errorf(\"callRemoteMethod: method %s does not exist\", method)\n\t}\n\n\tmm := mi.Methods[ordinal]\n\n\t// A void function must have request id of 0, whereas one with response params\n\t// should have a unique request id.\n\theader := bindings.MessageHeader{\n\t\tType: ordinal,\n\t\tFlags: bindings.MessageExpectsResponseFlag,\n\t\tRequestId: fs.ids.Count(),\n\t}\n\n\t// Now produce the *bindings.Message that we will send to the other side.\n\tinType, err := transcoder.MojomStructToVDLType(mm.Parameters, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tmessage, err := encodeMessageFromVom(header, argptrs, inType)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Otherwise, make a generic call with the message.\n\toutMessage, err := fs.callRemoteWithResponse(ctx, message)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Decode the *vom.RawBytes from the mojom bytes and mojom type.\n\toutType, err := transcoder.MojomStructToVDLType(*mm.ResponseParams, desc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttarget := util.StructSplitTarget()\n\tif err := transcoder.FromMojo(target, outMessage.Payload, outType); err != nil {\n\t\treturn nil, fmt.Errorf(\"transcoder.FromMojo failed: %v\", err)\n\t}\n\treturn target.Fields(), nil\n}", "func svcHandler()", "func newService(rcvr interface{}, guard Guard) *service {\n\ts := new(service)\n\ts.typ = reflect.TypeOf(rcvr)\n\ts.rcvr = reflect.ValueOf(rcvr)\n\ts.name = reflect.Indirect(s.rcvr).Type().Name()\n\ts.guard = guard\n\n\t// install the methods\n\ts.method = suitableMethods(s.typ, true)\n\n\treturn s\n}", "func buildServiceFromMethods(descr *dpb.FileDescriptorProto, renderer *Renderer) (err error) {\n\tmethods := renderer.Model.Methods\n\tserviceName := findValidServiceName(descr.MessageType, strings.Title(renderer.Package))\n\n\tservice := &dpb.ServiceDescriptorProto{\n\t\tName: &serviceName,\n\t}\n\tdescr.Service = []*dpb.ServiceDescriptorProto{service}\n\n\tfor _, method := range methods {\n\t\tmOptionsDescr := &dpb.MethodOptions{}\n\t\trequestBody := getRequestBodyForRequestParameters(method.ParametersTypeName, renderer.Model.Types)\n\t\thttpRule := getHttpRuleForMethod(method, requestBody)\n\t\tif err := proto.SetExtension(mOptionsDescr, annotations.E_Http, &httpRule); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif method.ParametersTypeName == \"\" {\n\t\t\tmethod.ParametersTypeName = \"google.protobuf.Empty\"\n\t\t\tshouldRenderEmptyImport = true\n\t\t}\n\t\tif method.ResponsesTypeName == \"\" {\n\t\t\tmethod.ResponsesTypeName = \"google.protobuf.Empty\"\n\t\t\tshouldRenderEmptyImport = true\n\t\t}\n\n\t\tmDescr := &dpb.MethodDescriptorProto{\n\t\t\tName: &method.HandlerName,\n\t\t\tInputType: &method.ParametersTypeName,\n\t\t\tOutputType: &method.ResponsesTypeName,\n\t\t\tOptions: mOptionsDescr,\n\t\t}\n\n\t\tservice.Method = append(service.Method, mDescr)\n\t}\n\treturn nil\n}", "func (server *Server) callService(conn *ConnDriver, seq uint64, service *service, methodType *methodType, argv, replyv reflect.Value) {\n\tfunction := methodType.method.Func\n\n\treturnValues := function.Call([]reflect.Value{service.rcvr, argv, replyv})\n\t// The return value for the method is an error.\n\terrInter := returnValues[0].Interface()\n\n\trespHeader := NewResponseHeader()\n\trespHeader.ReplyType = ReplyTypeData\n\trespHeader.Seq = seq\n\tif errInter != nil {\n\t\tswitch errInter.(type) {\n\t\tcase *Error:\n\t\t\tserver.replyCmd(conn, seq, errInter.(*Error), CmdTypeErr)\n\t\tcase Error:\n\t\t\te := errInter.(Error)\n\t\t\tserver.replyCmd(conn, seq, &e, CmdTypeErr)\n\t\tcase error:\n\t\t\tserver.replyCmd(conn, seq, &Error{500, ErrTypeLogic, errInter.(error).Error()}, CmdTypeErr)\n\t\t}\n\t\treturn\n\t}\n\tconn.Lock()\n\terr := server.SendFrame(conn, respHeader, replyv)\n\tconn.Unlock()\n\tif err != nil && !isNetError(err) {\n\t\tlog.Fatalln(\"encoding error:\" + err.Error())\n\t}\n\treturn\n}", "func (s *Server) call(req *Request) *Response {\n\t// TODO: simplfy this function, or split into several functions\n\tdot := strings.LastIndex(req.Method, \".\") // split req.Method like \"type.Method\"\n\tif dot < 0 {\n\t\terr := errors.New(\"rpc: service/method request ill-formed: \" + req.Method)\n\t\treturn NewResponse(req.ID, nil, NewJsonrpcErr(ParseErr, err.Error(), err))\n\t}\n\n\tserviceName := req.Method[:dot]\n\tmethodName := req.Method[dot+1:]\n\n\t// method existed or not\n\tsvci, ok := s.m.Load(serviceName)\n\tif !ok {\n\t\terr := errors.New(\"rpc: can't find service \" + req.Method)\n\t\treturn NewResponse(req.ID, nil, NewJsonrpcErr(MethodNotFound, err.Error(), nil))\n\t}\n\tsvc := svci.(*service)\n\tmtype := svc.method[methodName]\n\tif mtype == nil {\n\t\terr := errors.New(\"rpc: can't find method \" + req.Method)\n\t\treturn NewResponse(req.ID, nil, NewJsonrpcErr(MethodNotFound, err.Error(), nil))\n\t}\n\n\t// to prepare argv and replyv in reflect.Value\n\t// ref to `net/http/rpc`\n\targIsValue := false // if true, need to indirect before calling.\n\tvar argv reflect.Value\n\tif mtype.ArgType.Kind() == reflect.Ptr {\n\t\targv = reflect.New(mtype.ArgType.Elem())\n\t} else {\n\t\targv = reflect.New(mtype.ArgType)\n\t\targIsValue = true\n\t}\n\n\t// argv guaranteed to be a pointer now.\n\tif argIsValue {\n\t\targv = argv.Elem()\n\t}\n\n\tconvert(req.Params, argv.Interface())\n\t// fmt.Println(argv.Interface())\n\n\treplyv := reflect.New(mtype.ReplyType.Elem())\n\tswitch mtype.ReplyType.Elem().Kind() {\n\tcase reflect.Map:\n\t\treplyv.Elem().Set(reflect.MakeMap(mtype.ReplyType.Elem()))\n\tcase reflect.Slice:\n\t\treplyv.Elem().Set(reflect.MakeSlice(mtype.ReplyType.Elem(), 0, 0))\n\t}\n\n\treturn svc.call(mtype, req, argv, replyv)\n}", "func (_m *MockDispatchServer) mustEmbedUnimplementedDispatchServer() {\n\t_m.Called()\n}", "func (f *FakeInstance) Start(_ context.Context, _ string) error {\n\tpanic(\"implement me\")\n}", "func (c *serviceClass) defineOwnMethods() {\n\tc.defineSingletonMethods()\n\tc.definePodFinderMethods()\n}", "func (f FakeServiceGetter) Get(name string) (*v1.Service, error) {\n\treturn f.Svc, f.Err\n}", "func (s *Service) Listen() {\n\n}", "func hello(w http.ResponseWriter, r *http.Request) {\n\tw.Write([]byte(fake.Hello()))\n}", "func Test(c *gin.Context) {\n\tvar serviceTestDTO model.ServiceTest\n\n\terr := c.BindJSON(&serviceTestDTO)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\trefConf := genericServiceImpl.NewRefConf(\"dubbo-admin\", serviceTestDTO.Service, \"dubbo\")\n\ttime.Sleep(2 * time.Second)\n\tresp, err := refConf.\n\t\tGetRPCService().(*generic.GenericService).\n\t\tInvoke(\n\t\t\tc,\n\t\t\tserviceTestDTO.Method,\n\t\t\tserviceTestDTO.ParameterTypes,\n\t\t\t[]hessian.Object{\"A003\"}, // fixme\n\t\t)\n\trefConf.GetInvoker().Destroy()\n\tif err != nil {\n\t\tlogger.Error(\"Error do generic invoke for service test\", err)\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": err.Error()})\n\t\treturn\n\t}\n\tc.JSON(http.StatusOK, resp)\n}", "func (s serviceInjector) invoke(f interface{}) {\n\tif reflect.TypeOf(f).Kind() != reflect.Func {\n\t\t// log and return to prevent panic.\n\t\tlog.println(\"Cannot invoke non function type\")\n\t\treturn\n\t}\n\n\targs := make([]reflect.Value, reflect.TypeOf(f).NumIn())\n\tfor i := range args {\n\t\targType := reflect.TypeOf(f).In(i)\n\t\tif service, ok := s[argType]; ok {\n\t\t\targs[i] = reflect.ValueOf(service)\n\t\t} else {\n\t\t\t// set zero value\n\t\t\targs[i] = reflect.Zero(argType)\n\t\t}\n\t}\n\treflect.ValueOf(f).Call(args)\n}", "func TestCallToPublicService(t *testing.T) {\n\tt.Parallel()\n\n\tclients := Setup(t)\n\n\tt.Log(\"Creating a Service for the helloworld test app.\")\n\tnames := test.ResourceNames{\n\t\tService: test.ObjectNameForTest(t),\n\t\tImage: test.HelloWorld,\n\t}\n\n\ttest.EnsureTearDown(t, clients, &names)\n\n\tresources, err := v1test.CreateServiceReady(t, clients, &names)\n\tif err != nil {\n\t\tt.Fatalf(\"Failed to create initial Service: %v: %v\", names.Service, err)\n\t}\n\n\tif resources.Route.Status.URL.Host == \"\" {\n\t\tt.Fatalf(\"Route is missing .Status.URL: %#v\", resources.Route.Status)\n\t}\n\tif resources.Route.Status.Address == nil {\n\t\tt.Fatalf(\"Route is missing .Status.Address: %#v\", resources.Route.Status)\n\t}\n\n\tgatewayTestCases := []struct {\n\t\tname string\n\t\turl *url.URL\n\t\taccessibleExternally bool\n\t}{\n\t\t{\"local_address\", resources.Route.Status.Address.URL.URL(), false},\n\t\t{\"external_address\", resources.Route.Status.URL.URL(), true},\n\t}\n\n\tfor _, tc := range gatewayTestCases {\n\t\ttc := tc\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tt.Parallel()\n\t\t\tif !test.ServingFlags.DisableLogStream {\n\t\t\t\tcancel := logstream.Start(t)\n\t\t\t\tdefer cancel()\n\t\t\t}\n\t\t\ttestProxyToHelloworld(t, clients, tc.url, false /*inject*/, tc.accessibleExternally)\n\t\t})\n\t}\n}", "func TestServiceMethodNamesUnderscores(t *testing.T) {\n\ts := httptest.NewServer(NewHaberdasherV1Server(&HaberdasherService{}, nil))\n\tdefer s.Close()\n\n\tclient := NewHaberdasherV1ProtobufClient(s.URL, compatibilityTestClient{client: http.DefaultClient},\n\t\ttwirp.WithClientLiteralURLs(true))\n\that, err := client.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err != nil {\n\t\tt.Fatalf(\"compatible protobuf client err=%q\", err)\n\t}\n\tif hat.Size != 1 {\n\t\tt.Errorf(\"wrong hat size returned\")\n\t}\n\n\tcamelCasedClient := NewHaberdasherV1ProtobufClient(s.URL, compatibilityTestClient{client: http.DefaultClient},\n\t\ttwirp.WithClientLiteralURLs(false)) // default value, send CamelCased routes\n\t_, err = camelCasedClient.MakeHatV1(context.Background(), &MakeHatArgsV1_SizeV1{Inches: 1})\n\tif err == nil {\n\t\tt.Fatalf(\"expected error raised by the compatibilityTestClient because routes are camelcased. Got nil.\")\n\t}\n\tif err.Error() != \"twirp error internal: failed to do request: expected: /twirp/twirp.internal.twirptest.snake_case_names.Haberdasher_v1/MakeHat_v1, got: /twirp/twirp.internal.twirptest.snake_case_names.HaberdasherV1/MakeHatV1\" {\n\t\tt.Fatalf(\"expected error to be about the expected path, got err=%q\", err)\n\t}\n}", "func (t *liverpc) generateClient(file *descriptor.FileDescriptorProto, service *descriptor.ServiceDescriptorProto) {\n\tclientName := clientName(service)\n\tstructName := unexported(clientName)\n\tnewClientFunc := \"New\" + clientName\n\n\tt.P(`type `, structName, ` struct {`)\n\tt.P(` client *liverpc.Client`)\n\tt.P(`}`)\n\tt.P()\n\tt.P(`// `, newClientFunc, ` creates a client that implements the `, clientName, ` interface.`)\n\tt.P(`func `, newClientFunc, `(client *liverpc.Client) `, clientName, ` {`)\n\tt.P(` return &`, structName, `{`)\n\tt.P(` client: client,`)\n\tt.P(` }`)\n\tt.P(`}`)\n\tt.P()\n\n\tfor _, method := range service.Method {\n\t\tmethName := methodName(method)\n\t\tpkgName := pkgName(file)\n\n\t\tinputType := t.goTypeName(method.GetInputType())\n\t\toutputType := t.goTypeName(method.GetOutputType())\n\n\t\tparts := strings.Split(pkgName, \".\")\n\t\tif len(parts) < 2 {\n\t\t\tpanic(\"package name must contain at least to parts, eg: service.v1, get \" + pkgName + \"!\")\n\t\t}\n\t\tvStr := parts[len(parts)-1]\n\t\tif len(vStr) < 2 {\n\t\t\tpanic(\"package name must contain a valid version, eg: service.v1\")\n\t\t}\n\t\t_, err := strconv.Atoi(vStr[1:])\n\t\tif err != nil {\n\t\t\tpanic(\"package name must contain a valid version, eg: service.v1, get \" + vStr)\n\t\t}\n\n\t\trpcMethod := method.GetName()\n\t\trpcCtrl := service.GetName()\n\t\trpcCmd := rpcCtrl + \".\" + rpcMethod\n\n\t\tt.P(`func (c *`, structName, `) `, methName, `(ctx `, t.pkgs[\"context\"], `.Context, in *`, inputType, `, opts ...liverpc.CallOption) (*`, outputType, `, error) {`)\n\t\tt.P(` out := new(`, outputType, `)`)\n\t\tt.P(` err := doRPCRequest(ctx,c.client, `, vStr[1:], `, \"`, rpcCmd, `\", in, out, opts)`)\n\t\tt.P(` if err != nil {`)\n\t\tt.P(` return nil, err`)\n\t\tt.P(` }`)\n\t\tt.P(` return out, nil`)\n\t\tt.P(`}`)\n\t\tt.P()\n\t}\n}", "func (s Service) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\tvar err error\n\n\tswitch method {\n\tcase RPC.Service.Getinfo:\n\t\tresp.Set(s.Getinfo())\n\n\tcase RPC.Service.GetNewAddress:\n\t\tresp.Set(s.GetNewAddress())\n\n\tcase RPC.Service.ValidateAddress:\n\t\tvar args = struct {\n\t\t\tAddr string `json:\"addr\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.ValidateAddress(args.Addr))\n\n\tcase RPC.Service.ListTransactions:\n\t\tvar args = struct {\n\t\t\tAddr *string `json:\"addr\"`\n\t\t\tCount *int `json:\"count\"`\n\t\t\tSkip *int `json:\"skip\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\", \"count\", \"skip\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\t//zenrpc:addr=\"*\"\n\t\tif args.Addr == nil {\n\t\t\tvar v string = \"*\"\n\t\t\targs.Addr = &v\n\t\t}\n\n\t\t//zenrpc:count=300\n\t\tif args.Count == nil {\n\t\t\tvar v int = 300\n\t\t\targs.Count = &v\n\t\t}\n\n\t\t//zenrpc:skip=0\n\t\tif args.Skip == nil {\n\t\t\tvar v int = 0\n\t\t\targs.Skip = &v\n\t\t}\n\n\t\tresp.Set(s.ListTransactions(*args.Addr, *args.Count, *args.Skip))\n\n\tcase RPC.Service.SendToAddress:\n\t\tvar args = struct {\n\t\t\tAddr string `json:\"addr\"`\n\t\t\tAmount json.Number `json:\"amount\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"addr\", \"amount\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.SendToAddress(args.Addr, args.Amount))\n\n\tcase RPC.Service.GetRecords:\n\t\tvar args = struct {\n\t\t\tSTime int64 `json:\"sTime\"`\n\t\t\tETime int64 `json:\"eTime\"`\n\t\t}{}\n\n\t\tif zenrpc.IsArray(params) {\n\t\t\tif params, err = zenrpc.ConvertToObject([]string{\"sTime\", \"eTime\"}, params); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tif len(params) > 0 {\n\t\t\tif err := json.Unmarshal(params, &args); err != nil {\n\t\t\t\treturn zenrpc.NewResponseError(nil, zenrpc.InvalidParams, \"\", err.Error())\n\t\t\t}\n\t\t}\n\n\t\tresp.Set(s.GetRecords(args.STime, args.ETime))\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func newServiceNoPatchTest(name string, options ...controllertesting.ServiceOption) TableRow {\n\ttest := newDispatcherBasicTest(\"Existing Dispatcher Service, \" + name + \", No Patch\")\n\ttest.Objects = append(test.Objects,\n\t\tcontrollertesting.NewKafkaChannelDispatcherService(options...),\n\t\tcontrollertesting.NewKafkaChannelDispatcherDeployment())\n\treturn test\n}", "func (s *Test2) Call(reqBody *jModels.RequestBody, r *http.Request) (interface{}, *jModels.Error) {\n\tswitch reqBody.GetMethod() {\n\tcase \"NilArgs\":\n\t\tif reqBody.HasParams() {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInvalidParams, \"That method of service can't has param\", nil)\n\t\t}\n\t\tvar args jModels.NilArgs\n\t\tvar res Test2NilArgsResult\n\t\terr := s.NilArgs(args, &res)\n\t\tif err != nil {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInternalError, \"Internal error\", err.Error())\n\t\t}\n\t\treturn res, nil\n\tcase \"NilResult\":\n\t\tvar args models_7620940177658827552.Test2NilResultArgs\n\t\tif reqBody.HasParams() {\n\t\t\terr := json.Unmarshal(*reqBody.Params, &args)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInvalidParams, \"Can't unmarshal params to args structure'\", err.Error())\n\t\t\t}\n\t\t}\n\t\tvar res jModels.NilResult\n\t\terr := s.NilResult(args, &res)\n\t\tif err != nil {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInternalError, \"Internal error\", err.Error())\n\t\t}\n\t\treturn res, nil\n\tcase \"AnotherPackageResult\":\n\t\tvar args models.NilArgs\n\t\tif reqBody.HasParams() {\n\t\t\terr := json.Unmarshal(*reqBody.Params, &args)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInvalidParams, \"Can't unmarshal params to args structure'\", err.Error())\n\t\t\t}\n\t\t}\n\t\tvar res models.SomeModel\n\t\terr := s.AnotherPackageResult(args, &res)\n\t\tif err != nil {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInternalError, \"Internal error\", err.Error())\n\t\t}\n\t\treturn res, nil\n\tcase \"DoubleStarAnotherResult\":\n\t\tif reqBody.HasParams() {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInvalidParams, \"That method of service can't has param\", nil)\n\t\t}\n\t\tvar args jModels.NilArgs\n\t\tvar res *models.SomeModel\n\t\terr := s.DoubleStarAnotherResult(args, &res)\n\t\tif err != nil {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInternalError, \"Internal error\", err.Error())\n\t\t}\n\t\treturn res, nil\n\tcase \"DoubleStarResult\":\n\t\tif reqBody.HasParams() {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInvalidParams, \"That method of service can't has param\", nil)\n\t\t}\n\t\tvar args jModels.NilArgs\n\t\tvar res *Test2NilArgsResult\n\t\terr := s.DoubleStarResult(args, &res)\n\t\tif err != nil {\n\t\t\treturn nil, jModels.NewError(jModels.ErrorCodeInternalError, \"Internal error\", err.Error())\n\t\t}\n\t\treturn res, nil\n\tdefault:\n\t\treturn nil, jModels.NewError(jModels.ErrorCodeMethodNotFound, fmt.Sprintf(\"Unknown method '%s' for service '%s'\", reqBody.GetMethod(), \"Test2\"), nil)\n\t}\n}", "func (fs fakeService) Prepare(ctx *context.T, method string, numArgs int) (argptrs []interface{}, tags []*vdl.Value, _ error) {\n\tinargs := make([]*vom.RawBytes, numArgs)\n\tinptrs := make([]interface{}, len(inargs))\n\tfor i := range inargs {\n\t\tinptrs[i] = &inargs[i]\n\t}\n\treturn inptrs, nil, nil\n}", "func (poolMgr *Poolmgr) tapService(w http.ResponseWriter, r *http.Request) {\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\thttp.Error(w, \"Failed to read request\", 500)\n\t\treturn\n\t}\n\tsvcName := string(body)\n\tsvcHost := strings.TrimPrefix(svcName, \"http://\")\n\n\terr = poolMgr.fsCache.TouchByAddress(svcHost)\n\tif err != nil {\n\t\tlog.Printf(\"funcSvc tap error: %v\", err)\n\t\thttp.Error(w, \"Not found\", 404)\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusOK)\n}", "func call(method string, argsArray []string) {\n\tfmt.Println(\"------ start ------ \", method)\n\t// all args\n\tvar args [][]byte\n\targs = append(args, []byte(method))\n\tfmt.Printf(\"- args=[\")\n\tfmt.Printf(\"p0=%s\", method)\n\tif argsArray != nil {\n\t\tfor i := 0; i < len(argsArray); i++ {\n\t\t\targs = append(args, []byte(argsArray[i]))\n\t\t\tfmt.Printf(\",p%d=%s\", i+1, argsArray[i])\n\t\t}\n\t}\n\tfmt.Printf(\"]\")\n\tfmt.Println(\"\")\n\t// invoke\n\tresponse := stub.MockInvoke(\"uuid\", args)\n\tfmt.Printf(\"- status=\")\n\tfmt.Println(response.GetStatus())\n\tfmt.Printf(\"- error message=\")\n\tfmt.Println(response.GetMessage())\n\tfmt.Printf(\"- payload=\")\n\tfmt.Println(string(response.GetPayload()))\n\tfmt.Println(\"------ end ------ \")\n\tfmt.Println(\"\")\n}", "func testService() *corev1.Service {\n\treturn &corev1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tNamespace: \"default\",\n\t\t\tName: \"symbols\",\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"deploy\": \"sourcegraph\",\n\t\t\t},\n\t\t},\n\t\tSpec: corev1.ServiceSpec{\n\t\t\tType: corev1.ServiceTypeClusterIP,\n\t\t\tPorts: []corev1.ServicePort{\n\t\t\t\t{\n\t\t\t\t\tName: \"http\",\n\t\t\t\t\tPort: 3184,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tStatus: corev1.ServiceStatus{},\n\t}\n}", "func TestMakePublicService(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tsks *v1alpha1.ServerlessService\n\t\twant *corev1.Service\n\t}{{\n\t\tname: \"HTTP - serve\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tUID: \"1982\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolHTTP1,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"collie\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"collie\",\n\t\t\t\t\tUID: \"1982\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameHTTP1,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTPPort,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTPPort),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP - proxy\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tUID: \"1982\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tMode: v1alpha1.SKSOperationModeProxy,\n\t\t\t\tProtocolType: networking.ProtocolHTTP1,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"melon\",\n\t\t\t\tName: \"collie\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"collie\",\n\t\t\t\t\tserving.RevisionUID: \"1982\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"collie\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"collie\",\n\t\t\t\t\tUID: \"1982\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameHTTP1,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTPPort,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTPPort),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - serve\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - serve - no backends\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeServe,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}, {\n\t\tname: \"HTTP2 - proxy\",\n\t\tsks: &v1alpha1.ServerlessService{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tUID: \"1988\",\n\t\t\t\t// Those labels are propagated from the Revision->PA.\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tSpec: v1alpha1.ServerlessServiceSpec{\n\t\t\t\tProtocolType: networking.ProtocolH2C,\n\t\t\t\tMode: v1alpha1.SKSOperationModeProxy,\n\t\t\t},\n\t\t},\n\t\twant: &corev1.Service{\n\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\tNamespace: \"siamese\",\n\t\t\t\tName: \"dream\",\n\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t// Those should be propagated.\n\t\t\t\t\tserving.RevisionLabelKey: \"dream\",\n\t\t\t\t\tserving.RevisionUID: \"1988\",\n\t\t\t\t\tnetworking.SKSLabelKey: \"dream\",\n\t\t\t\t\tnetworking.ServiceTypeKey: \"Public\",\n\t\t\t\t},\n\t\t\t\tAnnotations: map[string]string{\n\t\t\t\t\t\"cherub\": \"rock\",\n\t\t\t\t},\n\t\t\t\tOwnerReferences: []metav1.OwnerReference{{\n\t\t\t\t\tAPIVersion: v1alpha1.SchemeGroupVersion.String(),\n\t\t\t\t\tKind: \"ServerlessService\",\n\t\t\t\t\tName: \"dream\",\n\t\t\t\t\tUID: \"1988\",\n\t\t\t\t\tController: ptr.Bool(true),\n\t\t\t\t\tBlockOwnerDeletion: ptr.Bool(true),\n\t\t\t\t}},\n\t\t\t},\n\t\t\tSpec: corev1.ServiceSpec{\n\t\t\t\tPorts: []corev1.ServicePort{{\n\t\t\t\t\tName: networking.ServicePortNameH2C,\n\t\t\t\t\tProtocol: corev1.ProtocolTCP,\n\t\t\t\t\tPort: networking.ServiceHTTP2Port,\n\t\t\t\t\tTargetPort: intstr.FromInt(networking.BackendHTTP2Port),\n\t\t\t\t}},\n\t\t\t},\n\t\t},\n\t}}\n\n\tfor _, test := range tests {\n\t\tt.Run(test.name, func(t *testing.T) {\n\t\t\tgot := MakePublicService(test.sks)\n\t\t\tif diff := cmp.Diff(test.want, got); diff != \"\" {\n\t\t\t\tt.Errorf(\"Public K8s Service mismatch (-want, +got) = %v\", diff)\n\t\t\t}\n\t\t})\n\t}\n}", "func suiteMethod(method reflect.Method) *MethodType {\n\tmtype := method.Type\n\tmname := method.Name\n\tinNum := mtype.NumIn()\n\toutNum := mtype.NumOut()\n\n\t// Method must be exported.\n\tif method.PkgPath != \"\" {\n\t\treturn nil\n\t}\n\n\tvar (\n\t\treplyType, ctxType reflect.Type\n\t\targsType []reflect.Type\n\t)\n\n\t// Reference is used to define service reference, and method with prefix 'XXX' is generated by triple pb tool.\n\t// SetGRPCServer is used for pb reflection.\n\t// They should not to be checked.\n\tif mname == \"Reference\" || mname == \"SetGRPCServer\" || strings.HasPrefix(mname, \"XXX\") {\n\t\treturn nil\n\t}\n\n\tif outNum != 1 && outNum != 2 {\n\t\tlogger.Warnf(\"method %s of mtype %v has wrong number of in out parameters %d; needs exactly 1/2\",\n\t\t\tmname, mtype.String(), outNum)\n\t\treturn nil\n\t}\n\n\t// The latest return type of the method must be error.\n\tif returnType := mtype.Out(outNum - 1); returnType != typeOfError {\n\t\tlogger.Debugf(`\"%s\" method will not be exported because its last return type %v doesn't have error`, mname, returnType)\n\t\treturn nil\n\t}\n\n\t// replyType\n\tif outNum == 2 {\n\t\treplyType = mtype.Out(0)\n\t\tif !isExportedOrBuiltinType(replyType) {\n\t\t\tlogger.Errorf(\"reply type of method %s not exported{%v}\", mname, replyType)\n\t\t\treturn nil\n\t\t}\n\t}\n\n\tindex := 1\n\n\t// ctxType\n\tif inNum > 1 && mtype.In(1).String() == \"context.Context\" {\n\t\tctxType = mtype.In(1)\n\t\tindex = 2\n\t}\n\n\tfor ; index < inNum; index++ {\n\t\targsType = append(argsType, mtype.In(index))\n\t\t// need not be a pointer.\n\t\tif !isExportedOrBuiltinType(mtype.In(index)) {\n\t\t\tlogger.Errorf(\"argument type of method %q is not exported %v\", mname, mtype.In(index))\n\t\t\treturn nil\n\t\t}\n\t}\n\n\treturn &MethodType{method: method, argsType: argsType, replyType: replyType, ctxType: ctxType}\n}", "func TestGetUserServicePatched (t *testing.T){\n\tuser1, err := GetUserService(user_01.SocialNumber)\n\tassert.Equal(t, 200, err.HTTPStatus)\n\tassert.Equal(t, user1.Name, new_name_user_01)\n}", "func (g *grpc) generateService(file *generator.FileDescriptor, service *pb.ServiceDescriptorProto, index int) {\n\tpath := fmt.Sprintf(\"6,%d\", index) // 6 means service.\n\n\torigServName := service.GetName()\n\tfullServName := origServName\n\tif pkg := file.GetPackage(); pkg != \"\" {\n\t\tfullServName = pkg + \".\" + fullServName\n\t}\n\tservName := generator.CamelCase(origServName)\n\tdeprecated := service.GetOptions().GetDeprecated()\n\n\tg.P()\n\tg.P(fmt.Sprintf(`// %sClient is the client API for %s service.\n//\n// For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.`, servName, servName))\n\n\t// Client interface.\n\tif deprecated {\n\t\tg.P(\"//\")\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"type \", servName, \"Client interface {\")\n\tfor i, method := range service.Method {\n\t\tg.gen.PrintComments(fmt.Sprintf(\"%s,2,%d\", path, i)) // 2 means method in a service.\n\t\tg.P(g.generateClientSignature(servName, method))\n\t}\n\tg.P(\"}\")\n\tg.P()\n\n\t// Client structure.\n\tg.P(\"type \", unexport(servName), \"Client struct {\")\n\tg.P(\"cc *\", grpcPkg, \".ClientConn\")\n\tg.P(\"}\")\n\tg.P()\n\n\t// NewClient factory.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"func New\", servName, \"Client (cc *\", grpcPkg, \".ClientConn) \", servName, \"Client {\")\n\tg.P(\"return &\", unexport(servName), \"Client{cc}\")\n\tg.P(\"}\")\n\tg.P()\n\n\tvar methodIndex, streamIndex int\n\tserviceDescVar := \"_\" + servName + \"_serviceDesc\"\n\t// Client method implementations.\n\tfor _, method := range service.Method {\n\t\tvar descExpr string\n\t\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\t\t// Unary RPC method\n\t\t\tdescExpr = fmt.Sprintf(\"&%s.Methods[%d]\", serviceDescVar, methodIndex)\n\t\t\tmethodIndex++\n\t\t} else {\n\t\t\t// Streaming RPC method\n\t\t\tdescExpr = fmt.Sprintf(\"&%s.Streams[%d]\", serviceDescVar, streamIndex)\n\t\t\tstreamIndex++\n\t\t}\n\t\tg.generateClientMethod(servName, fullServName, serviceDescVar, method, descExpr)\n\t}\n\n\t// Server interface.\n\tserverType := servName + \"Server\"\n\tg.P(\"// \", serverType, \" is the server API for \", servName, \" service.\")\n\tif deprecated {\n\t\tg.P(\"//\")\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"type \", serverType, \" interface {\")\n\tfor i, method := range service.Method {\n\t\tg.gen.PrintComments(fmt.Sprintf(\"%s,2,%d\", path, i)) // 2 means method in a service.\n\t\tg.P(g.generateServerSignature(servName, method))\n\t}\n\tg.P(\"}\")\n\tg.P()\n\n\t// Server Unimplemented struct for forward compatability.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.generateUnimplementedServer(servName, service)\n\n\t// Server registration.\n\tif deprecated {\n\t\tg.P(deprecationComment)\n\t}\n\tg.P(\"func Register\", servName, \"Server(s *\", grpcPkg, \".Server, srv \", serverType, \") {\")\n\tg.P(\"s.RegisterService(&\", serviceDescVar, `, srv)`)\n\tg.P(\"}\")\n\tg.P()\n\n\t// Server handler implementations.\n\tvar handlerNames []string\n\tfor _, method := range service.Method {\n\t\thname := g.generateServerMethod(servName, fullServName, method)\n\t\thandlerNames = append(handlerNames, hname)\n\t}\n\n\t// Service descriptor.\n\tg.P(\"var \", serviceDescVar, \" = \", grpcPkg, \".ServiceDesc {\")\n\tg.P(\"ServiceName: \", strconv.Quote(fullServName), \",\")\n\tg.P(\"HandlerType: (*\", serverType, \")(nil),\")\n\tg.P(\"Methods: []\", grpcPkg, \".MethodDesc{\")\n\tfor i, method := range service.Method {\n\t\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\t\tcontinue\n\t\t}\n\t\tg.P(\"{\")\n\t\tg.P(\"MethodName: \", strconv.Quote(method.GetName()), \",\")\n\t\tg.P(\"Handler: \", handlerNames[i], \",\")\n\t\tg.P(\"},\")\n\t}\n\tg.P(\"},\")\n\tg.P(\"Streams: []\", grpcPkg, \".StreamDesc{\")\n\tfor i, method := range service.Method {\n\t\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\t\tcontinue\n\t\t}\n\t\tg.P(\"{\")\n\t\tg.P(\"StreamName: \", strconv.Quote(method.GetName()), \",\")\n\t\tg.P(\"Handler: \", handlerNames[i], \",\")\n\t\tif method.GetServerStreaming() {\n\t\t\tg.P(\"ServerStreams: true,\")\n\t\t}\n\t\tif method.GetClientStreaming() {\n\t\t\tg.P(\"ClientStreams: true,\")\n\t\t}\n\t\tg.P(\"},\")\n\t}\n\tg.P(\"},\")\n\tg.P(\"Metadata: \\\"\", file.GetName(), \"\\\",\")\n\tg.P(\"}\")\n\tg.P()\n}", "func (client ListClient) FleetMethodResponder(resp *http.Response) (result FleetOKResponseType, err error) {\n err = autorest.Respond(\n resp,\n client.ByInspecting(),\n azure.WithErrorUnlessStatusCode(http.StatusOK),\n autorest.ByUnmarshallingJSON(&result),\n autorest.ByClosing())\n result.Response = autorest.Response{Response: resp}\n return\n }", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n //args := APIstub.GetFunctionAndParameters()\n args := APIstub.GetStringArgs()\n function := args[0]\n if function == \"addRecord\" {\n return s.addRecord(APIstub, args)\n } else if function == \"getRecord\" {\n return s.getRecord(APIstub, args)\n } else if function == \"encRecord\" {\n return s.encRecord(APIstub, args)\n } else if function == \"decRecord\" {\n return s.decRecord(APIstub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name \"+function)\n}", "func handleService(req typhon.Request) typhon.Response {\n\tparts := reService.FindStringSubmatch(req.URL.Path)\n\tif len(parts) != 3 {\n\t\treturn typhon.Response{Error: terrors.NotFound(\"bad_endpoint\", \"Unable to determine service endpoint.\", nil)}\n\t}\n\n\treturn handle(req, \"s-\"+parts[1], parts[2])\n}", "func (f *FakeInstance) Create(_ context.Context, _ *govultr.InstanceCreateReq) (*govultr.Instance, *http.Response, error) {\n\tpanic(\"implement me\")\n}", "func (socket *MockSocket) Listen() {\n}", "func TestService(t *testing.T) {\n\t// Create service to test\n\ts := res.NewService(\"foo\")\n\ts.Handle(\"bar.$id\",\n\t\tres.Access(res.AccessGranted),\n\t\tres.GetModel(func(r res.ModelRequest) {\n\t\t\tr.Model(struct {\n\t\t\t\tMessage string `json:\"msg\"`\n\t\t\t}{r.PathParam(\"id\")})\n\t\t}),\n\t)\n\n\t// Create test session\n\tc := restest.NewSession(t, s)\n\tdefer c.Close()\n\n\t// Test sending get request and validate response\n\tc.Get(\"foo.bar.42\").\n\t\tResponse().\n\t\tAssertModel(map[string]string{\"msg\": \"42\"})\n}", "func (m *Module) processService(srv pgs.Service, nameWithAlias func(n pgs.Entity) string) *ServiceData {\n\tsrvData := &ServiceData{\n\t\tName: m.ctx.Name(srv).String(),\n\t\tMethods: make([]*MethodData, 0, len(srv.Methods())),\n\t}\n\n\t// check service option: ServiceSkip\n\tsrvSkip := false\n\tm.must(srv.Extension(redact.E_ServiceSkip, &srvSkip))\n\tif srvSkip {\n\t\tsrvData.Skip = true\n\t\t// continue\n\t}\n\n\t// check internal service options\n\tsrvInternal := false\n\tm.must(srv.Extension(redact.E_InternalService, &srvInternal))\n\tsrvCode := uint32(codes.PermissionDenied) // default code\n\tif !m.must(srv.Extension(redact.E_InternalServiceCode, &srvCode)) {\n\t\tsrvCode = uint32(codes.PermissionDenied)\n\t}\n\tif srvCode > uint32(codes.Unauthenticated) { // 16\n\t\tm.Fail(invalidCodeErrMsg)\n\t}\n\tsrvErrMsg := \"\"\n\tif !m.must(srv.Extension(redact.E_InternalServiceErrMessage, &srvErrMsg)) {\n\t\tsrvErrMsg = defaultErrMsg\n\t}\n\n\t// methods\n\tfor _, meth := range srv.Methods() {\n\t\tin := meth.Input()\n\t\tout := meth.Output()\n\t\tmethData := &MethodData{\n\t\t\tName: m.ctx.Name(meth).String(),\n\t\t\tInput: nameWithAlias(in),\n\t\t\tOutput: m.processMessage(out, nameWithAlias),\n\t\t}\n\t\tsrvData.Methods = append(srvData.Methods, methData)\n\n\t\t// check method skip options\n\t\tmethSkip := false\n\t\tm.must(meth.Extension(redact.E_MethodSkip, &methSkip))\n\t\tif methSkip || srvSkip {\n\t\t\tmethData.Skip = true\n\t\t\tcontinue\n\t\t}\n\n\t\tmethInternal := false\n\t\tm.must(meth.Extension(redact.E_InternalMethod, &methInternal))\n\t\tmethCode := srvCode // serviceCode\n\t\tif !m.must(meth.Extension(redact.E_InternalMethodCode, &methCode)) {\n\t\t\tmethCode = srvCode\n\t\t}\n\t\tif methCode > uint32(codes.Unauthenticated) { // 16\n\t\t\tm.Fail(invalidCodeErrMsg)\n\t\t}\n\t\tmethErrMsg := srvErrMsg\n\t\tif !m.must(meth.Extension(redact.E_InternalMethodErrMessage, &methErrMsg)) {\n\t\t\tmethErrMsg = srvErrMsg\n\t\t}\n\n\t\t// apply format specifiers\n\t\tmethErrMsg = strings.ReplaceAll(methErrMsg, specifierMethod, methData.Name)\n\t\tmethErrMsg = strings.ReplaceAll(methErrMsg, specifierService, srvData.Name)\n\n\t\tmethData.ErrMessage = \"`\" + methErrMsg + \"`\"\n\t\tmethData.StatusCode = codes.Code(methCode).String()\n\t\tmethData.Internal = srvInternal || methInternal\n\t}\n\treturn srvData\n}", "func (Service) TooManyArguments(a, b string) {\n}", "func stub() {\n\tpanic( \"calling a stub.\")\n}", "func MakeEndpoint_FooService_PostHello(cli FooServiceClient) endpoint.Endpoint {\n\tendp := func (ctx context.Context, inp interface{}) (interface{}, error) {\n\t\treturn cli.PostHello(ctx, inp.(*HelloRequest))\n\t}\n\n\treturn endp\n}", "func FakeHandler() http.Handler {\n\tgin.SetMode(gin.TestMode)\n\n\te := gin.New()\n\n\te.GET(\"/api/v1/users\", getUsers)\n\te.GET(\"/api/v1/users/:user\", getUser)\n\te.POST(\"/api/v1/users\", createUser)\n\te.PUT(\"/api/v1/users/:user\", updateUser)\n\te.DELETE(\"/api/v1/users/:user\", deleteUser)\n\n\treturn e\n}", "func (s *SmartContract) Invoke(APIstub shim.ChaincodeStubInterface) sc.Response {\n\n\tfunction, args := APIstub.GetFunctionAndParameters()\n\tlogger.Infof(\"Function name is: %d\", function)\n\tlogger.Infof(\"Args length is : %d\", len(args))\n\n\tswitch function {\n\tcase \"queryById\":\n\t\treturn s.queryById(APIstub, args)\n\tcase \"createRecord\":\n\t\treturn s.createRecord(APIstub, args)\n\tdefault:\n\t\treturn shim.Error(\"Invalid Smart Contract function name.\")\n\t}\n}", "func CallServiceMethodWithReflect(g GymService, method string, param []interface{}) []reflect.Value {\n\tlength := len(param)\n\tvar in []reflect.Value\n\tif length > 0 {\n\t\tin = make([]reflect.Value, length)\n\t\tfor k, v := range param {\n\t\t\tin[k] = reflect.ValueOf(v)\n\t\t}\n\t} else {\n\t\tin = []reflect.Value{}\n\t}\n\tresults := reflect.ValueOf(g).MethodByName(method).Call(in)\n\treturn results\n}", "func payloadMethods(typ reflect.Type) map[string]reflect.Method {\n\tmethods := make(map[string]reflect.Method)\nLoopMethods:\n\tfor i := 0; i < typ.NumMethod(); i++ {\n\t\tmethod := typ.Method(i)\n\t\tmtype := method.Type\n\t\tmname := method.Name\n\t\tif method.PkgPath != \"\" {\n\t\t\tcontinue LoopMethods\n\t\t}\n\t\tswitch mtype.NumIn() {\n\t\tcase 2:\n\t\t\teventType := mtype.In(1)\n\t\t\tif eventType.Kind() != reflect.Ptr {\n\t\t\t\tlog.Println(\"method\", mname, \"takes wrong type of event:\", eventType)\n\t\t\t\tcontinue LoopMethods\n\t\t\t}\n\t\t\tevent, ok := payloads.Name(eventType.Elem())\n\t\t\tif !ok {\n\t\t\t\tlog.Println(\"method\", mname, \"takes wrong type of event:\", eventType)\n\t\t\t\tcontinue LoopMethods\n\t\t\t}\n\t\t\tif _, ok = methods[event]; ok {\n\t\t\t\tpanic(fmt.Sprintf(\"there is more than one method handling %v event\", eventType))\n\t\t\t}\n\t\t\tmethods[event] = method\n\t\tcase 3:\n\t\t\tif mtype.In(1).Implements(contextType) && mtype.In(2).Kind() == reflect.Ptr {\n\t\t\t\teventType := mtype.In(2)\n\t\t\t\tevent, ok := payloads.Name(eventType.Elem())\n\t\t\t\tif !ok {\n\t\t\t\t\tlog.Println(\"method\", mname, \"takes wrong type of event:\", eventType)\n\t\t\t\t\tcontinue LoopMethods\n\t\t\t\t}\n\t\t\t\tif _, ok = methods[event]; ok {\n\t\t\t\t\tpanic(fmt.Sprintf(\"there is more than one method handling %v event\", eventType))\n\t\t\t\t}\n\t\t\t\tmethods[event] = method\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif mtype.In(1).Kind() != reflect.String || mtype.In(2) != empty {\n\t\t\t\tlog.Println(\"wildcard method\", mname, \"takes wrong types of arguments\")\n\t\t\t\tcontinue LoopMethods\n\t\t\t}\n\t\t\tif _, ok := methods[\"*\"]; ok {\n\t\t\t\tpanic(\"there is more than one method handling all events\")\n\t\t\t}\n\t\t\tmethods[\"*\"] = method\n\t\tdefault:\n\t\t\tlog.Println(\"method\", mname, \"takes wrong number of arguments:\", mtype.NumIn())\n\t\t\tcontinue LoopMethods\n\t\t}\n\t}\n\treturn methods\n}", "func (g *grpc) generateServerMethodConcrete(servName string, method *pb.MethodDescriptorProto) {\n\theader := g.generateServerSignatureWithParamNames(servName, method)\n\tg.P(\"func (*Unimplemented\", servName, \"Server) \", header, \" {\")\n\tvar nilArg string\n\tif !method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\tnilArg = \"nil, \"\n\t}\n\tmethName := generator.CamelCase(method.GetName())\n\tstatusPkg := string(g.gen.AddImport(statusPkgPath))\n\tcodePkg := string(g.gen.AddImport(codePkgPath))\n\tg.P(\"return \", nilArg, statusPkg, `.Errorf(`, codePkg, `.Unimplemented, \"method `, methName, ` not implemented\")`)\n\tg.P(\"}\")\n}", "func (s PublicService) Invoke(ctx context.Context, method string, params json.RawMessage) zenrpc.Response {\n\tresp := zenrpc.Response{}\n\n\tswitch method {\n\tcase RPC.PublicService.GoPGVersions:\n\t\tresp.Set(s.GoPGVersions())\n\n\tcase RPC.PublicService.Modes:\n\t\tresp.Set(s.Modes())\n\n\tcase RPC.PublicService.SearchTypes:\n\t\tresp.Set(s.SearchTypes())\n\n\tcase RPC.PublicService.Types:\n\t\tresp.Set(s.Types())\n\n\tcase RPC.PublicService.DBTypes:\n\t\tresp.Set(s.DBTypes())\n\n\tcase RPC.PublicService.HTMLTypes:\n\t\tresp.Set(s.HTMLTypes())\n\n\tcase RPC.PublicService.Ping:\n\t\tresp.Set(s.Ping())\n\n\tdefault:\n\t\tresp = zenrpc.NewResponseError(nil, zenrpc.MethodNotFound, \"\", nil)\n\t}\n\n\treturn resp\n}", "func (o *FakeObject) Call(method string, args ...interface{}) Object {\n\treturn MakeFakeObject(o.Methods[method](args...))\n}", "func (p *printer) Service(service *descriptor.ServiceDescriptorProto, methodIndex int) {\n\tp.MaybeLeadingComments(service)\n\tdefer p.open(\"service %s\", service.GetName())()\n\n\tif methodIndex < 0 {\n\t\tfor i := range service.Method {\n\t\t\tp.Method(service.Method[i])\n\t\t}\n\t} else {\n\t\tp.Method(service.Method[methodIndex])\n\t\tif len(service.Method) > 1 {\n\t\t\tp.Printf(\"// other methods were omitted.\\n\")\n\t\t}\n\t}\n}", "func TestService_Handle_Inviter(t *testing.T) {\n\tmockStore := &mockstorage.MockStore{Store: make(map[string]mockstorage.DBEntry)}\n\tstoreProv := mockstorage.NewCustomMockStoreProvider(mockStore)\n\tk := newKMS(t, storeProv)\n\tprov := &protocol.MockProvider{\n\t\tStoreProvider: storeProv,\n\t\tServiceMap: map[string]interface{}{\n\t\t\tmediator.Coordination: &mockroute.MockMediatorSvc{},\n\t\t},\n\t\tCustomKMS: k,\n\t\tKeyTypeValue: kms.ED25519Type,\n\t\tKeyAgreementTypeValue: kms.X25519ECDHKWType,\n\t}\n\n\tctx := &context{\n\t\toutboundDispatcher: prov.OutboundDispatcher(),\n\t\tcrypto: &tinkcrypto.Crypto{},\n\t\tkms: k,\n\t\tkeyType: kms.ED25519Type,\n\t\tkeyAgreementType: kms.X25519ECDHKWType,\n\t}\n\n\tverPubKey, encPubKey := newSigningAndEncryptionDIDKeys(t, ctx)\n\n\tctx.vdRegistry = &mockvdr.MockVDRegistry{CreateValue: createDIDDocWithKey(verPubKey, encPubKey)}\n\n\tconnRec, err := connection.NewRecorder(prov)\n\trequire.NoError(t, err)\n\trequire.NotNil(t, connRec)\n\n\tctx.connectionRecorder = connRec\n\n\tdoc, err := ctx.vdRegistry.Create(testMethod, nil)\n\trequire.NoError(t, err)\n\n\ts, err := New(prov)\n\trequire.NoError(t, err)\n\n\tactionCh := make(chan service.DIDCommAction, 10)\n\terr = s.RegisterActionEvent(actionCh)\n\trequire.NoError(t, err)\n\n\tstatusCh := make(chan service.StateMsg, 10)\n\terr = s.RegisterMsgEvent(statusCh)\n\trequire.NoError(t, err)\n\n\tcompletedFlag := make(chan struct{})\n\trespondedFlag := make(chan struct{})\n\n\tgo msgEventListener(t, statusCh, respondedFlag, completedFlag)\n\n\tgo func() { service.AutoExecuteActionEvent(actionCh) }()\n\n\tinvitation := &Invitation{\n\t\tType: InvitationMsgType,\n\t\tID: randomString(),\n\t\tLabel: \"Bob\",\n\t\tRecipientKeys: []string{verPubKey},\n\t\tServiceEndpoint: \"http://alice.agent.example.com:8081\",\n\t}\n\n\terr = ctx.connectionRecorder.SaveInvitation(invitation.ID, invitation)\n\trequire.NoError(t, err)\n\n\tthid := randomString()\n\n\t// Invitation was previously sent by Alice to Bob.\n\t// Bob now sends a did-exchange Invitation\n\tpayloadBytes, err := json.Marshal(\n\t\t&Request{\n\t\t\tType: RequestMsgType,\n\t\t\tID: thid,\n\t\t\tLabel: \"Bob\",\n\t\t\tThread: &decorator.Thread{\n\t\t\t\tPID: invitation.ID,\n\t\t\t},\n\t\t\tDID: doc.DIDDocument.ID,\n\t\t\tDocAttach: unsignedDocAttach(t, doc.DIDDocument),\n\t\t})\n\trequire.NoError(t, err)\n\tmsg, err := service.ParseDIDCommMsgMap(payloadBytes)\n\trequire.NoError(t, err)\n\t_, err = s.HandleInbound(msg, service.NewDIDCommContext(doc.DIDDocument.ID, \"\", nil))\n\trequire.NoError(t, err)\n\n\tselect {\n\tcase <-respondedFlag:\n\tcase <-time.After(2 * time.Second):\n\t\trequire.Fail(t, \"didn't receive post event responded\")\n\t}\n\t// Alice automatically sends exchange Response to Bob\n\t// Bob replies with an ACK\n\tpayloadBytes, err = json.Marshal(\n\t\t&model.Ack{\n\t\t\tType: AckMsgType,\n\t\t\tID: randomString(),\n\t\t\tStatus: \"OK\",\n\t\t\tThread: &decorator.Thread{ID: thid},\n\t\t})\n\trequire.NoError(t, err)\n\n\tdidMsg, err := service.ParseDIDCommMsgMap(payloadBytes)\n\trequire.NoError(t, err)\n\n\t_, err = s.HandleInbound(didMsg, service.NewDIDCommContext(doc.DIDDocument.ID, \"\", nil))\n\trequire.NoError(t, err)\n\n\tselect {\n\tcase <-completedFlag:\n\tcase <-time.After(2 * time.Second):\n\t\trequire.Fail(t, \"didn't receive post event complete\")\n\t}\n\n\tvalidateState(t, s, thid, findNamespace(AckMsgType), (&completed{}).Name())\n}", "func (t *Deliverys) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\r\n function, args := stub.GetFunctionAndParameters()\r\n fmt.Println(\"invoke is running \" + function)\r\n\r\n // Handle different functions\r\n if function == \"createDelivery\" { //create a new Delivery\r\n return t.createDelivery(stub, args)\r\n\t}else if function == \"getDeliveryByPurchaseID\" { //find delivery for a particular purchase id using rich query\r\n return t.getDeliveryByPurchaseID(stub, args)\r\n }else if function == \"getAllDAPDelivery\" { //find delivery for a particular purchase id using rich query\r\n return t.getAllDAPDelivery(stub, args)\r\n } else if function == \"getAllDAPDeliveryDate\" { //find delivery for a particular purchase id using rich query\r\n return t.getAllDAPDeliveryDate(stub, args)\r\n }\r\n\t \r\n eventMessage := \"{ \\\"message\\\" : \\\"Received unknown function invocation\\\", \\\"code\\\" : \\\"503\\\"}\"\r\n err := stub.SetEvent(\"errEvent\", []byte(eventMessage))\r\n if err != nil {\r\n return shim.Error(err.Error())\r\n }\r\n fmt.Println(\"invoke did not find func: \" + function) //error\r\n return shim.Error(\"Received unknown function invocation\")\r\n}", "func (t *SimpleChaincode) Invoke(stub shim.ChaincodeStubInterface) pb.Response {\n\tfunction, args := stub.GetFunctionAndParameters()\n\t// tMap, _ := stub.GetTransient()\n\tfmt.Println(\" \")\n\tfmt.Println(\"starting invoke, for - \" + function)\n\n\t// Handle different functions\n\tswitch function {\n\tcase \"create_project\":\n\t\treturn create_project(stub, args)\n\tcase \"get_project_by_id\":\n\t\treturn get_project_by_id(stub, args)\n\tcase \"query_all_projects\":\n\t\treturn query_all_projects(stub, args)\n\tcase \"query_paging_projects\":\n\t\treturn query_paging_projects(stub, args)\n\tcase \"remove_project\":\n\t\treturn remove_project(stub, args)\n\tcase \"modify_project\":\n\t\treturn modify_project(stub, args)\n\tcase \"create_linear_workflow\":\n\t\treturn create_linear_workflow(stub, args)\n\tcase \"get_workflow_by_id\":\n\t\treturn get_workflow_by_id(stub, args)\n\tcase \"query_all_workflows\":\n\t\treturn query_all_workflows(stub, args)\n\tcase \"enable_or_disable_workflow\":\n\t\treturn enable_or_disable_workflow(stub, args)\n\tcase \"modify_workflow_def\":\n\t\treturn modify_workflow_def(stub, args)\n\tcase \"query_accessable_workflows\":\n\t\treturn query_accessable_workflows(stub, args)\n\tcase \"start_process\":\n\t\treturn start_process(stub, args)\n\tcase \"get_process_by_id\":\n\t\treturn get_process_by_id(stub, args)\n\tcase \"query_logs_by_process_id\":\n\t\treturn query_logs_by_process_id(stub, args)\n\tcase \"transfer_process\":\n\t\treturn transfer_process(stub, args)\n\tcase \"return_process\":\n\t\treturn return_process(stub, args)\n\tcase \"withdraw_process\":\n\t\treturn withdraw_process(stub, args)\n\tcase \"cancel_process\":\n\t\treturn cancel_process(stub, args)\n\tcase \"query_todo_process\":\n\t\treturn query_todo_process(stub, args)\n\tcase \"query_done_process\":\n\t\treturn query_done_process(stub, args)\n\tcase \"save_org_public_key\":\n\t\treturn save_org_public_key(stub, args)\n\tcase \"encrypt_data\":\n\t\treturn encrypt_data(stub, args)\n\tcase \"decrypt_data\":\n\t\treturn decrypt_data(stub, args)\n\tdefault:\n\t\t// error out\n\t\tfmt.Println(\"Received unknown invoke function name - \" + function)\n\t\treturn shim.Error(\"Received unknown invoke function name - '\" + function + \"'\")\n\t}\n}", "func TestCryptoSignerInterfaceBehavior(t *testing.T) {\n\tcs := NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.EmptyCryptoServiceInterfaceBehaviorTests(t, cs)\n\tinterfaces.CreateGetKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.CreateListKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.AddGetKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n\n\tcs = NewCryptoService(trustmanager.NewKeyMemoryStore(passphraseRetriever))\n\tinterfaces.AddListKeyCryptoServiceInterfaceBehaviorTests(t, cs, data.ECDSAKey)\n}", "func (m *CloudWatchLogsServiceMock) CreateNewServiceIfUnHealthy() {\n\n}", "func (m *MockConsulHealth) Service(service, tag string, passingOnly bool, q *api.QueryOptions) (entries []*api.ServiceEntry, meta *api.QueryMeta, err error) {\n\n\targs := m.Called(service, tag, passingOnly, q)\n\n\tentries = nil\n\tmeta = nil\n\terr = args.Error(2)\n\n\tif e := args.Get(0); e != nil {\n\t\tentries = e.(func() []*api.ServiceEntry)()\n\t}\n\n\tif m := args.Get(1); m != nil {\n\t\tmeta = m.(*api.QueryMeta)\n\t}\n\n\treturn\n}", "func (e ExampleType) ExampleMethod() {\n}", "func testWrappers(e exchange.IBotExchange) ([]string, error) {\n\tiExchange := reflect.TypeOf(&e).Elem()\n\tactualExchange := reflect.ValueOf(e)\n\terrType := reflect.TypeOf(common.ErrNotYetImplemented)\n\n\tcontextParam := reflect.TypeOf((*context.Context)(nil)).Elem()\n\n\tvar funcs []string\n\tfor x := 0; x < iExchange.NumMethod(); x++ {\n\t\tname := iExchange.Method(x).Name\n\t\tmethod := actualExchange.MethodByName(name)\n\t\tinputs := make([]reflect.Value, method.Type().NumIn())\n\n\t\tfor y := 0; y < method.Type().NumIn(); y++ {\n\t\t\tinput := method.Type().In(y)\n\n\t\t\tif input.Implements(contextParam) {\n\t\t\t\t// Need to deploy a context.Context value as nil value is not\n\t\t\t\t// checked throughout codebase. Cancelled to minimise external\n\t\t\t\t// calls and speed up operation.\n\t\t\t\tcancelled, cancelfn := context.WithTimeout(context.Background(), 0)\n\t\t\t\tcancelfn()\n\t\t\t\tinputs[y] = reflect.ValueOf(cancelled)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tinputs[y] = reflect.Zero(input)\n\t\t}\n\n\t\toutputs := method.Call(inputs)\n\t\tif method.Type().NumIn() == 0 {\n\t\t\t// Some empty functions will reset the exchange struct to defaults,\n\t\t\t// so turn off verbosity.\n\t\t\te.GetBase().Verbose = false\n\t\t}\n\n\t\tfor y := range outputs {\n\t\t\tincoming := outputs[y].Interface()\n\t\t\tif reflect.TypeOf(incoming) != errType {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\terr, ok := incoming.(error)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"%s type assertion failure for %v\", name, incoming)\n\t\t\t}\n\t\t\tif errors.Is(err, common.ErrNotYetImplemented) {\n\t\t\t\tfuncs = append(funcs, name)\n\t\t\t}\n\t\t\t// found error; there should not be another error in this slice.\n\t\t\tbreak\n\t\t}\n\t}\n\treturn funcs, nil\n}", "func (f *lazyCallReq) Service() []byte {\n\tl := f.Payload[_serviceLenIndex]\n\treturn f.Payload[_serviceNameIndex : _serviceNameIndex+int(l)]\n}", "func (g *grpc) generateClientSignature(servName string, method *pb.MethodDescriptorProto) string {\n\torigMethName := method.GetName()\n\tmethName := generator.CamelCase(origMethName)\n\tif reservedClientName[methName] {\n\t\tmethName += \"_\"\n\t}\n\treqArg := \", in *\" + g.typeName(method.GetInputType())\n\tif method.GetClientStreaming() {\n\t\treqArg = \"\"\n\t}\n\trespName := \"*\" + g.typeName(method.GetOutputType())\n\tif method.GetServerStreaming() || method.GetClientStreaming() {\n\t\trespName = servName + \"_\" + generator.CamelCase(origMethName) + \"Client\"\n\t}\n\treturn fmt.Sprintf(\"%s(ctx %s.Context%s, opts ...%s.CallOption) (%s, error)\", methName, contextPkg, reqArg, grpcPkg, respName)\n}", "func (mr *MockapprunnerDescriberMockRecorder) Service() *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Service\", reflect.TypeOf((*MockapprunnerDescriber)(nil).Service))\n}", "func (t *ManagePatient) Invoke(stub shim.ChaincodeStubInterface, function string, args []string) ([]byte, error) {\n fmt.Println(\"invoke is running \" + function)\n\n // Handle different functions\n if function == \"init\" { //initialize the chaincode state, used as reset\n return t.Init(stub, \"init\", args)\n } else if function == \"create_patient\" { //create a new Patient\n return t.create_patient(stub, args)\n }\n fmt.Println(\"invoke did not find func: \" + function) //error\n return nil, errors.New(\"Received unknown function invocation\")\n}", "func Test_IndexHandler(t *testing.T) {\n\tvar (\n\t\tversionMsg Service\n\t\tresp *http.Response\n\t)\n\n\tsvc := NewService()\n\n\tts := httptest.NewServer(svc.NewRouter(\"*\"))\n\tdefer ts.Close()\n\n\treq, _ := http.NewRequest(\"GET\", ts.URL+\"/\", nil)\n\n\toutputLog := helpers.CaptureOutput(func() {\n\t\tresp, _ = http.DefaultClient.Do(req)\n\t})\n\n\tif got, want := resp.StatusCode, 200; got != want {\n\t\tt.Fatalf(\"Invalid status code, got %d but want %d\", got, want)\n\t}\n\n\tdata, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\tt.Fatalf(\"Got an error when reading body: %s\", err.Error())\n\t}\n\n\terr = json.Unmarshal(data, &versionMsg)\n\tif err != nil {\n\t\tt.Fatalf(\"Got an error when parsing json: %s\", err.Error())\n\t}\n\tif got, want := versionMsg.Version, svc.Version; got != want {\n\t\tt.Fatalf(\"Wrong version return, got %s but want %s\", got, want)\n\t}\n\tif got, want := versionMsg.Name, svc.Name; got != want {\n\t\tt.Fatalf(\"Wrong version return, got %s but want %s\", got, want)\n\t}\n\n\tmatched, err := regexp.MatchString(`uri=/ `, outputLog)\n\tif matched != true || err != nil {\n\t\tt.Fatalf(\"request is not logged :\\n%s\", outputLog)\n\t}\n}", "func (o *FakeObject) New(args ...interface{}) Object { return o.Invoke(args) }", "func (v *ServiceGenerator_Generate_Args) MethodName() string {\n\treturn \"generate\"\n}", "func (s *Service) Get(name string) Method {\n\treturn s.methodManager.Get(name)\n}", "func newMyService(g greeter, f factory) builder {\n\treturn func(s *MyService) {\n\t\ts.greeter = g\n\t\ts.f = f\n\t}\n}", "func (k *Kloud) coreMethods(r *kite.Request, fn machineFunc) (result interface{}, reqErr error) {\n\t// calls with zero arguments causes args to be nil. Check it that we\n\t// don't get a beloved panic\n\tif r.Args == nil {\n\t\treturn nil, NewError(ErrNoArguments)\n\t}\n\n\tk.Log.Debug(\"solo: calling %q by %q with %q\", r.Username, r.Method, r.Args.Raw)\n\n\tvar args struct {\n\t\tMachineId string\n\t\tProvider string\n\t\tDebug bool\n\t}\n\n\tif err := r.Args.One().Unmarshal(&args); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif args.MachineId == \"\" {\n\t\treturn nil, NewError(ErrMachineIdMissing)\n\t}\n\n\tif args.Provider == \"\" {\n\t\treturn nil, NewError(ErrProviderIsMissing)\n\t}\n\n\t// Lock the machine id so no one else can access it. It means this\n\t// kloud instance is now responsible for this machine id. Its basically\n\t// a distributed lock. It's unlocked when there is an error or if the\n\t// method call is finished (unlocking is done inside the responsible\n\t// method calls).\n\tif r.Method != \"info\" {\n\t\tif err := k.Locker.Lock(args.MachineId); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// if something goes wrong after step reset the document which is was\n\t\t// set in the by previous step by Locker.Lock(). If there is no error,\n\t\t// the lock will be unlocked in the respective method function.\n\t\tdefer func() {\n\t\t\tif reqErr != nil {\n\t\t\t\t// otherwise that means Locker.Lock or something else in\n\t\t\t\t// ControlFunc failed. Reset the lock again so it can be acquired by\n\t\t\t\t// others.\n\t\t\t\tk.Locker.Unlock(args.MachineId)\n\t\t\t}\n\t\t}()\n\t}\n\n\tprovider, ok := k.providers[args.Provider]\n\tif !ok {\n\t\treturn nil, NewError(ErrProviderNotFound)\n\t}\n\n\tp, ok := provider.(Provider)\n\tif !ok {\n\t\treturn nil, NewError(ErrProviderNotImplemented)\n\t}\n\n\tctx := request.NewContext(context.Background(), r)\n\t// add publicKeys to be deployed to the machine, the machine provider is\n\t// responsible of deploying it to the machine while building it.\n\tif k.PublicKeys != nil {\n\t\tctx = publickeys.NewContext(ctx, k.PublicKeys)\n\t}\n\n\tif k.ContextCreator != nil {\n\t\tctx = k.ContextCreator(ctx)\n\t}\n\n\t// if debug is enabled, generate TraceID and pass it with the context\n\tif args.Debug {\n\t\tctx = k.setTraceID(r.Username, r.Method, ctx)\n\t}\n\n\t// old events are not needed anymore, so we're just going to remove them.\n\tk.cleanupEventers(args.MachineId)\n\n\t// each method has his own unique eventer\n\teventId := r.Method + \"-\" + args.MachineId\n\tev := k.NewEventer(eventId)\n\tctx = eventer.NewContext(ctx, ev)\n\n\tmachine, err := p.Machine(ctx, args.MachineId)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstater, ok := machine.(Stater)\n\tif !ok {\n\t\treturn nil, NewError(ErrStaterNotImplemented)\n\t}\n\n\tif stater.ProviderName() != args.Provider {\n\t\tk.Log.Debug(\"want provider %q, got %q\", stater.ProviderName(), args.Provider)\n\n\t\treturn nil, NewError(ErrProviderIsWrong)\n\t}\n\n\t// Check if the given method is in valid methods of that current state. For\n\t// example if the method is \"build\", and the state is \"stopped\" than this\n\t// will return an error.\n\tif !methodIn(r.Method, stater.State().ValidMethods()...) {\n\t\treturn nil, fmt.Errorf(\"%s not allowed for current state '%s'. Allowed methods are: %v\",\n\t\t\tr.Method, strings.ToLower(stater.State().String()), stater.State().ValidMethods())\n\t}\n\n\tpair, ok := states[r.Method]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"no state pair available for %s\", r.Method)\n\t}\n\n\ttags := []string{\n\t\t\"instanceId:\" + args.MachineId,\n\t\t\"provider:\" + args.Provider,\n\t}\n\n\tctx = k.traceRequest(ctx, tags)\n\n\tev.Push(&eventer.Event{\n\t\tMessage: r.Method + \" started\",\n\t\tStatus: pair.start,\n\t})\n\n\t// Start our core method in a goroutine to not block it for the client\n\t// side. However we do return an event id which is an unique for tracking\n\t// the current status of the running method.\n\tgo func() {\n\t\tfinalEvent := &eventer.Event{\n\t\t\tMessage: r.Method + \" finished\",\n\t\t\tStatus: pair.final,\n\t\t\tPercentage: 100,\n\t\t}\n\n\t\tk.Log.Info(\"[%s] ======> %s started (requester: %s, provider: %s)<======\",\n\t\t\targs.MachineId, strings.ToUpper(r.Method), r.Username, args.Provider)\n\t\tstart := time.Now()\n\t\terr := fn(ctx, machine)\n\t\tif err != nil {\n\t\t\t// don't pass the error directly to the eventer, mask it to avoid\n\t\t\t// error leaking to the client. We just log it here.\n\t\t\tk.Log.Error(\"[%s] ======> %s finished with error: '%s' (requester: %s, provider: %s) <======\",\n\t\t\t\targs.MachineId, strings.ToUpper(r.Method), err, r.Username, args.Provider)\n\n\t\t\tfinalEvent.Error = strings.ToTitle(r.Method) + \" failed. Please contact support.\"\n\n\t\t\t// however, eventerErr is an error we want to pass explicitly to\n\t\t\t// the client side\n\t\t\tif eventerErr, ok := err.(*EventerError); ok {\n\t\t\t\tfinalEvent.Error = eventerErr.Error()\n\t\t\t}\n\n\t\t\tfinalEvent.Status = stater.State() // fallback to to old state\n\t\t} else {\n\t\t\tk.Log.Info(\"[%s] ======> %s finished (time: %s, requester: %s, provider: %s) <======\",\n\t\t\t\targs.MachineId, strings.ToUpper(r.Method), time.Since(start), r.Username, args.Provider)\n\t\t}\n\n\t\tev.Push(finalEvent)\n\t\tk.Locker.Unlock(args.MachineId)\n\t\tk.send(ctx)\n\t}()\n\n\treturn ControlResult{\n\t\tEventId: eventId,\n\t}, nil\n}", "func Service() typhon.Service {\n\treturn Proxy\n}", "func (c *Caller) Start(gclient GRPCClient) error {\n\t//target is for the naming finder,example etcd:///test.example.com\n\t//the grpc will use the naming server of \"etcd\" for name resolver\n\ttarget := c.Resolver + \":///\" + c.Domain\n\tclient, err := grpc.Dial(target, grpc.WithInsecure(), grpc.WithBalancerName(c.Balance))\n\tif err != nil {\n\t\treturn err\n\t}\n\tc.grpcClient = client\n\tc.Client = gclient.NewClient(client)\n\tctype := reflect.TypeOf(c.Client)\n\tfor i := 0; i < ctype.NumMethod(); i++ {\n\t\tm := ctype.Method(i)\n\t\tc.Method[m.Name] = m\n\t\tfmt.Printf(\"%s: %v: %v\\n\", m.Name, m.Type, m.Func)\n\t}\n\t// for _, name := range methodName {\n\t// \t_, ok := c.Method[name]\n\t// \tif !ok {\n\t// \t\tvalue := c.getFunctionByName(name)\n\t// \t\tif !value.IsValid() {\n\t// \t\t\tclient.Close()\n\t// \t\t\treturn ErrCanNotFoundFunc\n\t// \t\t}\n\t// \t\tc.Method[name] = value\n\t// \t}\n\t// }\n\treturn nil\n}", "func (p *PodmanTestIntegration) StartRemoteService() {\n}", "func (c *FakeTiFlashClient) fakeAPI(actionType ActionType, action *Action) (interface{}, error) {\n\tif reaction, ok := c.reactions[actionType]; ok {\n\t\tresult, err := reaction(action)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\t}\n\treturn nil, &NotFoundReaction{actionType}\n}", "func (m *MethodDescriptor) GetService() *ServiceDescriptor { return m.Service }", "func init() {\n\tmakeFunc := func(base func([]reflect.Value) []reflect.Value, fptr interface{}) {\n\t\tfn := reflect.ValueOf(fptr).Elem()\n\t\tv := reflect.MakeFunc(fn.Type(), base)\n\t\tfn.Set(v)\n\t}\n\n\t// getAll(Repository) (int, string)\n\tgetAll := func(in []reflect.Value) []reflect.Value {\n\t\tvalues := in[0].MethodByName(\"GetAll\").Call([]reflect.Value{})\n\t\t// values is []reflect.Value returned by reflect.Call.\n\t\t// Since GetAll only returns interface{}, we just want the first object in values\n\t\tjsonResponse := string(jsonEncode(values[0].Interface()))\n\t\treturn genericHandlerReturn(http.StatusFound, jsonResponse)\n\t}\n\n\tmakeFunc(getAll, &GetAllUnits)\n\n\t/*func AddUnit(rw http.ResponseWriter, u Unit, repo IUnitRepository) (int, string) {\n\t\trepo.Add(&u)\n\t\trw.Header().Set(\"Location\", fmt.Sprintf(\"/unit/%d\", u.Id))\n\t\treturn http.StatusCreated, \"\"\n\t}\n\t// add(http.ResponseWriter, entity, Repository) (int, string)\n\tadd := func(in []reflect.Value) []reflect.Value {\n\t\tin[2].MethodByName(\"Add\").Call([]reflect.Value{in[1]})\n\t\theader := in[0].MethodByName(\"Header\").Call(nil)\n\t\tlocation := reflect.ValueOf(\"Location\")\n\t\tlocationValue := reflect.ValueOf(fmt.Sprintf(\"/unit/%d\", in[1].FieldByName(\"Id\")))\n\t\treflect.ValueOf(header).MethodByName(\"Set\").Call([]reflect.Value{location, locationValue})\n\t\treturn genericHandlerReturn(http.StatusCreated, \"\")\n\t}\n\n\tmakeFunc(add, &AddUnit)\n\t*/\n\n\t// get(martini.Params, Repository) (int, string)\n\tget := func(in []reflect.Value) []reflect.Value {\n\t\tparams := in[0].Interface().(martini.Params)\n\t\tid, err := strconv.Atoi(params[\"id\"])\n\n\t\tif err != nil {\n\t\t\treturn notFoundGeneric()\n\t\t}\n\n\t\tinGet := []reflect.Value{reflect.ValueOf(id)}\n\t\tvalues := in[1].MethodByName(\"Get\").Call(inGet)\n\n\t\tif values[0].IsNil() {\n\t\t\treturn notFoundGeneric()\n\t\t}\n\n\t\tjsonResponse := string(jsonEncode(values[0].Interface()))\n\t\treturn []reflect.Value{reflect.ValueOf(http.StatusOK), reflect.ValueOf(jsonResponse)}\n\t}\n\n\tmakeFunc(get, &GetUnit)\n\n}", "func serviceMutateFn(ctx context.Context, newService *corev1.Service, client client.Client) func() error {\n\treturn func() error {\n\t\t// TODO: probably nothing has to be done, check v1 implementation of CreateOrUpdate\n\t\t//existingService := existing.(*corev1.Service)\n\t\t//if !reflect.DeepEqual(newService.Spec, existingService.Spec) {\n\t\t//\treturn client.Update(ctx, existing)\n\t\t//}\n\t\treturn nil\n\t}\n}", "func (_m *AuthServer) mustEmbedUnimplementedAuthServer() {\n\t_m.Called()\n}", "func (s *helloServiceImpl) SayHello(name string) string {\n\treturn \"Hello\" + name\n}", "func TestSetGetGoodArgsFull(t *testing.T) {\n\tfmt.Println(\"Entering the test method for SetGetGoodArgsFull\")\n\tprovcc := new(SimpleAsset)\n\tstub := shim.NewMockStub(\"ANY_PARAM\", provcc)\n\n\t// Testing the init. It always return true. No parameters in init. \n\t\n\tcheckInit(t, stub, [][]byte{[]byte(\"init\")})\n\n\tres := stub.MockInvoke(\"1\", [][]byte{[]byte(\"set\"), []byte(\"S52fkpF2rCEArSuwqyDA9tVjawUdrkGzbNQLaa7xJfA=\"),\n\t[]byte(\"agentInfo.atype\"),[]byte(\"1.2.3.4\"),\n\t[]byte(\"agentInfo.id\"),[]byte(\"agentidentifier\"),\n\t[]byte(\"agentinfo.name\"),[]byte(\"7.8.9\"),\n\t[]byte(\"agentinfo.idp\"),[]byte(\"urn:tiani-spirit:sts\"),\n\t[]byte(\"locationInfo.id\"),[]byte(\"urn:oid:1.2.3\"),\n\t[]byte(\"locationInfo.name\"),[]byte(\"General Hospital\"),\n\t[]byte(\"locationInfo.locality\"),[]byte(\"Nashville, TN\"),\n\t[]byte(\"locationInfo.docid\"),[]byte(\"1.2.3\"),\n\t[]byte(\"action\"),[]byte(\"ex:CREATE\"),\n\t[]byte(\"date\"),[]byte(\"2017-11-21T10:29:49.816Z\"),\n\t[]byte(\"digest1\"),[]byte(\"E0nioxbCYD5AlzGWXDDDl0Gt5AAKv3ppKt4XMhE1rfo\"),\n\t[]byte(\"digest2\"),[]byte(\"xLrbWN5QJBJUAsdevfrxGlN3o0p8VZMnFFnV9iMll5o\"),\n\t[]byte(\"digest3\"),[]byte(\"THIS_IS_DIGEST_3\"),\n\t[]byte(\"digest4\"),[]byte(\"THIS_IS_DIGEST_4\")})\n\n\tif res.Status != shim.OK {\n\t\tfmt.Println(\"Invoke failed\", string(res.Message))\n\t\tt.FailNow()\n\t}\n\t\n\tresGet := stub.MockInvoke(\"1\", [][]byte{[]byte(\"get\"), []byte(\"S52fkpF2rCEArSuwqyDA9tVjawUdrkGzbNQLaa7xJfA=\")})\n\tif resGet.Status != shim.OK {\n\t\tfmt.Println(\"Invoke failed\", string(resGet.Message))\n\t\tt.FailNow()\n\t}\n}", "func mockNoopStore(id string, key string, value interface{}) {}", "func APIMockTypesBuildFromService(t model.MockTypes) *APIMockTypes {\n\tm := APIMockTypes{}\n\tm.BoolType = BoolBool(t.BoolType)\n\tm.BoolPtrType = BoolPtrBoolPtr(t.BoolPtrType)\n\tm.IntType = IntInt(t.IntType)\n\tm.IntPtrType = IntPtrIntPtr(t.IntPtrType)\n\tm.StringType = StringString(t.StringType)\n\tm.StringPtrType = StringPtrStringPtr(t.StringPtrType)\n\tm.Uint64Type = Uint64Int(t.Uint64Type)\n\tm.Uint64PtrType = Uint64PtrIntPtr(t.Uint64PtrType)\n\tm.Float64Type = Float64Float64(t.Float64Type)\n\tm.Float64PtrType = Float64PtrFloat64Ptr(t.Float64PtrType)\n\tm.RuneType = RuneInt(t.RuneType)\n\tm.RunePtrType = RunePtrIntPtr(t.RunePtrType)\n\treturn &m\n}", "func newHelloService() HelloService {\n\treturn &helloServiceImpl{}\n}", "func Decode_FooService_PostHello(ctx context.Context, req *http.Request) (interface{}, error) {\n\tvar ret HelloRequest\n\n\tqry := req.URL.Query()\n\t_ = qry\n\n\n\tif buff, err := ioutil.ReadAll(req.Body); err == nil {\n\t\tif err := runtime.Decode(ctx, &ret.Who, string(buff)); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif val := qry.Get(\"who\"); val != \"\" {\n\t\tif err := runtime.Decode(ctx, &ret.Who, val); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tparts := strings.Split(req.URL.Path, \"/\")\n\tif len(parts) < 2 {\n\t\treturn nil, errors.New(\"Missing Parameters.\")\n\t}\n\n\treturn &ret, nil\n}", "func (mr *MockProviderMockRecorder) Service(arg0 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Service\", reflect.TypeOf((*MockProvider)(nil).Service), arg0)\n}", "func (_m *MockHandler) ServeHTTP(_a0 http.ResponseWriter, _a1 *http.Request) {\n\t_m.Called(_a0, _a1)\n}", "func service() typhon.Service {\r\n\trouter := typhon.Router{}\r\n\trouter.GET(\"/oxcross\", serveResponse)\r\n\trouter.GET(\"/healthz\", serveResponse)\r\n\r\n\tsvc := router.Serve().Filter(typhon.ErrorFilter).Filter(typhon.H2cFilter)\r\n\r\n\treturn svc\r\n}", "func setupFakeClient(url string) *Client {\n\treturn &Client{\n\t\tServiceEndpoint: ServiceEndpoint{\n\t\t\tRequestURL: url,\n\t\t\tDocsURL: \"some-docs-url\",\n\t\t},\n\t}\n}", "func TestHandlerInterface(t *testing.T) {\n\tvar h Handler\n\th = &testHandler{}\n\th.Work(nil)\n}", "func TestHandler_OK(t *testing.T) {\n\tnow, _ := clock.ParseRFC3339(\"2000-01-01T00:00:00Z\")\n\tinputMachineID := \"AAAAAAAA-A00A-1234-1234-5864377B4831\"\n\ttimeProvider := clock.FrozenTimeProvider{\n\t\tCurrent: now,\n\t}\n\tvar request = events.APIGatewayProxyRequest{\n\t\tHTTPMethod: \"POST\",\n\t\tResource: \"/preflight/{machine_id}\",\n\t\tPathParameters: map[string]string{\"machine_id\": inputMachineID},\n\t\tHeaders: map[string]string{\"Content-Type\": \"application/json\"},\n\t\tBody: `{\n\t\"os_build\":\"20D5029f\",\n\t\"santa_version\":\"2021.1\",\n\t\"hostname\":\"my-awesome-macbook-pro.attlocal.net\",\n\t\"transitive_rule_count\":0,\n\t\"os_version\":\"11.2\",\n\t\"certificate_rule_count\":2,\n\t\"client_mode\":\"MONITOR\",\n\t\"serial_num\":\"C02123456789\",\n\t\"binary_rule_count\":3,\n\t\"primary_user\":\"nobody\",\n\t\"compiler_rule_count\":0\n}`,\n\t}\n\tmockedConfigurationFetcher := &MockDynamodb{}\n\n\tconfig := machineconfiguration.MachineConfiguration{\n\t\tClientMode: types.Lockdown,\n\t\tBatchSize: 37,\n\t\tUploadLogsURL: \"/aaa\",\n\t\tEnableBundles: true,\n\t\tAllowedPathRegex: \"\",\n\t\tCleanSync: false,\n\t}\n\n\treturnedConfig, err := attributevalue.MarshalMap(config)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tmockedConfigurationFetcher.On(\"GetItem\", mock.Anything, mock.Anything).Return(&awsdynamodb.GetItemOutput{\n\t\tItem: returnedConfig,\n\t}, nil)\n\n\tmockedStateTracking := &MockDynamodb{}\n\tmockedStateTracking.On(\"GetItem\", mock.Anything, mock.Anything).Return(&awsdynamodb.GetItemOutput{\n\t\tItem: nil,\n\t}, nil)\n\n\t// mockedStateTracking.On(\"PutItem\", mock.MatchedBy(func(item interface{}) bool {\n\tmockedStateTracking.On(\"PutItem\", mock.MatchedBy(func(syncState syncstate.SyncStateRow) bool {\n\t\treturn syncState.MachineID == inputMachineID && syncState.BatchSize == 37 && syncState.LastCleanSync == \"2000-01-01T00:00:00Z\" && syncState.FeedSyncCursor == \"2000-01-01T00:00:00Z\"\n\t})).Return(&awsdynamodb.PutItemOutput{}, nil)\n\n\tmockedStateTracking.On(\"PutItem\", mock.MatchedBy(func(sensorData sensordata.SensorData) bool {\n\t\treturn sensorData.OSBuild == \"20D5029f\" && sensorData.SerialNum == \"C02123456789\" && sensorData.MachineID == inputMachineID && sensorData.PrimaryUser == \"nobody\" && sensorData.BinaryRuleCount == 3 && sensorData.CompilerRuleCount == 0\n\t})).Return(&awsdynamodb.PutItemOutput{}, nil)\n\n\th := &PostPreflightHandler{\n\t\ttimeProvider: timeProvider,\n\t\tmachineConfigurationService: machineconfiguration.GetMachineConfigurationService(mockedConfigurationFetcher, timeProvider),\n\t\tstateTrackingService: getStateTrackingService(mockedStateTracking, timeProvider),\n\t\tcleanSyncService: getCleanSyncService(timeProvider),\n\t}\n\n\tresp, err := h.Handle(request)\n\n\tassert.Empty(t, err)\n\tassert.Equal(t, 200, resp.StatusCode)\n\n\t// Ensure that the response matches the configuration returned\n\tassert.Equal(t, `{\"client_mode\":\"LOCKDOWN\",\"blocked_path_regex\":\"\",\"allowed_path_regex\":\"\",\"batch_size\":37,\"enable_bundles\":true,\"enable_transitive_rules\":false,\"clean_sync\":true,\"upload_logs_url\":\"/aaa\"}`, resp.Body)\n}", "func TestNewBeerService_ReturnsBeerInteractor(t *testing.T) {\n\tt.Parallel()\n\tinteractor := &mocks.BeerInteractor{}\n\tassert.NotNil(t, adapters.NewBeerService(interactor))\n}", "func TestEndpointCase45(t *testing.T) {\n\tvar params = EndpointParameters{\n\t\tUseDualStack: ptr.Bool(false),\n\t\tUseFIPS: ptr.Bool(false),\n\t\tEndpoint: ptr.String(\"https://example.com\"),\n\t}\n\n\tresolver := NewDefaultEndpointResolverV2()\n\tresult, err := resolver.ResolveEndpoint(context.Background(), params)\n\t_, _ = result, err\n\n\tif err != nil {\n\t\tt.Fatalf(\"expect no error, got %v\", err)\n\t}\n\n\turi, _ := url.Parse(\"https://example.com\")\n\n\texpectEndpoint := smithyendpoints.Endpoint{\n\t\tURI: *uri,\n\t\tHeaders: http.Header{},\n\t\tProperties: smithy.Properties{},\n\t}\n\n\tif e, a := expectEndpoint.URI, result.URI; e != a {\n\t\tt.Errorf(\"expect %v URI, got %v\", e, a)\n\t}\n\n\tif diff := cmp.Diff(expectEndpoint.Headers, result.Headers); diff != \"\" {\n\t\tt.Errorf(\"expect headers to match\\n%s\", diff)\n\t}\n\n\tif diff := cmp.Diff(expectEndpoint.Properties, result.Properties,\n\t\tcmp.AllowUnexported(smithy.Properties{}),\n\t); diff != \"\" {\n\t\tt.Errorf(\"expect properties to match\\n%s\", diff)\n\t}\n}", "func (s *SmartContract) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n \n fn, args := stub.GetFunctionAndParameters()\n\n if fn == \"AddComponent\" {\n return s.AddComponent(stub, args)\n } else if fn == \"InitLedger\" {\n return s.InitLedger(stub)\n } else if fn == \"QueryCar\" {\n return s.QueryCar(stub, args)\n } else if fn == \"QueryComponent\" {\n return s.QueryComponent(stub, args)\n }\n\n return shim.Error(\"Invalid Smart Contract function name.\")\n \n}", "func callMethod(v *reflect.Value, method string, params []interface{}) {\n f := (*v).MethodByName(method) //get method instance of structure \n if f.IsValid() {\n args := make([]reflect.Value, len(params)) //make a map as args for method \n for k, param := range params {\n args[k] = reflect.ValueOf(param)\n }\n // call method here \n ret := f.Call(args)\n if ret[0].Kind() == reflect.String {\n fmt.Printf(\"%s Called result: %s\\n\", method, ret[0].String())\n }\n } else {\n fmt.Println(\"can't call \" + method)\n }\n fmt.Println(\"\")\n}", "func mockMetadataEndpoint(w http.ResponseWriter, r *http.Request, data string) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tfmt.Fprint(w, data)\n}", "func newClientMethod(r *raml.Resource, rd *resourceDef, m *raml.Method, methodName, lang string) (methodInterface, error) {\n\tmethod := newMethod(r, rd, m, methodName)\n\n\tmethod.ResourcePath = paramizingURI(method.Endpoint)\n\n\tname := normalizeURITitle(method.Endpoint)\n\n\tmethod.ReqBody = setBodyName(m.Bodies, name+methodName, \"ReqBody\")\n\n\tswitch lang {\n\tcase langGo:\n\t\tgcm := goClientMethod{method: &method}\n\t\terr := gcm.setup(methodName)\n\t\treturn gcm, err\n\tcase langPython:\n\t\tpcm := pythonClientMethod{method: method}\n\t\tpcm.setup()\n\t\treturn pcm, nil\n\tdefault:\n\t\tpanic(\"invalid language:\" + lang)\n\n\t}\n}", "func (_SushiV2Factory *SushiV2FactoryRaw) Call(opts *bind.CallOpts, result *[]interface{}, method string, params ...interface{}) error {\n\treturn _SushiV2Factory.Contract.SushiV2FactoryCaller.contract.Call(opts, result, method, params...)\n}", "func (ri reflectInvoker) Signature(ctx *context.T, call ServerCall) ([]signature.Interface, error) {\n\treturn signature.CopyInterfaces(ri.sig), nil\n}", "func (m *Method) Method() {\n\n}" ]
[ "0.6673551", "0.6281075", "0.59397185", "0.57743317", "0.5699266", "0.55664563", "0.5489243", "0.5443471", "0.5373476", "0.53463346", "0.5330238", "0.53104776", "0.5306657", "0.5282867", "0.5250428", "0.5248784", "0.5234313", "0.5202553", "0.5201545", "0.5193338", "0.5143856", "0.51385117", "0.5137643", "0.51198906", "0.5103954", "0.5101805", "0.50619817", "0.5060108", "0.5030312", "0.5015583", "0.5003007", "0.4999757", "0.49940765", "0.49817163", "0.4956316", "0.49544668", "0.49495295", "0.49383163", "0.49339467", "0.49322858", "0.49268028", "0.4926376", "0.49199885", "0.49174312", "0.4914023", "0.49126637", "0.48974058", "0.48955482", "0.4893027", "0.489224", "0.48915628", "0.488919", "0.4888412", "0.48705265", "0.48675793", "0.48581123", "0.48455074", "0.48406804", "0.48343328", "0.48185375", "0.48163745", "0.48154396", "0.48077625", "0.4804713", "0.47982648", "0.4793185", "0.4792063", "0.47917518", "0.47828233", "0.47807983", "0.47795033", "0.4769155", "0.476094", "0.47585914", "0.47563583", "0.47558364", "0.47549364", "0.47544464", "0.4745933", "0.4744829", "0.47434053", "0.4736199", "0.47308835", "0.4720543", "0.4720145", "0.47191054", "0.4713585", "0.47124884", "0.4712423", "0.47084957", "0.47027788", "0.47018933", "0.4700151", "0.46964896", "0.46946782", "0.46878955", "0.46865642", "0.46848553", "0.46811345", "0.4679865" ]
0.6707497
0
The fake service will never need to glob.
func (fs fakeService) Globber() *rpc.GlobState { log.Printf("Fake Service Globber???") return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (badGlob3) Glob__(*context.T, rpc.GlobServerCall) {}", "func (badGlob4) Glob__(*context.T, rpc.GlobServerCall, *glob.Glob) {}", "func (badGlob2) Glob__(*context.T) {}", "func (badGlob5) Glob__(*context.T, rpc.ServerCall, *glob.Glob) error { return nil }", "func (badGlob6) Glob__() error { return nil }", "func (badGlob1) Glob__() {}", "func (cs *copyState) glob(pattern string) (files []manager.RepositoryEntry) {\n\tif pattern == \"\" {\n\t\tcs.state.Exitf(\"empty path name\")\n\t}\n\n\t// Path on local machine?\n\tif isLocal(pattern) {\n\t\tfor _, repoPath := range cs.state.GlobLocal(subcmd.Tilde(pattern)) {\n\t\t\tfiles = append(files, manager.RepositoryEntry{\n\t\t\t\tName: path.Base(repoPath),\n\t\t\t\tEnabled: true,\n\t\t\t\tLastUpdate: time.Now(),\n\t\t\t\tPath: repoPath,\n\t\t\t})\n\t\t}\n\t\treturn files\n\t}\n\n\t// Extra check to catch use of relative path on local machine.\n\tif !strings.Contains(pattern, \"@\") {\n\t\tcs.state.Exitf(\"local pattern not qualified path: %s\", pattern)\n\t}\n\n\treturn files\n}", "func BenchmarkService(b *testing.B) {\n\tfp := protofile\n\n\treader, _ := os.Open(fp)\n\tdefer reader.Close()\n\n\tparser := proto.NewParser(reader)\n\tdefinition, _ := parser.Parse()\n\n\tvar (\n\t\tcurrentService *proto.Service\n\t\tcurrentTemplateMap map[string]map[string]string\n\t)\n\n\thandlerService := proto.WithService(func(s *proto.Service) {\n\t\tclearRandoms()\n\t\tif rpc != \"\" && s.Name != rpc {\n\t\t\treturn\n\t\t}\n\t\tcurrentService = s\n\t\tif m, ok := Templates[s.Name]; ok {\n\t\t\tcurrentTemplateMap = m\n\t\t}\n\t})\n\n\thandlerRPC := proto.WithRPC(func(r *proto.RPC) {\n\t\tfor _, action := range []string{\"create\", \"update\", \"delete\"} {\n\t\t\tif m, ok := currentTemplateMap[action]; ok {\n\t\t\t\tif t, ok := m[r.Name]; ok {\n\n\t\t\t\t\tb.Run(fmt.Sprintf(\"%s\", r.Name), func(b *testing.B) {\n\t\t\t\t\t\t// if ovs { // skip unvalid service: useful when various services are defined in a single proto file\n\t\t\t\t\t\t// \ttokens := strings.Split(service, \".\")\n\t\t\t\t\t\t// \tsuffix := tokens[len(tokens)-1]\n\t\t\t\t\t\t// \ttcr := strings.ToLower(currentService)\n\t\t\t\t\t\t// \tif !strings.HasPrefix(tcr, suffix) {\n\t\t\t\t\t\t// \t\treturn\n\t\t\t\t\t\t// \t}\n\t\t\t\t\t\t// }\n\n\t\t\t\t\t\t// Gather error count\n\t\t\t\t\t\tvar gotError chan bool\n\t\t\t\t\t\tvar done chan int\n\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\tgotError = make(chan bool)\n\t\t\t\t\t\t\tdone = make(chan int)\n\t\t\t\t\t\t\tgo func() {\n\t\t\t\t\t\t\t\terrNb := 0\n\t\t\t\t\t\t\t\tfor {\n\t\t\t\t\t\t\t\t\tselect {\n\t\t\t\t\t\t\t\t\tcase nb := <-done:\n\t\t\t\t\t\t\t\t\t\tresults := float32(errNb) / float32(nb) * 100\n\t\t\t\t\t\t\t\t\t\tfmt.Printf(\"Benchmarking %s - %.00f%% error rate\\n\", currentService.Name+\".\"+r.Name, results)\n\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\tcase <-gotError:\n\t\t\t\t\t\t\t\t\t\terrNb++\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}()\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\t// Retrieve the test template\n\t\t\t\t\t\ttmpl, err := template.New(r.Name).Parse(t)\n\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tfor i := 0; i < b.N; i++ {\n\t\t\t\t\t\t\tif r.StreamsRequest && !r.StreamsReturns {\n\t\t\t\t\t\t\t\terr := benchStreamClient(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else if !r.StreamsRequest && r.StreamsReturns {\n\t\t\t\t\t\t\t\terr := benchStreamServer(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\terr := benchCall(b, action, currentService, r, tmpl)\n\t\t\t\t\t\t\t\tif err != nil {\n\t\t\t\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\t\t\t\tgotError <- true\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif logFailedCalls {\n\t\t\t\t\t\t\tdone <- b.N\n\t\t\t\t\t\t}\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t})\n\n\tproto.Walk(definition, handlerService, handlerRPC)\n}", "func glob(pattern string) []string {\n\tpattern, err := homedir.Expand(pattern)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tpaths, err := filepath.Glob(pattern)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif len(paths) == 0 {\n\t\tlog.Fatalf(\"%s: no such file or directory\", pattern)\n\t}\n\treturn paths\n}", "func testSingleFileGet(t *testing.T, tg *siatest.TestGroup) {\n\t// Grab the first of the group's renters\n\trenter := tg.Renters()[0]\n\t// Upload file, creating a piece for each host in the group\n\tdataPieces := uint64(1)\n\tparityPieces := uint64(len(tg.Hosts())) - dataPieces\n\tfileSize := 100 + siatest.Fuzz()\n\t_, _, err := renter.UploadNewFileBlocking(fileSize, dataPieces, parityPieces)\n\tif err != nil {\n\t\tt.Fatal(\"Failed to upload a file for testing: \", err)\n\t}\n\n\tfiles, err := renter.Files()\n\tif err != nil {\n\t\tt.Fatal(\"Failed to get renter files: \", err)\n\t}\n\n\tvar file modules.FileInfo\n\tfor _, f := range files {\n\t\tfile, err = renter.File(f.SiaPath)\n\t\tif err != nil {\n\t\t\tt.Fatal(\"Failed to request single file\", err)\n\t\t}\n\t\tif file != f {\n\t\t\tt.Fatal(\"Single file queries does not match file previously requested.\")\n\t\t}\n\t}\n}", "func (s *Server) Service() string { return \"file\" }", "func (allGlobberObject) Glob__(*context.T, rpc.GlobServerCall, *glob.Glob) error {\n\treturn nil\n}", "func (self *Path) GlobIt() []string {\n\tos.Chdir(self.srcPath)\n\tf, err := filepath.Glob(\"*.fcs\")\n\n\tcheck(err)\n\n\treturn f\n\n}", "func Service_public_files(fname string) ([]byte, string, error) {\n\td, err := fs.Get(fname)\n\tif nil != err {\n\t\treturn nil, \"\", err\n\t}\n\treturn d, http.DetectContentType(d), err\n}", "func mockAlwaysRun() bool { return true }", "func (m mockServicesGenerator) GenerateFileContent() {\n\tif len(m.file.Services) == 0 {\n\t\treturn\n\t}\n\tfor _, service := range m.file.Services {\n\t\tm.genService(service)\n\t}\n}", "func generateServiceImplFile(pdArr []ProtoData, option string) error {\n\tdirPath := filepath.Join(appPath)\n\t_, fileErr := os.Stat(dirPath)\n\tif fileErr != nil {\n\t\tos.MkdirAll(dirPath, os.ModePerm)\n\t}\n\tfor _, pd := range pdArr {\n\t\tconnectorFile := filepath.Join(appPath, strings.Split(protoFileName, \".\")[0]+\".\"+pd.RegServiceName+\".\"+option+\".grpcservice.go\")\n\t\tf, err := os.Create(connectorFile)\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"error: \", err)\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close()\n\t\tpd.Option = option\n\t\tif strings.Compare(option, \"server\") == 0 {\n\t\t\terr = registryServerTemplate.Execute(f, pd)\n\t\t} else {\n\t\t\terr = registryClientTemplate.Execute(f, pd)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (badGlobChildren3) GlobChildren__(*context.T, rpc.GlobChildrenServerCall) {}", "func Test_DeviceService_GetAll_Empty(t *testing.T) {\n\th := TestHelper{}\n\trep := new(mocks.IDeviceRepository)\n\trepAuth := new(mocks.IDeviceAuthRepository)\n\ts := h.CreateTestDeviceService(rep, repAuth)\n\n\tdevices := make([]models.Device, 0)\n\trep.On(\"GetAll\").Return(devices, nil)\n\n\tall, err := s.GetAll()\n\n\tassert.NoError(t, err)\n\tassert.Equal(t, len(all), len(devices))\n}", "func (_m *IService) StartupPurge() {\n\t_m.Called()\n}", "func (fsOnDisk) Glob(pattern string) ([]string, error) {\n\tvar result []string\n\tallFilePaths, err := filepath.Glob(pattern)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif IsHiddenFilePath(pattern) {\n\t\tresult = allFilePaths\n\t} else {\n\t\tresult = RemoveHiddenFiles(allFilePaths)\n\t}\n\treturn result, nil\n}", "func (fs *FakeFilesystem) Use() func() {\n\t// create the new fake fs root dir in /tmp/sriov...\n\ttmpDir, err := os.MkdirTemp(\"\", \"k8s-rdma-shared-dev-plugin-\")\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"error creating fake root dir: %s\", err.Error()))\n\t}\n\tfs.RootDir = tmpDir\n\n\tfor _, dir := range fs.Dirs {\n\t\tosErr := os.MkdirAll(path.Join(fs.RootDir, dir), 0700)\n\t\tif osErr != nil {\n\t\t\tpanic(fmt.Errorf(\"error creating fake directory: %s\", osErr.Error()))\n\t\t}\n\t}\n\tfor filename, body := range fs.Files {\n\t\tioErr := os.WriteFile(path.Join(fs.RootDir, filename), body, 0600)\n\t\tif ioErr != nil {\n\t\t\tpanic(fmt.Errorf(\"error creating fake file: %s\", ioErr.Error()))\n\t\t}\n\t}\n\tfor link, target := range fs.Symlinks {\n\t\tosErr := os.Symlink(target, path.Join(fs.RootDir, link))\n\t\tif osErr != nil {\n\t\t\tpanic(fmt.Errorf(\"error creating fake symlink: %s\", osErr.Error()))\n\t\t}\n\t}\n\terr = os.MkdirAll(path.Join(fs.RootDir, \"usr/share/hwdata\"), 0700)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"error creating fake directory: %s\", err.Error()))\n\t}\n\n\t// TODO: Remove writing pci.ids file once ghw is mocked\n\t// This is to fix the CI failure where ghw lib fails to\n\t// unzip pci.ids file downloaded from internet.\n\tpciData, err := os.ReadFile(\"/usr/share/hwdata/pci.ids\")\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"error reading file: %s\", err.Error()))\n\t}\n\terr = os.WriteFile(path.Join(fs.RootDir, \"usr/share/hwdata/pci.ids\"), pciData, 0600)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"error creating fake file: %s\", err.Error()))\n\t}\n\n\tsysNetDevices = path.Join(fs.RootDir, \"/sys/class/net\")\n\n\treturn func() {\n\t\t// remove temporary fake fs\n\t\terr := os.RemoveAll(fs.RootDir)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Errorf(\"error tearing down fake filesystem: %s\", err.Error()))\n\t\t}\n\t}\n}", "func TestDeviceFiles(ctx context.Context, s *testing.State, pattern string) {\n\tconst (\n\t\tdir = \"/dev/snd\"\n\t\tmode = 0660\n\t)\n\n\tfiles, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\ts.Fatal(\"Failed to list files at /dev/snd: \", err)\n\t}\n\n\tif f, err := os.Create(filepath.Join(s.OutDir(), \"ls.txt\")); err != nil {\n\t\ts.Error(\"Failed to open output file: \", err)\n\t} else {\n\t\tdefer f.Close()\n\t\tcmd := testexec.CommandContext(ctx, \"ls\", \"-l\", dir)\n\t\tcmd.Stdout = f\n\t\tcmd.Stderr = f\n\t\tif err := cmd.Run(); err != nil {\n\t\t\ts.Errorf(\"Failed to run ls on %v: %v\", dir, err)\n\t\t}\n\t}\n\n\tcheck := func(ps string) {\n\t\tp := regexp.MustCompile(ps)\n\t\tfound := false\n\t\tfor _, fi := range files {\n\t\t\tif p.MatchString(fi.Name()) {\n\t\t\t\tif fi.Mode()&0777 != mode {\n\t\t\t\t\ts.Errorf(\"%s: permission mismatch: expected %o, actually %o\", fi.Name(), mode, fi.Mode())\n\t\t\t\t}\n\t\t\t\tfound = true\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\ts.Errorf(\"No file matched %s\", ps)\n\t\t}\n\t}\n\n\tcheck(`^controlC\\d+$`)\n\tcheck(pattern)\n}", "func mockNeverRun() bool { return false }", "func glob(dir, pattern string) ([]string, error) {\n\tm := make([]string, 0)\n\tfi, err := pkger.Stat(dir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !fi.IsDir() {\n\t\treturn nil, err\n\t}\n\td, err := pkger.Open(dir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer d.Close()\n\n\tnames, _ := d.Readdir(-1)\n\n\tfor _, n := range names {\n\t\tmatched, err := filepath.Match(pattern, n.Name())\n\t\tif err != nil {\n\t\t\treturn m, err\n\t\t}\n\t\tif matched {\n\t\t\tm = append(m, dir+\"/\"+n.Name())\n\t\t}\n\t}\n\treturn m, nil\n}", "func TestGetUserServicePatched (t *testing.T){\n\tuser1, err := GetUserService(user_01.SocialNumber)\n\tassert.Equal(t, 200, err.HTTPStatus)\n\tassert.Equal(t, user1.Name, new_name_user_01)\n}", "func getFiles(ctx *context) (filesOut []os.FileInfo, errOut error) {\n\tpattern := filepath.Base(*ctx.src)\n\tfiles, err := ioutil.ReadDir(filepath.Dir(*ctx.src))\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfor _, file := range files {\n\t\tif res, err := filepath.Match(strings.ToLower(pattern), strings.ToLower(file.Name())); res {\n\t\t\tif err != nil {\n\t\t\t\terrOut = err\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfilesOut = append(filesOut, file)\n\t\t\tctx.estimatesize += uint64(file.Size())\n\t\t\t// fmt.Printf(\"prise en compte de %s\", file.Name())\n\t\t}\n\t}\n\treturn filesOut, nil\n}", "func (badGlobChildren5) GlobChildren__(*context.T, rpc.GlobChildrenServerCall) error { return nil }", "func (sfs *SuiteFS) SampleFiles(tb testing.TB, testDir string) []*File {\n\ttb.Helper()\n\n\tvfs := sfs.vfsSetup\n\n\tfiles := GetSampleFiles()\n\tfor _, file := range files {\n\t\tpath := vfs.Join(testDir, file.Path)\n\n\t\terr := vfs.WriteFile(path, file.Content, file.Mode)\n\t\tif err != nil {\n\t\t\ttb.Fatalf(\"WriteFile %s : want error to be nil, got %v\", path, err)\n\t\t}\n\t}\n\n\treturn files\n}", "func getJSONFiles(dir string) (jsonFiles []string, err error) {\r\n\tfileMask := filepath.Join(dir, \"*.json\")\r\n\tjsonFiles, err = filepath.Glob(fileMask)\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\treturn jsonFiles, nil\r\n}", "func (suite *LabelManagerTestSuite) TestLabelManagerHasLabelWildcardService() {\n\tlabelManager := NewLabelManager(&LabelManagerConfig{\n\t\tEntries: []*LabelManagerConfigEntry{\n\t\t\t{Procedures: []string{\"*:Procedure1\"}, Labels: []string{\"label1\", \"label2\"}},\n\t\t\t{Procedures: []string{\"*:Procedure2\"}, Labels: []string{\"label3\", \"label4\"}},\n\t\t\t{Procedures: []string{\"*:Procedure3*\"}, Labels: []string{\"label5\", \"label6\"}},\n\t\t},\n\t})\n\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure1\", \"label1\"))\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure1\", \"label2\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure1\", \"label1\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure1\", \"label2\"))\n\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure2\", \"label3\"))\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure2\", \"label4\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure2\", \"label3\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure2\", \"label4\"))\n\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure3\", \"label5\"))\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure3\", \"label6\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure3\", \"label5\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure3\", \"label6\"))\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure34\", \"label5\"))\n\tsuite.True(labelManager.HasLabel(\"testService1::Procedure34\", \"label6\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure34\", \"label5\"))\n\tsuite.True(labelManager.HasLabel(\"testService2::Procedure34\", \"label6\"))\n\n\tsuite.False(labelManager.HasLabel(\"testService1::Procedure1\", \"label3\"))\n\tsuite.False(labelManager.HasLabel(\"testService1::Procedure1\", \"label4\"))\n\tsuite.False(labelManager.HasLabel(\"testService2::Procedure1\", \"label3\"))\n\tsuite.False(labelManager.HasLabel(\"testService2::Procedure1\", \"label4\"))\n\n\tsuite.False(labelManager.HasLabel(\"testService1::Procedure2\", \"label1\"))\n\tsuite.False(labelManager.HasLabel(\"testService1::Procedure2\", \"label2\"))\n\tsuite.False(labelManager.HasLabel(\"testService2::Procedure2\", \"label1\"))\n\tsuite.False(labelManager.HasLabel(\"testService2::Procedure2\", \"label2\"))\n}", "func (badGlobChildren2) GlobChildren__(*context.T) {}", "func TestGenerateComposefileGlobs(t *testing.T) {\n\tt.Parallel()\n\tglobs := strings.Join([]string{filepath.Join(generateComposeBaseDir, \"globs\", \"**\", \"docker-compose.yml\"), filepath.Join(generateComposeBaseDir, \"globs\", \"docker-compose.yml\")}, \",\")\n\tflags := []string{fmt.Sprintf(\"--compose-file-globs=%s\", globs)}\n\tcomposefiles := []string{filepath.Join(generateComposeBaseDir, \"globs\", \"image\", \"docker-compose.yml\"), filepath.Join(generateComposeBaseDir, \"globs\", \"docker-compose.yml\")}\n\tvar tOs []generateTestObject\n\tfor _, composefile := range composefiles {\n\t\ttO := generateTestObject{\n\t\t\tfilePath: filepath.ToSlash(composefile),\n\t\t\twantImages: []generate.ComposefileImage{\n\t\t\t\t{Image: generate.Image{Name: \"busybox\", Tag: \"latest\"}, ServiceName: \"svc\", Dockerfile: \"\"},\n\t\t\t},\n\t\t\ttestFn: checkGenerateComposefile,\n\t\t}\n\t\ttOs = append(tOs, tO)\n\t}\n\ttestGenerate(t, flags, tOs)\n}", "func EchoServer(impl EchoServerMethods) EchoServerStub {\n\tstub := implEchoServerStub{\n\t\timpl: impl,\n\t}\n\t// Initialize GlobState; always check the stub itself first, to handle the\n\t// case where the user has the Glob method defined in their VDL source.\n\tif gs := rpc.NewGlobState(stub); gs != nil {\n\t\tstub.gs = gs\n\t} else if gs := rpc.NewGlobState(impl); gs != nil {\n\t\tstub.gs = gs\n\t}\n\treturn stub\n}", "func mockProcessFinderZeroDaemon(name string) ([]process.Process, error) {\n\treturn []process.Process{}, nil\n\n}", "func globSources(ctx context.Context, path string, extensions []string) (srcSet, error) {\n\tvar allSourcePaths []string\n\tfor _, extension := range extensions {\n\t\tpattern := \"*.\" + extension\n\t\tmatched, err := platform.Glob(ctx, filepath.Join(path, pattern))\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"glob(%s): %s\", pattern, err)\n\t\t}\n\t\tallSourcePaths = append(allSourcePaths, matched...)\n\t}\n\tsrcs := make(srcSet)\n\tfor _, p := range allSourcePaths {\n\t\tfileName := filepath.Base(p)\n\t\tif isTempFile(fileName) {\n\t\t\tcontinue // Ignore editor swap/backup files.\n\t\t}\n\t\t// Try platform.Stat as a fallback, for Google file systems.\n\t\t_, err := platform.Stat(ctx, p)\n\t\tif os.IsNotExist(err) {\n\t\t\tplatform.Infof(\"platform.Glob returned non-existent file (dangling symlink?). Ignoring %q.\", p)\n\t\t\tcontinue\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"cannot stat platform.Glob result %q: %v\", p, err)\n\t\t}\n\t\tp, err := filepath.Rel(path, p)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"filepath.Rel(%s, %s): %v\", path, p, err)\n\t\t}\n\t\tsrcs[p] = true\n\t}\n\treturn srcs, nil\n}", "func TestLogGlob(t *testing.T) {\n\tdir, err := ioutil.TempDir(\"\", \"dircacheserverlog\")\n\tif err != nil {\n\t\tt.Fatal(\"creating test directory\")\n\t}\n\tdefer os.RemoveAll(dir)\n\tl, err := openLog(config.SetUserName(config.New(), testUser), dir, 1000000)\n\tif err != nil {\n\t\tt.Fatal(\"creating test log\")\n\t}\n\n\t// Log the glob entry.\n\tvar entries []*upspin.DirEntry\n\tfor i := 0; i < 10; i++ {\n\t\tde := mkDirEntry(fmt.Sprintf(\"[email protected]/a/b/c/%d\", i))\n\t\tde.Sequence = int64(upspin.SeqBase + i)\n\t\tentries = append(entries, de)\n\t}\n\tl.logGlobRequest(\"[email protected]/a/b/c/*\", nil, entries)\n\n\t// Check for individual entries.\n\tnentries, err, ok := l.lookupGlob(\"[email protected]/a/b/c/*\")\n\tif err != nil {\n\t\tt.Fatalf(\"lookupGlob: %v\", err)\n\t}\n\tif !ok {\n\t\tt.Fatalf(\"lookupGlob not found\")\n\t}\n\tif len(nentries) != len(entries) {\n\t\tt.Fatalf(\"lookupGlob missing entries: %d instead of %d\", len(nentries), len(entries))\n\t}\nl:\n\tfor _, ode := range entries {\n\t\tfor _, nde := range nentries {\n\t\t\tif reflect.DeepEqual(nde, ode) {\n\t\t\t\tcontinue l\n\t\t\t}\n\t\t}\n\t\tt.Fatalf(\"lookupGlob missing %v\", *ode)\n\t}\n\tl.close()\n\tt.Log(\"reopening log\")\n\n\t// Reopen, and ensure the glob services.\n\tl, err = openLog(config.SetUserName(config.New(), testUser), dir, 1000000)\n\tif err != nil {\n\t\tt.Fatal(\"creating test log\")\n\t}\n\tnentries, err, ok = l.lookupGlob(\"[email protected]/a/b/c/*\")\n\tif err != nil {\n\t\tt.Fatalf(\"lookupGlob: %v\", err)\n\t}\n\tif !ok {\n\t\tt.Fatalf(\"lookupGlob not found\")\n\t}\n\tif len(nentries) != len(entries) {\n\t\tt.Fatalf(\"lookupGlob (after reopen) missing entries: %d instead of %d\", len(nentries), len(entries))\n\t}\nl2:\n\tfor _, ode := range entries {\n\t\tfor _, nde := range nentries {\n\t\t\tif reflect.DeepEqual(nde, ode) {\n\t\t\t\tcontinue l2\n\t\t\t}\n\t\t}\n\t\tt.Fatalf(\"lookupGlob (after reopen) missing %v\", *ode)\n\t}\n\tl.close()\n}", "func (g *Generator) FService() *GeneratorFile {\n\treturn g.F(FILEID_SERVICE)\n}", "func Test_searchWord_FindMockFile_MockFileFound(t *testing.T) {\n\t//afero can be used here to mock filesystem\n}", "func mustGlobFiles(pattern string) []string {\n\tmatches, err := filepath.Glob(pattern)\n\tif matches == nil {\n\t\tpanic(\"Unable to locate required template files in working dir that match: \" + pattern)\n\t} else if err != nil {\n\t\tpanic(err) // Programmer error, bad glob pattern\n\t}\n\treturn matches\n}", "func testResourceAll(t *testing.T, s *Service) {\n\tres, err := s.ResourceAll(context.TODO())\n\tif err != nil {\n\t\tt.Logf(\"testResourceAll error(%v) \\n\", err)\n\t\treturn\n\t}\n\tt.Logf(\"testResourceAll res: %+v \\n\", res)\n}", "func runService(ctx context.Context, targetDir string, service string) {\n\t// Save and restore lsater current working dir\n\twd, err := os.Getwd()\n\tcheck(err)\n\tdefer os.Chdir(wd)\n\n\t// Build the server if needed\n\t_, err = os.Stat(\"./\" + service)\n\tif os.IsNotExist(err) {\n\t\tout, err := exec.Command(\"go\", \"build\", \".\").CombinedOutput()\n\t\tlog.Println(out)\n\t\tcheck(err)\n\t}\n\n\tcmd := exec.CommandContext(ctx, \"./\"+service)\n\terr = cmd.Start()\n\tcheck(err)\n}", "func TestFowler(t *testing.T) {\n\tfiles, err := filepath.Glob(\"testdata/*.dat\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tfor _, file := range files {\n\t\tt.Log(file)\n\t\ttestFowler(t, file)\n\t}\n}", "func serverMock() *httptest.Server {\n\thandler := http.NewServeMux()\n\thandler.HandleFunc(\"/planetary/apod\", picturesMock)\n\tsrv := httptest.NewServer(handler)\n\treturn srv\n}", "func mockFilters(filterNames []string) []filters.Spec {\n\tfs := make([]filters.Spec, len(filterNames))\n\tfor i, filterName := range filterNames {\n\t\tfs[i] = &filtertest.Filter{\n\t\t\tFilterName: filterName,\n\t\t}\n\t}\n\treturn fs\n}", "func testFullCycleFilesystem(t *testing.T) {\n\tt.Log(\"TODO\")\n}", "func ScoreKeeperServer(impl ScoreKeeperServerMethods) ScoreKeeperServerStub {\n\tstub := implScoreKeeperServerStub{\n\t\timpl: impl,\n\t}\n\t// Initialize GlobState; always check the stub itself first, to handle the\n\t// case where the user has the Glob method defined in their VDL source.\n\tif gs := rpc.NewGlobState(stub); gs != nil {\n\t\tstub.gs = gs\n\t} else if gs := rpc.NewGlobState(impl); gs != nil {\n\t\tstub.gs = gs\n\t}\n\treturn stub\n}", "func start() {\n\n\tif LOGERR != nil {\n\t\tpanic(LOGERR)\n\t}\n\n\tservicePath := &CONFIG.ServiceConfig.Path\n\tremoveSlashes(servicePath)\n\n\tdb := new(DB)\n\tdb.getClient()\n\n\tLOGGER.Warning(fmt.Sprintf(\"Scanning %s\\n\", *servicePath))\n\n\tif _, err := os.Stat(\"/\" + *servicePath); err != nil {\n\t\tif crErr := os.Mkdir(\"/\"+*servicePath, 0755); crErr != nil {\n\t\t\tLOGGER.Crit(fmt.Sprintf(\"Scanning %s failed - directory does not exist and is not creatable\\n\", *servicePath))\n\t\t\tfmt.Printf(\"Scanning %s failed - directory does not exist and is not creatable\\n\", *servicePath)\n\t\t\tusage(1)\n\t\t}\n\t}\n\n\trunningServices := make(map[string]*Service)\n\n\t// Loop knownServices and services in directory\n\t// If differ, decide which to remove or add\n\tfor {\n\t\tservicesInDir := readServiceDir(servicePath)\n\t\tdb.createNewServicesIfNeeded(&servicesInDir, servicePath)\n\t\tknownServices := db.getServices()\n\n\t\tfor serviceName, service := range knownServices {\n\t\t\tserviceName := serviceName\n\t\t\tservice := service\n\n\t\t\tsrvDone := make(chan error, 1)\n\n\t\t\t_, ok := runningServices[serviceName]\n\t\t\tif ok != true {\n\t\t\t\t// service is not yet running\n\t\t\t\t// so start it and a logger\n\t\t\t\tgo func() {\n\t\t\t\t\terr1 := updateServicePaths(&knownServices, servicePath)\n\t\t\t\t\terr2 := removeServiceBefore(&servicesInDir, serviceName)\n\t\t\t\t\tif err1 == nil && err2 == nil {\n\t\t\t\t\t\tLOGGER.Debug(fmt.Sprintf(\"%s not yet running\\n\", serviceName))\n\t\t\t\t\t\ttime.Sleep(1 * time.Second)\n\t\t\t\t\t\tsv := new(ServiceHandler)\n\t\t\t\t\t\tsv.mutex = &sync.Mutex{}\n\t\t\t\t\t\tsv.service = service\n\t\t\t\t\t\tsv.startService(srvDone, runningServices, serviceName)\n\t\t\t\t\t}\n\t\t\t\t}()\n\t\t\t} else {\n\t\t\t\t// the service is running\n\t\t\t\t// but might have been removed manually (rm)\n\t\t\t\terr := removeServiceAfter(&servicesInDir, serviceName, runningServices[serviceName], srvDone)\n\t\t\t\tif err == nil {\n\t\t\t\t\tLOGGER.Debug(fmt.Sprintf(\"%s already running\\n\", serviceName))\n\t\t\t\t} else {\n\t\t\t\t\tdelete(runningServices, serviceName)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\ttime.Sleep(5 * time.Second)\n\t}\n\n\tLOGGER.Warning(\"exiting\")\n}", "func (poolMgr *Poolmgr) tapService(w http.ResponseWriter, r *http.Request) {\n\tbody, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\thttp.Error(w, \"Failed to read request\", 500)\n\t\treturn\n\t}\n\tsvcName := string(body)\n\tsvcHost := strings.TrimPrefix(svcName, \"http://\")\n\n\terr = poolMgr.fsCache.TouchByAddress(svcHost)\n\tif err != nil {\n\t\tlog.Printf(\"funcSvc tap error: %v\", err)\n\t\thttp.Error(w, \"Not found\", 404)\n\t\treturn\n\t}\n\tw.WriteHeader(http.StatusOK)\n}", "func (mock *EnvironmentMock) UseServiceCalls() []struct {\n\tName string\n} {\n\tvar calls []struct {\n\t\tName string\n\t}\n\tlockEnvironmentMockUseService.RLock()\n\tcalls = mock.calls.UseService\n\tlockEnvironmentMockUseService.RUnlock()\n\treturn calls\n}", "func (o *globObject) Glob__(ctx *context.T, call rpc.GlobServerCall, g *glob.Glob) error {\n\to.globLoop(call, \"\", g, o.n)\n\treturn nil\n}", "func generateDummyFiles(ctx *Context) (err error) {\n\terr = ioutil.WriteFile(filepath.Join(ctx.CGenDir, \"_dummy.h\"), []byte(\"// Dummy File\"), 0755)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = ioutil.WriteFile(filepath.Join(ctx.CPPGenDir, \"_dummy.h\"), []byte(\"// Dummy File\"), 0755)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = ioutil.WriteFile(filepath.Join(ctx.CPPGenDir, \"_dummy.cpp\"), []byte(\"// Dummy File\"), 0755)\n\treturn\n}", "func (r *ReconcileTFAnalytics) handleAlarmGenService() (bool, error) {\n\t// Define a new AlarmGen service object\n\tsvcmService := newServicesForAlarmGen(r.instance)\n\t// Set TFAnalytics instance as the owner and controller\n\tif err := controllerutil.SetControllerReference(r.instance, svcmService, r.scheme); err != nil {\n\t\treturn false, err\n\t}\n\t// Check if this AlarmGen Service already exists\n\tfoundSvcmService := &corev1.Service{}\n\terr := r.client.Get(context.TODO(), types.NamespacedName{Name: svcmService.Name, Namespace: svcmService.Namespace}, foundSvcmService)\n\tif err != nil && errors.IsNotFound(err) {\n\t\tr.reqLogger.Info(\"Creating a new AlarmGen Service\", \"Service.Name\", svcmService.Name)\n\t\terr = r.client.Create(context.TODO(), svcmService)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\t// Service has been created successfully - don't requeue\n\t\treturn false, nil\n\t} else if err != nil {\n\t\treturn false, err\n\t}\n\t// Service already exists - don't requeue\n\tr.reqLogger.Info(\"Skip reconcile: AlarmGen Service already exists\", \"Service.Name\", foundSvcmService.Name)\n\treturn false, nil\n}", "func newServiceNoPatchTest(name string, options ...controllertesting.ServiceOption) TableRow {\n\ttest := newDispatcherBasicTest(\"Existing Dispatcher Service, \" + name + \", No Patch\")\n\ttest.Objects = append(test.Objects,\n\t\tcontrollertesting.NewKafkaChannelDispatcherService(options...),\n\t\tcontrollertesting.NewKafkaChannelDispatcherDeployment())\n\treturn test\n}", "func GlobPatternsWithLoader(fSys filesys.FileSystem, ldr ifc.Loader, patterns []string) ([]string, error) {\n\tvar result []string\n\tfor _, pattern := range patterns {\n\t\tfiles, err := fSys.Glob(pattern)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif len(files) == 0 {\n\t\t\tloader, err := ldr.New(pattern)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"%s has no match: %w\", pattern, err)\n\t\t\t} else {\n\t\t\t\tresult = append(result, pattern)\n\t\t\t\tif loader != nil {\n\t\t\t\t\tloader.Cleanup()\n\t\t\t\t}\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, files...)\n\t}\n\treturn result, nil\n}", "func (m *SDMock) SetupHandlers() {\n\theaders := make(map[string]string)\n\trawHeaders, err := os.ReadFile(filepath.Join(\"testdata\", m.directory, \"headers.yml\"))\n\trequire.NoError(m.t, err)\n\tyaml.Unmarshal(rawHeaders, &headers)\n\n\tprefix := \"/\"\n\tif v, ok := headers[\"Api-Version\"]; ok {\n\t\tprefix += \"v\" + v + \"/\"\n\t}\n\n\tfor _, path := range []string{\"_ping\", \"networks/\", \"services/\", \"nodes/\", \"nodes\", \"services\", \"tasks\", \"containers/\"} {\n\t\tp := path\n\t\thandler := prefix + p\n\t\tif p == \"_ping\" {\n\t\t\thandler = \"/\" + p\n\t\t}\n\t\tm.Mux.HandleFunc(handler, func(w http.ResponseWriter, r *http.Request) {\n\t\t\t// The discovery should only call each API endpoint once.\n\t\t\tm.calls[r.RequestURI]++\n\t\t\tif m.calls[r.RequestURI] != 1 {\n\t\t\t\tw.WriteHeader(http.StatusTooManyRequests)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor k, v := range headers {\n\t\t\t\tw.Header().Add(k, v)\n\t\t\t}\n\t\t\tparts := strings.Split(r.RequestURI, \"/\")\n\t\t\tvar f string\n\t\t\tif strings.HasSuffix(p, \"/\") {\n\t\t\t\tf = filepath.Join(p[:len(p)-1], strutil.SanitizeLabelName(parts[len(parts)-1]))\n\t\t\t} else {\n\t\t\t\tquery := strings.Split(parts[len(parts)-1], \"?\")\n\t\t\t\tf = query[0] + \".json\"\n\t\t\t\tif len(query) == 2 {\n\t\t\t\t\th := sha1.New()\n\t\t\t\t\th.Write([]byte(query[1]))\n\t\t\t\t\t// Avoing long filenames for Windows.\n\t\t\t\t\tf += \"__\" + base64.URLEncoding.EncodeToString(h.Sum(nil))[:10]\n\t\t\t\t}\n\t\t\t}\n\t\t\tif response, err := os.ReadFile(filepath.Join(\"testdata\", m.directory, f+\".json\")); err == nil {\n\t\t\t\tw.Header().Add(\"content-type\", \"application/json\")\n\t\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\t\tw.Write(response)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif response, err := os.ReadFile(filepath.Join(\"testdata\", m.directory, f)); err == nil {\n\t\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\t\tw.Write(response)\n\t\t\t\treturn\n\t\t\t}\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t})\n\t}\n}", "func Glob(merge bool) error {\n\tvar paths []path\n\t// get paths for wich Go code for JSON schemas should be generated.\n\tfor _, p := range strings.Split(os.Getenv(\"GOPATH\"),\n\t\tstring(os.PathListSeparator)) {\n\t\tif p == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tpaths = append(paths, globGopath(p)...)\n\t}\n\tch, ret := make(chan path, len(paths)), make(chan error)\n\tfor _, r := range paths {\n\t\tch <- r\n\t}\n\tdefer close(ch)\n\tfor n := min(runtime.GOMAXPROCS(-1), len(paths)); n > 0; n-- {\n\t\tgo func() {\n\t\t\tfor c := range ch {\n\t\t\t\tret <- New(merge).Generate(c.in, c.out)\n\t\t\t}\n\t\t}()\n\t}\n\tvar e error\n\tfor _ = range paths {\n\t\tif err := <-ret; err != nil {\n\t\t\te = err\n\t\t}\n\t}\n\treturn e\n}", "func TestDirStar(t *testing.T) {\n\t// Init test files\n\tcurDir, _ := os.Getwd()\n\ttmpDir := filepath.Join(curDir, \"./tmp\")\n\tdefer os.RemoveAll(tmpDir)\n\tmakeTmpFiles(tmpDir, []string{\n\t\t\".git\",\n\t\t\"app.js\",\n\t\t\"package.json\",\n\t\t\"src/router.js\",\n\t\t\"src/store.js\",\n\t\t\"src/api/home.js\",\n\t\t\"src/api/user.js\",\n\t\t\"src/api/test.js\",\n\t})\n\n\tpatterns := []string{\n\t\t\"src/**/*\",\n\t}\n\n\t// Match the patterns\n\tfiles := Match(patterns, Option{BaseDir: tmpDir})\n\t// Expected match files:\n\texpected := []string{\n\t\t\"src/router.js\",\n\t\t\"src/store.js\",\n\t\t\"src/api/home.js\",\n\t\t\"src/api/user.js\",\n\t\t\"src/api/test.js\",\n\t}\n\tif checkFiles(tmpDir, files, expected) {\n\t\tt.Errorf(\"files not match, expected %v, but got %v\", expected, files)\n\t}\n}", "func prepareService(path string) *service.Service {\n\tpath, didDownload, err := downloadServiceIfNeeded(path)\n\tutils.HandleError(err)\n\tif didDownload {\n\t\tdefer os.RemoveAll(path)\n\t\tfmt.Printf(\"%s Service downloaded with success\\n\", aurora.Green(\"✔\"))\n\t}\n\timportedService, err := importer.From(path)\n\thandleValidationError(err)\n\tutils.HandleError(err)\n\timageHash, err := buildDockerImage(path)\n\tutils.HandleError(err)\n\tfmt.Printf(\"%s Image built with success\\n\", aurora.Green(\"✔\"))\n\tinjectConfigurationInDependencies(importedService, imageHash)\n\treturn importedService\n}", "func (badGlobChildren1) GlobChildren__() {}", "func (m *CloudWatchLogsServiceMock) CreateNewServiceIfUnHealthy() {\n\n}", "func handleService(req typhon.Request) typhon.Response {\n\tparts := reService.FindStringSubmatch(req.URL.Path)\n\tif len(parts) != 3 {\n\t\treturn typhon.Response{Error: terrors.NotFound(\"bad_endpoint\", \"Unable to determine service endpoint.\", nil)}\n\t}\n\n\treturn handle(req, \"s-\"+parts[1], parts[2])\n}", "func listFilesWithName(name string) []*drive.File {\r\n // set up the call\r\n list := fs.List()\r\n query := \"name = '\" + name + \"'\"\r\n list.Q(query)\r\n list.Fields(\"nextPageToken, files(id, name, parents, mimeType)\")\r\n // make the call\r\n random := strconv.Itoa(rand.Int())\r\n \r\n // this is terrible but works. If the server gives an error, just try again\r\n var files *drive.FileList = nil\r\n var err error = errors.New(\"empty error\")\r\n for err != nil {\r\n files, err = list.Do(googleapi.QuotaUser(random))\r\n }\r\n // return the contained array\r\n return files.Files\r\n}", "func createMockServer() (*httptest.Server, error) {\n\tdata, err := os.ReadFile(\"testdata/metadata_service.json\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Note(ederst): source of inspiration https://clavinjune.dev/en/blogs/mocking-http-call-in-golang-a-better-way/\n\tmockServer := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tif strings.HasSuffix(r.URL.Path, \"/openstack/latest/meta_data.json\") {\n\t\t\tmockMetadataEndpoint(w, r, string(data))\n\t\t} else {\n\t\t\thttp.NotFoundHandler().ServeHTTP(w, r)\n\t\t}\n\t}))\n\n\treturn mockServer, nil\n}", "func GetFileList(uuid string) {\n owlh, err := ndb.GetStapServerInformation(uuid)\n if err != nil {\n logs.Error(\"Error retrieving stap server information\")\n }\n logs.Error(\"Get file list for \" + owlh[\"name\"] + \" - \" + owlh[\"ip\"])\n file_list := GetFileListSSH(uuid, owlh, owlh[\"pcap_path\"])\n for file := range file_list {\n var validOutput = regexp.MustCompile(`\\.pcap+`)\n if validOutput.MatchString(file_list[file]) {\n logs.Notice(\"Change remote file owned\")\n OwnerOwlh(uuid, owlh, file_list[file])\n\n logs.Notice(\"Copy files using sftp command and remove it!!\")\n TransportFile(uuid, owlh, file_list[file])\n\n logs.Warn(\"File list completed!\")\n }\n }\n\n}", "func DischargerServer(impl DischargerServerMethods) DischargerServerStub {\n\tstub := implDischargerServerStub{\n\t\timpl: impl,\n\t}\n\t// Initialize GlobState; always check the stub itself first, to handle the\n\t// case where the user has the Glob method defined in their VDL source.\n\tif gs := rpc.NewGlobState(stub); gs != nil {\n\t\tstub.gs = gs\n\t} else if gs := rpc.NewGlobState(impl); gs != nil {\n\t\tstub.gs = gs\n\t}\n\treturn stub\n}", "func main() {\n\n\t// Prepare some dependencies:\n\tlogger := logrus.New()\n\tstorer := new(storageMocks.FakeStorer)\n\n\t// Program the storer mock to respond with _something_:\n\tstorer.CreateCruftReturns(\"12345\", nil)\n\tstorer.ReadCruftReturns(nil, storage.ErrNotFound)\n\n\t// Inject the dependencies into a new Handler:\n\thandler := serviceHandler.New(logger, storer)\n\n\t// Make a new GRPC Server (usually I would have this in a common / shared library, and pre-load it with middleware built from our logger / instrumenter / tracer interfaces):\n\tgrpcServer := grpc.NewServer()\n\n\t// Register our Handler and GRPC Server with our generated service-proto code:\n\tserviceProto.RegisterExampleServer(grpcServer, handler)\n\n\t// Listen for connections:\n\tlistener, err := net.Listen(\"tcp\", listenAddress)\n\tif err != nil {\n\t\tlogger.Fatalf(\"Unable to start GRPC server on TCP address %s\", listenAddress)\n\t}\n\n\t// Start the GRPC server:\n\tif err := grpcServer.Serve(listener); err != nil {\n\t\tlogger.Fatalf(\"Unable to start the GRPC server: %v\", err)\n\t}\n}", "func jsonFilenames(t *testing.T) []string {\n\tmatches, err := filepath.Glob(filepath.Join(jsonDir, \"*.json\"))\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\treturn matches\n}", "func Glob(filePattern string) ([]string, error) {\n\treturn zglob.Glob(filePattern)\n}", "func (s *SidecarApi) oneServiceHandler(response http.ResponseWriter, req *http.Request, params map[string]string) {\n\tdefer req.Body.Close()\n\n\tresponse.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tresponse.Header().Set(\"Access-Control-Allow-Methods\", \"GET\")\n\tresponse.Header().Set(\"Content-Type\", \"application/json\")\n\n\tif params[\"extension\"] != \"json\" {\n\t\tsendJsonError(response, 404, \"Not Found - Invalid content type extension\")\n\t\treturn\n\t}\n\n\tname, ok := params[\"name\"]\n\tif !ok {\n\t\tsendJsonError(response, 404, \"Not Found - No service name provided\")\n\t\treturn\n\t}\n\n\tif s.state == nil {\n\t\tsendJsonError(response, 500, \"Internal Server Error - Something went terribly wrong\")\n\t\treturn\n\t}\n\n\tvar instances []*service.Service\n\t// Enter critical section\n\ts.state.RLock()\n\tdefer s.state.RUnlock()\n\ts.state.EachService(func(hostname *string, id *string, svc *service.Service) {\n\t\tif svc.Name == name {\n\t\t\tinstances = append(instances, svc)\n\t\t}\n\t})\n\n\t// Did we have any entries for this service in the catalog?\n\tif len(instances) == 0 {\n\t\tsendJsonError(response, 404, fmt.Sprintf(\"no instances of %s found\", name))\n\t\treturn\n\t}\n\n\tclusterName := \"\"\n\tif s.list != nil {\n\t\tclusterName = s.list.ClusterName()\n\t}\n\n\t// Everything went fine, we found entries for this service.\n\t// Send the json back.\n\tsvcInstances := make(map[string][]*service.Service)\n\tsvcInstances[name] = instances\n\tresult := ApiServices{\n\t\tServices: svcInstances,\n\t\tClusterName: clusterName,\n\t}\n\n\tjsonBytes, err := json.MarshalIndent(&result, \"\", \" \")\n\tif err != nil {\n\t\tlog.Errorf(\"Error marshaling state in oneServiceHandler: %s\", err.Error())\n\t\tsendJsonError(response, 500, \"Internal server error\")\n\t\treturn\n\t}\n\n\t_, err = response.Write(jsonBytes)\n\tif err != nil {\n\t\tlog.Errorf(\"Error writing one service response to client: %s\", err)\n\t}\n}", "func extractServices(param *types.Proto, protos []*descriptor.FileDescriptorProto) {\n\tsvcTmp := make([]types.Service, 0, 0)\n\tvar p *descriptor.FileDescriptorProto\n\tfor _, p = range protos {\n\t\tif generate, _ := inArray(p.GetName(), param.FilesToGenerate); generate {\n\t\t\tfor _, svc := range p.GetService() {\n\t\t\t\ts := types.Service{Name: svc.GetName()}\n\t\t\t\tmethods := make([]types.MethodTemplate, len(svc.Method))\n\t\t\t\tfor j, method := range svc.Method {\n\t\t\t\t\ttipe := methodTypeStandard\n\t\t\t\t\tif method.GetServerStreaming() && !method.GetClientStreaming() {\n\t\t\t\t\t\ttipe = methodTypeServerStream\n\t\t\t\t\t} else if !method.GetServerStreaming() && method.GetClientStreaming() {\n\t\t\t\t\t\ttipe = methodTypeClientStream\n\t\t\t\t\t} else if method.GetServerStreaming() && method.GetClientStreaming() {\n\t\t\t\t\t\ttipe = methodTypeBidirectional\n\t\t\t\t\t}\n\t\t\t\t\t_, pkg := getGoPackage(p)\n\t\t\t\t\tmethods[j] = types.MethodTemplate{\n\t\t\t\t\t\tName: strings.Title(*method.Name),\n\t\t\t\t\t\tServiceName: svc.GetName(),\n\t\t\t\t\t\tInput: getMessageType(pkg, protos, p.GetDependency(), method.GetInputType()),\n\t\t\t\t\t\tOutput: getMessageType(pkg, protos, p.GetDependency(), method.GetOutputType()),\n\t\t\t\t\t\tMethodType: tipe,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\ts.Methods = methods\n\t\t\t\tsvcTmp = append(svcTmp, s)\n\t\t\t}\n\t\t}\n\n\t}\n\tif len(svcTmp) != 0 {\n\t\t_, pkg := getGoPackage(p)\n\t\tpath, outFile := getOutFile(p, pkg)\n\t\tparam.Services = svcTmp\n\t\tparam.Package = pkg\n\t\tparam.PackagePath = path\n\t\tparam.OutFile = outFile\n\t}\n}", "func TestFileServer(t *testing.T) {\n\tdir, err := ioutil.TempDir(\"\", \"cachetest\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tif !keepTempDir {\n\t\tdefer os.RemoveAll(dir)\n\t} else {\n\t\tprintln(dir)\n\t}\n\n\tfs := http.FileServer(http.Dir(\"testdata\"))\n\tsrvActual := httptest.NewServer(Handler(dir, fs))\n\tsrvExpected := httptest.NewServer(fs)\n\tdefer srvActual.Close()\n\tdefer srvExpected.Close()\n\n\tvar wg sync.WaitGroup\n\tfor i := 0; i < 10; i++ {\n\t\twg.Add(1)\n\t\tgo func() {\n\t\t\tdefer wg.Done()\n\t\t\tresp, err := http.Head(srvActual.URL + \"/file.txt\")\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tresp.Body.Close()\n\t\t\trespExpected, err := http.Head(srvExpected.URL + \"/file.txt\")\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\trespExpected.Body.Close()\n\t\t\tif !headersEqual(resp.Header, respExpected.Header) {\n\t\t\t\tt.Log(resp.Request.Method, resp.StatusCode, resp.Header)\n\t\t\t\tt.Log(respExpected.Request.Method, respExpected.StatusCode, respExpected.Header)\n\t\t\t\tt.Fatal(\"headers not equal\")\n\t\t\t}\n\t\t\tresp, err = http.Get(srvActual.URL + \"/file.txt\")\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tdefer resp.Body.Close()\n\t\t\tbuf := bytes.NewBuffer(make([]byte, 0, 256))\n\t\t\t_, err = io.Copy(buf, resp.Body)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\trespExpected, err = http.Get(srvExpected.URL + \"/file.txt\")\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tbufExpected := bytes.NewBuffer(make([]byte, 0, 256))\n\t\t\tdefer respExpected.Body.Close()\n\t\t\t_, err = io.Copy(bufExpected, respExpected.Body)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tif !headersEqual(resp.Header, respExpected.Header) {\n\t\t\t\tt.Log(resp.Request.Method, resp.StatusCode, resp.Header)\n\t\t\t\tt.Log(respExpected.Request.Method, respExpected.StatusCode, respExpected.Header)\n\t\t\t\tt.Fatal(\"headers not equal\")\n\t\t\t}\n\t\t\tif !bytes.Equal(buf.Bytes(), bufExpected.Bytes()) {\n\t\t\t\tt.Fatal(\"bodies not equal\")\n\t\t\t}\n\t\t}()\n\t}\n\twg.Wait()\n}", "func collectNonGoFiles(baseDir string, logger *log.Logger) ([]string, error) {\n\tfiles := make([]string, 0)\n\n\terr := filepath.Walk(baseDir, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Ignore directories.\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Ignore all Go files.\n\t\tif strings.HasSuffix(info.Name(), \".go\") {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Ignore perserved files.\n\t\tif isPreservedFile(info.Name()) {\n\t\t\treturn nil\n\t\t}\n\n\t\tfiles = append(files, path)\n\n\t\treturn nil\n\t})\n\n\treturn files, err\n}", "func TestRunGen(t *testing.T) {\n\tls, _ := ioutil.ReadDir(\".\")\n\tprotos := make(map[string]os.FileInfo)\n\tpbgos := make(map[string]os.FileInfo)\n\tfor _, f := range ls {\n\t\tn := f.Name()\n\t\tif strings.HasSuffix(n, \".proto\") {\n\t\t\tprotos[strings.TrimSuffix(n, \".proto\")] = f\n\t\t} else if strings.HasSuffix(n, \".pb.go\") {\n\t\t\tpbgos[strings.TrimSuffix(n, \".pb.go\")] = f\n\t\t}\n\t}\n\tif len(protos) != len(pbgos) {\n\t\tt.Error(\"Please do a clean regeneration of proto code\")\n\t}\n\tfor n, info := range protos {\n\t\tif info.ModTime().After(pbgos[n].ModTime()) {\n\t\t\tt.Error(\"Please do a clean regeneration of proto code\")\n\t\t}\n\t}\n}", "func fakeGcp() (*compute.Service, error) {\n\tclient := &http.Client{}\n\thttpmock.ActivateNonDefault(client)\n\treturn compute.NewService(context.Background(), option.WithoutAuthentication(), option.WithHTTPClient(client))\n}", "func (ds *DNSService) processWildcardMatches() {\n\twildcards := make(map[string]*dnsWildcard)\nloop:\n\tfor {\n\t\tselect {\n\t\tcase req := <-ds.wildcards:\n\t\t\tr := req.Req\n\t\t\treq.Ans <- matchesWildcard(r.Name, r.Domain, r.Address, wildcards)\n\t\tcase <-ds.Quit():\n\t\t\tbreak loop\n\t\t}\n\t}\n}", "func testServiceSignalReceiver(cmd cmdType, t *testing.T) {\n\texpectedCmd := cmd.toServiceSignal()\n\tserviceCmd := <-globalServiceSignalCh\n\tif serviceCmd != expectedCmd {\n\t\tt.Errorf(\"Expected service command %v but received %v\", expectedCmd, serviceCmd)\n\t}\n}", "func (process *Process) matchFiles(fs Filesystem, globs []string, loadContents bool) ([]cwl.FileDir, error) {\n\t// it's important this slice isn't nil, because the outputEval field\n\t// expects it to be non-null during expression evaluation.\n\tfiles := []cwl.FileDir{}\n\n\t// resolve all the globs into file objects.\n\tfor _, pattern := range globs {\n\t\tmatches, err := fs.Glob(pattern)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to execute glob: %s\", err)\n\t\t}\n\n\t\tfor _, m := range matches {\n\t\t\t// TODO handle directories\n\t\t\tif m.ClassName() == \"Directory\" {\n\t\t\t\tv := cwl.Directory{\n\t\t\t\t\tClassBase: cwl.ClassBase{\"Directory\"},\n\t\t\t\t\tLocation: m.Location,\n\t\t\t\t\tPath: m.Path,\n\t\t\t\t}\n\t\t\t\tvar loadListingType = cwl.NO_LISTING\n\t\t\t\tif rll := process.tool.RequiresLoadListing(); rll != nil {\n\t\t\t\t\tloadListingType = rll.LoadListing\n\t\t\t\t}\n\t\t\t\td, err := process.resolveDir(v, loadListingType)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t\tfiles = append(files, cwl.NewFileDir(&d))\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tv := cwl.File{\n\t\t\t\tClassBase: cwl.ClassBase{\"File\"},\n\t\t\t\t//Class: \"File\",\n\t\t\t\tLocation: m.Location,\n\t\t\t\tPath: m.Path,\n\t\t\t\t//File: cwl.File{\n\t\t\t\tChecksum: m.Checksum,\n\t\t\t\tSize: m.Size,\n\t\t\t\t//},\n\t\t\t}\n\t\t\tf, err := process.resolveFile(v, loadContents)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tfiles = append(files, cwl.NewFileDir(&f))\n\t\t}\n\t}\n\treturn files, nil\n}", "func (badGlobChildren4) GlobChildren__(*context.T, rpc.GlobChildrenServerCall, *glob.Element) {}", "func simpleMockMongoFilesInstanceWithFilename(command, fname string) *MongoFiles {\n\treturn &MongoFiles{\n\t\tToolOptions: toolOptions,\n\t\tInputOptions: &InputOptions{},\n\t\tStorageOptions: &StorageOptions{GridFSPrefix: \"fs\", DB: testDB},\n\t\tCommand: command,\n\t\tFileName: fname,\n\t}\n}", "func TestGetFileInof(t *testing.T) {\n\tfmt.Println(\"Get File info test\")\n\tcc := getClientConnection()\n\n\t// when done the connection will be close.\n\tdefer cc.Close()\n\n\t// Create a new client service...\n\tc := filepb.NewFileServiceClient(cc)\n\n\trqst := &filepb.GetFileInfoRequest{\n\t\tPath: \"C:\\\\Temp\\\\Cargo\\\\WebApp\\\\Cargo\\\\Apps\\\\BrisOutil\\\\Upload\\\\515\\\\NGEN3603.JPG\",\n\t\tThumnailHeight: 256,\n\t\tThumnailWidth: 256,\n\t}\n\n\trsp, err := c.GetFileInfo(context.Background(), rqst)\n\tif err != nil {\n\t\tlog.Fatalf(\"error while testing get file info: %v\", err)\n\t}\n\n\tlog.Println(\"Response form Get file info response :\", string(rsp.Data))\n}", "func globGopath(gopath string) (paths []path) {\n\tinter := fsutil.Intersect(filepath.Join(gopath, \"src\"),\n\t\tfilepath.Join(gopath, \"schema\"))\n\tfor i := range inter {\n\t\tpaths = append(paths, path{filepath.Join(gopath, \"schema\", inter[i]),\n\t\t\tfilepath.Join(gopath, \"src\", inter[i])})\n\t}\n\treturn\n}", "func (p *Project) Glob(pattern string) (paths []string, err error) {\n\tprefix := p.BaseDir + string(filepath.Separator)\n\tfullPattern := prefix + pattern\n\tpaths, err = zglob.Glob(fullPattern)\n\tif err != nil {\n\t\treturn\n\t}\n\tprefixLen := len(prefix)\n\tfor n, fullpath := range paths {\n\t\tpaths[n] = fullpath[prefixLen:]\n\t}\n\treturn\n}", "func FetchAllFlightsService() (f *models.AllFlights) {\n\n\tfile, err := os.OpenFile(\"database/flightList.json\", os.O_RDWR|os.O_APPEND, 0666)\n\tcheckError(err)\n\n\tvar sampleflight models.AllFlights\n\tb, err := ioutil.ReadAll(file)\n\tcheckError(err)\n\n\tif len(b) != 0 {\n\t\terr := json.Unmarshal(b, &sampleflight.Flights)\n\t\tcheckError(err)\n\t\treturn &sampleflight\n\t}\n\n\treturn &sampleflight\n\n}", "func Glob(path ...string) []string {\n\tf, _ := filepathx.Glob(filepath.Join(path...))\n\treturn f\n}", "func loadGlobFS(tpl *template.Template, lcfg LoadConfig, pattern template.TrustedSource, fsys embed.FS) (*template.Template, error) {\n\tfilenames, err := fs.Glob(fsys, pattern.String())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(filenames) == 0 {\n\t\treturn nil, fmt.Errorf(\"pattern matches no files: %#q\", pattern.String())\n\t}\n\tvar tts []template.TrustedSource\n\tfor _, fn := range filenames {\n\t\t// The pattern expanded from a trusted source, so the expansion is still trusted.\n\t\ttts = append(tts, uncheckedconversions.TrustedSourceFromStringKnownToSatisfyTypeContract(fn))\n\t}\n\treturn loadFilesFS(tpl, lcfg, fsys, tts...)\n}", "func (s *perfSuite) openGlob(pattern string) (readers []io.Reader) {\n\tassert := s.NewAssert()\n\n\ts.Pause(func() {\n\t\tglob, err := filepath.Glob(pattern)\n\t\tassert.NoError(err)\n\t\treaders = make([]io.Reader, len(glob))\n\t\tfor i, m := range glob {\n\t\t\tr, err := os.Open(m)\n\t\t\tassert.NoError(err)\n\t\t\treaders[i] = r\n\t\t}\n\t})\n\treturn\n}", "func tempFile(base, suffix string, mode os.FileMode) (f *os.File, err error) {\n\tfor i := 0; i < 10000; i++ {\n\t\tname := base + suffix + strconv.Itoa(rand.Intn(1e9))\n\t\tf, err = os.OpenFile(name, os.O_RDWR|os.O_CREATE|os.O_EXCL, mode)\n\t\tif os.IsExist(err) {\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn\n}", "func loadService(c context.Context, path string) (service, error) {\n\turl, err := url.Parse(path)\n\tif err == nil && url.Scheme != \"\" {\n\t\tlog.Fields{\n\t\t\t\"url\": path,\n\t\t}.Infof(c, \"Identified path as service URL.\")\n\t\treturn &remoteDiscoveryService{\n\t\t\turl: *url,\n\t\t}, nil\n\t}\n\tlog.Fields{\n\t\tlog.ErrorKey: err,\n\t\t\"value\": path,\n\t}.Debugf(c, \"Path did not parse as URL. Trying local filesystem options.\")\n\n\tyamlPath := \"\"\n\tst, err := os.Stat(path)\n\tswitch {\n\tcase os.IsNotExist(err):\n\t\tlog.Fields{\n\t\t\t\"path\": path,\n\t\t}.Debugf(c, \"Path does not exist. Maybe it's a Go path?\")\n\n\t\t// Not a filesysem path. Perhaps it's a Go package on GOPATH?\n\t\tpkgPath, err := getPackagePath(path)\n\t\tif err != nil {\n\t\t\tlog.Fields{\n\t\t\t\t\"path\": path,\n\t\t\t}.Debugf(c, \"Could not resolve package path.\")\n\t\t\treturn nil, fmt.Errorf(\"could not resolve path [%s]\", path)\n\t\t}\n\t\tpath = pkgPath\n\n\tcase err != nil:\n\t\treturn nil, fmt.Errorf(\"failed to stat [%s]: %s\", path, err)\n\n\tcase st.IsDir():\n\t\tbreak\n\n\tdefault:\n\t\t// \"path\" is a path to a non-directory. Use its parent directory.\n\t\tyamlPath, err = filepath.Abs(path)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not get absolute path for YAML config [%s]: %s\", path, err)\n\t\t}\n\t\tpath = filepath.Dir(path)\n\t}\n\n\t// \"path\" is a directory. Does its `app.yaml` exist?\n\tif yamlPath == \"\" {\n\t\tyamlPath = filepath.Join(path, \"app.yaml\")\n\t}\n\n\tif _, err = os.Stat(yamlPath); err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to stat YAML config at [%s]: %s\", yamlPath, err)\n\t}\n\n\tconfigData, err := os.ReadFile(yamlPath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to read YAML config at [%s]: %s\", yamlPath, err)\n\t}\n\n\tconfig := appYAML{}\n\tif err := yaml.Unmarshal(configData, &config); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to Unmarshal YAML config from [%s]: %s\", yamlPath, err)\n\t}\n\n\tswitch config.Runtime {\n\tcase \"go\":\n\t\tif config.VM {\n\t\t\treturn &discoveryTranslateService{\n\t\t\t\tdir: path,\n\t\t\t}, nil\n\t\t}\n\t\treturn &devAppserverService{\n\t\t\tprerun: func(c context.Context) error {\n\t\t\t\treturn checkBuild(c, path)\n\t\t\t},\n\t\t\targs: []string{\"goapp\", \"serve\", yamlPath},\n\t\t}, nil\n\n\tcase \"python27\":\n\t\treturn &devAppserverService{\n\t\t\targs: []string{\"dev_appserver.py\", yamlPath},\n\t\t}, nil\n\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"don't know how to load service runtime [%s]\", config.Runtime)\n\t}\n}", "func GetDataFiles(dataDir string, pattern string) []string {\n\tfilenames, err := filepath.Glob(filepath.Join(dataDir, pattern))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn filenames\n}", "func TestGenerateDockerfileGlobs(t *testing.T) {\n\tt.Parallel()\n\tglobs := strings.Join([]string{filepath.Join(generateDockerBaseDir, \"globs\", \"**\", \"Dockerfile\"), filepath.Join(generateDockerBaseDir, \"globs\", \"Dockerfile\")}, \",\")\n\tflags := []string{fmt.Sprintf(\"--dockerfile-globs=%s\", globs)}\n\tdockerfiles := []string{filepath.Join(generateDockerBaseDir, \"globs\", \"globs\", \"Dockerfile\"), filepath.Join(generateDockerBaseDir, \"globs\", \"Dockerfile\")}\n\tvar tOs []generateTestObject\n\tfor _, dockerfile := range dockerfiles {\n\t\ttO := generateTestObject{\n\t\t\tfilePath: filepath.ToSlash(dockerfile),\n\t\t\twantImages: []generate.DockerfileImage{\n\t\t\t\t{Image: generate.Image{Name: \"busybox\", Tag: \"latest\"}},\n\t\t\t},\n\t\t\ttestFn: checkGenerateDockerfile,\n\t\t}\n\t\ttOs = append(tOs, tO)\n\t}\n\ttestGenerate(t, flags, tOs)\n}", "func TestMayaAnyK8sGenerateServiceTemplated(t *testing.T) {\n\ttests := map[string]MockMayaAnyK8s{\n\t\t\"templated service\": {\n\t\t\tkind: \"Service\",\n\t\t\tapiVersion: \"v1\",\n\t\t\towner: \"pv-123-abc\",\n\t\t\tsuffixName: \"-svc\",\n\t\t\tisError: false,\n\t\t\tyaml: `\napiVersion: {{.APIVersion}}\nkind: {{.Kind}}\nmetadata:\n name: {{.Owner}}-svc\nspec:\n ports:\n - name: api\n port: 5656\n protocol: TCP\n targetPort: 5656\n selector:\n name: maya-apiserver\n sessionAffinity: None\n`},\n\t}\n\n\tfor name, mock := range tests {\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\tma := mock.NewMayaAnyK8s()\n\t\t\ts, err := ma.GenerateService()\n\n\t\t\tif mock.isError && s != nil {\n\t\t\t\tt.Fatalf(\"Expected: 'nil service' Actual: '%v'\", s)\n\t\t\t}\n\t\t\tmock.TestObjectMeta(s.ObjectMeta, err, t)\n\t\t\tmock.TestTypeMeta(s.TypeMeta, err, t)\n\t\t})\n\t}\n}", "func GetAllSources(sys System) []Service {\n\tsrvs := []Service{\n\t\tNewAlienVault(sys),\n\t\tNewArchiveIt(sys),\n\t\tNewArchiveToday(sys),\n\t\tNewArquivo(sys),\n\t\tNewAsk(sys),\n\t\tNewBaidu(sys),\n\t\tNewBinaryEdge(sys),\n\t\tNewBing(sys),\n\t\tNewBufferOver(sys),\n\t\tNewCensys(sys),\n\t\tNewCertSpotter(sys),\n\t\tNewCIRCL(sys),\n\t\tNewCommonCrawl(sys),\n\t\tNewCrtsh(sys),\n\t\tNewDNSDB(sys),\n\t\tNewDNSDumpster(sys),\n\t\tNewDNSTable(sys),\n\t\tNewDogpile(sys),\n\t\tNewEntrust(sys),\n\t\tNewExalead(sys),\n\t\tNewGitHub(sys),\n\t\tNewGoogle(sys),\n\t\tNewGoogleCT(sys),\n\t\tNewHackerOne(sys),\n\t\tNewHackerTarget(sys),\n\t\tNewIPToASN(sys),\n\t\tNewIPv4Info(sys),\n\t\tNewLoCArchive(sys),\n\t\tNewMnemonic(sys),\n\t\tNewNetcraft(sys),\n\t\tNewNetworksDB(sys),\n\t\tNewOpenUKArchive(sys),\n\t\tNewPassiveTotal(sys),\n\t\tNewPastebin(sys),\n\t\tNewPTRArchive(sys),\n\t\tNewRADb(sys),\n\t\tNewRiddler(sys),\n\t\tNewRobtex(sys),\n\t\tNewSiteDossier(sys),\n\t\tNewSecurityTrails(sys),\n\t\tNewShadowServer(sys),\n\t\tNewShodan(sys),\n\t\tNewSpyse(sys),\n\t\tNewSublist3rAPI(sys),\n\t\tNewTeamCymru(sys),\n\t\tNewThreatCrowd(sys),\n\t\tNewTwitter(sys),\n\t\tNewUKGovArchive(sys),\n\t\tNewUmbrella(sys),\n\t\tNewURLScan(sys),\n\t\tNewViewDNS(sys),\n\t\tNewVirusTotal(sys),\n\t\tNewWayback(sys),\n\t\tNewWhoisXML(sys),\n\t\tNewYahoo(sys),\n\t}\n\n\t// Filtering in-place - https://github.com/golang/go/wiki/SliceTricks\n\ti := 0\n\tfor _, s := range srvs {\n\t\tif shouldEnable(s.String(), sys.Config()) {\n\t\t\tsrvs[i] = s\n\t\t\ti++\n\t\t}\n\t}\n\tsrvs = srvs[:i]\n\treturn srvs\n}", "func TestExtensionIgnored(t *testing.T) {\n\n\ttestUUID := \"e724bba2-266f-434d-aaf2-935d2b405aee.arbitrary\"\n\ttemplate := \"\"\n\n\t_, err := RMFiler(\"../testfiles/\"+testUUID, template)\n\tif err != nil {\n\t\tt.Errorf(\"Could not open file %v\", err)\n\t}\n}", "func (s *Service) Listen() {\n\n}", "func (g *Group) Glob(patterns ...string) error {\n\tfiles := make([]string, 0, 8)\n\tfor _, pattern := range patterns {\n\t\tpattern = filepath.Join(g.dir, pattern)\n\t\tmatches, err := filepath.Glob(pattern)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor _, m := range matches {\n\t\t\tm, err = filepath.Rel(g.dir, m)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif filepath.Base(m)[0] == '.' {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfiles = append(files, m)\n\t\t}\n\t}\n\treturn g.Files(files...)\n}", "func startService() {\n\n box := packr.NewBox(\"./templates\")\n\n http.Handle(\"/\", http.FileServer(box))\n http.HandleFunc(\"/upload/\", ReceiveFile)\n\n log.Println(\"starting http service...\")\n if err := http.ListenAndServe(\":80\", nil); err != nil {\n log.Fatal(err)\n }\n}", "func FilesFromPatterns(patterns ...string) ([]string, error) {\n\tfiles := make(map[string]bool)\n\tassets, _, err := glob.Glob(patterns)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor _, asset := range assets {\n\t\tif strings.Contains(asset.Path, \"/vendor/\") {\n\t\t\t// skip vendor directory\n\t\t\tcontinue\n\t\t}\n\t\tif strings.HasSuffix(asset.Path, \"_test.go\") {\n\t\t\t// skip unit tests\n\t\t\tcontinue\n\t\t}\n\n\t\tlog.Printf(\"found file: %s\", asset.Path)\n\t\tfiles[asset.Path] = true\n\t}\n\n\tuniqs := make([]string, 0, len(files))\n\tfor file := range files {\n\t\tuniqs = append(uniqs, file)\n\t}\n\treturn uniqs, nil\n}", "func testsFilter(info os.FileInfo) bool {\n\treturn !strings.HasSuffix(info.Name(), \"_test.go\")\n}", "func removeTestFiles(defDir string) {\n\t// svcout dir\n\tos.RemoveAll(filepath.Join(defDir, \"metaverse\"))\n\t// service dir\n\tos.RemoveAll(filepath.Join(defDir, \"test-service\"))\n\t// where the binaries are compiled to\n\tos.RemoveAll(filepath.Join(defDir, \"bin\"))\n\t// Remove all the .pb.go files which may remain\n\tdirs, _ := ioutil.ReadDir(defDir)\n\tfor _, d := range dirs {\n\t\tif strings.HasSuffix(d.Name(), \".pb.go\") {\n\t\t\tos.RemoveAll(filepath.Join(defDir, d.Name()))\n\t\t}\n\t}\n}" ]
[ "0.636986", "0.604599", "0.6015841", "0.59520227", "0.5733951", "0.5687662", "0.5567919", "0.5378951", "0.53750557", "0.5268664", "0.52668816", "0.52246433", "0.5176961", "0.5127348", "0.50312924", "0.50312763", "0.5011737", "0.50043464", "0.5001042", "0.4997822", "0.49969277", "0.49902552", "0.4988971", "0.49844784", "0.4970165", "0.49649915", "0.4961967", "0.4947835", "0.49396136", "0.49339056", "0.49044263", "0.49019593", "0.48880103", "0.48838416", "0.48759514", "0.48687077", "0.48680553", "0.48347914", "0.4833624", "0.48305124", "0.48252016", "0.4813693", "0.48069656", "0.4806648", "0.48055562", "0.48029444", "0.48014084", "0.47925285", "0.478567", "0.47811466", "0.47753245", "0.47619048", "0.47557968", "0.47529823", "0.47515818", "0.4750266", "0.4742369", "0.47389102", "0.47360644", "0.47327724", "0.47307792", "0.47299924", "0.47214392", "0.47192442", "0.4718327", "0.47157702", "0.47025558", "0.4695268", "0.46924216", "0.46872017", "0.46868256", "0.46802342", "0.46781495", "0.46779886", "0.4677127", "0.46757838", "0.46732694", "0.46647146", "0.46632963", "0.46616933", "0.46580133", "0.46579748", "0.4652649", "0.46516165", "0.4651489", "0.4647591", "0.4644067", "0.46434098", "0.4635412", "0.4632629", "0.4631421", "0.46305165", "0.46304423", "0.4627057", "0.4622962", "0.4621977", "0.46202952", "0.46179315", "0.46162412", "0.46150935" ]
0.62766236
1
/ instance of a MCIS
func NewMCIS(ns string, name string) *MCIS { return &MCIS{ Model: Model{Name: name, Namespace: ns}, VMs: []VM{}, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewIS(c Cluster) *IS {\n\treturn &IS{\n\t\tCluster: c,\n\t}\n}", "func NewIskratelMsan() *IskratelMsan {\r\n var t = &IskratelMsan{}\r\n\r\n return t\r\n}", "func NewXmlmcInstance(servername string) *XmlmcInstStruct {\n\tndb := new(XmlmcInstStruct)\n\t//-- TK Add Support for passing in instance name\n\tmatchedURL, err := regexp.MatchString(`(?i)(http|https)(?-i):\\/\\/([\\w\\-_]+(?:(?:\\.[\\w\\-_]+)+))([\\w\\-\\.,@?^=%&amp;:/~\\+#]*[\\w\\-\\@?^=%&amp;/~\\+#])?`, servername)\n\t//-- Catch Error\n\tif err != nil {\n\t\tlog.Fatal(\"Unable to Parse server Name\")\n\t}\n\t//-- If URL then just use it\n\tif matchedURL {\n\t\tndb.server = servername\n\t\tndb.DavEndpoint = strings.Replace(servername, \"/xmlmc/\", \"/dav/\", 1)\n\t} else {\n\t\t//-- Else look it up\n\t\tserverZoneInfo, ziErr := GetZoneInfo(servername)\n\t\tndb.FileError = ziErr\n\t\tif ziErr != nil {\n\t\t\treturn ndb\n\t\t}\n\t\tif serverZoneInfo.Zoneinfo.Endpoint != \"\" {\n\t\t\tndb.server = serverZoneInfo.Zoneinfo.Endpoint + \"xmlmc/\"\n\t\t\tndb.DavEndpoint = serverZoneInfo.Zoneinfo.Endpoint + \"dav/\"\n\t\t}\n\t\tif serverZoneInfo.Zoneinfo.Stream != \"\" {\n\t\t\tndb.stream = serverZoneInfo.Zoneinfo.Stream\n\t\t}\n\t}\n\tndb.transport = &http.Transport{\n\t\tProxy: http.ProxyFromEnvironment,\n\t}\n\n\tndb.userAgent = \"Go-http-client/1.1\"\n\tndb.timeout = 30\n\tndb.jsonresp = false\n\treturn ndb\n}", "func New() MME {\n\t// TODO: Implement this!\n\toperationCosts = make(map[rpcs.Operation]int)\n\toperationCosts[rpcs.SMS] = -1\n\toperationCosts[rpcs.Call] = -5\n\toperationCosts[rpcs.Load] = 10\n\tm := new(mme)\n\tm.state = make(map[uint64]rpcs.MMEState)\n\tm.stateMutex = new(sync.Mutex)\n\treturn m\n}", "func (n NetworkTypeMobile) construct() NetworkTypeClass { return &n }", "func New() MME {\n\tvar m MME = new(mme)\n\treturn m\n}", "func (i InputBackgroundRemote) construct() InputBackgroundClass { return &i }", "func newFirstMima() *Mima {\n\tm := new(Mima)\n\tkey := randomKey()\n\tm.Username = RandomString64()\n\tm.Password = util.Base64Encode(key[:])\n\tm.CreatedAt = util.TimeNow()\n\treturn m\n}", "func init() {\n\tinstance = GetInstance()\n}", "func init() {\n\ti = New()\n}", "func New(logger *zap.Logger, conn *grpc.ClientConn, sensors []*config.Sensor, outChan telemetry.ExtDSChan) telemetry.NMI {\n\tvar metrics = make(map[string]status.Metrics)\n\n\tmetrics[\"gRPCDataTotal\"] = status.NewCounter(\"arista_gnmi_grpc_data_total\", \"\")\n\tmetrics[\"dropsTotal\"] = status.NewCounter(\"arista_gnmi_drops_total\", \"\")\n\tmetrics[\"errorsTotal\"] = status.NewCounter(\"arista_gnmi_errors_total\", \"\")\n\tmetrics[\"processNSecond\"] = status.NewGauge(\"arista_gnmi_process_nanosecond\", \"\")\n\n\tstatus.Register(status.Labels{\"host\": conn.Target()}, metrics)\n\n\treturn &GNMI{\n\t\tlogger: logger,\n\t\tconn: conn,\n\t\tsubscriptions: telemetry.GetGNMISubscriptions(sensors),\n\t\tpathOutput: telemetry.GetPathOutput(sensors),\n\t\tdefaultOutput: telemetry.GetDefaultOutput(sensors),\n\t\tdataChan: make(chan *gpb.SubscribeResponse, 100),\n\t\toutChan: outChan,\n\t\tmetrics: metrics,\n\t}\n}", "func New() cs.CS {\n\t// create root constraint system\n\tcircuit := cs.New()\n\n\t// declare secret and public inputs\n\tpreImage := circuit.SECRET_INPUT(\"pi\")\n\thash := circuit.PUBLIC_INPUT(\"h\")\n\n\t// hash function\n\tmimc := mimc.NewMiMC(\"seed\")\n\n\t// specify constraints\n\t// mimc(preImage) == hash\n\tcircuit.MUSTBE_EQ(hash, mimc.Hash(&circuit, preImage))\n\n\treturn circuit\n}", "func (i InputCheckPasswordSRP) construct() InputCheckPasswordSRPClass { return &i }", "func New(store storage.Store) *Scim {\n\treturn &Scim{\n\t\tstore: store,\n\t}\n}", "func New(s *service.Canal) (engine *bm.Engine, err error) {\n\tcs = s\n\tengine = bm.DefaultServer(conf.Conf.BM)\n\n\tinitRouter(engine)\n\terr = engine.Start()\n\treturn\n}", "func (n NetworkTypeMobileRoaming) construct() NetworkTypeClass { return &n }", "func Init() (err error) {\n\n\t// Retrieve Sandbox name from environment variable\n\tsandboxNameEnv := strings.TrimSpace(os.Getenv(\"MEEP_SANDBOX_NAME\"))\n\tif sandboxNameEnv != \"\" {\n\t\tsandboxName = sandboxNameEnv\n\t}\n\tif sandboxName == \"\" {\n\t\terr = errors.New(\"MEEP_SANDBOX_NAME env variable not set\")\n\t\tlog.Error(err.Error())\n\t\treturn err\n\t}\n\tlog.Info(\"MEEP_SANDBOX_NAME: \", sandboxName)\n\n\t// Get MEP name\n\tmepNameEnv := strings.TrimSpace(os.Getenv(\"MEEP_MEP_NAME\"))\n\tif mepNameEnv != \"\" {\n\t\tmepName = mepNameEnv\n\t}\n\tlog.Info(\"MEEP_MEP_NAME: \", mepName)\n\n\t// hostUrl is the url of the node serving the resourceURL\n\t// Retrieve public url address where service is reachable, if not present, use Host URL environment variable\n\thostUrl, err = url.Parse(strings.TrimSpace(os.Getenv(\"MEEP_PUBLIC_URL\")))\n\tif err != nil || hostUrl == nil || hostUrl.String() == \"\" {\n\t\thostUrl, err = url.Parse(strings.TrimSpace(os.Getenv(\"MEEP_HOST_URL\")))\n\t\tif err != nil {\n\t\t\thostUrl = new(url.URL)\n\t\t}\n\t}\n\tlog.Info(\"MEEP_HOST_URL: \", hostUrl)\n\n\t// Create new active scenario model\n\tmodelCfg := mod.ModelCfg{\n\t\tName: \"activeScenario\",\n\t\tNamespace: sandboxName,\n\t\tModule: moduleName,\n\t\tUpdateCb: nil,\n\t\tDbAddr: redisAddr,\n\t}\n\tactiveModel, err = mod.NewModel(modelCfg)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create model: \", err.Error())\n\t\treturn err\n\t}\n\tlog.Info(\"Active Scenario Model created\")\n\n\t// Create message queue\n\tmqLocal, err = mq.NewMsgQueue(mq.GetLocalName(sandboxName), moduleName, sandboxName, redisAddr)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create Message Queue with error: \", err)\n\t\treturn err\n\t}\n\tlog.Info(\"Message Queue created\")\n\n\t// Create Swagger API Manager\n\tmep := \"\"\n\tif mepName != defaultMepName {\n\t\tmep = mepName\n\t}\n\tapiMgr, err = sam.NewSwaggerApiMgr(moduleName, sandboxName, mep, mqLocal)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create Swagger API Manager. Error: \", err)\n\t\treturn err\n\t}\n\tlog.Info(\"Swagger API Manager created\")\n\n\t// Initialize Service Management\n\terr = sm.Init(sandboxName, mepName, hostUrl, mqLocal, &mutex)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Initialize App Support\n\terr = as.Init(sandboxName, mepName, hostUrl, mqLocal, &mutex)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func NewMNSTopic(name string, client MNSClient, /*qps ...int32*/) AliMNSTopic {\n\tif name == \"\" {\n\t\tpanic(\"ali_mns: topic name could not be empty\")\n\t}\n\n\ttopic := new(MNSTopic)\n\ttopic.client = client\n\ttopic.name = name\n\ttopic.stopChan = make(chan bool)\n\t//topic.qpsLimit = DefaultTopicQPSLimit\n\ttopic.decoder = NewAliMNSDecoder()\n\n\t//if qps != nil && len(qps) == 1 && qps[0] > 0 {\n\t//\ttopic.qpsLimit = qps[0]\n\t//}\n\n\tproxyURL := \"\"\n\ttopicProxyEnvKey := PROXY_PREFIX + strings.Replace(strings.ToUpper(name), \"-\", \"_\", -1)\n\tif url := os.Getenv(topicProxyEnvKey); url != \"\" {\n\t\tproxyURL = url\n\t}\n\n\tclient.SetProxy(proxyURL)\n\n\t//topic.qpsMonitor = NewQPSMonitor(5)\n\n\treturn topic\n}", "func New(c string, i io.Reader, o io.Writer) (M, error) {\n\tins, err := compiler.New(c).Compile()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tm := &machine{\n\t\tinput: i,\n\t\toutput: o,\n\t\tinstructions: ins,\n\t\tbuf: make([]byte, 1),\n\t\tinstructionLength: uint8(len(ins)),\n\t}\n\treturn m, nil\n}", "func (s *Saiyan) constructor(name string, power int) *Saiyan {\n\treturn &Saiyan{\n\t\tName: name,\n\t\tPower: power,\n\t}\n}", "func newCPU() *Mos6502 {\n\treturn &Mos6502{\n\t\tsp: 0xFF,\n\t\tpc: 0,\n\t\ta: 0,\n\t\tx: 0,\n\t\ty: 0,\n\t\tps: flagDefault,\n\t}\n}", "func (m *InstanceMetadata) init() *ec2metadata.EC2Metadata {\n\t// Locate this instance\n\ttimeoutDuration, _ := time.ParseDuration(\"30s\")\n\ttimeout := time.After(timeoutDuration)\n\ti := 0\n\tfor {\n\t\ti++\n\t\tselect {\n\t\tdefault:\n\t\t\tlog.Printf(\"[INFO] Connecting metadata service (%d)\", i)\n\t\t\tsess, err := session.NewSession()\n\t\t\tif err != nil {\n\t\t\t\tfmt.Println(\"failed to create session,\", err)\n\t\t\t\tpanic(\"failed to create session\")\n\t\t\t}\n\n\t\t\tm.client = ec2metadata.New(sess)\n\t\t\tm.document, _ = m.client.GetInstanceIdentityDocument()\n\t\t\tif m.client.Available() {\n\t\t\t\tlog.Printf(\"[INFO] AWS EC2 instance detected via default metadata API endpoint\")\n\t\t\t\treturn m.client\n\t\t\t}\n\t\tcase <-timeout:\n\t\t\tpanic(\"AWS metadata service connection failed\")\n\t\t}\n\t}\n}", "func NewEmbeddedSIMActivationCode()(*EmbeddedSIMActivationCode) {\n m := &EmbeddedSIMActivationCode{\n }\n m.backingStore = ie8677ce2c7e1b4c22e9c3827ecd078d41185424dd9eeb92b7d971ed2d49a392e.BackingStoreFactoryInstance();\n m.SetAdditionalData(make(map[string]any))\n return m\n}", "func newISISLSP() *isisLSP {\n\treturn &isisLSP{\n\t\tLSP: &oc.Lsp{\n\t\t\tTlv: map[oc.E_OpenconfigIsisLsdbTypes_ISIS_TLV_TYPE]*oc.Lsp_Tlv{},\n\t\t},\n\t\trawTLVs: []*rawTLV{},\n\t}\n}", "func (mon *monDat) init() {\n\n\tmcf := mon.cf\n\n\tif mcf.Port == 0 {\n\t\tdl.Fatal(\"monitor port not specified\")\n\t}\n\tif mcf.Protocol == \"\" {\n\t\tmcf.Protocol = \"yb\"\n\t}\n\tif mcf.System == \"\" {\n\t\tmcf.System = DEFSYS\n\t}\n\tif mcf.Addr == \"\" {\n\t\tmcf.Addr = LOCALHOST\n\t}\n\tif mcf.Environment == \"\" {\n\t\tmcf.Environment = mon.pdb.Env()\n\t}\n\n\tinfo := myinfo.GetInfo(mcf.Hostname)\n\n\tif mcf.Hostname == \"\" {\n\t\tmcf.Hostname = info.Hostname\n\t}\n\tif mcf.Datacenter == \"\" {\n\t\tmcf.Datacenter = info.Datacenter\n\t}\n\tif mcf.Rack == \"\" {\n\t\tmcf.Rack = info.Rack\n\t}\n\tif mcf.Id == \"\" {\n\t\tmcf.Id = info.ServerId(mcf.System, mcf.Environment, mcf.Port)\n\t}\n\n\tmon.id = mcf.Id\n\tmon.endpoint = net.JoinHostPort(mcf.Addr, fmt.Sprintf(\"%d\", mcf.Port))\n\n\tif mcf.Addr == LOCALHOST {\n\t\tninfo := myinfo.Network(mcf.Datacenter, mcf.Port)\n\n\t\tfor _, ni := range ninfo {\n\t\t\tdl.Debug(\"intf %s [%s]\", ni.Addr, ni.Dom)\n\n\t\t\ta := ni.Addr\n\t\t\td := ni.Dom\n\n\t\t\tmon.netinfo = append(mon.netinfo, &kibitz.NetInfo{\n\t\t\t\tAddr: a,\n\t\t\t\tNatdom: d,\n\t\t\t})\n\t\t}\n\t} else {\n\t\ta := mon.endpoint\n\t\td := \"\"\n\t\tmon.netinfo = []*kibitz.NetInfo{{Addr: a, Natdom: d}}\n\t}\n\n\tswitch mcf.Protocol {\n\tcase \"yb\":\n\tcase \"ac\":\n\tcase \"http\":\n\t\tif mcf.URL == \"\" {\n\t\t\tdl.Fatal(\"monitor url required for protocol http\")\n\t\t}\n\tdefault:\n\t\tdl.Fatal(\"unknown monitor protocol '%s'\", mcf.Protocol)\n\t}\n}", "func New(c *conf.Config, d *dao.Dao) *Mis {\n\tm := &Mis{\n\t\tc: c,\n\t\tdao: d,\n\t}\n\treturn m\n}", "func New() *SMS {\n\t// Initialize config and client\n\tif client == nil {\n\t\tcfg = config.Get().SMS\n\t\tclient = infobip.ClientWithBasicAuth(cfg.Username, cfg.Password)\n\t}\n\n\treturn &SMS{\n\t\tFrom: cfg.FromName,\n\t}\n}", "func New(c *conf.Config) (s *Service) {\n\ts = &Service{\n\t\tconf: c,\n\t\tdao: dao.New(c),\n\t\tarcRPC: arcCli.New2(c.ArchiveRPC),\n\t\tassRPC: assCli.New(c.AssistRPC),\n\t\tcoinRPC: coinCli.New(c.CoinRPC),\n\t\trelRPC: relCli.New(c.RelationRPC),\n\t\tlocationRPC: locCli.New(c.LocationRPC),\n\t\tcache: fanout.New(\"cache\", fanout.Worker(1), fanout.Buffer(1024)),\n\t\trealname: make(map[int64]int64),\n\t\tarcTypes: make(map[int16]int16),\n\t\tbroadcastChan: make(chan *broadcast, 1024),\n\t\tlocalCache: make(map[string][]byte),\n\t\tseqDmArg: &seqMdl.ArgBusiness{BusinessID: c.Seq.DM.BusinessID, Token: c.Seq.DM.Token},\n\t\tseqSubtitleArg: &seqMdl.ArgBusiness{BusinessID: c.Seq.Subtitle.BusinessID, Token: c.Seq.Subtitle.Token},\n\t\tseqRPC: seqCli.New2(c.SeqRPC),\n\t\tassistLogChan: make(chan *assMdl.ArgAssistLogAdd, 1024),\n\t\tdmOperationLogSvc: infoc.New(c.Infoc2),\n\t\topsLogCh: make(chan *oplog.Infoc, 1024),\n\t\tmoniOidMap: make(map[int64]struct{}),\n\t\tmemberRPC: memberCli.New(c.MemberRPC),\n\t\tfigureRPC: figureCli.New(c.FigureRPC),\n\t\tspyRPC: spyCli.New(c.SpyRPC),\n\t\tgarbageDanmu: c.Switch.GarbageDanmu,\n\t\tbroadcastLimit: c.BroadcastLimit.Limit,\n\t\tbroadcastlimitInterval: c.BroadcastLimit.Interval,\n\t\tlocalViewCache: make(map[string]*model.ViewDm),\n\t\taidSheild: make(map[int64]struct{}),\n\t\tmidsSheild: make(map[int64]struct{}),\n\t}\n\tfor idStr, cid := range s.conf.Realname.Threshold {\n\t\tids, err := xstr.SplitInts(idStr)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tfor _, id := range ids {\n\t\t\tif _, ok := s.realname[id]; !ok {\n\t\t\t\ts.realname[id] = cid\n\t\t\t}\n\t\t}\n\t}\n\taccountRPC, err := accountCli.NewClient(c.AccountRPC)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ts.accountRPC = accountRPC\n\tugcPayRPC, err := ugcPayCli.NewClient(c.UgcPayRPC)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ts.ugcPayRPC = ugcPayRPC\n\tfilterRPC, err := filterCli.NewClient(s.conf.FilterRPC)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ts.filterRPC = filterRPC\n\tseasonRPC, err := seasonCli.NewClient(c.SeasonRPC)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"seasonCli.NewClient.error(%v)\", err))\n\t}\n\ts.seasonRPC = seasonRPC\n\tthumbupRPC, err := thumbupApi.NewClient(c.ThumbupRPC)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ts.thumbupRPC = thumbupRPC\n\tsubtitleLans, err := s.dao.SubtitleLans(context.Background())\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ts.subtitleLans = model.SubtitleLans(subtitleLans)\n\tgo s.broadcastproc()\n\tgo s.archiveTypeproc()\n\tgo s.localcacheproc()\n\tgo s.assistLogproc()\n\tgo s.oplogproc()\n\tgo s.monitorproc()\n\tgo s.viewProc()\n\tgo s.shieldProc()\n\treturn\n}", "func New(username string, password string, host string) (ipmi *Ipmi, err error) {\n\tipmi = &Ipmi{\n\t\tUsername: username,\n\t\tPassword: password,\n\t\tHost: host,\n\t}\n\n\tipmi.ipmitool, err = exec.LookPath(\"ipmitool\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ipmi, err\n}", "func MIScrape(phrase string, logger *log.Logger) error {\r\n\tres, err := scrapers.MarketIntelligence(phrase, logger)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\terr = dal.MarketIntelSave(res, phrase)\r\n\treturn err\r\n}", "func init(){\n\t\n\tMemClient.mc = memcache.New(server)\n\n\tif MemClient.mc == nil {\n\t\tlog.Panic(\"Error: unable to initialize memcache \\n\")\n\t}\n\n\tMemClient.mc.Timeout = time.Duration(memcacheTimeOut) * time.Duration(time.Millisecond)\n}", "func CreateInterface(config *MemifConfig, callbacks *MemifCallbacks) (memif *Memif, err error) {\n\tcontext.lock.Lock()\n\tdefer context.lock.Unlock()\n\n\tif !context.initialized {\n\t\treturn nil, ErrNotInit\n\t}\n\n\tlog.WithField(\"ifName\", config.IfName).Debug(\"Creating a new memif interface\")\n\n\tlog2RingSize := config.Log2RingSize\n\tif log2RingSize == 0 {\n\t\tlog2RingSize = 10\n\t}\n\n\tbufferSize := config.BufferSize\n\tif bufferSize <= 0 {\n\t\tbufferSize = 2048\n\t}\n\n\t// Create memif-wrapper for Go-libmemif.\n\tmemif = &Memif{\n\t\tMemifMeta: config.MemifMeta,\n\t\tcallbacks: &MemifCallbacks{},\n\t\tifIndex: context.nextMemifIndex,\n\t\tringSize: 1 << log2RingSize,\n\t\tbufferSize: int(bufferSize),\n\t}\n\n\t// Initialize memif callbacks.\n\tif callbacks != nil {\n\t\tmemif.callbacks.OnConnect = callbacks.OnConnect\n\t\tmemif.callbacks.OnDisconnect = callbacks.OnDisconnect\n\t}\n\n\t// Initialize memif-global interrupt channel.\n\tmemif.intCh = make(chan uint8, 1<<6)\n\tmemif.intErrCh = make(chan error, 1<<6)\n\n\t// Initialize event file descriptor for stopping Rx/Tx queue polling.\n\tmemif.stopQPollFd = int(C.eventfd(0, C.EFD_NONBLOCK))\n\tif memif.stopQPollFd < 0 {\n\t\treturn nil, ErrSyscall\n\t}\n\n\t// Initialize memif input arguments.\n\targs := &C.govpp_memif_conn_args_t{}\n\tsockargs := &C.memif_socket_args_t{}\n\n\tC.strncpy((*C.char)(unsafe.Pointer(&sockargs.path)), C.CString(config.MemifMeta.SocketFilename), C.strlen(C.CString(config.MemifMeta.SocketFilename)))\n\tC.strncpy((*C.char)(unsafe.Pointer(&sockargs.app_name)), C.CString(config.MemifMeta.AppName), C.strlen(C.CString(config.MemifMeta.AppName)))\n\n\tif !config.IsMaster {\n\t\tsockargs.connection_request_timer.it_value.tv_sec = 2\n\t\tsockargs.connection_request_timer.it_value.tv_nsec = 0\n\t\tsockargs.connection_request_timer.it_interval.tv_sec = 2\n\t\tsockargs.connection_request_timer.it_interval.tv_nsec = 0\n\t}\n\t/*\n\t\tsockargs.connection_request_timer.it_value.tv_sec = 2\n\t\tsockargs.connection_request_timer.it_value.tv_nsec = 0\n\t\tsockargs.connection_request_timer.it_interval.tv_sec = 2\n\t\tsockargs.connection_request_timer.it_interval.tv_nsec = 0\n\t*/\n\t// - socket file name\n\tif config.SocketFilename != \"\" {\n\t\tlog.WithField(\"name\", config.SocketFilename).Debug(\"A new memif socket was created\")\n\t\t//errCode := C.govpp_memif_create_socket(&memif.sHandle, sockargs)\n\t\terrCode := C.govpp_memif_create_socket(&socketHandler, sockargs)\n\t\tif getMemifError(int(errCode)) != nil {\n\t\t\terr = getMemifError(int(errCode))\n\t\t\treturn nil, err\n\t\t}\n\t}\n\targs.socket = socketHandler\n\t// - interface ID\n\targs.interface_id = C.uint32_t(config.ConnID)\n\t// - interface name\n\tif config.IfName != \"\" {\n\t\targs.interface_name = C.CString(config.IfName)\n\t\tdefer C.free(unsafe.Pointer(args.interface_name))\n\t}\n\t// - mode\n\tswitch config.Mode {\n\tcase IfModeEthernet:\n\t\targs.mode = C.MEMIF_INTERFACE_MODE_ETHERNET\n\tcase IfModeIP:\n\t\targs.mode = C.MEMIF_INTERFACE_MODE_IP\n\tcase IfModePuntInject:\n\t\targs.mode = C.MEMIF_INTERFACE_MODE_PUNT_INJECT\n\tdefault:\n\t\targs.mode = C.MEMIF_INTERFACE_MODE_ETHERNET\n\t}\n\t// - secret\n\tif config.Secret != \"\" {\n\t\targs.secret = C.CString(config.Secret)\n\t\tdefer C.free(unsafe.Pointer(args.secret))\n\t}\n\t// - master/slave flag + number of Rx/Tx queues\n\tif config.IsMaster {\n\t\targs.num_s2m_rings = C.uint8_t(config.NumRxQueues)\n\t\targs.num_m2s_rings = C.uint8_t(config.NumTxQueues)\n\t\targs.is_master = C.uint8_t(1)\n\t} else {\n\t\targs.num_s2m_rings = C.uint8_t(config.NumTxQueues)\n\t\targs.num_m2s_rings = C.uint8_t(config.NumRxQueues)\n\t\targs.is_master = C.uint8_t(0)\n\t}\n\t// - buffer size\n\targs.buffer_size = C.uint16_t(config.BufferSize)\n\t// - log_2(ring size)\n\targs.log2_ring_size = C.uint8_t(config.Log2RingSize)\n\n\t// Create memif in C-libmemif.\n\terrCode := C.govpp_memif_create(&memif.cHandle, args, unsafe.Pointer(uintptr(memif.ifIndex)))\n\tif getMemifError(int(errCode)) != nil {\n\t\terr = getMemifError(int(errCode))\n\t\treturn nil, err\n\t}\n\n\t// Register the new memif.\n\tcontext.memifs[memif.ifIndex] = memif\n\tcontext.nextMemifIndex++\n\tlog.WithField(\"ifName\", config.IfName).Debug(\"A new memif interface was created\")\n\n\treturn memif, nil\n}", "func (px *Paxos) getInstance(seq int) Instance {\n px.mu.Lock()\n defer px.mu.Unlock()\n\n if _, ok := px.Instances[seq]; !ok {\n px.Instances[seq] = Instance { NullProposal(), \n NullProposal(), \n nil, \n Pending }\n if seq > px.maxSeq {\n px.maxSeq = seq\n }\n }\n\n px.freeMemory()\n\n return px.Instances[seq]\n}", "func (c Client) Scim() scim.Client {\n\treturn scim.NewClient(c...)\n}", "func Instance() Service {\n\tif c == nil {\n\t\tc = &configService{\n\t\t\tdataMap: make(map[string]interface{}),\n\t\t}\n\t}\n\treturn c\n}", "func (i InputBackgroundLocal) construct() InputBackgroundClass { return &i }", "func (self *PhysicsP2) CreateMaterialI(args ...interface{}) *PhysicsP2Material{\n return &PhysicsP2Material{self.Object.Call(\"createMaterial\", args)}\n}", "func NewMDC(s *Store) error {\n\tvar mdc MetadataClient\n\tswitch testtype {\n\tcase \"unit\":\n\t\tmdc = &MockedMetadataClient{data: make(map[string][]byte)}\n\tcase \"integration\":\n\t\tmdc = NewMetadataClient(defaults.DefaultConfig().Metadata)\n\t}\n\treturn s.initMetadataClient(mdc)\n}", "func Instance() Desktop {\n\treturn instance\n}", "func newOpenGNMIStr(channel string) *Instruction {\n\treturn &Instruction{\n\t\tType: OpenGNMIStrInst,\n\t\tName: \"OpenGNMIStream\",\n\t\tChannel: channel,\n\t}\n}", "func (n NetworkTypeWiFi) construct() NetworkTypeClass { return &n }", "func (p *ProxMox) CreateInstance(ctx *lepton.Context) error {\n\n\tvar err error\n\n\tconfig := ctx.Config()\n\n\tnextid := p.getNextID()\n\n\tp.instanceName = config.RunConfig.InstanceName\n\n\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\n\tp.imageName = config.CloudConfig.ImageName\n\n\tp.arch = \"x86_64\"\n\tif config.TargetConfig[\"Arch\"] != \"\" {\n\t\tp.arch = config.TargetConfig[\"Arch\"]\n\t}\n\n\tp.machine = \"q35\"\n\tif config.TargetConfig[\"Machine\"] != \"\" {\n\t\tp.machine = config.TargetConfig[\"Machine\"]\n\t}\n\n\tp.sockets = \"1\"\n\tif config.TargetConfig[\"Sockets\"] != \"\" {\n\t\tsocketsInt, err := strconv.Atoi(config.TargetConfig[\"Sockets\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif socketsInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Sockets can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.sockets = config.TargetConfig[\"Sockets\"]\n\t}\n\n\tp.cores = \"1\"\n\tif config.TargetConfig[\"Cores\"] != \"\" {\n\t\tcoresInt, err := strconv.Atoi(config.TargetConfig[\"Cores\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif coresInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Cores can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.cores = config.TargetConfig[\"Cores\"]\n\t}\n\n\tp.numa = \"0\"\n\tif config.TargetConfig[\"Numa\"] != \"\" {\n\t\tif config.TargetConfig[\"Numa\"] != \"0\" && config.TargetConfig[\"Numa\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Numa can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.numa = config.TargetConfig[\"Numa\"]\n\t}\n\n\t// Memory\n\n\tp.memory = \"512\"\n\tif config.TargetConfig[\"Memory\"] != \"\" {\n\t\tmemoryInt, err := lepton.RAMInBytes(config.TargetConfig[\"Memory\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmemoryInt = memoryInt / 1024 / 1024\n\t\tp.memory = strconv.FormatInt(memoryInt, 10)\n\t}\n\n\t// Main storage\n\n\tp.storageName = \"local-lvm\"\n\tif config.TargetConfig[\"StorageName\"] != \"\" {\n\t\tp.storageName = config.TargetConfig[\"StorageName\"]\n\t}\n\n\t// Iso storage\n\n\tp.isoStorageName = \"local\"\n\tif config.TargetConfig[\"IsoStorageName\"] != \"\" {\n\t\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\t}\n\n\t// Bridge prefix\n\n\tp.bridgePrefix = \"vmbr\"\n\tif config.TargetConfig[\"BridgePrefix\"] != \"\" {\n\t\tp.bridgePrefix = config.TargetConfig[\"BridgePrefix\"]\n\t}\n\n\t// Onboot\n\n\tp.onboot = \"0\"\n\tif config.TargetConfig[\"Onboot\"] != \"\" {\n\t\tif config.TargetConfig[\"Onboot\"] != \"0\" && config.TargetConfig[\"Onboot\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Onboot can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.onboot = config.TargetConfig[\"Onboot\"]\n\t}\n\n\t// Protection\n\n\tp.protection = \"0\"\n\tif config.TargetConfig[\"Protection\"] != \"\" {\n\t\tif config.TargetConfig[\"Protection\"] != \"0\" && config.TargetConfig[\"Protection\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Protection can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.protection = config.TargetConfig[\"Protection\"]\n\t}\n\n\t// These two preventive checks here, because Proxmox will not return\n\t// an error if the storage is missing and a misconfigured instance will be created.\n\n\terr = p.CheckStorage(p.storageName, \"images\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.CheckStorage(p.isoStorageName, \"iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := url.Values{}\n\tdata.Set(\"vmid\", nextid)\n\tdata.Set(\"name\", p.instanceName)\n\tdata.Set(\"name\", p.imageName)\n\tdata.Set(\"machine\", p.machine)\n\tdata.Set(\"sockets\", p.sockets)\n\tdata.Set(\"cores\", p.cores)\n\tdata.Set(\"numa\", p.numa)\n\tdata.Set(\"memory\", p.memory)\n\tdata.Set(\"onboot\", p.onboot)\n\tdata.Set(\"protection\", p.protection)\n\tdata.Set(\"serial0\", \"socket\")\n\n\t// Configuring network interfaces\n\n\tnics := config.RunConfig.Nics\n\tfor i := 0; i < len(nics); i++ {\n\t\tis := strconv.Itoa(i)\n\t\tbrName := nics[i].BridgeName\n\t\tif brName == \"\" {\n\t\t\tbrName = p.bridgePrefix + is\n\t\t}\n\n\t\terr = p.CheckBridge(brName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif nics[i].IPAddress != \"\" {\n\t\t\tcidr := \"24\"\n\n\t\t\tif nics[i].NetMask != \"\" {\n\t\t\t\tcidrInt := lepton.CCidr(nics[i].NetMask)\n\t\t\t\tcidr = strconv.FormatInt(int64(cidrInt), 10)\n\t\t\t}\n\n\t\t\tif nics[i].Gateway != \"\" {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr+\",\"+\"gw=\"+nics[i].Gateway)\n\t\t\t} else {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr)\n\t\t\t}\n\t\t} else {\n\t\t\tdata.Set(\"ipconfig\"+is, \"dhcp\")\n\t\t}\n\n\t\tdata.Set(\"net\"+is, \"model=virtio,bridge=\"+brName)\n\t}\n\tif len(nics) == 0 {\n\t\t// single dhcp nic\n\t\tdata.Set(\"net0\", \"model=virtio,bridge=vmbr0\")\n\t}\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", bytes.NewBufferString(data.Encode()))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tdebug := false\n\tif debug {\n\t\tfmt.Println(string(body))\n\t}\n\n\terr = p.CheckResultType(body, \"createinstance\", \"file=\"+p.isoStorageName+\":iso/\"+p.imageName+\".iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.addVirtioDisk(ctx, nextid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.movDisk(ctx, nextid)\n\n\treturn err\n}", "func newMachine(s SSHInfo) *Machine {\n\tm := Machine{\n\t\tConnectionErrors: make([]string, 0),\n\t\tStreamData: make([]Stream, 0),\n\t\tSSHInfo: s,\n\t}\n\tif m.Extras == nil {\n\t\tm.Extras = make(map[string]interface{}, 0)\n\t}\n\treturn &m\n}", "func Constructor() MedianFinder {\n return MedianFinder{}\n}", "func multiRegToInstance(request *register.MultiRequest) *registry.Instance {\n\tinst := &registry.Instance{\n\t\tId: request.GetInstanceId(),\n\t\tHostname: request.GetHostname(),\n\t\tMachineClass: request.GetMachineClass(),\n\t\tName: request.GetService().GetName(),\n\t\tDescription: request.GetService().GetDescription(),\n\t\tVersion: request.GetService().GetVersion(),\n\t\tAzName: request.GetAzName(),\n\t\tSource: request.GetService().GetSource(),\n\t\tOwnerEmail: request.GetService().GetOwnerEmail(),\n\t\tOwnerMobile: request.GetService().GetOwnerMobile(),\n\t\tOwnerTeam: request.GetService().GetOwnerTeam(),\n\t\tEndpoints: make([]*registry.Endpoint, 0),\n\t}\n\tfor _, endpoint := range request.GetEndpoints() {\n\t\tinst.Endpoints = append(inst.Endpoints, &registry.Endpoint{\n\t\t\tName: endpoint.GetName(),\n\t\t\tSubscribe: endpoint.GetSubscribe(),\n\t\t\tSla: registry.Sla{\n\t\t\t\tMean: uint32(endpoint.GetMean()),\n\t\t\t\tUpper95: uint32(endpoint.GetUpper95()),\n\t\t\t},\n\t\t})\n\t}\n\treturn inst\n}", "func New() *Ica {\n\tica := &Ica{}\n\n\treturn ica\n}", "func (i InlineQueryPeerTypePM) construct() InlineQueryPeerTypeClass { return &i }", "func (s AuthSentCodeTypeSms) construct() AuthSentCodeTypeClass { return &s }", "func NewIPAM() *IPAM {\n\treturn &IPAM{}\n}", "func (m Hash) New(seed string) hash.Hash {\n\tswitch m {\n\tcase MIMC_BN254:\n\t\treturn bn254.NewMiMC(seed)\n\tcase MIMC_BLS12_381:\n\t\treturn bls381.NewMiMC(seed)\n\tcase MIMC_BLS12_377:\n\t\treturn bls377.NewMiMC(seed)\n\tcase MIMC_BW6_761:\n\t\treturn bw761.NewMiMC(seed)\n\tcase MIMC_BLS24_315:\n\t\treturn bls315.NewMiMC(seed)\n\tcase MIMC_BW6_633:\n\t\treturn bw633.NewMiMC(seed)\n\tdefault:\n\t\tpanic(\"Unknown mimc ID\")\n\t}\n}", "func (i *Info) Instance(r *http.Request) (*sessions.Session, error) {\n\treturn i.store.Get(r, i.Name)\n}", "func (p *Plugin) Instance(roomId string, broker Broker) *Instance {\n\ti := Instance{\n\t\tPlugin: p,\n\t\tRoomId: roomId,\n\t\tBroker: broker,\n\t\tRegex: p.Regex,\n\t}\n\n\treturn &i\n}", "func init() {\n\tmodules.Register(\"k6/x/mllp\", new(MLLP))\n}", "func (config Service) Open(log gopi.Logger) (gopi.Driver, error) {\n\tlog.Debug(\"<grpc.service.mihome>Open{ server=%v mihome=%v }\", config.Server, config.MiHome)\n\n\tthis := new(service)\n\tthis.log = log\n\tthis.mihome = config.MiHome\n\tthis.pubsub = nil\n\n\t// Register service with server\n\tconfig.Server.Register(this)\n\n\t// Reset the radio\n\tif err := this.mihome.ResetRadio(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Start goroutine for capturing events from mihome\n\tthis.startCapture()\n\n\t// Success\n\treturn this, nil\n}", "func NewClient(c diam.Conn) {\n // Build CCR\n\n parser, _ := diamdict.NewParser()\n parser.Load(bytes.NewReader(diamdict.DefaultXML))\n parser.Load(bytes.NewReader(diamdict.CreditControlXML))\n\n m := diam.NewRequest(257, 0, parser)\n // Add AVPs\n m.NewAVP(\"Origin-Host\", 0x40, 0x00, Identity)\n m.NewAVP(\"Origin-Realm\", 0x40, 0x00, Realm)\n m.NewAVP(\"Origin-State-Id\", 0x40, 0x00, diamtype.Unsigned32(rand.Uint32()))\n m.NewAVP(\"Auth-Application-Id\", 0x40, 0x00, AuthApplicationId)\n laddr := c.LocalAddr()\n ip, _, _ := net.SplitHostPort(laddr.String())\n m.NewAVP(\"Host-IP-Address\", 0x40, 0x0, diamtype.Address(net.ParseIP(ip)))\n m.NewAVP(\"Vendor-Id\", 0x40, 0x0, VendorId)\n m.NewAVP(\"Product-Name\", 0x00, 0x0, ProductName)\n\n log.Printf(\"Sending message to %s\", c.RemoteAddr().String())\n log.Println(m.String())\n // Send message to the connection\n if _, err := m.WriteTo(c); err != nil {\n log.Fatal(\"Write failed:\", err)\n }\n\n m = diam.NewRequest(272, 4, parser)\n // Add AVPs\n m.NewAVP(\"Session-Id\", 0x40, 0x00, diamtype.UTF8String(fmt.Sprintf(\"%v\", rand.Uint32())))\n m.NewAVP(\"Origin-Host\", 0x40, 0x00, Identity)\n m.NewAVP(\"Origin-Realm\", 0x40, 0x00, Realm)\n m.NewAVP(\"Destination-Realm\", 0x40, 0x00, DestinationRealm)\n m.NewAVP(\"Auth-Application-Id\", 0x40, 0x0, AuthApplicationId)\n m.NewAVP(\"CC-Request-Type\", 0x40, 0x0, CCRequestType)\n m.NewAVP(\"Service-Context-Id\", 0x40, 0x0, ServiceContextId)\n m.NewAVP(\"Service-Identifier\", 0x40, 0x0, ServiceIdentifier)\n m.NewAVP(\"CC-Request-Number\", 0x40, 0x0, CCRequestNumber)\n m.NewAVP(\"Requested-Action\", 0x40, 0x0, RequestedAction)\n m.NewAVP(\"Subscription-Id\", 0x40, 0x00, &diam.Grouped{\n AVP: []*diam.AVP{\n // Subscription-Id-Type\n diam.NewAVP(450, 0x40, 0x0, SubscriptionIdType),\n // Subscription-Id-Data\n diam.NewAVP(444, 0x40, 0x0, SubscriptionIdData),\n },\n })\n m.NewAVP(\"Service-Parameter-Info\", 0x40, 0x00, &diam.Grouped{\n AVP: []*diam.AVP{\n // Service-Parameter-Type\n diam.NewAVP(441, 0x40, 0x0, ServiceParameterType1),\n // Service-Parameter-Value\n diam.NewAVP(442, 0x40, 0x0, ServiceParameterValue1),\n },\n })\n m.NewAVP(\"Service-Parameter-Info\", 0x40, 0x00, &diam.Grouped{\n AVP: []*diam.AVP{\n // Service-Parameter-Type\n diam.NewAVP(441, 0x40, 0x0, ServiceParameterType2),\n // Service-Parameter-Value\n diam.NewAVP(442, 0x40, 0x0, ServiceParameterValue2),\n },\n })\n\n log.Printf(\"Sending message to %s\", c.RemoteAddr().String())\n log.Println(m.String())\n // Send message to the connection\n if _, err := m.WriteTo(c); err != nil {\n log.Fatal(\"Write failed:\", err)\n }\n}", "func (n NotifyBroadcasts) construct() NotifyPeerClass { return &n }", "func Constructor() MedianFinder {\n return MedianFinder{small:&PQue{}, big: &PQue{}}\n}", "func MClassic(c pcsc.Card) (mifare.Classic, error) {\n\n\tmc := &mifareClassic{\n\t\tCard: c,\n\t}\n\treturn mc, nil\n}", "func (mio *Mio) Create(id string, sz uint64, anyPool ...string) error {\n if mio.obj != nil {\n return errors.New(\"object is already opened\")\n }\n if err := mio.objNew(id); err != nil {\n return err\n }\n pool, err := checkPool(anyPool)\n if err != nil {\n return err\n }\n\n lid := C.m0_layout_find_by_objsz(C.instance, pool, C.ulong(sz))\n if lid <= 0 {\n return fmt.Errorf(\"could not find layout: rc=%v\", lid)\n }\n C.m0_obj_init(mio.obj, &C.container.co_realm, &mio.objID, C.ulong(lid))\n\n var op *C.struct_m0_op\n rc := C.m0_entity_create(pool, &mio.obj.ob_entity, &op)\n if rc != 0 {\n return fmt.Errorf(\"failed to create object: %d\", rc)\n }\n C.m0_op_launch(&op, 1)\n rc = C.m0_op_wait(op, bits(C.M0_OS_FAILED,\n C.M0_OS_STABLE), C.M0_TIME_NEVER)\n if rc == 0 {\n rc = C.m0_rc(op)\n }\n C.m0_op_fini(op)\n C.m0_op_free(op)\n\n if rc != 0 {\n return fmt.Errorf(\"create op failed: %d\", rc)\n }\n\n return mio.open(sz)\n}", "func (i InlineQueryPeerTypeBotPM) construct() InlineQueryPeerTypeClass { return &i }", "func NewComponentScaleMinMaxI(args ...interface{}) *ComponentScaleMinMax {\n return &ComponentScaleMinMax{js.Global.Get(\"Phaser\").Get(\"Component\").Get(\"ScaleMinMax\").New(args)}\n}", "func SETMI(mr operand.Op) { ctx.SETMI(mr) }", "func Instance() *Cache {\n\tif client == nil {\n\t\tclient = new()\n\t}\n\treturn client\n}", "func init() {\n\tif err := mb.Registry.AddMetricSet(\"haproxy\", \"info\", New); err != nil {\n\t\tpanic(err)\n\t}\n}", "func Constructor() MedianFinder {\n\treturn MedianFinder{}\n}", "func Constructor() MedianFinder {\n\treturn MedianFinder{}\n}", "func init() {\n\tdao.Server = \"mongodb://shivam:[email protected]:25294/shayona-store\"\n\tdao.Database = \"shayona-store\"\n\tdao.Connect()\n}", "func Inst() *Torpedo {\n\treturn instance\n}", "func (i InlineQueryPeerTypeSameBotPM) construct() InlineQueryPeerTypeClass { return &i }", "func (i InlineQueryPeerTypeBroadcast) construct() InlineQueryPeerTypeClass { return &i }", "func New() *M {\n\tc := &M{}\n\tc.Component()\n\tc.items = make([]*js.Object, 0)\n\treturn c\n}", "func (i InputInlineQueryResultGame) construct() InputInlineQueryResultClass { return &i }", "func (self *PhysicsP2) CreateContactMaterialI(args ...interface{}) *PhysicsP2ContactMaterial{\n return &PhysicsP2ContactMaterial{self.Object.Call(\"createContactMaterial\", args)}\n}", "func (m *Mongo) Instance() (*mgo.Database, *mgo.Session) {\n\ts := m.Session.Copy()\n\treturn s.DB(m.Settings.Database), s\n}", "func (m *Motifini) Run() error {\n\tvar err error\n\tm.Info.Println(\"Connecting to SecuritySpy:\", m.Conf.SecuritySpy.URL)\n\tif m.SSpy, err = securityspy.GetServer(m.Conf.SecuritySpy); err != nil {\n\t\treturn err\n\t}\n\tm.ProcessEventStream()\n\tdefer m.SSpy.Events.Stop(true)\n\n\tm.Info.Println(\"Opening Subscriber Database:\", m.Conf.Global.StateFile)\n\tif m.Subs, err = subscribe.GetDB(m.Conf.Global.StateFile); err != nil {\n\t\treturn errors.Wrap(err, \"sub state\")\n\t}\n\n\tm.Msgs = &messenger.Messenger{\n\t\tSSpy: m.SSpy,\n\t\tSubs: m.Subs,\n\t\tConf: m.Conf.Imessage,\n\t\tTempDir: m.Conf.Global.TempDir,\n\t\tInfo: log.New(os.Stdout, \"[MSGS] \", m.Info.Flags()),\n\t\tDebug: m.Debug,\n\t\tError: m.Error,\n\t}\n\tif err := messenger.New(m.Msgs); err != nil {\n\t\treturn err\n\t}\n\n\tm.HTTP = &webserver.Config{\n\t\tSSpy: m.SSpy,\n\t\tSubs: m.Subs,\n\t\tMsgs: m.Msgs,\n\t\tInfo: log.New(os.Stdout, \"[HTTP] \", m.Info.Flags()),\n\t\tDebug: m.Debug,\n\t\tError: m.Error,\n\t\tTempDir: m.Conf.Global.TempDir,\n\t\tAllowedTo: m.Conf.Global.AllowedTo,\n\t\tPort: m.Conf.Global.Port,\n\t}\n\tgo m.waitForSignal()\n\treturn webserver.Start(m.HTTP)\n}", "func Constructor() MedianFinder {\n\n}", "func New(\n\tctx context.Context,\n\tlogger logr.Logger,\n\tclock clock.Clock,\n\tc client.Client,\n\tnamespace string,\n\tidentity string,\n\trotation Config,\n) (\n\tInterface,\n\terror,\n) {\n\tm := &manager{\n\t\tstore: make(secretStore),\n\t\tclock: clock,\n\t\tlogger: logger.WithValues(\"namespace\", namespace),\n\t\tclient: c,\n\t\tnamespace: namespace,\n\t\tidentity: identity,\n\t\tlastRotationInitiationTimes: make(nameToUnixTime),\n\t}\n\n\tif err := m.initialize(ctx, rotation); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func new() exampleInterface {\n\treturn config{}\n}", "func (self *PhysicsP2) PxmiI(args ...interface{}) int{\n return self.Object.Call(\"pxmi\", args).Int()\n}", "func (config Service) Open(log gopi.Logger) (gopi.Driver, error) {\n\tlog.Debug(\"<grpc.service.mihome.Open>{ server=%v mihome=%v }\", config.Server, config.MiHome)\n\n\t// Check for bad input parameters\n\tif config.Server == nil || config.MiHome == nil {\n\t\treturn nil, gopi.ErrBadParameter\n\t}\n\n\tthis := new(service)\n\tthis.log = log\n\tthis.mihome = config.MiHome\n\n\t// Init queue\n\tif err := this.queue.Init(log, config); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Register service with GRPC server\n\tpb.RegisterMiHomeServer(config.Server.(grpc.GRPCServer).GRPCServer(), this)\n\n\t// Success\n\treturn this, nil\n}", "func NewCKMS(keyID string) *CKMS {\n\tif keyID == \"\" {\n\t\tlog.Fatal(\"missing keyID\")\n\t}\n\tregionTmp := strings.Split(keyID, \"arn:aws:kms:\")\n\tregion := \"us-west-1\"\n\tif len(regionTmp) > 1 {\n\t\tindex := strings.Index(regionTmp[1], \":\")\n\t\tif index >= 0 {\n\t\t\tregion = regionTmp[1][:index]\n\t\t}\n\t}\n\t// Initialize a session that the SDK uses to load\n\tsess, err := session.NewSession(\n\t\t&aws.Config{\n\t\t\tRegion: aws.String(region),\n\t\t},\n\t)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to initialize session:%v\\n\", err)\n\t}\n\t// Create KMS service client\n\tsvc := kms.New(sess)\n\tsm := secretsmanager.New(sess)\n\n\treturn &CKMS{\n\t\tsvc: svc,\n\t\tsm: sm,\n\t\tkeyID: keyID,\n\t}\n}", "func (wf *Wavefront) ManagedInst() *Inst {\n\treturn wf.inst\n}", "func init() {\n\tmodules.Register(\"k6/x/memcached\", new(Memcached))\n}", "func (t *Ticker) NewMFI(inTimePeriod int32) *MFI {\n\tcalculator := &mfiCalculator{\n\t\tTicker: t,\n\t\tFlowIdx: 0,\n\t\tMaxFlowIdx: inTimePeriod - 1,\n\t\tPeriod: inTimePeriod,\n\t\tMoneyFlow: make([]moneyFlow, inTimePeriod),\n\t}\n\treturn &MFI{\n\t\tCalculator: calculator,\n\t}\n}", "func (c *client) Instance(bridge, instance string) Interface {\n\treturn &internalClient{\n\t\tsvc: c.services,\n\t\tbridge: bridge,\n\t\tinstance: instance,\n\t}\n}", "func Instance() (err error) {\n\tdsn := getDBConfig()\n\n\tfmt.Println(\"DB dsn adress:\" + dsn)\n\n\tDbConnect, err = gorm.Open(getDBDriver(), dsn)\n\n\tif err != nil {\n\t\tfmt.Printf(\"sql.Open func append faild:%v\", err)\n\t\treturn err\n\t}\n\tDbConnect.SingularTable(true)\n\n\tDbConnect.LogMode(true)\n\treturn nil\n}", "func Instance(config StoreConfig) *Controller {\n\tonce.Do(func() {\n\t\tc = &Controller{}\n\t\tc.storedjourneyService = storedjourneyservice.Instance()\n\t\tc.finished = config.Channel\n\t\tc.metricsService = metricsservice.Instance()\n\t\tc.storagedriver = storage.NewStdoutStorage()\n\t\tc.config = config\n\t\tc.queue = queueservice.Instance()\n\t})\n\treturn c\n}", "func Init(cfg SbiCfg) (err error) {\n\n\t// Create new SBI instance\n\tsbi = new(LocServSbi)\n\tsbi.sandboxName = cfg.SandboxName\n\tsbi.updateUserInfoCB = cfg.UserInfoCb\n\tsbi.updateZoneInfoCB = cfg.ZoneInfoCb\n\tsbi.updateAccessPointInfoCB = cfg.ApInfoCb\n\tsbi.updateScenarioNameCB = cfg.ScenarioNameCb\n\tsbi.cleanUpCB = cfg.CleanUpCb\n\n\t// Create message queue\n\tsbi.mqLocal, err = mq.NewMsgQueue(mq.GetLocalName(sbi.sandboxName), moduleName, sbi.sandboxName, cfg.RedisAddr)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create Message Queue with error: \", err)\n\t\treturn err\n\t}\n\tlog.Info(\"Message Queue created\")\n\n\t// Create new active scenario model\n\tmodelCfg := mod.ModelCfg{\n\t\tName: \"activeScenario\",\n\t\tNamespace: sbi.sandboxName,\n\t\tModule: moduleName,\n\t\tUpdateCb: nil,\n\t\tDbAddr: cfg.RedisAddr,\n\t}\n\tsbi.activeModel, err = mod.NewModel(modelCfg)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create model: \", err.Error())\n\t\treturn err\n\t}\n\tlog.Info(\"Active Scenario Model created\")\n\n\t// Connect to Postgis DB\n\tsbi.pc, err = postgis.NewConnector(geModuleName, sbi.sandboxName, postgisUser, postgisPwd, cfg.PostgisHost, cfg.PostgisPort)\n\tif err != nil {\n\t\tlog.Error(\"Failed to create postgis connector with error: \", err.Error())\n\t\treturn err\n\t}\n\tlog.Info(\"Postgis Connector created\")\n\n\t// Initialize service\n\tprocessActiveScenarioUpdate()\n\n\treturn nil\n}", "func New(memoryInterface MemoryReadWriter) *CPU {\n\tcpu := &CPU{mem: memoryInterface}\n\t// DEBUG this is an exceptionally slow clock\n\tcpu.Clock = time.NewTicker(time.Microsecond).C\n\tcpu.readyForStart = true\n\treturn cpu\n}", "func New(id ID, ch swarm.Chunk) *SOC {\n\treturn &SOC{\n\t\tid: id,\n\t\tchunk: ch,\n\t}\n}", "func (a *ACMEInstance) CreateInstance(challenge, dir, domains, domainIPs, dnsManager string, revoke bool, serverK server.KeyInstance, clientK client.KeyInstanceCert) {\n\n\tif ok := supportedChallenges[challenge]; !ok {\n\t\tfmt.Fprintln(os.Stderr, \"unsupported challengeType: \", challenge)\n\t\tos.Exit(2)\n\t}\n\ta.challengeType = challenge\n\ta.dirURL = dir\n\ta.domainList = strings.Split(domains, \";\")\n\n\tipList := strings.Split(domainIPs, \";\")\n\tipDom := make(map[string]string)\n\tfor n, domain := range a.domainList {\n\t\tipDom[domain] = ipList[n]\n\t}\n\ta.ipForDomain = ipDom\n\n\tif dnsManager == \"azuredns\" {\n\t\ta.dnsManager = dnsmanager.AZUREDNSCredentials{Login: \"dummyuser\",\n\t\t\tPassword: \"1234\",\n\t\t\tSubscriptionID: \"9fa587f1-4961-48a6-b6f6-ec69c6d724f1\",\n\t\t\tResourceGroups: \"fileTransfer\",\n\t\t\tAuthorization: \"Bearer eyJ0eXAiOiJKV1QiLCJhbGciOiJSUzI1NiIsIng1dCI6Imh1Tjk1SXZQZmVocTM0R3pCRFoxR1hHaXJuTSIsImtpZCI6Imh1Tjk1SXZQZmVocTM0R3pCRFoxR1hHaXJuTSJ9.eyJhdWQiOiJodHRwczovL21hbmFnZW1lbnQuY29yZS53aW5kb3dzLm5ldC8iLCJpc3MiOiJodHRwczovL3N0cy53aW5kb3dzLm5ldC9mMDhhODUxZS1hYzVjLTQ5NGItODk0MS00N2U2YTI2NTc4MWQvIiwiaWF0IjoxNTk3MDY5Njc5LCJuYmYiOjE1OTcwNjk2NzksImV4cCI6MTU5NzA3MzU3OSwiYWNyIjoiMSIsImFpbyI6IkFTUUEyLzhRQUFBQWJPNHQrWW5pYVp2NEUzRG4rYUdGOGxVc25RZzhJQ1VHNS9yZitLMEFrcWM9IiwiYW1yIjpbInB3ZCJdLCJhcHBpZCI6IjdmNTlhNzczLTJlYWYtNDI5Yy1hMDU5LTUwZmM1YmIyOGI0NCIsImFwcGlkYWNyIjoiMiIsImZhbWlseV9uYW1lIjoiTWVpZXIiLCJnaXZlbl9uYW1lIjoiRmlsaXAiLCJncm91cHMiOlsiNzA5YzZmM2YtNTEwYy00NTgwLTlkMGYtYzQ1OWJiMTcyMDE3Il0sImlwYWRkciI6IjUxLjE1NC41My4xNjUiLCJuYW1lIjoiRmlsaXAgTWVpZXIiLCJvaWQiOiJmMGNhMzk3Ni1lNjBlLTQzZDItYTM0ZS0wMDZhZGRlNDVhYmIiLCJwdWlkIjoiMTAwMzIwMDA2QjI2MUZDNyIsInJoIjoiMC5BVHdBSG9XSzhGeXNTMG1KUVVmbW9tVjRIWE9uV1gtdkxweENvRmxRX0Z1eWkwUThBTTQuIiwic2NwIjoidXNlcl9pbXBlcnNvbmF0aW9uIiwic3ViIjoiMzlKdHpVVUFCOFNYNWRWRVBOYWJsZzFiUkp5MkhGdW43TEZDRktxamRlVSIsInRpZCI6ImYwOGE4NTFlLWFjNWMtNDk0Yi04OTQxLTQ3ZTZhMjY1NzgxZCIsInVuaXF1ZV9uYW1lIjoiZmlsaXAubWVpZXJAOGRheXNhd2Vlay5jYyIsInVwbiI6ImZpbGlwLm1laWVyQDhkYXlzYXdlZWsuY2MiLCJ1dGkiOiJ1aXBUU3VVTmhrYXdqckpyNzRSSkFBIiwidmVyIjoiMS4wIiwid2lkcyI6WyI2MmU5MDM5NC02OWY1LTQyMzctOTE5MC0wMTIxNzcxNDVlMTAiXSwieG1zX3RjZHQiOjE1NjgxODgyNzZ9.nzU-Yj1uwpqPqcUJVE7iKunPtTCwFbQ4_pE-EiAiVGWpzi6A2e1t9YAW6s2BqnHVltNDO2xBJRZgyjSuetCekuX_nxvwZIU4hDppl5lrt6O85-PtQrYR34DOa05O2fg7a53lhP_b5uSy3XexZpqwNvpbC0dqAictuv59kN6rlZQyUoP_J70jVx-WhXwGQNpgn9uDs11SDgxioKIgrDh0rA1q0kJxJ-4pLbO6l2B2KfL0lrkSJinKFPslwhRhKTHFhqVbdSkiFV7gyK-Kc13iGzXUiB2aHu3M6B-Yy5fmRoF4SMFqFJelLvlctDPUiLK8b9_qQpX60aYIEnbur4amtg\",\n\t\t}\n\t}\n\n\ta.revokeCert = revoke\n\n\ta.challengeStatusMap = make(map[string]ChallengeStatus)\n\ta.challangeCreated = make(map[string]bool)\n\n\ta.serverKey = serverK //used to communicate with ACME server\n\ta.clientKey = clientK\n\n}", "func startMnm(ctx *cli.Context) error {\n\tconfPath := path.Join(ctx.GlobalString(\"config-dir\"), \"config.json\")\n\n\tconfigContents, err := ioutil.ReadFile(confPath)\n\tif err != nil {\n\t\tlog.Fatal(\"Unable to read configuration file: \", err)\n\t}\n\n\tconfig := Config{}\n\terr = json.Unmarshal(configContents, &config)\n\tif err != nil {\n\t\tlog.Fatal(\"Unable to parse configuration file: \", err)\n\t}\n\n\taddress := ctx.GlobalString(\"address\")\n\tm := mnm{hashMap: make(map[string]*minio.Client), address: address, auth: config.Auth}\n\n\tfor _, host := range config.Endpoints {\n\t\turl, err := url.Parse(host.URL)\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Unable to parse %s: %s\", host.URL, err)\n\t\t}\n\t\tclient, err := minio.New(url.Host, host.AccessKey, host.SecretKey, url.Scheme == \"https\")\n\t\tif err != nil {\n\t\t\tlog.Fatal(\"Minio client init failed: \", err)\n\t\t}\n\t\tsum := fmt.Sprintf(\"%x\", md5.Sum([]byte(url.Host)))\n\t\tm.hashMap[sum] = client\n\t\tm.hashes = append(m.hashes, sum)\n\t}\n\n\tr := mux.NewRouter()\n\tr.Methods(\"GET\").Path(\"/mnm/{hash}/{object:.+}\").Queries(\"auth\", \"{auth:.*}\").HandlerFunc(m.get)\n\tr.Methods(\"GET\").Path(\"/mnm/{hash}/{object:.+}\").HandlerFunc(m.get)\n\tr.Methods(\"PUT\").Path(\"/mnm/{object:.+}\").Queries(\"auth\", \"{auth:.*}\").HandlerFunc(m.put)\n\tr.Methods(\"PUT\").Path(\"/mnm/{object:.+}\").HandlerFunc(m.put)\n\n\treturn http.ListenAndServe(address, r)\n}", "func New() *inMem {\n\tim := inMem{\n\t\tstorage: make(map[string]interface{}),\n\t\tcloseChanel: make(chan struct{}),\n\t}\n\treturn &im\n}", "func New(ctx context.Context, trace bool, testURL string, resolverURLs string, credentialsManagerChannel <-chan interface{}, opts ...Option) (vos.Interface, error) {\n\tinst := &instance{periodicDiskMonitorTime: periodicDiskMonitorTime}\n\n\t// Run options\n\tfor _, opt := range opts {\n\t\tif opt != nil {\n\t\t\topt(inst)\n\t\t}\n\t}\n\n\tupdateGoMaxProcs()\n\tdisableMinioBrowser()\n\n\tvar credentialsManager vminio.CredentialsManager\n\n\tselect {\n\tcase item := <-credentialsManagerChannel:\n\t\tif mgr, ok := item.(vminio.CredentialsManager); ok {\n\t\t\tcredentialsManager = mgr\n\t\t} else {\n\t\t\tlog.Errorf(\"Unexpected item found in credentials manager channel, item: %v\", item)\n\t\t\treturn nil, errors.New(\"unexpected item found in credentials manager channel\")\n\t\t}\n\t}\n\n\tminioCreds, err := credentialsManager.GetCredentials()\n\tif err != nil {\n\t\tlog.Errorf(\"Unable to get MINIO credentials, error: %s\", err.Error())\n\t\treturn nil, errors.Wrap(err, \"unable to get MINIO credentials\")\n\t}\n\n\tos.Setenv(globals.MinioAccessKeyName, minioCreds.AccessKey)\n\tos.Setenv(globals.MinioSecretKeyName, minioCreds.SecretKey)\n\tif minioCreds.OldAccessKey != \"\" {\n\t\tos.Setenv(globals.MinioOldAccessKeyName, minioCreds.OldAccessKey)\n\t}\n\tif minioCreds.OldSecretKey != \"\" {\n\t\tos.Setenv(globals.MinioOldSecretKeyName, minioCreds.OldSecretKey)\n\t}\n\tvar envVars string\n\t// remove secrets out from log info\n\tfor _, envVar := range os.Environ() {\n\t\tif !strings.Contains(envVar, \"MINIO_ACCESS_KEY\") &&\n\t\t\t!strings.Contains(envVar, \"MINIO_SECRET_KEY\") {\n\t\t\tenvVars = fmt.Sprintf(\"%s %s\", envVars, envVar)\n\t\t}\n\t}\n\tlog.Infof(\"minio env: %s\", envVars)\n\n\tif trace {\n\t\tos.Setenv(\"MINIO_HTTP_TRACE\", \"/dev/stdout\")\n\t\tlog.Infof(\"minio enabled API tracing\")\n\t}\n\n\tupdateAPIThrottlingParams()\n\n\tlog.Infof(\"minio args: %+v\", inst.bootupArgs)\n\n\tgo minio.Main(inst.bootupArgs)\n\ttime.Sleep(10 * time.Second)\n\n\turl := k8s.GetPodIP() + \":\" + globals.VosMinioPort\n\tif testURL != \"\" {\n\t\turl = testURL + \":\" + globals.VosMinioPort\n\t}\n\n\tlog.Infof(\"connecting to minio at [%v]\", url)\n\n\tsecureMinio := true\n\tif testURL != \"\" {\n\t\tsecureMinio = false\n\t}\n\tmclient, err := miniogo.New(url, minioCreds.AccessKey, minioCreds.SecretKey, secureMinio)\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create client (%s)\", err)\n\t\treturn nil, errors.Wrap(err, \"Failed to create Client\")\n\t}\n\tdefTr := http.DefaultTransport.(*http.Transport)\n\ttlcc, tlsc, err := getTLSConfig(testURL, false)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefTr.TLSClientConfig = tlcc\n\n\titlcc, _, err := getTLSConfig(testURL, true)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tmclient.SetCustomTransport(defTr)\n\tclient := &storeImpl{BaseBackendClient: mclient}\n\tinst.Init(client)\n\n\tadminClient, err := minioclient.NewPinnedAdminClient(url,\n\t\tminioclient.WithCredentialsManager(credentialsManager), minioclient.WithTLSConfig(itlcc))\n\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create admin client (%s)\", err)\n\t\treturn nil, errors.Wrap(err, \"Failed to create admin client\")\n\t}\n\n\tgo monitorAndRecoverMinio(ctx, adminClient)\n\n\tgrpcBackend, err := newGrpcServer(inst, client)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to start grpc listener\")\n\t}\n\n\thttpBackend, err := newHTTPHandler(inst, client, adminClient)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to start http listener\")\n\t}\n\n\t// For simplicity all nodes in the cluster check if the default buckets exist,\n\t// if not, try to create the buckets. all nodes in the cluster try this till\n\t// all default buckets are created. A little inefficient but simple and a rare\n\t// operation (only on a create of a new cluster)\n\terr = inst.createDefaultBuckets(client)\n\tif err != nil {\n\t\tlog.Errorf(\"failed to create buckets (%+v)\", err)\n\t\treturn nil, errors.Wrap(err, \"failed to create buckets\")\n\t}\n\n\tfor {\n\t\tonline, err := adminClient.IsClusterOnline(ctx)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"minio cluster is not online yet (%+v)\", err)\n\t\t\tcontinue\n\t\t}\n\n\t\tif online {\n\t\t\terr := inst.setBucketLifecycle()\n\t\t\tif err != nil {\n\t\t\t\tlog.Errorf(\"failed to set bucket lifecycle (%+v)\", err)\n\t\t\t\treturn nil, errors.Wrap(err, \"failed to set bucket lifecycle\")\n\t\t\t}\n\t\t\tbreak\n\t\t}\n\t}\n\n\t_, err = inst.createDiskUpdateWatcher(inst.bucketDiskThresholds,\n\t\tinst.periodicDiskMonitorTime, DiskPaths, flowlogsDiskThresholdCriticalEventReportingPeriod)\n\tif err != nil {\n\t\tlog.Errorf(\"failed to start disk watcher (%+v)\", err)\n\t\treturn nil, errors.Wrap(err, \"failed to start disk watcher\")\n\t}\n\n\t// Register all plugins\n\tplugins.RegisterPlugins(inst)\n\tgrpcBackend.start(ctx)\n\thttpBackend.start(ctx, globals.VosHTTPPort, minioCreds, tlsc)\n\tinst.initTsdbClient(resolverURLs, minioCreds.AccessKey, minioCreds.SecretKey)\n\tlog.Infof(\"Initialization complete\")\n\t<-ctx.Done()\n\treturn inst, nil\n}", "func init() {\n\tinputs.Add(\"nmon_poweragent\", newNmonServer)\n}", "func OrganizationManagerInstance() (*OrganizationManager) {\n if organizationManager != nil {\n return organizationManager\n } else {\n // Get our database connection.\n dao := database.Instance()\n organizationManager = &OrganizationManager{dao}\n return organizationManager\n }\n}", "func NewStandalone(mi *mesos.MasterInfo) *Standalone {\n\tlog.V(2).Infof(\"creating new standalone detector for %+v\", mi)\n\tstand := &Standalone{\n\t\tch: make(chan *mesos.MasterInfo),\n\t\ttr: &http.Transport{},\n\t\tinitial: mi,\n\t\tdone: make(chan struct{}),\n\t\tleaderSyncInterval: defaultMesosLeaderSyncInterval,\n\t\thttpClientTimeout: defaultMesosHttpClientTimeout,\n\t\tassumedMasterPort: defaultMesosMasterPort,\n\t}\n\tstand.poller = stand._poller\n\tstand.fetchPid = stand._fetchPid\n\treturn stand\n}", "func NewIosDeviceType()(*IosDeviceType) {\n m := &IosDeviceType{\n }\n m.SetAdditionalData(make(map[string]interface{}));\n return m\n}", "func NewMosn() *Mosn {\n\tlog.StartLogger.Infof(\"[mosn start] create an empty mosn structure\")\n\tm := &Mosn{\n\t\tUpgrade: UpgradeData{},\n\t}\n\treturn m\n}", "func StartMME(OaiObj Oai) {\n\t// Start Mme\n\tstartMme(OaiObj)\n}" ]
[ "0.61289626", "0.5952556", "0.5817913", "0.554648", "0.5344804", "0.53326446", "0.5241001", "0.52322817", "0.50988203", "0.50717044", "0.50496864", "0.50247544", "0.4993466", "0.49884257", "0.49783844", "0.49596447", "0.4942577", "0.492635", "0.49250606", "0.49086714", "0.48879352", "0.48814258", "0.487238", "0.48693675", "0.48418027", "0.48410416", "0.4836538", "0.48251858", "0.48220325", "0.48196763", "0.48140654", "0.48106152", "0.48053092", "0.4803759", "0.4803552", "0.47859928", "0.47837907", "0.47835216", "0.47554544", "0.47535184", "0.47323862", "0.4728537", "0.47272384", "0.47271112", "0.4721439", "0.47156426", "0.47131062", "0.4712544", "0.47100118", "0.47069207", "0.46895027", "0.46878374", "0.46770105", "0.46731222", "0.4668282", "0.46620235", "0.46563372", "0.46544078", "0.46455806", "0.46426654", "0.46426374", "0.46362314", "0.4632305", "0.46297124", "0.46269307", "0.46269307", "0.46203563", "0.46154714", "0.46138135", "0.46013793", "0.46006137", "0.45958742", "0.45917925", "0.45913425", "0.45901582", "0.45799884", "0.4575789", "0.45627007", "0.45620075", "0.4561077", "0.45504755", "0.45460898", "0.45425245", "0.4539391", "0.45386794", "0.45362276", "0.453082", "0.45257828", "0.4520687", "0.45132026", "0.45106062", "0.45035505", "0.4499953", "0.44995263", "0.44946545", "0.44930598", "0.44870505", "0.44852617", "0.44835922", "0.44813302" ]
0.7207688
0
/ instance of a VM
func NewVM(namespace string, name string, mcisName string) *VM { return &VM{ Model: Model{Name: name, Namespace: namespace}, McisName: mcisName, UserAccount: VM_USER_ACCOUNT, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *crostiniActivation) VM() interface{} {\n\treturn c.container\n}", "func (c *Compute) Instance(name, disk string) (string, error) {\n\tinstance, err := c.Instances.Get(c.Project, c.Zone, name).Do()\n\tif err == nil {\n\t\tlog.Printf(\"found existing instance: %q\", instance.SelfLink)\n\t\treturn instance.SelfLink, nil\n\t}\n\tlog.Printf(\"not found, creating new instance: %q\", name)\n\top, err := c.Instances.Insert(c.Project, c.Zone, &compute.Instance{\n\t\tName: name,\n\t\tDescription: \"gorogoro vm\",\n\t\tMachineType: c.Prefix + *machine,\n\t\tDisks: []*compute.AttachedDisk{\n\t\t\t{\n\t\t\t\tBoot: true,\n\t\t\t\tType: \"PERSISTENT\",\n\t\t\t\tMode: \"READ_WRITE\",\n\t\t\t\tSource: disk,\n\t\t\t},\n\t\t},\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t&compute.AccessConfig{Type: \"ONE_TO_ONE_NAT\"},\n\t\t\t\t},\n\t\t\t\tNetwork: c.Prefix + \"/global/networks/default\",\n\t\t\t},\n\t\t},\n\t}).Do()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"instance insert api call failed: %v\", err)\n\t}\n\tif err := c.wait(op); err != nil {\n\t\treturn \"\", fmt.Errorf(\"instance insert operation failed: %v\", err)\n\t}\n\tlog.Printf(\"instance created: %q\", op.TargetLink)\n\treturn op.TargetLink, nil\n}", "func (t Test) VM() error {\n\tmg.Deps(t.FixtureVMImages)\n\treturn sh.RunWithV(ENV, \"go\", \"test\", \"-v\", \"-tags=vm_integration\", \"./integration/...\")\n}", "func (r *Finder) VM(ref *base.Ref) (object interface{}, err error) {\n\tvm := &VM{}\n\terr = r.ByRef(vm, *ref)\n\tif err == nil {\n\t\tref.ID = vm.ID\n\t\tref.Name = vm.Name\n\t\tobject = vm\n\t}\n\n\treturn\n}", "func (c *Controller) VM(name string) *VMInterface {\n\treturn newVMInterface(c, name)\n}", "func NewVM() VM {\n\tcpu := CPU{0, 0, MEMSIZE - 1}\n\treturn VM{cpu: cpu}\n}", "func vmQemuInstantiate(s *state.State, args db.InstanceArgs) *vmQemu {\n\tvm := &vmQemu{\n\t\tstate: s,\n\t\tid: args.ID,\n\t\tproject: args.Project,\n\t\tname: args.Name,\n\t\tdescription: args.Description,\n\t\tephemeral: args.Ephemeral,\n\t\tarchitecture: args.Architecture,\n\t\tdbType: args.Type,\n\t\tsnapshot: args.Snapshot,\n\t\tcreationDate: args.CreationDate,\n\t\tlastUsedDate: args.LastUsedDate,\n\t\tprofiles: args.Profiles,\n\t\tlocalConfig: args.Config,\n\t\tlocalDevices: args.Devices,\n\t\tstateful: args.Stateful,\n\t\tnode: args.Node,\n\t\texpiryDate: args.ExpiryDate,\n\t}\n\n\t// Cleanup the zero values.\n\tif vm.expiryDate.IsZero() {\n\t\tvm.expiryDate = time.Time{}\n\t}\n\n\tif vm.creationDate.IsZero() {\n\t\tvm.creationDate = time.Time{}\n\t}\n\n\tif vm.lastUsedDate.IsZero() {\n\t\tvm.lastUsedDate = time.Time{}\n\t}\n\n\treturn vm\n}", "func NewVM(c *config.Config, m *metrics.Metrics) (v VM) {\n\tv.Config = c\n\tv.Metrics = m\n\treturn\n}", "func NewVM() (*VM, error) {\n\tclient, err := cutil.NewDockerClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tVM := &VM{Client: client}\n\treturn VM, nil\n}", "func NewVM(config *VMConfig) *VM {\n\tvm := &VM{\n\t\tcfg: config,\n\t\tvalueid: ValueGo + 1,\n\t\tvalues: make(map[Ref]*Value),\n\t}\n\tRegisterBuiltins(vm.cfg.Global)\n\tvm.initDefaultValue()\n\treturn vm\n}", "func vmQemuCreate(s *state.State, args db.InstanceArgs) (Instance, error) {\n\t// Create the instance struct.\n\tvm := &vmQemu{\n\t\tstate: s,\n\t\tid: args.ID,\n\t\tproject: args.Project,\n\t\tname: args.Name,\n\t\tnode: args.Node,\n\t\tdescription: args.Description,\n\t\tephemeral: args.Ephemeral,\n\t\tarchitecture: args.Architecture,\n\t\tdbType: args.Type,\n\t\tsnapshot: args.Snapshot,\n\t\tstateful: args.Stateful,\n\t\tcreationDate: args.CreationDate,\n\t\tlastUsedDate: args.LastUsedDate,\n\t\tprofiles: args.Profiles,\n\t\tlocalConfig: args.Config,\n\t\tlocalDevices: args.Devices,\n\t\texpiryDate: args.ExpiryDate,\n\t}\n\n\t// Cleanup the zero values.\n\tif vm.expiryDate.IsZero() {\n\t\tvm.expiryDate = time.Time{}\n\t}\n\n\tif vm.creationDate.IsZero() {\n\t\tvm.creationDate = time.Time{}\n\t}\n\n\tif vm.lastUsedDate.IsZero() {\n\t\tvm.lastUsedDate = time.Time{}\n\t}\n\n\tctxMap := log.Ctx{\n\t\t\"project\": args.Project,\n\t\t\"name\": vm.name,\n\t\t\"ephemeral\": vm.ephemeral,\n\t}\n\n\tlogger.Info(\"Creating instance\", ctxMap)\n\n\trevert := true\n\tdefer func() {\n\t\tif !revert {\n\t\t\treturn\n\t\t}\n\n\t\tvm.Delete()\n\t}()\n\n\t// Load the config.\n\terr := vm.init()\n\tif err != nil {\n\t\tlogger.Error(\"Failed creating instance\", ctxMap)\n\t\treturn nil, err\n\t}\n\n\t// Validate expanded config\n\terr = containerValidConfig(s.OS, vm.expandedConfig, false, true)\n\tif err != nil {\n\t\tlogger.Error(\"Failed creating instance\", ctxMap)\n\t\treturn nil, err\n\t}\n\n\terr = containerValidDevices(s, s.Cluster, vm.Name(), vm.expandedDevices, true)\n\tif err != nil {\n\t\tlogger.Error(\"Failed creating instance\", ctxMap)\n\t\treturn nil, errors.Wrap(err, \"Invalid devices\")\n\t}\n\n\t// Retrieve the instance's storage pool\n\t_, rootDiskDevice, err := shared.GetRootDiskDevice(vm.expandedDevices.CloneNative())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif rootDiskDevice[\"pool\"] == \"\" {\n\t\treturn nil, fmt.Errorf(\"The instances's root device is missing the pool property\")\n\t}\n\n\tstoragePool := rootDiskDevice[\"pool\"]\n\n\t// Get the storage pool ID for the instance.\n\tpoolID, pool, err := s.Cluster.StoragePoolGet(storagePool)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Fill in any default volume config.\n\tvolumeConfig := map[string]string{}\n\terr = storagePools.VolumeFillDefault(storagePool, volumeConfig, pool)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create a new database entry for the instance's storage volume.\n\t_, err = s.Cluster.StoragePoolVolumeCreate(args.Project, args.Name, \"\", db.StoragePoolVolumeTypeVM, false, poolID, volumeConfig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !vm.IsSnapshot() {\n\t\t// Update MAAS.\n\t\terr = vm.maasUpdate(nil)\n\t\tif err != nil {\n\t\t\tlogger.Error(\"Failed creating instance\", ctxMap)\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Add devices to instance.\n\t\tfor k, m := range vm.expandedDevices {\n\t\t\terr = vm.deviceAdd(k, m)\n\t\t\tif err != nil && err != device.ErrUnsupportedDevType {\n\t\t\t\treturn nil, errors.Wrapf(err, \"Failed to add device '%s'\", k)\n\t\t\t}\n\t\t}\n\t}\n\n\tlogger.Info(\"Created instance\", ctxMap)\n\tvm.state.Events.SendLifecycle(vm.project, \"virtual-machine-created\",\n\t\tfmt.Sprintf(\"/1.0/virtual-machines/%s\", vm.name), nil)\n\n\trevert = false\n\treturn vm, nil\n}", "func (g *GctScriptManager) New() *VM {\n\tif VMSCount.Len() >= int32(g.GetMaxVirtualMachines()) {\n\t\tif g.config.Verbose {\n\t\t\tlog.Warnf(log.GCTScriptMgr, \"GCTScript MaxVirtualMachines (%v) hit, unable to start further instances\",\n\t\t\t\tg.GetMaxVirtualMachines())\n\t\t}\n\t\treturn nil\n\t}\n\tVMSCount.add()\n\tvm := g.NewVM()\n\tif vm == nil {\n\t\tVMSCount.remove()\n\t} else {\n\t\tAllVMSync.Store(vm.ID, vm)\n\t}\n\treturn vm\n}", "func (m *Manager) CreateVM(name string, image string, size string, region string, sshKey core.SSHKey) (core.VMInfo, error) {\n\tvar vmInfo core.VMInfo\n\n\t// generate a strong root password. we will through this away\n\t// TODO: should really disable password login for root\n\t// TODO: should check if it is actually enabled\n\trootPassword, err := password.Generate(64, 10, 10, false, false)\n\tif err != nil {\n\t\treturn vmInfo, err\n\t}\n\n\tcreateOptions := linodego.InstanceCreateOptions{\n\t\tRegion: region,\n\t\tType: size,\n\t\tLabel: name,\n\t\tImage: image,\n\t\tRootPass: rootPassword,\n\t}\n\n\tcreateOptions.AuthorizedKeys = append(createOptions.AuthorizedKeys, sshKey.GetPublicKey())\n\tcreateOptions.Tags = append(createOptions.Tags, \"eezhee\")\n\tnewInstance, err := m.api.CreateInstance(context.Background(), createOptions)\n\tif err != nil {\n\t\treturn vmInfo, err\n\t}\n\n\t// see if vm ready\n\tstatus := newInstance.Status\n\tfor status != linodego.InstanceRunning {\n\t\t// wait a bit\n\t\ttime.Sleep(2 * time.Second)\n\n\t\tinstanceInfo, err := m.api.GetInstance(context.Background(), newInstance.ID)\n\t\tif err != nil {\n\t\t\treturn vmInfo, err\n\t\t}\n\n\t\tstatus = instanceInfo.Status\n\n\t}\n\n\t// TODO - instanceInfo has the latest info - new Instance is stale\n\tvmInfo, _ = convertVMInfoToGenericFormat(*newInstance)\n\n\treturn vmInfo, nil\n}", "func NewMachine(opts machine.InitOptions) (machine.VM, error) {\n\tvmConfigDir, err := machine.GetConfDir(vmtype)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvm := new(MachineVM)\n\tif len(opts.Name) > 0 {\n\t\tvm.Name = opts.Name\n\t}\n\tignitionFile := filepath.Join(vmConfigDir, vm.Name+\".ign\")\n\tvm.IgnitionFilePath = ignitionFile\n\n\t// An image was specified\n\tif len(opts.ImagePath) > 0 {\n\t\tvm.ImagePath = opts.ImagePath\n\t}\n\n\t// Assign remote user name. if not provided, use default\n\tvm.RemoteUsername = opts.Username\n\tif len(vm.RemoteUsername) < 1 {\n\t\tvm.RemoteUsername = defaultRemoteUser\n\t}\n\n\t// Add a random port for ssh\n\tport, err := utils.GetRandomPort()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvm.Port = port\n\n\tvm.CPUs = opts.CPUS\n\tvm.Memory = opts.Memory\n\n\t// Look up the executable\n\texecPath, err := exec.LookPath(QemuCommand)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcmd := append([]string{execPath})\n\t// Add memory\n\tcmd = append(cmd, []string{\"-m\", strconv.Itoa(int(vm.Memory))}...)\n\t// Add cpus\n\tcmd = append(cmd, []string{\"-smp\", strconv.Itoa(int(vm.CPUs))}...)\n\t// Add ignition file\n\tcmd = append(cmd, []string{\"-fw_cfg\", \"name=opt/com.coreos/config,file=\" + vm.IgnitionFilePath}...)\n\t// Add qmp socket\n\tmonitor, err := NewQMPMonitor(\"unix\", vm.Name, defaultQMPTimeout)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvm.QMPMonitor = monitor\n\tcmd = append(cmd, []string{\"-qmp\", monitor.Network + \":/\" + monitor.Address + \",server=on,wait=off\"}...)\n\n\t// Add network\n\t// Right now the mac address is hardcoded so that the host networking gives it a specific IP address. This is\n\t// why we can only run one vm at a time right now\n\tcmd = append(cmd, []string{\"-netdev\", \"socket,id=vlan,fd=3\", \"-device\", \"virtio-net-pci,netdev=vlan,mac=5a:94:ef:e4:0c:ee\"}...)\n\tsocketPath, err := getRuntimeDir()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvirtualSocketPath := filepath.Join(socketPath, \"podman\", vm.Name+\"_ready.sock\")\n\t// Add serial port for readiness\n\tcmd = append(cmd, []string{\n\t\t\"-device\", \"virtio-serial\",\n\t\t\"-chardev\", \"socket,path=\" + virtualSocketPath + \",server=on,wait=off,id=\" + vm.Name + \"_ready\",\n\t\t\"-device\", \"virtserialport,chardev=\" + vm.Name + \"_ready\" + \",name=org.fedoraproject.port.0\"}...)\n\tvm.CmdLine = cmd\n\treturn vm, nil\n}", "func NewVM(bc *bytecode.BC) *VM {\n\tbc.AddInst(instruction.NewInst(opcode.Match, 0, nil, nil), bc.N)\n\n\treturn &VM{\n\t\tbc: *bc,\n\t\tthreads: []*Thread{},\n\t}\n}", "func (p *plugin) Provision(spec instance.Spec) (*instance.ID, error) {\n\n\tvar properties map[string]interface{}\n\n\tif spec.Properties != nil {\n\t\tif err := spec.Properties.Decode(&properties); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Invalid instance properties: %s\", err)\n\t\t}\n\t}\n\n\tnewVM, err := parseParameters(properties, p)\n\tif err != nil {\n\t\tlog.Error(\"Problems Whilst Parsting\", \"err\", err)\n\t\treturn nil, err\n\t}\n\n\terr = setInternalStructures(p.vC, p.vCenterInternals)\n\tif err != nil {\n\t\tlog.Error(\"Problem whilst setting Internal Config\", \"err\", err)\n\t\treturn nil, err\n\t}\n\n\tif *p.vC.networkName != \"\" {\n\t\tfindNetwork(p.vC, p.vCenterInternals)\n\t}\n\n\t// Use the VMware plugin data in order to provision a new VM server\n\tvmName := instance.ID(fmt.Sprintf(newVM.vmPrefix+\"-%d\", rand.Int63()))\n\tif spec.Tags != nil {\n\t\tlog.Info(\"Provisioning\", \"vm\", string(vmName), \"group\", spec.Tags[group.GroupTag])\n\t} else {\n\t\tlog.Info(\"Provisioning\", \"vm\", string(vmName))\n\t}\n\n\t// Spawn a goroutine to provision in the background\n\tgo func() {\n\t\tnewVM.vmName = string(vmName)\n\t\tvar newInstanceError error\n\t\tif newVM.vmTemplate != \"\" {\n\t\t\tlog.Info(\"Cloning new instance\", \"template\", newVM.vmTemplate)\n\t\t\tnewInstanceError = cloneNewInstance(p, &newVM, spec)\n\t\t} else {\n\t\t\tnewInstanceError = createNewVMInstance(p, &newVM, spec)\n\t\t}\n\t\tif newInstanceError != nil {\n\t\t\tlog.Warn(\"Error creating\", \"vm\", newVM.vmName)\n\t\t\tlog.Error(\"vCenter problem\", \"err\", newInstanceError)\n\t\t}\n\t}()\n\n\tvar newInstance provisioningFSM\n\tnewInstance.instanceName = string(vmName)\n\tnewInstance.timer = time.Now().Add(time.Minute * 10) // Ten Minute timeout\n\n\t// duplicate the tags for the instance\n\tnewInstance.tags = make(map[string]string)\n\tfor k, v := range spec.Tags {\n\t\tnewInstance.tags[k] = v\n\t}\n\tnewInstance.tags[\"infrakit.state\"] = \"Provisioning\"\n\tp.fsm = append(p.fsm, newInstance)\n\n\tlog.Debug(\"FSM\", \"Count\", len(p.fsm))\n\n\treturn &vmName, nil\n}", "func (n *Network) NewVM(nopts ...NetworkOpt) Device {\n\tif n == nil {\n\t\treturn nil\n\t}\n\n\tnewNum := atomic.AddUint32(&n.numVMs, 1)\n\tnum := newNum - 1\n\n\t// MAC for the virtualized NIC.\n\t//\n\t// This is from the range of locally administered address ranges.\n\tmac := net.HardwareAddr{0x0e, 0x00, 0x00, 0x00, 0x00, byte(num)}\n\tdevID := fmt.Sprintf(\"vm%d\", num)\n\n\targs := []string{\"-device\", fmt.Sprintf(\"e1000,netdev=%s,mac=%s\", devID, mac)}\n\t// Note: QEMU in CircleCI seems to in solve cases fail when using just ':1234' format.\n\t// It fails with \"address resolution failed for :1234: Name or service not known\"\n\t// hinting that this is somehow related to DNS resolution. To work around this,\n\t// we explicitly bind to 127.0.0.1 (IPv6 [::1] is not parsed correctly by QEMU).\n\tif num != 0 {\n\t\targs = append(args, \"-netdev\", fmt.Sprintf(\"socket,id=%s,connect=127.0.0.1:%d\", devID, n.port))\n\t} else {\n\t\targs = append(args, \"-netdev\", fmt.Sprintf(\"socket,id=%s,listen=127.0.0.1:%d\", devID, n.port))\n\t}\n\n\tfor _, opt := range nopts {\n\t\targs = append(args, opt(devID)...)\n\t}\n\treturn networkImpl{args}\n}", "func ShowInstance(id string) (*Runtime, error) {\n\tctx := context.Background()\n\tinfo, err := rpc.Show(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar r Runtime\n\terr = mapstructure.Decode(info, &r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &r, nil\n}", "func NewVm(id string, name string) *Vm {\n\tthis := Vm{}\n\tthis.Id = id\n\tthis.Name = name\n\treturn &this\n}", "func New(img Image, dump string, shrink bool, term *Term) *VM {\n\tvm := VM{\n\t\timg,\n\t\tdump,\n\t\tshrink,\n\t\tmake(map[int32]chan int32),\n\t\tmake(map[int32]*os.File),\n\t\tterm,\n\t}\n\treturn &vm\n}", "func cmdInit() {\n\tif state := status(B2D.VM); state != vmUnregistered {\n\t\tlog.Fatalf(\"%s already exists.\\n\")\n\t}\n\n\tif ping(fmt.Sprintf(\"localhost:%d\", B2D.DockerPort)) {\n\t\tlog.Fatalf(\"DOCKER_PORT=%d on localhost is occupied. Please choose another none.\", B2D.DockerPort)\n\t}\n\n\tif ping(fmt.Sprintf(\"localhost:%d\", B2D.SSHPort)) {\n\t\tlog.Fatalf(\"SSH_HOST_PORT=%d on localhost is occupied. Please choose another one.\", B2D.SSHPort)\n\t}\n\n\tlog.Printf(\"Creating VM %s...\", B2D.VM)\n\tif err := vbm(\"createvm\", \"--name\", B2D.VM, \"--register\"); err != nil {\n\t\tlog.Fatalf(\"failed to create vm: %s\", err)\n\t}\n\n\tif err := vbm(\"modifyvm\", B2D.VM,\n\t\t\"--ostype\", \"Linux26_64\",\n\t\t\"--cpus\", fmt.Sprintf(\"%d\", runtime.NumCPU()),\n\t\t\"--memory\", fmt.Sprintf(\"%d\", B2D.Memory),\n\t\t\"--rtcuseutc\", \"on\",\n\t\t\"--acpi\", \"on\",\n\t\t\"--ioapic\", \"on\",\n\t\t\"--hpet\", \"on\",\n\t\t\"--hwvirtex\", \"on\",\n\t\t\"--vtxvpid\", \"on\",\n\t\t\"--largepages\", \"on\",\n\t\t\"--nestedpaging\", \"on\",\n\t\t\"--firmware\", \"bios\",\n\t\t\"--bioslogofadein\", \"off\",\n\t\t\"--bioslogofadeout\", \"off\",\n\t\t\"--bioslogodisplaytime\", \"0\",\n\t\t\"--biosbootmenu\", \"disabled\",\n\t\t\"--boot1\", \"dvd\"); err != nil {\n\t\tlog.Fatal(\"failed to modify vm: %s\", err)\n\t}\n\n\tlog.Printf(\"Setting VM networking\")\n\tif err := vbm(\"modifyvm\", B2D.VM, \"--nic1\", \"nat\", \"--nictype1\", \"virtio\", \"--cableconnected1\", \"on\"); err != nil {\n\t\tlog.Fatalf(\"failed to modify vm: %s\", err)\n\t}\n\n\tif err := vbm(\"modifyvm\", B2D.VM,\n\t\t\"--natpf1\", fmt.Sprintf(\"ssh,tcp,127.0.0.1,%d,,22\", B2D.SSHPort),\n\t\t\"--natpf1\", fmt.Sprintf(\"docker,tcp,127.0.0.1,%d,,4243\", B2D.DockerPort)); err != nil {\n\t\tlog.Fatalf(\"failed to modify vm: %s\", err)\n\t}\n\tlog.Printf(\"Port forwarding [ssh]: host tcp://127.0.0.1:%d --> guest tcp://0.0.0.0:22\", B2D.SSHPort)\n\tlog.Printf(\"Port forwarding [docker]: host tcp://127.0.0.1:%d --> guest tcp://0.0.0.0:4243\", B2D.DockerPort)\n\n\tif _, err := os.Stat(B2D.ISO); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\tcmdDownload()\n\t\t} else {\n\t\t\tlog.Fatalf(\"failed to open ISO image: %s\", err)\n\t\t}\n\t}\n\n\tif _, err := os.Stat(B2D.Disk); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\terr := makeDiskImage(B2D.Disk, B2D.DiskSize)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatalf(\"failed to create disk image: %s\", err)\n\t\t\t}\n\t\t} else {\n\t\t\tlog.Fatalf(\"failed to open disk image: %s\", err)\n\t\t}\n\t}\n\n\tlog.Printf(\"Setting VM disks\")\n\tif err := vbm(\"storagectl\", B2D.VM, \"--name\", \"SATA\", \"--add\", \"sata\", \"--hostiocache\", \"on\"); err != nil {\n\t\tlog.Fatalf(\"failed to add storage controller: %s\", err)\n\t}\n\n\tif err := vbm(\"storageattach\", B2D.VM, \"--storagectl\", \"SATA\", \"--port\", \"0\", \"--device\", \"0\", \"--type\", \"dvddrive\", \"--medium\", B2D.ISO); err != nil {\n\t\tlog.Fatalf(\"failed to attach storage device: %s\", err)\n\t}\n\n\tif err := vbm(\"storageattach\", B2D.VM, \"--storagectl\", \"SATA\", \"--port\", \"1\", \"--device\", \"0\", \"--type\", \"hdd\", \"--medium\", B2D.Disk); err != nil {\n\t\tlog.Fatalf(\"failed to attach storage device: %s\", err)\n\t}\n\n\tlog.Printf(\"Done.\")\n\tlog.Printf(\"You can now type `%s up` and wait for the VM to start.\", os.Args[0])\n}", "func RegisteredVM(name, init string) (*otto.Otto, error) {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\n\tname = strings.ToLower(name)\n\tvm, ok := registry[name]\n\n\t// create new VM\n\tif !ok {\n\t\tvm = otto.New()\n\t\tif err := setConsole(vm, name); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif init != \"\" {\n\t\t\tif _, err := vm.Run(init); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\n\t\tif name != \"\" {\n\t\t\tregistry[name] = vm\n\t\t}\n\t}\n\n\treturn vm, nil\n}", "func NewVM(programName string, functionTable semantic.FunctionTable, classTable semantic.ClassTable, quads []quads.Quad) VirtualMachine {\n\tgmb := NewMemory(functionTable[programName].VarsSize, functionTable[programName].TempSize, functionTable[programName].ObjSize)\n\treturn VirtualMachine{\n\t\tglobalMemBlock: gmb,\n\t\tconstantMemBlock: MakeConstantMemory(),\n\t\tcurrentSelf: nil,\n\t\tquads: quads,\n\t\tprogramName: programName,\n\t\tfunctionTable: functionTable,\n\t\tclassTable: classTable,\n\t}\n}", "func HandleInstanceCreate(w rest.ResponseWriter, r *rest.Request) {\n\t// get ima\n\tima := Ima{}\n\terr := r.DecodeJsonPayload(&ima)\n\tif err != nil {\n\t\trest.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tif ima.Ima == \"\" {\n\t\trest.Error(w, \"ima required\", 400)\n\t\treturn\n\t}\n\tif ima.Mem == 0 {\n\t\trest.Error(w, \"memory required\", 400)\n\t\treturn\n\t}\n\tif ima.Cpu == 0 {\n\t\trest.Error(w, \"cpu required\", 400)\n\t\treturn\n\t}\n\n\t// start the instance\n\tos := getImaOs(ima.Ima)\n\tswitch os {\n\tcase \"freebsd\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tgo startFreeBSDVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tcase \"linux\":\n\t\t// clone ima to instance\n\t\tinstanceid := allocateInstanceId()\n\t\tcloneIma(ima.Ima, instanceid)\n\n\t\t// create network interface and bring up\n\t\ttap := allocateTap()\n\t\tif tap == \"\" {\n\t\t\treturn\n\t\t}\n\t\tsaveTap(tap, instanceid)\n\t\tbridge := findBridge()\n\t\taddTapToBridge(tap, bridge)\n\t\tbridgeUp(bridge)\n\n\t\t//nmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tsaveCpu(ima.Cpu, instanceid)\n\t\tsaveMem(ima.Mem, instanceid)\n\t\tbhyveDestroy(instanceid)\n\t\tnmdm := \"/dev/nmdm-\" + instanceid + \"-A\"\n\t\tgo startLinuxVM(nmdm, ima.Cpu, ima.Mem, tap, instanceid)\n\t\tw.WriteJson(&instanceid)\n\tdefault:\n\t\trest.Error(w, \"unknown OS\", 400)\n\t}\n}", "func (v *MachineVM) Start(name string, _ machine.StartOptions) error {\n\tvar (\n\t\tconn net.Conn\n\t\terr error\n\t\tqemuSocketConn net.Conn\n\t\twait time.Duration = time.Millisecond * 500\n\t)\n\n\tif err := v.startHostNetworking(); err != nil {\n\t\treturn errors.Errorf(\"unable to start host networking: %q\", err)\n\t}\n\n\trtPath, err := getRuntimeDir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If the temporary podman dir is not created, create it\n\tpodmanTempDir := filepath.Join(rtPath, \"podman\")\n\tif _, err := os.Stat(podmanTempDir); os.IsNotExist(err) {\n\t\tif mkdirErr := os.MkdirAll(podmanTempDir, 0755); mkdirErr != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tqemuSocketPath, _, err := v.getSocketandPid()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// If the qemusocketpath exists and the vm is off/down, we should rm\n\t// it before the dial as to avoid a segv\n\tif err := os.Remove(qemuSocketPath); err != nil && !errors.Is(err, os.ErrNotExist) {\n\t\tlogrus.Warn(err)\n\t}\n\tfor i := 0; i < 6; i++ {\n\t\tqemuSocketConn, err = net.Dial(\"unix\", qemuSocketPath)\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(wait)\n\t\twait++\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfd, err := qemuSocketConn.(*net.UnixConn).File()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tattr := new(os.ProcAttr)\n\tfiles := []*os.File{os.Stdin, os.Stdout, os.Stderr, fd}\n\tattr.Files = files\n\tlogrus.Debug(v.CmdLine)\n\tcmd := v.CmdLine\n\n\t// Disable graphic window when not in debug mode\n\t// Done in start, so we're not suck with the debug level we used on init\n\tif logrus.GetLevel() != logrus.DebugLevel {\n\t\tcmd = append(cmd, \"-display\", \"none\")\n\t}\n\n\t_, err = os.StartProcess(v.CmdLine[0], cmd, attr)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfmt.Println(\"Waiting for VM ...\")\n\tsocketPath, err := getRuntimeDir()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// The socket is not made until the qemu process is running so here\n\t// we do a backoff waiting for it. Once we have a conn, we break and\n\t// then wait to read it.\n\tfor i := 0; i < 6; i++ {\n\t\tconn, err = net.Dial(\"unix\", filepath.Join(socketPath, \"podman\", v.Name+\"_ready.sock\"))\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(wait)\n\t\twait++\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = bufio.NewReader(conn).ReadString('\\n')\n\treturn err\n}", "func New(tagName ...string) *VM {\n\tif len(tagName) == 0 {\n\t\ttagName = append(tagName, \"\")\n\t}\n\treturn &VM{\n\t\ttagName: tagName[0],\n\t\tstructJar: make(map[int32]*structVM, 256),\n\t}\n}", "func (conf *Configuration) VirtualMachine(name string) (*VirtualMachine, error) {\n\tctx := context.NewContext(conf.Timeout)\n\tdefer ctx.Cancel()\n\n\treturn conf.VirtualMachineWithContext(ctx, name)\n}", "func (self *Provisioner) BindVM(vms []tumblebug.VM) ([]*model.Node, error) {\n\n\tnodes := []*model.Node{}\n\tfor _, vm := range vms {\n\n\t\t// validate created vm\n\t\tif vm.Status == tumblebug.VMSTATUS_FAILED {\n\t\t\tstatus := app.Status{}\n\t\t\tif err := json.Unmarshal([]byte(vm.SystemMessage), &status); err != nil {\n\t\t\t\tstatus.Message = vm.SystemMessage\n\t\t\t}\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Failed to create a vm (status=%s, cause='%s')\", vm.Status, status.Message))\n\t\t} else if vm.PublicIP == \"\" && self.Cluster.ServiceType == app.ST_MULTI {\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Failed to create a vm (status=%s, cause='unbounded public-ip')\", vm.Status))\n\t\t} else if vm.PrivateIP == \"\" && self.Cluster.ServiceType == app.ST_SINGLE {\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Failed to create a vm (status=%s, cause='unbounded private-ip')\", vm.Status))\n\t\t}\n\n\t\tvar machine *Machine\n\n\t\tif self.leader.Name == vm.Name {\n\t\t\tmachine = self.leader.Machine\n\t\t} else {\n\t\t\t_, exists := self.ControlPlaneMachines[vm.Name]\n\t\t\tif exists {\n\t\t\t\tmachine = self.ControlPlaneMachines[vm.Name].Machine\n\t\t\t} else {\n\t\t\t\t_, exists = self.WorkerNodeMachines[vm.Name]\n\t\t\t\tif exists {\n\t\t\t\t\tmachine = self.WorkerNodeMachines[vm.Name].Machine\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif machine != nil {\n\t\t\tmachine.PublicIP = vm.PublicIP\n\t\t\tmachine.PrivateIP = vm.PrivateIP\n\t\t\tmachine.Username = lang.NVL(vm.UserAccount, tumblebug.VM_USER_ACCOUNT)\n\t\t\tmachine.Region = lang.NVL(vm.Region.Region, machine.Region) // region, zone 공백인 경우가 간혹 있음\n\t\t\tmachine.Zone = lang.NVL(vm.Region.Zone, machine.Zone)\n\t\t\tmachine.Spec = vm.CspViewVmDetail.VMSpecName\n\t\t\tnodes = append(nodes, machine.NewNode())\n\t\t\tnameInCsp, err := vm.GetNameInCsp()\n\t\t\tif err != nil {\n\t\t\t\tlogger.Infof(\"[%s.%s.%s] nameInCsp is empty\", vm.Namespace, vm.McisName, vm.Name)\n\t\t\t}\n\t\t\tmachine.NameInCsp = lang.NVL(nameInCsp, \"\")\n\t\t} else {\n\t\t\treturn nil, errors.New(fmt.Sprintf(\"Can't be found node by name '%s'\", vm.Name))\n\t\t}\n\t}\n\n\treturn nodes, nil\n}", "func (e *Engine) SSHVM(namespace, id, user, key string, term *termutil.Terminal) error {\n\tcontainer, err := e.docker.Inspect(id)\n\tif err != nil {\n\t\tfullName := internal.GenerateContainerName(namespace, id)\n\n\t\tcontainer, err = e.docker.Inspect(fullName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tip := container.NetworkSettings.IPAddress\n\tkeyPath := homedir.ExpandPath(key)\n\n\tprivateKey, err := ioutil.ReadFile(keyPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsigner, err := ssh.ParsePrivateKey(privateKey)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconfig := ssh.ClientConfig{\n\t\tUser: user,\n\t\tAuth: []ssh.AuthMethod{\n\t\t\tssh.PublicKeys(signer),\n\t\t},\n\t\tHostKeyCallback: ssh.InsecureIgnoreHostKey(), // nolint: gosec\n\t}\n\tconfig.SetDefaults()\n\n\tconn, err := ssh.Dial(\"tcp\", ip+\":22\", &config)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer conn.Close()\n\n\tsess, err := conn.NewSession()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer sess.Close()\n\n\tsess.Stdin = term.In()\n\tsess.Stdout = term.Out()\n\tsess.Stderr = term.Err()\n\n\tsz, err := term.GetWinsize()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = term.MakeRaw()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer handleError(term.Restore())\n\n\terr = sess.RequestPty(os.Getenv(\"TERM\"), int(sz.Height), int(sz.Width), nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = sess.Shell()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// If our terminal window changes, signal the ssh connection\n\tstopch := make(chan struct{})\n\tdefer close(stopch)\n\n\tgo func() {\n\t\tsigch := make(chan os.Signal, 1)\n\t\tsignal.Notify(sigch, syscall.SIGWINCH)\n\n\t\tdefer signal.Stop(sigch)\n\t\tdefer close(sigch)\n\touter:\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-sigch:\n\t\t\t\tsz, err := term.GetWinsize()\n\t\t\t\tif err == nil {\n\t\t\t\t\thandleError(sess.WindowChange(int(sz.Height), int(sz.Width)))\n\t\t\t\t}\n\t\t\tcase <-stopch:\n\t\t\t\tbreak outer\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn sess.Wait()\n}", "func createVM(c context.Context, v *crimson.VM) (*crimson.VM, error) {\n\tif err := validateVMForCreation(v); err != nil {\n\t\treturn nil, err\n\t}\n\tip, _ := common.ParseIPv4(v.Ipv4)\n\ttx, err := database.Begin(c)\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to begin transaction\").Err()\n\t}\n\tdefer tx.MaybeRollback(c)\n\n\thostnameID, err := model.AssignHostnameAndIP(c, tx, v.Name, ip)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// vms.hostname_id, vms.physical_host_id, and vms.os_id are NOT NULL as above.\n\t_, err = tx.ExecContext(c, `\n\t\tINSERT INTO vms (hostname_id, physical_host_id, os_id, description, deployment_ticket, state)\n\t\tVALUES (\n\t\t\t?,\n\t\t\t(SELECT p.id FROM physical_hosts p, hostnames h WHERE p.hostname_id = h.id AND h.name = ?),\n\t\t\t(SELECT id FROM oses WHERE name = ?),\n\t\t\t?,\n\t\t\t?,\n\t\t\t?\n\t\t)\n\t`, hostnameID, v.Host, v.Os, v.Description, v.DeploymentTicket, v.State)\n\tif err != nil {\n\t\tswitch e, ok := err.(*mysql.MySQLError); {\n\t\tcase !ok:\n\t\t\t// Type assertion failed.\n\t\tcase e.Number == mysqlerr.ER_BAD_NULL_ERROR && strings.Contains(e.Message, \"'physical_host_id'\"):\n\t\t\t// e.g. \"Error 1048: Column 'physical_host_id' cannot be null\".\n\t\t\treturn nil, status.Errorf(codes.NotFound, \"physical host %q does not exist\", v.Host)\n\t\tcase e.Number == mysqlerr.ER_BAD_NULL_ERROR && strings.Contains(e.Message, \"'os_id'\"):\n\t\t\t// e.g. \"Error 1048: Column 'os_id' cannot be null\".\n\t\t\treturn nil, status.Errorf(codes.NotFound, \"operating system %q does not exist\", v.Os)\n\t\t}\n\t\treturn nil, errors.Annotate(err, \"failed to create VM\").Err()\n\t}\n\n\tvms, err := listVMs(c, tx, &crimson.ListVMsRequest{\n\t\tNames: []string{v.Name},\n\t})\n\tif err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to fetch created VM\").Err()\n\t}\n\n\tif err := tx.Commit(); err != nil {\n\t\treturn nil, errors.Annotate(err, \"failed to commit transaction\").Err()\n\t}\n\treturn vms[0], nil\n}", "func NewVM() *VM {\n\treturn &VM{newVMState(), newInterpreter()}\n}", "func createVM(c context.Context, payload proto.Message) error {\n\ttask, ok := payload.(*tasks.CreateVM)\n\tswitch {\n\tcase !ok:\n\t\treturn errors.Reason(\"unexpected payload type %T\", payload).Err()\n\tcase task.GetId() == \"\":\n\t\treturn errors.Reason(\"ID is required\").Err()\n\tcase task.GetConfig() == \"\":\n\t\treturn errors.Reason(\"config is required\").Err()\n\t}\n\tvm := &model.VM{\n\t\tID: task.Id,\n\t\tConfig: task.Config,\n\t\tConfigured: clock.Now(c).Unix(),\n\t\tHostname: fmt.Sprintf(\"%s-%d-%s\", task.Prefix, task.Index, getSuffix(c)),\n\t\tIndex: task.Index,\n\t\tLifetime: task.Lifetime,\n\t\tPrefix: task.Prefix,\n\t\tRevision: task.Revision,\n\t\tSwarming: task.Swarming,\n\t\tTimeout: task.Timeout,\n\t}\n\tif task.Attributes != nil {\n\t\tvm.Attributes = *task.Attributes\n\t\t// TODO(crbug/942301): Auto-select zone if zone is unspecified.\n\t\tvm.Attributes.SetZone(vm.Attributes.GetZone())\n\t\tvm.IndexAttributes()\n\t}\n\t// createVM is called repeatedly, so do a fast check outside the transaction.\n\t// In most cases, this will skip the more expensive transactional check.\n\tswitch err := datastore.Get(c, vm); {\n\tcase err == datastore.ErrNoSuchEntity:\n\tcase err != nil:\n\t\treturn errors.Annotate(err, \"failed to fetch VM\").Err()\n\tdefault:\n\t\treturn nil\n\t}\n\treturn datastore.RunInTransaction(c, func(c context.Context) error {\n\t\tswitch err := datastore.Get(c, vm); {\n\t\tcase err == datastore.ErrNoSuchEntity:\n\t\tcase err != nil:\n\t\t\treturn errors.Annotate(err, \"failed to fetch VM\").Err()\n\t\tdefault:\n\t\t\treturn nil\n\t\t}\n\t\tif err := datastore.Put(c, vm); err != nil {\n\t\t\treturn errors.Annotate(err, \"failed to store VM\").Err()\n\t\t}\n\t\treturn nil\n\t}, nil)\n}", "func (ic *Context) SpawnVM() *vm.VM {\n\tv := vm.NewWithTrigger(ic.Trigger)\n\tv.GasLimit = -1\n\tv.SyscallHandler = ic.SyscallHandler\n\tic.VM = v\n\treturn v\n}", "func (h *Host) NewVM(name string) (*VM, error) {\n\tfinder := h.Finder()\n\n\tvm, err := finder.VirtualMachine(h.Ctx(), name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn h.makeVM(name, vm), nil\n}", "func (cce *CCEClient) Create(ctx context.Context, cluster *clusterv1.Cluster, machine *clusterv1.Machine) error {\n\tglog.V(4).Infof(\"Create machine: %+v\", machine.Name)\n\tinstance, err := cce.instanceIfExists(cluster, machine)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif instance != nil {\n\t\tglog.Infof(\"Skipped creating a VM that already exists, instanceID %s\", instance.InstanceID)\n\t}\n\n\tmachineCfg, err := machineProviderFromProviderConfig(machine.Spec.ProviderSpec)\n\tif err != nil {\n\t\tglog.Errorf(\"parse machine config err: %s\", err.Error())\n\t\treturn err\n\t}\n\tglog.V(4).Infof(\"machine config: %+v\", machineCfg)\n\n\tbccArgs := &bcc.CreateInstanceArgs{\n\t\tName: machine.Name,\n\t\tImageID: machineCfg.ImageID, // ubuntu-16.04-amd64\n\t\tBilling: billing.Billing{\n\t\t\tPaymentTiming: \"Postpaid\",\n\t\t},\n\t\tCPUCount: machineCfg.CPUCount,\n\t\tMemoryCapacityInGB: machineCfg.MemoryCapacityInGB,\n\t\tAdminPass: machineCfg.AdminPass,\n\t\tPurchaseCount: 1,\n\t\tInstanceType: \"N3\", // Normal 3\n\t\tNetworkCapacityInMbps: 1, //EIP bandwidth\n\t}\n\n\t// TODO support different regions\n\tinstanceIDs, err := cce.computeService.Bcc().CreateInstances(bccArgs, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(instanceIDs) != 1 {\n\t\treturn fmt.Errorf(\"CreateVMError\")\n\t}\n\n\tglog.Infof(\"Created a new VM, instanceID %s\", instanceIDs[0])\n\tif machine.ObjectMeta.Annotations == nil {\n\t\tmachine.ObjectMeta.Annotations = map[string]string{}\n\t}\n\tif cluster.ObjectMeta.Annotations == nil {\n\t\tcluster.ObjectMeta.Annotations = map[string]string{}\n\t}\n\tmachine.ObjectMeta.Annotations[TagInstanceID] = instanceIDs[0]\n\tmachine.ObjectMeta.Annotations[TagInstanceStatus] = \"Created\"\n\tmachine.ObjectMeta.Annotations[TagInstanceAdminPass] = machineCfg.AdminPass\n\tmachine.ObjectMeta.Annotations[TagKubeletVersion] = machine.Spec.Versions.Kubelet\n\n\ttoken, err := cce.getKubeadmToken()\n\tif err != nil {\n\t\tglog.Errorf(\"getKubeadmToken err: %+v\", err)\n\t\treturn err\n\t}\n\n\tif machineCfg.Role == \"master\" {\n\t\tcluster.ObjectMeta.Annotations[TagMasterInstanceID] = instanceIDs[0]\n\t\tcluster.ObjectMeta.Annotations[TagClusterToken] = token\n\t\tmachine.ObjectMeta.Annotations[TagInstanceRole] = \"master\"\n\t} else {\n\t\tmachine.ObjectMeta.Annotations[TagInstanceRole] = \"node\"\n\t}\n\n\tglog.V(4).Infof(\"new machine: %+v, annotation %+v\", machine.Name, machine.Annotations)\n\tcce.client.Update(context.Background(), cluster)\n\tcce.client.Update(context.Background(), machine)\n\n\t// TODO rewrite\n\tgo cce.postCreate(ctx, cluster, machine)\n\treturn nil\n}", "func Run(name string) (*Motto, otto.Value, error) {\n vm := New()\n v, err := vm.Run(name)\n\n return vm, v, err\n}", "func testInstance() Instance {\n\treturn Instance{\n\t\tCreated: true,\n\t\tImageID: \"ami-7172b611\",\n\t\tInstanceID: \"i-1234567890abcdef0\",\n\t\tInstanceType: \"t2.nano\",\n\t\tSubnetID: \"subnet-1234567890abcdef0\",\n\t\tKeyPairName: \"bastion-test\",\n\t\tSecurityGroupID: \"sg-1234567890abcdef0\",\n\t\tPublicIPAddress: \"8.8.8.8\",\n\t\tPrivateIPAddress: \"10.0.0.1\",\n\t\tSSHUser: \"ec2-user\",\n\t}\n}", "func create_virtual_machine(){\n parameters := make([]string, 0)\n parameters = append(parameters, \"apply\")\n parameters = append(parameters, \"--auto-approve\")\n\n execute_command(\"terraform\", parameters)\n}", "func (p *provider) Create(machine *v1alpha1.Machine, providerData *cloudprovidertypes.ProviderData, userdata string) (instance.Instance, error) {\n\tconfig, _, err := p.getConfig(machine.Spec.ProviderSpec)\n\tif err != nil {\n\t\treturn nil, newError(common.InvalidConfigurationMachineError, \"failed to parse MachineSpec: %v\", err)\n\t}\n\n\tapiClient := getClient(config.Token)\n\tctx, cancel := context.WithTimeout(context.Background(), anxtypes.CreateRequestTimeout)\n\tdefer cancel()\n\n\tstatus, err := getStatus(machine.Status.ProviderStatus)\n\tif err != nil {\n\t\treturn nil, newError(common.InvalidConfigurationMachineError, \"failed to get machine status: %v\", err)\n\t}\n\n\tif status.ProvisioningID == \"\" {\n\t\tips, err := apiClient.VSphere().Provisioning().IPs().GetFree(ctx, config.LocationID, config.VlanID)\n\t\tif err != nil {\n\t\t\treturn nil, newError(common.InvalidConfigurationMachineError, \"failed to get ip pool: %v\", err)\n\t\t}\n\t\tif len(ips) < 1 {\n\t\t\treturn nil, newError(common.InsufficientResourcesMachineError, \"no ip address is available for this machine\")\n\t\t}\n\n\t\tipID := ips[0].Identifier\n\t\tnetworkInterfaces := []anxvm.Network{{\n\t\t\tNICType: anxtypes.VmxNet3NIC,\n\t\t\tIPs: []string{ipID},\n\t\t\tVLAN: config.VlanID,\n\t\t}}\n\n\t\tvm := apiClient.VSphere().Provisioning().VM().NewDefinition(\n\t\t\tconfig.LocationID,\n\t\t\t\"templates\",\n\t\t\tconfig.TemplateID,\n\t\t\tmachine.ObjectMeta.Name,\n\t\t\tconfig.CPUs,\n\t\t\tconfig.Memory,\n\t\t\tconfig.DiskSize,\n\t\t\tnetworkInterfaces,\n\t\t)\n\n\t\tvm.Script = base64.StdEncoding.EncodeToString(\n\t\t\t[]byte(fmt.Sprintf(\"anexia: true\\n\\n%s\", userdata)),\n\t\t)\n\n\t\tsshKey, err := ssh.NewKey()\n\t\tif err != nil {\n\t\t\treturn nil, newError(common.CreateMachineError, \"failed to generate ssh key: %v\", err)\n\t\t}\n\t\tvm.SSH = sshKey.PublicKey\n\n\t\tprovisionResponse, err := apiClient.VSphere().Provisioning().VM().Provision(ctx, vm)\n\t\tif err != nil {\n\t\t\treturn nil, newError(common.CreateMachineError, \"instance provisioning failed: %v\", err)\n\t\t}\n\n\t\tstatus.ProvisioningID = provisionResponse.Identifier\n\t\tstatus.IPAllocationID = ipID\n\t\tif err := updateStatus(machine, status, providerData.Update); err != nil {\n\t\t\treturn nil, newError(common.UpdateMachineError, \"machine status update failed: %v\", err)\n\t\t}\n\t}\n\n\tinstanceID, err := apiClient.VSphere().Provisioning().Progress().AwaitCompletion(ctx, status.ProvisioningID)\n\tif err != nil {\n\t\treturn nil, newError(common.CreateMachineError, \"instance provisioning failed: %v\", err)\n\t}\n\n\tstatus.InstanceID = instanceID\n\tif err := updateStatus(machine, status, providerData.Update); err != nil {\n\t\treturn nil, newError(common.UpdateMachineError, \"machine status update failed: %v\", err)\n\t}\n\n\treturn p.Get(machine, providerData)\n}", "func (c *Client) queryInstance(instanceIP string) (*cvm.Instance, error) {\n\treq := cvm.NewDescribeInstancesRequest()\n\treq.Filters = []*cvm.Filter{\n\t\t{\n\t\t\tName: common.StringPtr(\"private-ip-address\"),\n\t\t\tValues: common.StringPtrs([]string{instanceIP}),\n\t\t},\n\t}\n\n\tblog.V(2).Infof(\"DescribeInstances req: %s\", req.ToJsonString())\n\n\tresp, err := c.cvmClient.DescribeInstances(req)\n\tif err != nil {\n\t\tblog.Errorf(\"DescribeInstances failed, err %s\", err.Error())\n\t\treturn nil, fmt.Errorf(\"DescribeInstances failed, err %s\", err.Error())\n\t}\n\n\tif len(resp.Response.InstanceSet) == 0 {\n\t\tblog.Errorf(\"cvm with %s not found\", err.Error())\n\t\treturn nil, fmt.Errorf(\"cvm with %s not found\", err.Error())\n\t}\n\n\tins := resp.Response.InstanceSet[0]\n\tif ins == nil {\n\t\treturn nil, fmt.Errorf(\"no vm info\")\n\t}\n\tif ins.Placement == nil || ins.VirtualPrivateCloud == nil {\n\t\treturn nil, fmt.Errorf(\"vm info lost Placement or VirtualPrivateCloud\")\n\t}\n\tif ins.Placement.Zone == nil || ins.VirtualPrivateCloud.VpcId == nil ||\n\t\tins.VirtualPrivateCloud.SubnetId == nil {\n\t\treturn nil, fmt.Errorf(\"vm info lost Zone or VpcId or SubentId\")\n\t}\n\n\treturn ins, nil\n}", "func NewRandomVirtualMachine(vmi *v1.VirtualMachineInstance, running bool) *v1.VirtualMachine {\n\tname := vmi.Name\n\tnamespace := vmi.Namespace\n\tvm := &v1.VirtualMachine{\n\t\tObjectMeta: v12.ObjectMeta{\n\t\t\tName: name,\n\t\t\tNamespace: namespace,\n\t\t},\n\t\tSpec: v1.VirtualMachineSpec{\n\t\t\tRunning: running,\n\t\t\tTemplate: &v1.VirtualMachineInstanceTemplateSpec{\n\t\t\t\tObjectMeta: v12.ObjectMeta{\n\t\t\t\t\tLabels: map[string]string{\"name\": name},\n\t\t\t\t\tName: name,\n\t\t\t\t\tNamespace: namespace,\n\t\t\t\t},\n\t\t\t\tSpec: vmi.Spec,\n\t\t\t},\n\t\t},\n\t}\n\treturn vm\n}", "func initializeVM(jobName string, args []types.JobArgs) *otto.Otto {\n\tcontext := &types.JobInstanceContext{\n\t\tJobName: jobName,\n\t}\n\tvm := NewVMWithPlugins(context)\n\t// Setting job arguments in VM\n\tfor _, arg := range args {\n\t\tvm.Set(arg.Name, arg.Value)\n\t}\n\treturn vm\n}", "func (c *Client) CreateInstance(name, zone, machineType, ipxeURL string) error {\n\treq := &packngo.DeviceCreateRequest{\n\t\tHostname: name,\n\t\tPlan: machineType,\n\t\tProjectID: c.projectID,\n\t\tFacility: []string{zone},\n\t\tIPXEScriptURL: ipxeURL,\n\t\tOS: \"custom_ipxe\",\n\t\tDescription: \"eden test vm\",\n\t\tBillingCycle: \"hourly\",\n\t}\n\t_, _, err := c.client.Devices.Create(req)\n\treturn err\n}", "func NewVMWithSingletons(ctx context.Context, t testing.TB, bs ipldcbor.IpldBlockstore) *VM {\n\tlookup := map[cid.Cid]runtime.VMActor{}\n\tfor _, ba := range exported.BuiltinActors() {\n\t\tlookup[ba.Code()] = ba\n\t}\n\n\tstore := adt.WrapBlockStore(ctx, bs)\n\tvm := NewVM(ctx, lookup, store)\n\n\tsystemState, err := system.ConstructState(store)\n\trequire.NoError(t, err)\n\tinitializeActor(ctx, t, vm, systemState, builtin.SystemActorCodeID, builtin.SystemActorAddr, big.Zero())\n\n\tinitState, err := initactor.ConstructState(store, \"scenarios\")\n\trequire.NoError(t, err)\n\tinitializeActor(ctx, t, vm, initState, builtin.InitActorCodeID, builtin.InitActorAddr, big.Zero())\n\n\trewardState := reward.ConstructState(abi.NewStoragePower(0))\n\tinitializeActor(ctx, t, vm, rewardState, builtin.RewardActorCodeID, builtin.RewardActorAddr, reward.StorageMiningAllocationCheck)\n\n\tcronState := cron.ConstructState(cron.BuiltInEntries())\n\tinitializeActor(ctx, t, vm, cronState, builtin.CronActorCodeID, builtin.CronActorAddr, big.Zero())\n\n\tpowerState, err := power.ConstructState(store)\n\trequire.NoError(t, err)\n\tinitializeActor(ctx, t, vm, powerState, builtin.StoragePowerActorCodeID, builtin.StoragePowerActorAddr, big.Zero())\n\n\tmarketState, err := market.ConstructState(store)\n\trequire.NoError(t, err)\n\tinitializeActor(ctx, t, vm, marketState, builtin.StorageMarketActorCodeID, builtin.StorageMarketActorAddr, big.Zero())\n\n\t// this will need to be replaced with the address of a multisig actor for the verified registry to be tested accurately\n\tinitializeActor(ctx, t, vm, &account.State{Address: VerifregRoot}, builtin.AccountActorCodeID, VerifregRoot, big.Zero())\n\tvrState, err := verifreg.ConstructState(store, VerifregRoot)\n\trequire.NoError(t, err)\n\tinitializeActor(ctx, t, vm, vrState, builtin.VerifiedRegistryActorCodeID, builtin.VerifiedRegistryActorAddr, big.Zero())\n\n\t// burnt funds\n\tinitializeActor(ctx, t, vm, &account.State{Address: builtin.BurntFundsActorAddr}, builtin.AccountActorCodeID, builtin.BurntFundsActorAddr, big.Zero())\n\n\t_, err = vm.checkpoint()\n\trequire.NoError(t, err)\n\n\treturn vm\n}", "func (r *VmwareMapper) MapVM(targetVmName *string, vmSpec *kubevirtv1.VirtualMachine) (*kubevirtv1.VirtualMachine, error) {\n\tif vmSpec.Spec.Template == nil {\n\t\tvmSpec.Spec.Template = &kubevirtv1.VirtualMachineInstanceTemplateSpec{}\n\t}\n\t// Map annotations\n\tvmSpec.ObjectMeta.Annotations = r.mapAnnotations()\n\t// Map labels like vm tags\n\tvmSpec.ObjectMeta.Labels = r.mapLabels(vmSpec.ObjectMeta.Labels)\n\t// Set Namespace\n\tvmSpec.ObjectMeta.Namespace = r.namespace\n\n\t// Map name\n\tif targetVmName == nil {\n\t\tvmSpec.ObjectMeta.GenerateName = vmNamePrefix\n\t} else {\n\t\tvmSpec.ObjectMeta.Name = *targetVmName\n\t}\n\n\tif vmSpec.Spec.Template == nil {\n\t\tvmSpec.Spec.Template = &kubevirtv1.VirtualMachineInstanceTemplateSpec{}\n\t}\n\n\ttrue_ := true\n\tfalse_ := false\n\tvmSpec.Spec.Running = &false_\n\n\t// Map hostname\n\thostname, _ := utils.NormalizeName(r.vmProperties.Guest.HostName)\n\t// if this is a FQDN, split off the first subdomain and use it as the hostname.\n\tnameParts := strings.Split(hostname, \".\")\n\tvmSpec.Spec.Template.Spec.Hostname = nameParts[0]\n\n\tvmSpec.Spec.Template.Spec.Domain.Machine = kubevirtv1.Machine{Type: q35}\n\tvmSpec.Spec.Template.Spec.Domain.CPU = r.mapCPUTopology()\n\tvmSpec.Spec.Template.Spec.Domain.Firmware = r.mapFirmware()\n\tvmSpec.Spec.Template.Spec.Domain.Features = r.mapFeatures()\n\treservations, err := r.mapResourceReservations()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvmSpec.Spec.Template.Spec.Domain.Resources = reservations\n\n\t// Map clock\n\tvmSpec.Spec.Template.Spec.Domain.Clock = r.mapClock(r.hostProperties)\n\n\t// remove any default networks/interfaces from the template\n\tvmSpec.Spec.Template.Spec.Networks = []kubevirtv1.Network{}\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Interfaces = []kubevirtv1.Interface{}\n\n\tif r.mappings != nil && r.mappings.NetworkMappings != nil {\n\t\t// Map networks\n\t\tvmSpec.Spec.Template.Spec.Networks, err = r.mapNetworks()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnetworkToType := r.mapNetworksToTypes(vmSpec.Spec.Template.Spec.Networks)\n\t\tvmSpec.Spec.Template.Spec.Domain.Devices.Interfaces, err = r.mapNetworkInterfaces(networkToType)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// if there are no interfaces defined, force NetworkInterfaceMultiQueue to false\n\t// https://github.com/kubevirt/common-templates/issues/186\n\tif len(vmSpec.Spec.Template.Spec.Domain.Devices.Interfaces) > 0 {\n\t\tvmSpec.Spec.Template.Spec.Domain.Devices.NetworkInterfaceMultiQueue = &true_\n\t} else {\n\t\tvmSpec.Spec.Template.Spec.Domain.Devices.NetworkInterfaceMultiQueue = &false_\n\t}\n\n\tos, _ := r.osFinder.FindOperatingSystem(r.vmProperties)\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Inputs = r.mapInputDevice(os)\n\tvmSpec.Spec.Template.Spec.Domain.Devices.Disks = []kubevirtv1.Disk{}\n\treturn vmSpec, nil\n}", "func getInstance(params martini.Params, r render.Render) {\n\tname := params[\"name\"]\n\n\tif !instanceExists(name) {\n\t\tfmt.Println(\"Instance with specified name does not exist in provision table\")\n\t\tr.Text(400, \"Bad Request\")\n\t\treturn\n\t}\n\n\tdbinfo, err := getDBInfo(name)\n\tif err != nil {\n\t\toutput500Error(r, err)\n\t\treturn\n\t}\n\tr.JSON(200, map[string]string{\"NEPTUNE_DATABASE_URL\": dbinfo.Endpoint, \"NEPTUNE_ACCESS_KEY\": dbinfo.AccessKeyID, \"NEPTUNE_SECRET_KEY\": dbinfo.SecretAccessKey, \"NEPTUNE_REGION\": os.Getenv(\"REGION\")})\n}", "func (m *Manager) ListVMs() (vmInfo []core.VMInfo, err error) {\n\n\tinstances, err := m.api.ListInstances(context.Background(), nil)\n\tif err != nil {\n\t\treturn vmInfo, err\n\t}\n\n\tfor _, instance := range instances {\n\t\tif len(instance.Tags) > 0 {\n\t\t\tfor _, tag := range instance.Tags {\n\t\t\t\tif strings.Compare(tag, \"eezhee\") == 0 {\n\t\t\t\t\t// we created this VM\n\t\t\t\t\tinfo, _ := convertVMInfoToGenericFormat(instance)\n\t\t\t\t\tvmInfo = append(vmInfo, info)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tfmt.Println(instance.ID, \" \", instance.Label, \" \", instance.IPv4)\n\t\t// check tag. did we create it\n\t\t// if so, convert to generic format\n\t\t// add to results\n\t}\n\n\treturn vmInfo, nil\n}", "func startHost(api libmachine.API, mc config.ClusterConfig, n config.Node) (*host.Host, bool) {\n\texists, err := api.Exists(mc.Name)\n\tif err != nil {\n\t\texit.WithError(\"Failed to check if machine exists\", err)\n\t}\n\n\thost, err := machine.StartHost(api, mc, n)\n\tif err != nil {\n\t\texit.WithError(\"Unable to start VM. Please investigate and run 'minikube delete' if possible\", err)\n\t}\n\treturn host, exists\n}", "func (f *fakeEC2BuildletClient) StartNewVM(ctx context.Context, buildEnv *buildenv.Environment, hconf *dashboard.HostConfig, vmName, hostType string, opts *buildlet.VMOpts) (buildlet.Client, error) {\n\t// check required params\n\tif opts == nil || opts.TLS.IsZero() {\n\t\treturn nil, errors.New(\"TLS keypair is not set\")\n\t}\n\tif buildEnv == nil {\n\t\treturn nil, errors.New(\"invalid build enviornment\")\n\t}\n\tif hconf == nil {\n\t\treturn nil, errors.New(\"invalid host configuration\")\n\t}\n\tif vmName == \"\" || hostType == \"\" {\n\t\treturn nil, fmt.Errorf(\"invalid vmName: %q and hostType: %q\", vmName, hostType)\n\t}\n\tif opts.DeleteIn == 0 {\n\t\t// Note: This implements a short default in the rare case the caller doesn't care.\n\t\topts.DeleteIn = 30 * time.Minute\n\t}\n\tif !f.createVMRequestSuccess {\n\t\treturn nil, fmt.Errorf(\"unable to create instance %s: creation disabled\", vmName)\n\t}\n\tcondRun := func(fn func()) {\n\t\tif fn != nil {\n\t\t\tfn()\n\t\t}\n\t}\n\tcondRun(opts.OnInstanceRequested)\n\tif !f.VMCreated {\n\t\treturn nil, errors.New(\"error waiting for instance to exist: vm existence disabled\")\n\t}\n\n\tcondRun(opts.OnInstanceCreated)\n\n\tif !f.buildletCreated {\n\t\treturn nil, errors.New(\"error waiting for buildlet: buildlet creation disabled\")\n\t}\n\n\tif opts.OnGotEC2InstanceInfo != nil {\n\t\topts.OnGotEC2InstanceInfo(&cloud.Instance{\n\t\t\tCPUCount: 4,\n\t\t\tCreatedAt: time.Time{},\n\t\t\tDescription: \"sample vm\",\n\t\t\tID: \"id-\" + instanceName(\"random\", 4),\n\t\t\tIPAddressExternal: \"127.0.0.1\",\n\t\t\tIPAddressInternal: \"127.0.0.1\",\n\t\t\tImageID: \"image-x\",\n\t\t\tName: vmName,\n\t\t\tSSHKeyID: \"key-15\",\n\t\t\tSecurityGroups: nil,\n\t\t\tState: \"running\",\n\t\t\tTags: map[string]string{\n\t\t\t\t\"foo\": \"bar\",\n\t\t\t},\n\t\t\tType: \"yy.large\",\n\t\t\tZone: \"zone-a\",\n\t\t})\n\t}\n\treturn &buildlet.FakeClient{}, nil\n}", "func runMachineFromVmCode() {\n\tfilename := \"\"\n\tif len(os.Args) == 1 {\n\t\tfilename = \"./sample/vm/func.vm\"\n\t} else {\n\t\tfilename = os.Args[1]\n\t}\n\n\t_, err := os.Stat(filename)\n\tif err != nil {\n\t\tif len(os.Args) == 1 {\n\t\t\tfmt.Print(\"[ERROR] filename is not given\")\n\t\t} else {\n\t\t\tfmt.Printf(\"[ERROR] given filename(%s) is not confirm. please confirm you have correct path\", filename)\n\t\t}\n\t}\n\n\tvirtualMachine := vm.New()\n\tasmProgram := virtualMachine.FromFile(filename)\n\n\tassm := assembler.New()\n\tbinaryProgram := assm.FromString(asmProgram)\n\n\tmachine := computer.NewComputer()\n\tcomputer.LoadPresetBinaryProgram(machine)(binaryProgram)\n\tmachine.Run()\n}", "func (r *AzureMachineReconciler) findVM(scope *scope.MachineScope, ams *azureMachineService) (*infrav1.VM, error) {\n\tvar vm *infrav1.VM\n\n\t// If the ProviderID is populated, describe the VM using its name and resource group name.\n\tvm, err := ams.VMIfExists(scope.GetVMID())\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"failed to query AzureMachine VM\")\n\t}\n\n\treturn vm, nil\n}", "func NewVirtualMachine(ctx *pulumi.Context,\n\tname string, args *VirtualMachineArgs, opts ...pulumi.ResourceOption) (*VirtualMachine, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.LabName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'LabName'\")\n\t}\n\tif args.ResourceGroupName == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'ResourceGroupName'\")\n\t}\n\tif args.AllowClaim == nil {\n\t\targs.AllowClaim = pulumi.BoolPtr(false)\n\t}\n\tif args.DisallowPublicIpAddress == nil {\n\t\targs.DisallowPublicIpAddress = pulumi.BoolPtr(false)\n\t}\n\tif args.OwnerObjectId == nil {\n\t\targs.OwnerObjectId = pulumi.StringPtr(\"dynamicValue\")\n\t}\n\tif args.StorageType == nil {\n\t\targs.StorageType = pulumi.StringPtr(\"labStorageType\")\n\t}\n\taliases := pulumi.Aliases([]pulumi.Alias{\n\t\t{\n\t\t\tType: pulumi.String(\"azure-nextgen:devtestlab/v20180915:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-native:devtestlab:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-nextgen:devtestlab:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-native:devtestlab/v20150521preview:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-nextgen:devtestlab/v20150521preview:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-native:devtestlab/v20160515:VirtualMachine\"),\n\t\t},\n\t\t{\n\t\t\tType: pulumi.String(\"azure-nextgen:devtestlab/v20160515:VirtualMachine\"),\n\t\t},\n\t})\n\topts = append(opts, aliases)\n\tvar resource VirtualMachine\n\terr := ctx.RegisterResource(\"azure-native:devtestlab/v20180915:VirtualMachine\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func IsVM(name string) bool {\n\tif IsKIC(name) || BareMetal(name) {\n\t\treturn false\n\t}\n\treturn true\n}", "func (p *ProxMox) StartInstance(ctx *lepton.Context, instanceID string) error {\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu/\"+instanceID+\"/status/start\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\t_, err = io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func ThisVMI(vmi *virtv1.VirtualMachineInstance) func() (*virtv1.VirtualMachineInstance, error) {\n\treturn func() (p *virtv1.VirtualMachineInstance, err error) {\n\t\tvirtClient, err := kubecli.GetKubevirtClient()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tp, err = virtClient.VirtualMachineInstance(vmi.Namespace).Get(vmi.Name, &k8smetav1.GetOptions{})\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn nil, nil\n\t\t}\n\t\treturn\n\t}\n}", "func createVmRun(c *Cli) error {\n\tconfig := libvirt.Config{\n\t\tURI: c.v.GetStringSlice(configKeyHosts)[0],\n\t}\n\n\tclient, err := config.Client()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = libvirt.DomainCreate(client, c.createCmd.createVmCmd.name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Instance() Desktop {\n\treturn instance\n}", "func (v *MachineVM) Init(opts machine.InitOptions) error {\n\tvar (\n\t\tkey string\n\t)\n\tsshDir := filepath.Join(homedir.Get(), \".ssh\")\n\t// GetConfDir creates the directory so no need to check for\n\t// its existence\n\tvmConfigDir, err := machine.GetConfDir(vmtype)\n\tif err != nil {\n\t\treturn err\n\t}\n\tjsonFile := filepath.Join(vmConfigDir, v.Name) + \".json\"\n\tv.IdentityPath = filepath.Join(sshDir, v.Name)\n\n\tswitch opts.ImagePath {\n\tcase \"testing\", \"stable\", \"\":\n\t\t// Get image as usual\n\t\tdd, err := machine.NewFcosDownloader(vmtype, v.Name, opts.ImagePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.ImagePath = dd.Get().LocalUncompressedFile\n\t\tif err := dd.DownloadImage(); err != nil {\n\t\t\treturn err\n\t\t}\n\tdefault:\n\t\t// The user has provided an alternate image which can be a file path\n\t\t// or URL.\n\t\tg, err := machine.NewGenericDownloader(vmtype, v.Name, opts.ImagePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv.ImagePath = g.Get().LocalUncompressedFile\n\t\tif err := g.DownloadImage(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\t// Add arch specific options including image location\n\tv.CmdLine = append(v.CmdLine, v.addArchOptions()...)\n\n\t// Add location of bootable image\n\tv.CmdLine = append(v.CmdLine, \"-drive\", \"if=virtio,file=\"+v.ImagePath)\n\t// This kind of stinks but no other way around this r/n\n\tif len(opts.IgnitionPath) < 1 {\n\t\turi := machine.SSHRemoteConnection.MakeSSHURL(\"localhost\", \"/run/user/1000/podman/podman.sock\", strconv.Itoa(v.Port), v.RemoteUsername)\n\t\tif err := machine.AddConnection(&uri, v.Name, filepath.Join(sshDir, v.Name), opts.IsDefault); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\turiRoot := machine.SSHRemoteConnection.MakeSSHURL(\"localhost\", \"/run/podman/podman.sock\", strconv.Itoa(v.Port), \"root\")\n\t\tif err := machine.AddConnection(&uriRoot, v.Name+\"-root\", filepath.Join(sshDir, v.Name), opts.IsDefault); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tfmt.Println(\"An ignition path was provided. No SSH connection was added to Podman\")\n\t}\n\t// Write the JSON file\n\tb, err := json.MarshalIndent(v, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := ioutil.WriteFile(jsonFile, b, 0644); err != nil {\n\t\treturn err\n\t}\n\n\t// User has provided ignition file so keygen\n\t// will be skipped.\n\tif len(opts.IgnitionPath) < 1 {\n\t\tkey, err = machine.CreateSSHKeys(v.IdentityPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\t// Run arch specific things that need to be done\n\tif err := v.prepare(); err != nil {\n\t\treturn err\n\t}\n\n\toriginalDiskSize, err := getDiskSize(v.ImagePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Resize the disk image to input disk size\n\t// only if the virtualdisk size is less than\n\t// the given disk size\n\tif opts.DiskSize<<(10*3) > originalDiskSize {\n\t\tresize := exec.Command(\"qemu-img\", []string{\"resize\", v.ImagePath, strconv.Itoa(int(opts.DiskSize)) + \"G\"}...)\n\t\tresize.Stdout = os.Stdout\n\t\tresize.Stderr = os.Stderr\n\t\tif err := resize.Run(); err != nil {\n\t\t\treturn errors.Errorf(\"error resizing image: %q\", err)\n\t\t}\n\t}\n\t// If the user provides an ignition file, we need to\n\t// copy it into the conf dir\n\tif len(opts.IgnitionPath) > 0 {\n\t\tinputIgnition, err := ioutil.ReadFile(opts.IgnitionPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn ioutil.WriteFile(v.IgnitionFilePath, inputIgnition, 0644)\n\t}\n\t// Write the ignition file\n\tign := machine.DynamicIgnition{\n\t\tName: opts.Username,\n\t\tKey: key,\n\t\tVMName: v.Name,\n\t\tWritePath: v.IgnitionFilePath,\n\t}\n\treturn machine.NewIgnitionFile(ign)\n}", "func (jsonVM *jsonVM) toVM(project string, opts *providerOpts) (ret *vm.VM) {\n\tvar vmErrors []error\n\tvar err error\n\n\t// Check \"lifetime\" label.\n\tvar lifetime time.Duration\n\tif lifetimeStr, ok := jsonVM.Labels[\"lifetime\"]; ok {\n\t\tif lifetime, err = time.ParseDuration(lifetimeStr); err != nil {\n\t\t\tvmErrors = append(vmErrors, vm.ErrNoExpiration)\n\t\t}\n\t} else {\n\t\tvmErrors = append(vmErrors, vm.ErrNoExpiration)\n\t}\n\n\t// lastComponent splits a url path and returns only the last part. This is\n\t// used because some of the fields in jsonVM are defined using URLs like:\n\t// \"https://www.googleapis.com/compute/v1/projects/cockroach-shared/zones/us-east1-b/machineTypes/n1-standard-16\"\n\t// We want to strip this down to \"n1-standard-16\", so we only want the last\n\t// component.\n\tlastComponent := func(url string) string {\n\t\ts := strings.Split(url, \"/\")\n\t\treturn s[len(s)-1]\n\t}\n\n\t// Extract network information\n\tvar publicIP, privateIP, vpc string\n\tif len(jsonVM.NetworkInterfaces) == 0 {\n\t\tvmErrors = append(vmErrors, vm.ErrBadNetwork)\n\t} else {\n\t\tprivateIP = jsonVM.NetworkInterfaces[0].NetworkIP\n\t\tif len(jsonVM.NetworkInterfaces[0].AccessConfigs) == 0 {\n\t\t\tvmErrors = append(vmErrors, vm.ErrBadNetwork)\n\t\t} else {\n\t\t\t_ = jsonVM.NetworkInterfaces[0].AccessConfigs[0].Name // silence unused warning\n\t\t\tpublicIP = jsonVM.NetworkInterfaces[0].AccessConfigs[0].NatIP\n\t\t\tvpc = lastComponent(jsonVM.NetworkInterfaces[0].Network)\n\t\t}\n\t}\n\n\tmachineType := lastComponent(jsonVM.MachineType)\n\tzone := lastComponent(jsonVM.Zone)\n\tremoteUser := config.SharedUser\n\tif !opts.useSharedUser {\n\t\t// N.B. gcloud uses the local username to log into instances rather\n\t\t// than the username on the authenticated Google account but we set\n\t\t// up the shared user at cluster creation time. Allow use of the\n\t\t// local username if requested.\n\t\tremoteUser = config.OSUser.Username\n\t}\n\treturn &vm.VM{\n\t\tName: jsonVM.Name,\n\t\tCreatedAt: jsonVM.CreationTimestamp,\n\t\tErrors: vmErrors,\n\t\tDNS: fmt.Sprintf(\"%s.%s.%s\", jsonVM.Name, zone, project),\n\t\tLifetime: lifetime,\n\t\tPrivateIP: privateIP,\n\t\tProvider: ProviderName,\n\t\tProviderID: jsonVM.Name,\n\t\tPublicIP: publicIP,\n\t\tRemoteUser: remoteUser,\n\t\tVPC: vpc,\n\t\tMachineType: machineType,\n\t\tZone: zone,\n\t\tProject: project,\n\t}\n}", "func NewMachine(ctx context.Context, cfg Config, opts ...Opt) (*Machine, error) {\n\tm := &Machine{\n\t\texitCh: make(chan struct{}),\n\t}\n\n\tif cfg.VMID == \"\" {\n\t\trandomID, err := uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to create random ID for VMID\")\n\t\t}\n\t\tcfg.VMID = randomID.String()\n\t}\n\n\tm.Handlers = defaultHandlers\n\n\tif cfg.JailerCfg != nil {\n\t\tm.Handlers.Validation = m.Handlers.Validation.Append(JailerConfigValidationHandler)\n\t\tif err := jail(ctx, m, &cfg); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tm.Handlers.Validation = m.Handlers.Validation.Append(ConfigValidationHandler)\n\t\tm.cmd = configureBuilder(defaultFirecrackerVMMCommandBuilder, cfg).Build(ctx)\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\n\tif m.logger == nil {\n\t\tlogger := log.New()\n\n\t\tm.logger = log.NewEntry(logger)\n\t}\n\n\tif m.client == nil {\n\t\tm.client = NewClient(cfg.SocketPath, m.logger, false)\n\t}\n\n\tif cfg.ForwardSignals == nil {\n\t\tcfg.ForwardSignals = []os.Signal{\n\t\t\tos.Interrupt,\n\t\t\tsyscall.SIGQUIT,\n\t\t\tsyscall.SIGTERM,\n\t\t\tsyscall.SIGHUP,\n\t\t\tsyscall.SIGABRT,\n\t\t}\n\t}\n\n\tm.machineConfig = cfg.MachineCfg\n\tm.Cfg = cfg\n\n\tif cfg.NetNS == \"\" && cfg.NetworkInterfaces.cniInterface() != nil {\n\t\tm.Cfg.NetNS = m.defaultNetNSPath()\n\t}\n\n\tm.logger.Debug(\"Called NewMachine()\")\n\treturn m, nil\n}", "func NewMachine(ctx *context.MachineContext, client client.Client, namespace string, sshKeys *ssh.ClusterNodeSshKeys) (*Machine, error) {\n\tmachine := &Machine{\n\t\tclient: client,\n\t\tnamespace: namespace,\n\t\tmachineContext: ctx,\n\t\tvmiInstance: nil,\n\t\tvmInstance: nil,\n\t\tsshKeys: sshKeys,\n\t\tgetCommandExecutor: ssh.NewVMCommandExecutor,\n\t}\n\n\tnamespacedName := types.NamespacedName{Namespace: namespace, Name: ctx.KubevirtMachine.Name}\n\tvm := &kubevirtv1.VirtualMachine{}\n\tvmi := &kubevirtv1.VirtualMachineInstance{}\n\n\t// Get the active running VMI if it exists\n\terr := client.Get(ctx.Context, namespacedName, vmi)\n\tif err != nil {\n\t\tif !apierrors.IsNotFound(err) {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tmachine.vmiInstance = vmi\n\t}\n\n\t// Get the top level VM object if it exists\n\terr = client.Get(ctx.Context, namespacedName, vm)\n\tif err != nil {\n\t\tif !apierrors.IsNotFound(err) {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tmachine.vmInstance = vm\n\t}\n\n\treturn machine, nil\n}", "func (c *Controller) ensureVM(vm *vmapi.VirtualMachine) error {\n\tlogger := c.logger.WithFields(logrus.Fields{\n\t\t\"virtual-machine\": vm.Name,\n\t\t\"namespace\": vm.Namespace,\n\t})\n\n\tinstance, err := c.gceClient.InstancesGet(c.config.Project, string(c.config.Zone), vm.ObjectMeta.Name)\n\tif instance != nil {\n\t\tif _, err := c.kubeClient.CoreV1().Secrets(vm.Namespace).Get(vm.Name, meta.GetOptions{}); err != nil {\n\t\t\tif kerrors.IsNotFound(err) {\n\t\t\t\tlogger.Infof(\"Regenerating SSH key for existing VM.\")\n\t\t\t\treturn c.refreshSSHKey(vm, logger)\n\t\t\t}\n\t\t\treturn fmt.Errorf(\"failed to check for existance of secret: %v\", err)\n\t\t}\n\t\tlogger.Infof(\"Skipped creating a VM that is already created.\")\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\tif gerr, ok := err.(*googleapi.Error); ok && gerr.Code != http.StatusNotFound {\n\t\t\treturn fmt.Errorf(\"failed to check for existance of virtual machine: %v\", err)\n\t\t}\n\t}\n\n\treturn c.createNewVM(vm, logger)\n}", "func reachableFromVM(from echo.Instance) match.Matcher {\n\tif !from.Config().IsVM() {\n\t\treturn match.Any\n\t}\n\treturn match.NotExternal\n}", "func New(q *ast.Questionaire, toFrontend, fromFrontend chan *fe.Event) {\n\tv := &vm{\n\t\tquestionaire: q,\n\t\tsend: toFrontend,\n\t\treceive: fromFrontend,\n\t}\n\tv.loop()\n}", "func cmdStart() {\n\tswitch state := status(B2D.VM); state {\n\tcase vmUnregistered:\n\t\tlog.Fatalf(\"%s is not registered.\", B2D.VM)\n\tcase vmRunning:\n\t\tlog.Printf(\"%s is already running.\", B2D.VM)\n\tcase vmPaused:\n\t\tlog.Printf(\"Resuming %s\", B2D.VM)\n\t\tif err := vbm(\"controlvm\", B2D.VM, \"resume\"); err != nil {\n\t\t\tlog.Fatalf(\"Failed to resume vm: %s\", err)\n\t\t}\n\t\taddr := fmt.Sprintf(\"localhost:%d\", B2D.SSHPort)\n\t\tif err := read(addr); err != nil {\n\t\t\tlog.Fatalf(\"Failed to connect to SSH port at %s: %s\", addr, err)\n\t\t}\n\t\tlog.Printf(\"Resumed.\")\n\tcase vmSaved, vmPoweroff, vmAborted:\n\t\tlog.Printf(\"Starting %s...\", B2D.VM)\n\t\tif err := vbm(\"startvm\", B2D.VM, \"--type\", \"headless\"); err != nil {\n\t\t\tlog.Fatalf(\"Failed to start vm: %s\", err)\n\t\t}\n\t\tlog.Printf(\"Waiting for SSH server to start...\")\n\t\taddr := fmt.Sprintf(\"localhost:%d\", B2D.SSHPort)\n\t\tif err := read(addr); err != nil {\n\t\t\tlog.Fatalf(\"Failed to connect to SSH port at %s: %s\", addr, err)\n\t\t}\n\t\tlog.Printf(\"Started.\")\n\tdefault:\n\t\tlog.Fatalf(\"Cannot start %s from state %.\", B2D.VM, state)\n\t}\n\n\t// Check if $DOCKER_HOST ENV var is properly configured.\n\tDockerHost := getenv(\"DOCKER_HOST\", \"\")\n\tif DockerHost != fmt.Sprintf(\"tcp://localhost:%d\", B2D.DockerPort) {\n\t\tfmt.Printf(\"\\nTo connect the docker client to the Docker daemon, please set:\\n\")\n\t\tfmt.Printf(\"export DOCKER_HOST=tcp://localhost:%d\\n\\n\", B2D.DockerPort)\n\t}\n}", "func main() {\n\t// parse arguments\n\tvmConfigPathPtr := flag.String(\"v\", \"None\", \"required: path to vm config json file\")\n\tdbConfigPathPtr := flag.String(\"d\", \"None\", \"optional: path to db config file\")\n\tflag.Parse()\n\n\t// download json data\n\tbody, err := ioutil.ReadFile(*vmConfigPathPtr)\n\tec2utils.HandleError(err)\n\n\t// parse the json into structs\n\tvar config ec2utils.CreateVMInput\n\tjson.Unmarshal([]byte(body), &config)\n\n\t// region will be user's own aws config default region\n\tsess, err := session.NewSession()\n\tec2utils.HandleError(err)\n\n\tec2utils.CreateEC2Instance(sess, config, *dbConfigPathPtr)\n}", "func (vmops *VMOperation) CloneVM(newVMName string, poweron bool, host *mo.HostSystem, template *object.VirtualMachine, cluster *mo.ClusterComputeResource, datacenter *object.Datacenter, datastore *mo.Datastore) (*object.VirtualMachine, types.TaskInfoState) {\n\n\tctx := vmops.Context\n\tc := vmops.Vcenter.Client.Client\n\n\t//Get the Resourcepool\n\tresourcepool := cluster.ResourcePool\n\n\tdcfolder, err := datacenter.Folders(ctx)\n\tif err != nil {\n\t\texit(err)\n\t}\n\tvmfolder := dcfolder.VmFolder\n\n\trelocationSpec := types.VirtualMachineRelocateSpec{\n\n\t\tPool: resourcepool,\n\t}\n\thostref := host.Reference()\n\trelocationSpec.Host = &hostref\n\n\tdatastoreRef := datastore.Reference()\n\trelocationSpec.Datastore = &datastoreRef\n\n\tcloneSpec := &types.VirtualMachineCloneSpec{\n\t\tPowerOn: poweron,\n\t\tTemplate: false,\n\t}\n\n\tcloneSpec.Location = relocationSpec\n\n\ttask, _ := template.Clone(ctx, vmfolder, newVMName, *cloneSpec)\n\n\tinfo, err := task.WaitForResult(ctx, nil)\n\n\tif err != nil {\n\t\tfmt.Println(\"Task failed with error \", err)\n\t}\n\n\tfmt.Printf(\"%s Cloning completed with %s.\\n\", newVMName, info.State)\n\n\treturn object.NewVirtualMachine(c, info.Result.(types.ManagedObjectReference)), info.State\n\n}", "func New(ctx context.Context, local bool, instanceConfig config.InstanceConfig, startTime time.Time, version string, startSwarming bool) (*Machine, error) {\n\tstore, err := store.New(ctx, false, instanceConfig)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to build store instance.\")\n\t}\n\tsink, err := sink.New(ctx, local, instanceConfig)\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Failed to build sink instance.\")\n\t}\n\n\tkubernetesImage := os.Getenv(swarming.KubernetesImageEnvVar)\n\thostname, err := os.Hostname()\n\tif err != nil {\n\t\treturn nil, skerr.Wrapf(err, \"Could not determine hostname.\")\n\t}\n\tmachineID := os.Getenv(swarming.SwarmingBotIDEnvVar)\n\tif machineID == \"\" {\n\t\t// Fall back to hostname so we can track machines that\n\t\t// test_machine_monitor is running on that don't also run Swarming.\n\t\tmachineID = hostname\n\t}\n\n\treturn &Machine{\n\t\tdimensions: machine.SwarmingDimensions{},\n\t\tstore: store,\n\t\tsink: sink,\n\t\tadb: adb.New(),\n\t\tMachineID: machineID,\n\t\tHostname: hostname,\n\t\tKubernetesImage: kubernetesImage,\n\t\tVersion: version,\n\t\tstartTime: startTime,\n\t\tstartSwarming: startSwarming,\n\t\tinterrogateTimer: metrics2.GetFloat64SummaryMetric(\"bot_config_machine_interrogate_timer\", map[string]string{\"machine\": machineID}),\n\t\tinterrogateAndSendFailures: metrics2.GetCounter(\"bot_config_machine_interrogate_and_send_errors\", map[string]string{\"machine\": machineID}),\n\t\tstoreWatchArrivalCounter: metrics2.GetCounter(\"bot_config_machine_store_watch_arrival\", map[string]string{\"machine\": machineID}),\n\t}, nil\n}", "func (_class VTPMClass) GetVM(sessionID SessionRef, self VTPMRef) (_retval VMRef, _err error) {\n\tif IsMock {\n\t\treturn _class.GetVMMock(sessionID, self)\n\t}\t\n\t_method := \"VTPM.get_VM\"\n\t_sessionIDArg, _err := convertSessionRefToXen(fmt.Sprintf(\"%s(%s)\", _method, \"session_id\"), sessionID)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_selfArg, _err := convertVTPMRefToXen(fmt.Sprintf(\"%s(%s)\", _method, \"self\"), self)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_result, _err := _class.client.APICall(_method, _sessionIDArg, _selfArg)\n\tif _err != nil {\n\t\treturn\n\t}\n\t_retval, _err = convertVMRefToGo(_method + \" -> \", _result.Value)\n\treturn\n}", "func NewServiceInstance() (*simulator.Model, *simulator.Server, error) {\n\tmodel := simulator.VPX()\n\terr := model.Create()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn model, model.Service.NewServer(), nil\n}", "func create(vm *api.VM) error {\n\tlog.Infof(\"Creating VM %q with name %q...\", vm.GetUID(), vm.GetName())\n\tif err := ensureOCIImages(vm); err != nil {\n\t\treturn err\n\t}\n\tvmCreated.Inc()\n\t// Allocate and populate the overlay file\n\treturn dmlegacy.AllocateAndPopulateOverlay(vm)\n}", "func (vm *VirtualMachine) Start(client SkytapClient) (*VirtualMachine, error) {\n\tlog.WithFields(log.Fields{\"vmId\": vm.Id}).Info(\"Starting VM\")\n\n\treturn vm.ChangeRunstate(client, RunStateStart, RunStateStart)\n}", "func (i *Instances) InstanceID(name string) (string, error) {\n\t// Create context\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\n\t// Create vSphere client\n\tc, err := vsphereLogin(i.cfg, ctx)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer c.Logout(ctx)\n\n\tvm, err := getVirtualMachineByName(i.cfg, ctx, c, name)\n\n\tvar mvm mo.VirtualMachine\n\terr = getVirtualMachineManagedObjectReference(ctx, c, vm, \"summary\", &mvm)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif mvm.Summary.Runtime.PowerState == ActivePowerState {\n\t\treturn \"/\" + vm.InventoryPath, nil\n\t}\n\n\tif mvm.Summary.Config.Template == false {\n\t\tglog.Warning(\"VM %s, is not in %s state\", name, ActivePowerState)\n\t} else {\n\t\tglog.Warning(\"VM %s, is a template\", name)\n\t}\n\n\treturn \"\", cloudprovider.InstanceNotFound\n}", "func (g GCPClient) CreateInstance(name, image, zone, machineType string, disks Disks, data *string, nested, vtpm, replace bool) error {\n\tif replace {\n\t\tif err := g.DeleteInstance(name, zone, true); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tlog.Infof(\"Creating instance %s from image %s (type: %s in %s)\", name, image, machineType, zone)\n\n\tenabled := new(string)\n\t*enabled = \"1\"\n\n\tk, err := ssh.NewPublicKey(g.privKey.Public())\n\tif err != nil {\n\t\treturn err\n\t}\n\tsshKey := new(string)\n\t*sshKey = fmt.Sprintf(\"moby:%s moby\", string(ssh.MarshalAuthorizedKey(k)))\n\n\t// check provided image to be compatible with provided options\n\top, err := g.compute.Images.Get(g.projectName, image).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tuefiCompatible := false\n\tfor _, feature := range op.GuestOsFeatures {\n\t\tif feature != nil && feature.Type == uefiCompatibleFeature {\n\t\t\tuefiCompatible = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif vtpm && !uefiCompatible {\n\t\treturn fmt.Errorf(\"cannot use vTPM without UEFI_COMPATIBLE image\")\n\t}\n\t// we should check for nested\n\tvmxLicense := false\n\tfor _, license := range op.Licenses {\n\t\t// we omit hostname and version when define license\n\t\tif strings.HasSuffix(license, vmxImageLicence) {\n\t\t\tvmxLicense = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif nested && !vmxLicense {\n\t\treturn fmt.Errorf(\"cannot use nested virtualization without enable-vmx image\")\n\t}\n\n\tinstanceDisks := []*compute.AttachedDisk{\n\t\t{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: true,\n\t\t\tInitializeParams: &compute.AttachedDiskInitializeParams{\n\t\t\t\tSourceImage: fmt.Sprintf(\"global/images/%s\", image),\n\t\t\t},\n\t\t},\n\t}\n\n\tfor i, disk := range disks {\n\t\tvar diskName string\n\t\tif disk.Path != \"\" {\n\t\t\tdiskName = disk.Path\n\t\t} else {\n\t\t\tdiskName = fmt.Sprintf(\"%s-disk-%d\", name, i)\n\t\t}\n\t\tvar diskSizeGb int64\n\t\tif disk.Size == 0 {\n\t\t\tdiskSizeGb = int64(1)\n\t\t} else {\n\t\t\tdiskSizeGb = int64(convertMBtoGB(disk.Size))\n\t\t}\n\t\tdiskObj := &compute.Disk{Name: diskName, SizeGb: diskSizeGb}\n\t\tif vtpm {\n\t\t\tdiskObj.GuestOsFeatures = []*compute.GuestOsFeature{\n\t\t\t\t{Type: uefiCompatibleFeature},\n\t\t\t}\n\t\t}\n\t\tdiskOp, err := g.compute.Disks.Insert(g.projectName, zone, diskObj).Do()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := g.pollZoneOperationStatus(diskOp.Name, zone); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tinstanceDisks = append(instanceDisks, &compute.AttachedDisk{\n\t\t\tAutoDelete: true,\n\t\t\tBoot: false,\n\t\t\tSource: fmt.Sprintf(\"zones/%s/disks/%s\", zone, diskName),\n\t\t})\n\t}\n\n\tinstanceObj := &compute.Instance{\n\t\tMachineType: fmt.Sprintf(\"zones/%s/machineTypes/%s\", zone, machineType),\n\t\tName: name,\n\t\tDisks: instanceDisks,\n\t\tNetworkInterfaces: []*compute.NetworkInterface{\n\t\t\t{\n\t\t\t\tNetwork: \"global/networks/default\",\n\t\t\t\tAccessConfigs: []*compute.AccessConfig{\n\t\t\t\t\t{\n\t\t\t\t\t\tType: \"ONE_TO_ONE_NAT\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tMetadata: &compute.Metadata{\n\t\t\tItems: []*compute.MetadataItems{\n\t\t\t\t{\n\t\t\t\t\tKey: \"serial-port-enable\",\n\t\t\t\t\tValue: enabled,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"ssh-keys\",\n\t\t\t\t\tValue: sshKey,\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tKey: \"user-data\",\n\t\t\t\t\tValue: data,\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tif nested {\n\t\tinstanceObj.MinCpuPlatform = \"Intel Haswell\"\n\t}\n\tif vtpm {\n\t\tinstanceObj.ShieldedInstanceConfig = &compute.ShieldedInstanceConfig{EnableVtpm: true}\n\t}\n\n\t// Don't wait for operation to complete!\n\t// A headstart is needed as by the time we've polled for this event to be\n\t// completed, the instance may have already terminated\n\t_, err = g.compute.Instances.Insert(g.projectName, zone, instanceObj).Do()\n\tif err != nil {\n\t\treturn err\n\t}\n\tlog.Infof(\"Instance created\")\n\treturn nil\n}", "func newInstances(pod *Pod, prov provider.DataCenter, cfg *config.Instances) (*instances, error) {\n\tlog.Debug(\"Initializing Instances\")\n\n\ti := &instances{\n\t\tResources: resource.NewResources(),\n\t\tpod: pod,\n\t\tinstances: map[string]resource.Instance{},\n\t}\n\n\t// The reference to the network resource.\n\tnet := pod.Cluster().Compute().DataCenter().Network()\n\n\t// The availability zones available to these instances.\n\tavailabilityZones := net.AvailabilityZones()\n\n\t// The subnet group associated with these instances.\n\tsubnetGroup := net.SubnetGroups().Find(pod.SubnetGroup())\n\tif subnetGroup == nil {\n\t\treturn nil, fmt.Errorf(\"Cannot find subnet group %s configured for pod %s\", pod.SubnetGroup(), pod.Name())\n\t}\n\n\t// The keypair to be used with these instances.\n\tkeypair := pod.Cluster().Compute().KeyPair()\n\n\tn := 0\n\tfor _, conf := range *cfg {\n\t\t// Ensure the instance is uniquely named.\n\t\tif i.Find(conf.Name()) != nil {\n\t\t\treturn nil, fmt.Errorf(\"Instance name %q must be unique but is used multiple times\", conf.Name())\n\t\t}\n\n\t\t// The availability zone for this instance. Chosing via round robin. Always starting at 0.\n\t\taz := availabilityZones[n%len(availabilityZones)]\n\n\t\t// Get the subnet associated with the AZ.\n\t\tsubnetName := pod.SubnetGroup() + \"-\" + az\n\t\tsubnet := subnetGroup.Find(subnetName)\n\t\tif subnet == nil {\n\t\t\treturn nil, fmt.Errorf(\"Cannot find subnet %s configured for instance %s\", subnetName, conf.Name())\n\t\t}\n\n\t\tinstance, err := newInstance(pod, subnet, keypair, prov, conf)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ti.instances[instance.Name()] = instance\n\t\ti.Append(instance)\n\n\t\tn++\n\t}\n\treturn i, nil\n}", "func (o InstanceOutput) VmImage() VmImageResponseOutput {\n\treturn o.ApplyT(func(v *Instance) VmImageResponseOutput { return v.VmImage }).(VmImageResponseOutput)\n}", "func CTLinuxBuilderInstance(num int) (*gce.Instance, error) {\n\tvm, err := CT20170602(fmt.Sprintf(\"ct-linux-builder-%03d\", num), true /* useSSDDataDisk */)\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tvm.MachineType = gce.MACHINE_TYPE_HIGHMEM_64\n\treturn vm, nil\n}", "func (c *Cluster) createVMInNode(opts compute.VirtualMachine, nics []*template.NIC) (string, string, error) {\n\tnode, err := c.getNodeRegion(opts.Region)\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\tdefer node.Client.Client.Close()\n\tif opts.ClusterId != \"\" {\n\t\topts.TemplateName = node.template\n\t} else {\n\t\topts.TemplateName = opts.Image\n\t}\n\n\topts.T = node.Client\n\ttmp, err := opts.Compute()\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\tif opts.ForceNetwork {\n\t\ttmp.UserTemplate[0].Template.Nic = nics\n\t}\n\tres, err := opts.Create(tmp)\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\tvmid := res.(string)\n\treturn opts.Name, vmid, nil\n}", "func Inst() *Torpedo {\n\treturn instance\n}", "func (h *HatcheryVSphere) createVMModel(model sdk.Model) (*object.VirtualMachine, error) {\n\tctx := context.Background()\n\tlog.Info(ctx, \"Create vm model %s\", model.Name)\n\n\tvm, errV := h.finder.VirtualMachine(ctx, model.ModelVirtualMachine.Image)\n\tif errV != nil {\n\t\treturn vm, sdk.WrapError(errV, \"createVMModel> Cannot find virtual machine\")\n\t}\n\n\tannot := annotation{\n\t\tHatcheryName: h.Name(),\n\t\tWorkerModelLastModified: fmt.Sprintf(\"%d\", model.UserLastModified.Unix()),\n\t\tWorkerModelName: model.Name,\n\t\tModel: true,\n\t\tCreated: time.Now(),\n\t}\n\n\tcloneSpec, folder, errCfg := h.createVMConfig(vm, annot)\n\tif errCfg != nil {\n\t\treturn vm, sdk.WrapError(errCfg, \"createVMModel> cannot create VM configuration\")\n\t}\n\n\ttask, errC := vm.Clone(ctx, folder, model.Name+\"-tmp\", *cloneSpec)\n\tif errC != nil {\n\t\treturn vm, sdk.WrapError(errC, \"createVMModel> cannot clone VM\")\n\t}\n\n\tinfo, errWr := task.WaitForResult(ctx, nil)\n\tif errWr != nil || info.State == types.TaskInfoStateError {\n\t\treturn vm, sdk.WrapError(errWr, \"createVMModel> state in error\")\n\t}\n\n\tvm = object.NewVirtualMachine(h.vclient.Client, info.Result.(types.ManagedObjectReference))\n\n\tif _, errW := vm.WaitForIP(ctx); errW != nil {\n\t\treturn vm, sdk.WrapError(errW, \"createVMModel> cannot get an ip\")\n\t}\n\n\tif _, errS := h.launchClientOp(vm, model.ModelVirtualMachine.PreCmd+\"; \\n\"+model.ModelVirtualMachine.Cmd+\"; \\n\"+model.ModelVirtualMachine.PostCmd, nil); errS != nil {\n\t\tlog.Warning(ctx, \"createVMModel> cannot start program %s\", errS)\n\t\tannot := annotation{ToDelete: true}\n\t\tif annotStr, err := json.Marshal(annot); err == nil {\n\t\t\tvm.Reconfigure(ctx, types.VirtualMachineConfigSpec{\n\t\t\t\tAnnotation: string(annotStr),\n\t\t\t})\n\t\t}\n\t}\n\n\tctxTo, cancel := context.WithTimeout(ctx, 4*time.Minute)\n\tdefer cancel()\n\tif err := vm.WaitForPowerState(ctxTo, types.VirtualMachinePowerStatePoweredOff); err != nil {\n\t\treturn nil, sdk.WrapError(err, \"cannot wait for power state result\")\n\t}\n\tlog.Info(ctx, \"createVMModel> model %s is build\", model.Name)\n\n\tmodelFound, errM := h.getModelByName(ctx, model.Name)\n\tif errM == nil {\n\t\tif errD := h.deleteServer(modelFound); errD != nil {\n\t\t\tlog.Warning(ctx, \"createVMModel> Cannot delete previous model %s : %s\", model.Name, errD)\n\t\t}\n\t}\n\n\tctxTo, cancel = context.WithTimeout(ctx, 15*time.Second)\n\tdefer cancel()\n\ttask, errR := vm.Rename(ctxTo, model.Name)\n\tif errR != nil {\n\t\treturn vm, sdk.WrapError(errR, \"createVMModel> Cannot rename model %s\", model.Name)\n\t}\n\n\tctxTo, cancel = context.WithTimeout(ctx, 15*time.Second)\n\tdefer cancel()\n\tif _, err := task.WaitForResult(ctxTo, nil); err != nil {\n\t\treturn vm, sdk.WrapError(err, \"error on waiting result for vm renaming %s\", model.Name)\n\t}\n\n\treturn vm, nil\n}", "func LoadVMByName(name string) (machine.VM, error) {\n\tvm := new(MachineVM)\n\tvmConfigDir, err := machine.GetConfDir(vmtype)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tb, err := ioutil.ReadFile(filepath.Join(vmConfigDir, name+\".json\"))\n\tif os.IsNotExist(err) {\n\t\treturn nil, errors.Wrap(machine.ErrNoSuchVM, name)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\terr = json.Unmarshal(b, vm)\n\tlogrus.Debug(vm.CmdLine)\n\treturn vm, err\n}", "func Instance() serving.Serving {\n\treturn instance\n}", "func (v *VPSDigitalOcean) Launch(setting VPSsettings) (instances []Instance) {\n\ttokenSource := &TokenSource{\n\t\tAccessToken: setting.Token,\n\t}\n\toauthClient := oauth2.NewClient(oauth2.NoContext, tokenSource)\n\tclient := godo.NewClient(oauthClient)\n\tctx := context.TODO()\n\n\t// set default values\n\tif setting.Region == \"\" {\n\t\tsetting.Region = \"nyc1\"\n\t}\n\tif setting.Type == \"\" {\n\t\tsetting.Type = \"s-1vcpu-1gb\"\n\t}\n\tif setting.Image == \"\" {\n\t\tsetting.Image = \"ubuntu-16-04-x64\"\n\t}\n\tif setting.Amount == 0 {\n\t\tsetting.Amount = 1\n\t}\n\tv.Droplets = v.Droplets[:0]\n\tnames := v.getNewNames(setting.ProjectName, setting.Amount, client, ctx)\n\n\tcreateRequest := &godo.DropletMultiCreateRequest{\n\t\tNames: names,\n\t\tRegion: setting.Region,\n\t\tSize: setting.Type,\n\t\tImage: godo.DropletCreateImage{\n\t\t\tSlug: setting.Image,\n\t\t},\n\t\tIPv6: true,\n\t\tTags: []string{setting.ProjectName, \"CloudRoutines\"},\n\t}\n\n\tdroplets, _, err := client.Droplets.CreateMultiple(ctx, createRequest)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\tfmt.Printf(\"%+v\\n\", setting)\n\t\treturn\n\t}\n\n\tfor _, droplet := range droplets {\n\t\tinstances = append(instances, Instance{Cloud: setting.Cloud, Project: setting.ProjectName, Name: droplet.Name})\n\t}\n\n\tv.Droplets = append(v.Droplets, droplets...)\n\tfmt.Printf(\"%+v\\n\", v.Droplets)\n\treturn instances\n}", "func (h Helper) CurrentVM() *vm.VM {\n\treturn h.Vms[h.CurrentScript]\n}", "func (pv *Provisioner) cloneVirtualMachine(s *SessionContext, cluster *clusterv1.Cluster, machine *clusterv1.Machine) error {\n\t// Fetch the user-data for the cloud-init first, so that we can fail fast before even trying to connect to pv\n\tuserData, err := pv.getCloudInitUserData(cluster, machine)\n\tif err != nil {\n\t\t// err returned by the getCloudInitUserData would be of type RequeueAfterError in case kubeadm is not ready yet\n\t\treturn err\n\t}\n\tmetaData, err := pv.getCloudInitMetaData(cluster, machine)\n\tif err != nil {\n\t\t// err returned by the getCloudInitUserData would be of type RequeueAfterError in case kubeadm is not ready yet\n\t\treturn err\n\t}\n\tctx, cancel := context.WithCancel(*s.context)\n\tdefer cancel()\n\n\tvar spec types.VirtualMachineCloneSpec\n\tmachineConfig, err := vsphereutils.GetMachineProviderConfig(machine.Spec.ProviderConfig)\n\tif err != nil {\n\t\treturn err\n\t}\n\tglog.V(4).Infof(\"[cloneVirtualMachine]: Preparing clone spec for VM %s\", machine.Name)\n\n\tdc, err := s.finder.DatacenterOrDefault(ctx, machineConfig.MachineSpec.Datacenter)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts.finder.SetDatacenter(dc)\n\n\tglog.Infof(\"clone VM folder %s\", machineConfig.MachineSpec.VMFolder)\n\tvmFolder, err := s.finder.FolderOrDefault(ctx, machineConfig.MachineSpec.VMFolder)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tds, err := s.finder.DatastoreOrDefault(ctx, machineConfig.MachineSpec.Datastore)\n\tif err != nil {\n\t\treturn err\n\t}\n\tspec.Location.Datastore = types.NewReference(ds.Reference())\n\n\tpool, err := s.finder.ResourcePoolOrDefault(ctx, machineConfig.MachineSpec.ResourcePool)\n\tif err != nil {\n\t\treturn err\n\t}\n\tspec.Location.Pool = types.NewReference(pool.Reference())\n\tspec.PowerOn = true\n\n\tspec.Config = &types.VirtualMachineConfigSpec{}\n\tdiskUUIDEnabled := true\n\tspec.Config.Flags = &types.VirtualMachineFlagInfo{\n\t\tDiskUuidEnabled: &diskUUIDEnabled,\n\t}\n\tif machineConfig.MachineSpec.NumCPUs > 0 {\n\t\tspec.Config.NumCPUs = int32(machineConfig.MachineSpec.NumCPUs)\n\t}\n\tif machineConfig.MachineSpec.MemoryMB > 0 {\n\t\tspec.Config.MemoryMB = machineConfig.MachineSpec.MemoryMB\n\t}\n\tspec.Config.Annotation = fmt.Sprintf(\"Virtual Machine is part of the cluster %s managed by cluster-api\", cluster.Name)\n\tspec.Location.DiskMoveType = string(types.VirtualMachineRelocateDiskMoveOptionsMoveAllDiskBackingsAndAllowSharing)\n\tsrc, err := s.finder.VirtualMachine(ctx, machineConfig.MachineSpec.VMTemplate)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvmProps, err := Properties(src)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"error fetching virtual machine or template properties: %s\", err)\n\t}\n\n\tif machineConfig.MachineSpec.VsphereCloudInit {\n\t\t// In case of vsphere cloud-init datasource present, set the appropriate extraconfig options\n\t\tvar extraconfigs []types.BaseOptionValue\n\t\textraconfigs = append(extraconfigs, &types.OptionValue{Key: \"guestinfo.metadata\", Value: metaData})\n\t\textraconfigs = append(extraconfigs, &types.OptionValue{Key: \"guestinfo.metadata.encoding\", Value: \"base64\"})\n\t\textraconfigs = append(extraconfigs, &types.OptionValue{Key: \"guestinfo.userdata\", Value: userData})\n\t\textraconfigs = append(extraconfigs, &types.OptionValue{Key: \"guestinfo.userdata.encoding\", Value: \"base64\"})\n\t\tspec.Config.ExtraConfig = extraconfigs\n\t} else {\n\t\t// This case is to support backwords compatibility, where we are using the ubuntu cloud image ovf properties\n\t\t// to drive the cloud-init workflow. Once the vsphere cloud-init datastore is merged as part of the official\n\t\t// cloud-init, then we can potentially remove this flag from the spec as then all the native cloud images\n\t\t// available for the different distros will include this new datasource.\n\t\t// See (https://github.com/akutz/cloud-init-vmware-guestinfo/ - vmware cloud-init datasource) for details\n\t\tif vmProps.Config.VAppConfig == nil {\n\t\t\treturn fmt.Errorf(\"this source VM lacks a vApp configuration and cannot have vApp properties set on it\")\n\t\t}\n\t\tallProperties := vmProps.Config.VAppConfig.GetVmConfigInfo().Property\n\t\tvar props []types.VAppPropertySpec\n\t\tfor _, p := range allProperties {\n\t\t\tdefaultValue := \" \"\n\t\t\tif p.DefaultValue != \"\" {\n\t\t\t\tdefaultValue = p.DefaultValue\n\t\t\t}\n\t\t\tprop := types.VAppPropertySpec{\n\t\t\t\tArrayUpdateSpec: types.ArrayUpdateSpec{\n\t\t\t\t\tOperation: types.ArrayUpdateOperationEdit,\n\t\t\t\t},\n\t\t\t\tInfo: &types.VAppPropertyInfo{\n\t\t\t\t\tKey: p.Key,\n\t\t\t\t\tId: p.Id,\n\t\t\t\t\tValue: defaultValue,\n\t\t\t\t},\n\t\t\t}\n\t\t\tif p.Id == \"user-data\" {\n\t\t\t\tprop.Info.Value = userData\n\t\t\t}\n\t\t\tif p.Id == \"public-keys\" {\n\t\t\t\tprop.Info.Value, err = pv.GetSSHPublicKey(cluster)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t\tif p.Id == \"hostname\" {\n\t\t\t\tprop.Info.Value = machine.Name\n\t\t\t}\n\t\t\tprops = append(props, prop)\n\t\t}\n\t\tspec.Config.VAppConfig = &types.VmConfigSpec{\n\t\t\tProperty: props,\n\t\t}\n\t}\n\n\tl := object.VirtualDeviceList(vmProps.Config.Hardware.Device)\n\tdeviceSpecs := []types.BaseVirtualDeviceConfigSpec{}\n\tdisks := l.SelectByType((*types.VirtualDisk)(nil))\n\t// For the disks listed under the MachineSpec.Disks property, they are used\n\t// only for resizing a maching disk on the template. Currently, no new disk\n\t// is added. Only the matched disks via the DiskLabel are resized. If the\n\t// MachineSpec.Disks is specified but none of the disks matched to the disks\n\t// present in the VM Template then error is returned. This is to avoid the\n\t// case when the user did want to resize but accidentally passed a wrong\n\t// disk label. A 100% matching of disks in not enforced as the user might be\n\t// interested in resizing only a subset of disks and thus we don't want to\n\t// force the user to list all the disk and sizes if they don't want to change\n\t// all.\n\tdiskMap := func(diskSpecs []vsphereconfig.DiskSpec) map[string]int64 {\n\t\tdiskMap := make(map[string]int64)\n\t\tfor _, s := range diskSpecs {\n\t\t\tdiskMap[s.DiskLabel] = s.DiskSizeGB\n\t\t}\n\t\treturn diskMap\n\t}(machineConfig.MachineSpec.Disks)\n\tdiskChange := false\n\tfor _, dev := range disks {\n\t\tdisk := dev.(*types.VirtualDisk)\n\t\tif newSize, ok := diskMap[disk.DeviceInfo.GetDescription().Label]; ok {\n\t\t\tif disk.CapacityInBytes > vsphereutils.GiBToByte(newSize) {\n\t\t\t\treturn errors.New(\"[FATAL] Disk size provided should be more than actual disk size of the template. Please correct the machineSpec to proceed\")\n\t\t\t}\n\t\t\tglog.V(4).Infof(\"[cloneVirtualMachine] Resizing the disk \\\"%s\\\" to new size \\\"%d\\\"\", disk.DeviceInfo.GetDescription().Label, newSize)\n\t\t\tdiskChange = true\n\t\t\tdisk.CapacityInBytes = vsphereutils.GiBToByte(newSize)\n\t\t\tdiskspec := &types.VirtualDeviceConfigSpec{}\n\t\t\tdiskspec.Operation = types.VirtualDeviceConfigSpecOperationEdit\n\t\t\tdiskspec.Device = disk\n\t\t\tdeviceSpecs = append(deviceSpecs, diskspec)\n\t\t}\n\t}\n\tif !diskChange && len(machineConfig.MachineSpec.Disks) > 0 {\n\t\tglog.V(4).Info(\"[cloneVirtualMachine] No disks were resized while cloning from template\")\n\t\treturn fmt.Errorf(\"[FATAL] None of the disks specified in the MachineSpec matched with the disks on the template %s\", machineConfig.MachineSpec.VMTemplate)\n\t}\n\n\tnics := l.SelectByType((*types.VirtualEthernetCard)(nil))\n\t// Remove any existing nics on the source vm\n\tfor _, dev := range nics {\n\t\tnic := dev.(types.BaseVirtualEthernetCard).GetVirtualEthernetCard()\n\t\tnicspec := &types.VirtualDeviceConfigSpec{}\n\t\tnicspec.Operation = types.VirtualDeviceConfigSpecOperationRemove\n\t\tnicspec.Device = nic\n\t\tdeviceSpecs = append(deviceSpecs, nicspec)\n\t}\n\t// Add new nics based on the user info\n\tnicid := int32(-100)\n\tfor _, network := range machineConfig.MachineSpec.Networks {\n\t\tnetRef, err := s.finder.Network(ctx, network.NetworkName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tnic := types.VirtualVmxnet3{}\n\t\tnic.Key = nicid\n\t\tnic.Backing, err = netRef.EthernetCardBackingInfo(ctx)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tnicspec := &types.VirtualDeviceConfigSpec{}\n\t\tnicspec.Operation = types.VirtualDeviceConfigSpecOperationAdd\n\t\tnicspec.Device = &nic\n\t\tdeviceSpecs = append(deviceSpecs, nicspec)\n\t\tnicid--\n\t}\n\tspec.Config.DeviceChange = deviceSpecs\n\tpv.eventRecorder.Eventf(machine, corev1.EventTypeNormal, \"Creating\", \"Creating Machine %v\", machine.Name)\n\ttask, err := src.Clone(ctx, vmFolder, machine.Name, spec)\n\tglog.Infof(\"clone to folder: %s with spec %v\", vmFolder, spec)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn pv.setTaskRef(machine, task.Reference().Value)\n}", "func (p *ProxMox) CreateInstance(ctx *lepton.Context) error {\n\n\tvar err error\n\n\tconfig := ctx.Config()\n\n\tnextid := p.getNextID()\n\n\tp.instanceName = config.RunConfig.InstanceName\n\n\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\n\tp.imageName = config.CloudConfig.ImageName\n\n\tp.arch = \"x86_64\"\n\tif config.TargetConfig[\"Arch\"] != \"\" {\n\t\tp.arch = config.TargetConfig[\"Arch\"]\n\t}\n\n\tp.machine = \"q35\"\n\tif config.TargetConfig[\"Machine\"] != \"\" {\n\t\tp.machine = config.TargetConfig[\"Machine\"]\n\t}\n\n\tp.sockets = \"1\"\n\tif config.TargetConfig[\"Sockets\"] != \"\" {\n\t\tsocketsInt, err := strconv.Atoi(config.TargetConfig[\"Sockets\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif socketsInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Sockets can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.sockets = config.TargetConfig[\"Sockets\"]\n\t}\n\n\tp.cores = \"1\"\n\tif config.TargetConfig[\"Cores\"] != \"\" {\n\t\tcoresInt, err := strconv.Atoi(config.TargetConfig[\"Cores\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif coresInt < 1 {\n\t\t\treturn errors.New(\"Bad configuration option; Cores can only be set postitive starting from \\\"1\\\"\")\n\t\t}\n\t\tp.cores = config.TargetConfig[\"Cores\"]\n\t}\n\n\tp.numa = \"0\"\n\tif config.TargetConfig[\"Numa\"] != \"\" {\n\t\tif config.TargetConfig[\"Numa\"] != \"0\" && config.TargetConfig[\"Numa\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Numa can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.numa = config.TargetConfig[\"Numa\"]\n\t}\n\n\t// Memory\n\n\tp.memory = \"512\"\n\tif config.TargetConfig[\"Memory\"] != \"\" {\n\t\tmemoryInt, err := lepton.RAMInBytes(config.TargetConfig[\"Memory\"])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tmemoryInt = memoryInt / 1024 / 1024\n\t\tp.memory = strconv.FormatInt(memoryInt, 10)\n\t}\n\n\t// Main storage\n\n\tp.storageName = \"local-lvm\"\n\tif config.TargetConfig[\"StorageName\"] != \"\" {\n\t\tp.storageName = config.TargetConfig[\"StorageName\"]\n\t}\n\n\t// Iso storage\n\n\tp.isoStorageName = \"local\"\n\tif config.TargetConfig[\"IsoStorageName\"] != \"\" {\n\t\tp.isoStorageName = config.TargetConfig[\"IsoStorageName\"]\n\t}\n\n\t// Bridge prefix\n\n\tp.bridgePrefix = \"vmbr\"\n\tif config.TargetConfig[\"BridgePrefix\"] != \"\" {\n\t\tp.bridgePrefix = config.TargetConfig[\"BridgePrefix\"]\n\t}\n\n\t// Onboot\n\n\tp.onboot = \"0\"\n\tif config.TargetConfig[\"Onboot\"] != \"\" {\n\t\tif config.TargetConfig[\"Onboot\"] != \"0\" && config.TargetConfig[\"Onboot\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Onboot can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.onboot = config.TargetConfig[\"Onboot\"]\n\t}\n\n\t// Protection\n\n\tp.protection = \"0\"\n\tif config.TargetConfig[\"Protection\"] != \"\" {\n\t\tif config.TargetConfig[\"Protection\"] != \"0\" && config.TargetConfig[\"Protection\"] != \"1\" {\n\t\t\treturn errors.New(\"Bad configuration option; Protection can only be set to \\\"0\\\" or \\\"1\\\"\")\n\t\t}\n\t\tp.protection = config.TargetConfig[\"Protection\"]\n\t}\n\n\t// These two preventive checks here, because Proxmox will not return\n\t// an error if the storage is missing and a misconfigured instance will be created.\n\n\terr = p.CheckStorage(p.storageName, \"images\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.CheckStorage(p.isoStorageName, \"iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata := url.Values{}\n\tdata.Set(\"vmid\", nextid)\n\tdata.Set(\"name\", p.instanceName)\n\tdata.Set(\"name\", p.imageName)\n\tdata.Set(\"machine\", p.machine)\n\tdata.Set(\"sockets\", p.sockets)\n\tdata.Set(\"cores\", p.cores)\n\tdata.Set(\"numa\", p.numa)\n\tdata.Set(\"memory\", p.memory)\n\tdata.Set(\"onboot\", p.onboot)\n\tdata.Set(\"protection\", p.protection)\n\tdata.Set(\"serial0\", \"socket\")\n\n\t// Configuring network interfaces\n\n\tnics := config.RunConfig.Nics\n\tfor i := 0; i < len(nics); i++ {\n\t\tis := strconv.Itoa(i)\n\t\tbrName := nics[i].BridgeName\n\t\tif brName == \"\" {\n\t\t\tbrName = p.bridgePrefix + is\n\t\t}\n\n\t\terr = p.CheckBridge(brName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif nics[i].IPAddress != \"\" {\n\t\t\tcidr := \"24\"\n\n\t\t\tif nics[i].NetMask != \"\" {\n\t\t\t\tcidrInt := lepton.CCidr(nics[i].NetMask)\n\t\t\t\tcidr = strconv.FormatInt(int64(cidrInt), 10)\n\t\t\t}\n\n\t\t\tif nics[i].Gateway != \"\" {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr+\",\"+\"gw=\"+nics[i].Gateway)\n\t\t\t} else {\n\t\t\t\tdata.Set(\"ipconfig\"+is, \"ip=\"+nics[i].IPAddress+\"/\"+cidr)\n\t\t\t}\n\t\t} else {\n\t\t\tdata.Set(\"ipconfig\"+is, \"dhcp\")\n\t\t}\n\n\t\tdata.Set(\"net\"+is, \"model=virtio,bridge=\"+brName)\n\t}\n\tif len(nics) == 0 {\n\t\t// single dhcp nic\n\t\tdata.Set(\"net0\", \"model=virtio,bridge=vmbr0\")\n\t}\n\n\treq, err := http.NewRequest(\"POST\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", bytes.NewBufferString(data.Encode()))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tdebug := false\n\tif debug {\n\t\tfmt.Println(string(body))\n\t}\n\n\terr = p.CheckResultType(body, \"createinstance\", \"file=\"+p.isoStorageName+\":iso/\"+p.imageName+\".iso\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.addVirtioDisk(ctx, nextid)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = p.movDisk(ctx, nextid)\n\n\treturn err\n}", "func (n *Node) StartVM(ctx context.Context, vm model.VM) error {\n\t// TODO: make port configurable\n\tconn, err := grpc.Dial(fmt.Sprintf(\"%s:8888\", n.host.Address), grpc.WithInsecure())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tuuid := &UUID{\n\t\tValue: vm.ID.String(),\n\t}\n\n\tvmConfig := &VmConfig{\n\t\tVmID: uuid,\n\t\tMemory: vm.Memory,\n\t\tVcpus: vm.VCPU,\n\t\tKernelImage: vm.KernelImage,\n\t\tRootFileSystem: vm.RootFileSystem,\n\t}\n\tclient := NewNodeClient(conn)\n\tresp, err := client.StartVM(ctx, vmConfig)\n\tif err != nil {\n\t\tlog.Error(\"grpc error: \", err)\n\t}\n\n\tlog.Debug(\"grpc response:\", resp)\n\n\treturn nil\n}", "func vmData(ctx context.Context, client *govmomi.Client) ([]mo.VirtualMachine, error) {\n\tm := view.NewManager(client.Client)\n\tvms := []mo.VirtualMachine{}\n\tview, err := m.CreateContainerView(ctx, client.ServiceContent.RootFolder, []string{\"VirtualMachine\"}, true)\n\tif err != nil {\n\t\treturn vms, err\n\t}\n\n\tdefer view.Destroy(ctx)\n\n\terr = view.Retrieve(ctx, []string{\"VirtualMachine\"}, []string{\"name\", \"summary\", \"runtime\"}, &vms)\n\tif err != nil {\n\t\treturn vms, err\n\t}\n\treturn vms, nil\n}", "func createCommandVM() {\n\tVmCommand = goja.New()\n\n\terr := VmCommand.Set(\"call\", Call)\n\tif err != nil {\n\t\t//TODO write to error log\n\t\tfmt.Println(err)\n\t}\n\terr = VmCommand.Set(\"printToLog\", PrintToLog)\n\tif err != nil {\n\t\tWriteErrorLog(err.Error(), \"\")\n\t}\n\terr = VmCommand.Set(\"callCompare\", CallCompare)\n\tif err != nil {\n\t\t//TODO write to error log\n\t\tfmt.Println(err)\n\t}\n\terr = VmCommand.Set(\"callContains\", CallContains)\n\tif err != nil {\n\t\t//TODO write to error log\n\t\tfmt.Println(err)\n\t}\n\terr = VmCommand.Set(\"printToConsole\", PrintToConsole)\n\tif err != nil {\n\t\t//TODO write to error log\n\t\tfmt.Println(err)\n\t}\n\terr = VmCommand.Set(\"shell\", Shell)\n\tif err != nil {\n\t\t//TODO write to error log\n\t\tfmt.Println(err)\n\t}\n}", "func (p *ProxMox) ListInstances(ctx *lepton.Context) error {\n\n\treq, err := http.NewRequest(\"GET\", p.apiURL+\"/api2/json/nodes/\"+p.nodeNAME+\"/qemu\", nil)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\ttr := &http.Transport{\n\t\tTLSClientConfig: &tls.Config{InsecureSkipVerify: true},\n\t}\n\tclient := &http.Client{Transport: tr}\n\n\treq.Header.Add(\"Authorization\", \"PVEAPIToken=\"+p.tokenID+\"=\"+p.secret)\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\tbody, err := io.ReadAll(resp.Body)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tir := &InstanceResponse{}\n\tjson.Unmarshal([]byte(body), ir)\n\n\ttable := tablewriter.NewWriter(os.Stdout)\n\ttable.SetHeader([]string{\"ID\", \"Name\", \"MainIP\", \"Status\", \"ImageID\"})\n\ttable.SetHeaderColor(\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor},\n\t\ttablewriter.Colors{tablewriter.Bold, tablewriter.FgCyanColor})\n\ttable.SetRowLine(true)\n\n\tfor _, instance := range ir.Data {\n\t\tvar row []string\n\t\trow = append(row, strconv.Itoa(instance.VMID))\n\t\trow = append(row, instance.Name)\n\t\trow = append(row, \"\")\n\t\trow = append(row, instance.Status)\n\t\trow = append(row, \"\")\n\t\ttable.Append(row)\n\t}\n\n\ttable.Render()\n\n\treturn nil\n}", "func Create(cfg *mgrconfig.Config, debug bool) (*Pool, error) {\n\ttyp, ok := vmimpl.Types[cfg.Type]\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"unknown instance type '%v'\", cfg.Type)\n\t}\n\tenv := &vmimpl.Env{\n\t\tName: cfg.Name,\n\t\tOS: cfg.TargetOS,\n\t\tArch: cfg.TargetVMArch,\n\t\tWorkdir: cfg.Workdir,\n\t\tImage: cfg.Image,\n\t\tSSHKey: cfg.SSHKey,\n\t\tSSHUser: cfg.SSHUser,\n\t\tDebug: debug,\n\t\tConfig: cfg.VM,\n\t}\n\timpl, err := typ.Ctor(env)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Pool{\n\t\timpl: impl,\n\t\tworkdir: env.Workdir,\n\t}, nil\n}", "func (vm *VirtualMachine) handleInstance(quad quads.Quad) {\n\tif strings.Contains(quad.Left.ID(), \"self_\") {\n\t\tstrElements := strings.Split(quad.Left.ID(), \"_\")\n\t\tif len(strElements) < 2 {\n\t\t\tlog.Fatalf(\"Error: (handleInstance) unexpected element id structure\")\n\t\t}\n\t\tobjInstanceAddr, err := strconv.Atoi(strElements[1])\n\t\tif err != nil {\n\t\t\tlog.Fatalf(\"Error: (handleInstance) couldn't cast objInstanceAddr to int\")\n\t\t}\n\t\tif objInstanceAddr == -1 {\n\t\t\treturn\n\t\t}\n\t}\n\n\tmemblock := vm.getMemBlockForAddr(quad.Left.GetAddr())\n\tvar ok bool\n\tvm.currentSelf, ok = memblock.Get(quad.Left.GetAddr()).(Memory)\n\tif !ok {\n\t\tlog.Fatalf(\n\t\t\t\"Error: (Run) quads.INSTANCE couldn't cast %v to Memory\",\n\t\t\tmemblock.Get(quad.Left.GetAddr()),\n\t\t)\n\t}\n}", "func (c *vmClient) Get(uid meta.UID) (*api.VM, error) {\n\tlog.Debugf(\"Client.Get; UID: %q, Kind: %s\", uid, api.KindVM)\n\tobject, err := c.storage.GetByID(api.KindVM, uid)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn object.(*api.VM), nil\n}", "func (c *Client) CreateInstance(displayName, availabilityDomain, compartmentID, nodeShape, nodeImageName, nodeSubnetID, sshUser, authorizedKeys string, nodeOCPUs, nodeMemoryInGBs int) (string, error) {\n\n\treq := identity.ListAvailabilityDomainsRequest{}\n\treq.CompartmentId = &compartmentID\n\tads, err := c.identityClient.ListAvailabilityDomains(context.Background(), req)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Just in case shortened or lower-case availability domain name was used\n\tlog.Debugf(\"Resolving availability domain from %s\", availabilityDomain)\n\tfor _, ad := range ads.Items {\n\t\tif strings.Contains(*ad.Name, strings.ToUpper(availabilityDomain)) {\n\t\t\tlog.Debugf(\"Availability domain %s\", *ad.Name)\n\t\t\tavailabilityDomain = *ad.Name\n\t\t}\n\t}\n\n\timageID, err := c.getImageID(compartmentID, nodeImageName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create the launch compute instance request\n\trequest := core.LaunchInstanceRequest{\n\t\tLaunchInstanceDetails: core.LaunchInstanceDetails{\n\t\t\tAvailabilityDomain: &availabilityDomain,\n\t\t\tCompartmentId: &compartmentID,\n\t\t\tShape: &nodeShape,\n\t\t\tCreateVnicDetails: &core.CreateVnicDetails{\n\t\t\t\tSubnetId: &nodeSubnetID,\n\t\t\t},\n\t\t\tDisplayName: &displayName,\n\t\t\tMetadata: map[string]string{\n\t\t\t\t\"ssh_authorized_keys\": authorizedKeys,\n\t\t\t\t\"user_data\": base64.StdEncoding.EncodeToString(createCloudInitScript(sshUser)),\n\t\t\t},\n\t\t\tSourceDetails: core.InstanceSourceViaImageDetails{\n\t\t\t\tImageId: imageID,\n\t\t\t},\n\t\t},\n\t}\n\n\tif nodeOCPUs > 0 {\n\t\toCPUs := float32(nodeOCPUs)\n\t\tmemoryInGBs := float32(nodeMemoryInGBs)\n\n\t\tLaunchInstanceShapeConfigDetails := core.LaunchInstanceShapeConfigDetails{\n\t\t\tOcpus: &oCPUs,\n\t\t\tMemoryInGBs: &memoryInGBs,\n\t\t}\n\t\trequest.ShapeConfig = &LaunchInstanceShapeConfigDetails\n\t}\n\n\tlog.Debugf(\"Launching instance with cloud-init: %s\", string(createCloudInitScript(sshUser)))\n\n\tcreateResp, err := c.computeClient.LaunchInstance(context.Background(), request)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// wait until lifecycle status is Running\n\tpollUntilRunning := func(r common.OCIOperationResponse) bool {\n\t\tif converted, ok := r.Response.(core.GetInstanceResponse); ok {\n\t\t\treturn converted.LifecycleState != core.InstanceLifecycleStateRunning\n\t\t}\n\t\treturn true\n\t}\n\n\t// create get instance request with a retry policy which takes a function\n\t// to determine shouldRetry or not\n\tpollingGetRequest := core.GetInstanceRequest{\n\t\tInstanceId: createResp.Instance.Id,\n\t\tRequestMetadata: helpers.GetRequestMetadataWithCustomizedRetryPolicy(pollUntilRunning),\n\t}\n\n\tinstance, pollError := c.computeClient.GetInstance(context.Background(), pollingGetRequest)\n\tif pollError != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn *instance.Id, nil\n}", "func newInstanceManager(c *instanceManagerConfig) *instanceManager {\n\n\tctx, cancel := context.WithCancel(c.Ctx)\n\ti := &instanceManager{\n\t\tlogger: c.Logger.With(\"driver\", c.ID.Name),\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tloader: c.Loader,\n\t\tstoreReattach: c.StoreReattach,\n\t\tfetchReattach: c.FetchReattach,\n\t\tpluginConfig: c.PluginConfig,\n\t\tid: c.ID,\n\t\tupdateNodeFromDriver: c.UpdateNodeFromDriver,\n\t\teventHandlerFactory: c.EventHandlerFactory,\n\t\tfirstFingerprintCh: make(chan struct{}),\n\t}\n\n\tgo i.run()\n\treturn i\n}", "func (vm VM) Name() string {\n\treturn vm.name\n}", "func (i *instances) InstanceID(ctx context.Context, nodeName types.NodeName) (string, error) {\n\tvm, err := i.discoverNodeByName(ctx, nodeName)\n\tif err != nil {\n\t\tklog.Errorf(\"Error trying to find VM: %v\", err)\n\t\treturn \"\", err\n\t}\n\tif vm == nil {\n\t\tklog.V(4).Info(\"instances.InstanceID() InstanceNotFound \", nodeName)\n\t\treturn \"\", cloudprovider.InstanceNotFound\n\t}\n\n\tif vm.Status.BiosUUID == \"\" {\n\t\treturn \"\", errBiosUUIDEmpty\n\t}\n\n\tklog.V(4).Infof(\"instances.InstanceID() called to get vm: %v uuid: %v\", nodeName, vm.Status.BiosUUID)\n\treturn vm.Status.BiosUUID, nil\n}", "func printVM(vm compute.VirtualMachine) {\n\tfmt.Println(\"\\n---------------------------------------------------------------\")\n\t//see https://godoc.org/github.com/Azure/azure-sdk-for-go/arm/compute#VirtualMachine for details\n\ttags := \"\\n\"\n\tif vm.Tags == nil {\n\t\ttags += \"\\t\\tNo tags yet\\n\"\n\t} else {\n\t\tfor key, value := range *vm.Tags {\n\t\t\ttags += fmt.Sprintf(\"\\t\\t%s = %s\\n\", key, *value)\n\t\t}\n\t}\n\tfmt.Printf(\"Virtual machine '%s'\\n\", *vm.Name)\n\n\t//find resourcegroup\n\tr, _ := regexp.Compile(\"resourceGroups/(.*)/providers\")\n\t//r.MatchString(*vm.ID)\n\tresoureGroup := r.FindStringSubmatch(*vm.ID)[1]\n\tfmt.Println(resoureGroup)\n\n\t//see https://godoc.org/github.com/Azure/azure-sdk-for-go/arm/compute#VirtualMachineProperties\n\t//statuses seems to be filled only by vmClient.Get(rg,vmName, compute.InstanceView)\n\tvar vmProps = *vm.VirtualMachineProperties\n\tvar statuses *[]compute.InstanceViewStatus\n\tif vmProps.InstanceView != nil {\n\t\tstatuses = (*vmProps.InstanceView).Statuses\n\t}\n\n\tif statuses != nil {\n\t\tfor _, status := range *statuses {\n\t\t\tfmt.Printf(\"Status %s, Message %s\", *status.Code, *status.DisplayStatus)\n\t\t}\n\t} else {\n\t\tfmt.Println(\"no status found, get detailed vm data\")\n\t\t//see https://godoc.org/github.com/Azure/azure-sdk-for-go/arm/compute#VirtualMachinesClient.Get\n\t\tdetailedVM, err := vmClient.Get(resoureGroup, *vm.Name, compute.InstanceView)\n\t\tfmt.Println(*detailedVM.Name)\n\t\t//TODO refactor status extraction to separate function\n\t\tif (*detailedVM.VirtualMachineProperties).InstanceView != nil {\n\t\t\tstatuses = (*(*detailedVM.VirtualMachineProperties).InstanceView).Statuses\n\t\t}\n\n\t\tif statuses != nil {\n\t\t\tfor _, status := range *statuses {\n\t\t\t\tfmt.Printf(\"Status %s, Message %s\", *status.Code, *status.DisplayStatus)\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\t//seems not to work, maybe recursion is not possible with go\n\t\t\tprintVM(detailedVM)\n\t\t} else {\n\t\t\tfmt.Println(err)\n\t\t}\n\t}\n\n\telements := map[string]interface{}{\n\t\t\"ID\": *vm.ID,\n\t\t\"Name\": *vm.Name,\n\t\t\"Type\": *vm.Type,\n\t\t\"Location\": *vm.Location,\n\t\t\"Tags\": tags,\n\t\t\"ProvisioningState\": *vmProps.ProvisioningState,\n\t\t\"Computername\": *(*vmProps.OsProfile).ComputerName,\n\t\t\"OS-TYpe\": (*(*vmProps.StorageProfile).OsDisk).OsType,\n\t\t\"VMSize\": (*vmProps.HardwareProfile).VMSize, //compare with const values from https://godoc.org/github.com/Azure/azure-sdk-for-go/arm/compute#VirtualMachineSizeTypes\n\t\t//\"OS-DiskName\": *(*(*vmProps.StorageProfile).OsDisk).Name,\n\t\t//\"Instance Status\": (*(*vmProps.InstanceView).Statuses),\n\t\t\"ResourceGroup\": resoureGroup,\n\t\t\"test\": statuses}\n\tfor k, v := range elements {\n\t\tfmt.Printf(\"\\t%s: %s\\n\", k, v)\n\t}\n\tfmt.Println(\"---------------------------------------------------------------\")\n}", "func NewInstance(name string, opts ...Option) (Instance, error) {\n\tif name == \"\" {\n\t\treturn nil, ErrNoName\n\t}\n\ti := &instance{\n\t\tname: name,\n\t\tbindAddr: \"127.0.0.1\",\n\t\tbindPort: 8100,\n\t\tshutdownCh: make(chan struct{}),\n\t\tpeers: make(map[string]*Peer),\n\t\tcaches: make(map[string]*cache),\n\t\tlogger: log.New(os.Stdout, \"huton\", log.LstdFlags),\n\t\traftApplicationTimeout: 10 * time.Second,\n\t\traftTransportTimeout: 10 * time.Second,\n\t\traftRetainSnapshotCount: 2,\n\t\tserfEventChannel: make(chan serf.Event, 256),\n\t}\n\tfor _, opt := range opts {\n\t\topt(i)\n\t}\n\ti.logger.Println(\"Initializing RPC server...\")\n\tif err := i.setupRPC(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\ti.logger.Println(\"Initializing Raft cluster...\")\n\tif err := i.setupRaft(); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tip := net.ParseIP(i.bindAddr)\n\traftAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 1,\n\t}\n\trpcAddr := &net.TCPAddr{\n\t\tIP: ip,\n\t\tPort: i.bindPort + 2,\n\t}\n\n\ti.logger.Println(\"Initializing Serf cluster...\")\n\tif err := i.setupSerf(raftAddr, rpcAddr); err != nil {\n\t\ti.Shutdown()\n\t\treturn i, err\n\t}\n\tgo i.handleEvents()\n\treturn i, nil\n}", "func CTLinuxBuilderInstance(num int) *gce.Instance {\n\tvm := CT20170602(fmt.Sprintf(\"ct-linux-builder-%03d\", num), true /* useSSDDataDisk */)\n\tvm.MachineType = \"custom-32-70400\"\n\treturn vm\n}" ]
[ "0.6743622", "0.6602082", "0.65942115", "0.65404797", "0.63559693", "0.63554317", "0.61643267", "0.61447084", "0.61269116", "0.6097864", "0.6093309", "0.6036116", "0.60246336", "0.5875924", "0.5862115", "0.5853817", "0.5840736", "0.5831631", "0.58141804", "0.5812497", "0.57965094", "0.57778597", "0.5734207", "0.56979847", "0.5687771", "0.56779206", "0.56670874", "0.5654225", "0.5644324", "0.56279707", "0.5605728", "0.5597695", "0.55608577", "0.5542789", "0.5535309", "0.5533635", "0.55232924", "0.5514745", "0.55034894", "0.54971594", "0.5480374", "0.5467639", "0.5457492", "0.5444423", "0.5439483", "0.5435907", "0.5410139", "0.540296", "0.5392327", "0.53920287", "0.53882337", "0.5387477", "0.53735065", "0.53711057", "0.5369002", "0.53432363", "0.5341065", "0.5319371", "0.5312364", "0.53016627", "0.5295776", "0.526263", "0.52547175", "0.52514124", "0.5245876", "0.5243414", "0.5238245", "0.52338976", "0.5230466", "0.5226097", "0.521234", "0.5208518", "0.5205629", "0.51962686", "0.5193644", "0.5184546", "0.51638013", "0.5160166", "0.515926", "0.51584387", "0.51525116", "0.5148451", "0.51433194", "0.5130339", "0.5126123", "0.51147604", "0.51135796", "0.50991595", "0.5095315", "0.5091645", "0.5074203", "0.50574565", "0.50552", "0.5051478", "0.50399274", "0.5030857", "0.50287473", "0.5024888", "0.5022444", "0.5019098" ]
0.579638
21
/ new instance of NLB
func NewNLB(ns string, mcisName string, groupId string, config string) *NLB { nlb := &NLB{ NLBBase: NLBBase{ Model: Model{Name: groupId, Namespace: ns}, Config: config, Type: "PUBLIC", Scope: "REGION", Listener: NLBProtocolBase{Protocol: "TCP", Port: "6443"}, TargetGroup: TargetGroup{NLBProtocolBase: NLBProtocolBase{Protocol: "TCP", Port: "6443"}, MCIS: mcisName, VmGroupId: groupId}, }, HealthChecker: HealthCheck{ NLBProtocolBase: NLBProtocolBase{Protocol: "TCP", Port: "22"}, Interval: "default", Threshold: "default", Timeout: "default", }, } if strings.Contains(config, string(app.CSP_NCPVPC)) || strings.Contains(config, string(app.CSP_AZURE)) { nlb.HealthChecker.Timeout = "-1" } if strings.Contains(nlb.NLBBase.Config, string(app.CSP_GCP)) { nlb.HealthChecker.NLBProtocolBase.Protocol = "HTTP" nlb.HealthChecker.NLBProtocolBase.Port = "80" } return nlb }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New(ringWeight int) LoadBalancer {\n\t// TODO: Implement this!\n\tnewLB := new(loadBalancer)\n\tnewLB.sortedNames = make([]MMENode, 0)\n\tnewLB.weight = ringWeight\n\tnewLB.hashRing = NewRing()\n\tif 7 == 2 {\n\t\tfmt.Println(ringWeight)\n\t}\n\treturn newLB\n}", "func NewLB(addr string) *LB {\n\treturn &LB{\n\t\tServer: &http.Server{\n\t\t\tAddr: addr,\n\t\t},\n\t\tlf: lockfree.New(),\n\t}\n}", "func NewLLRB(name string, setts s.Settings) *LLRB {\n\tllrb := &LLRB{name: name, finch: make(chan struct{})}\n\tllrb.logprefix = fmt.Sprintf(\"LLRB [%s]\", name)\n\tllrb.inittxns()\n\n\tsetts = make(s.Settings).Mixin(Defaultsettings(), setts)\n\tllrb.readsettings(setts)\n\tllrb.setts = setts\n\n\tllrb.nodearena = malloc.NewArena(llrb.memcapacity, llrb.allocator)\n\tllrb.valarena = malloc.NewArena(llrb.memcapacity, llrb.allocator)\n\n\t// statistics\n\tllrb.h_upsertdepth = lib.NewhistorgramInt64(10, 100, 10)\n\n\tinfof(\"%v started ...\\n\", llrb.logprefix)\n\tllrb.logarenasettings()\n\treturn llrb\n}", "func newIIDBT(s BaseSentence) (IIDBT, error) {\n\tp := newParser(s, PrefixIIDBT)\n\treturn IIDBT{\n\t\tBaseSentence: s,\n\t\tDepthFeet: p.Float64(0, \"depth (feet)\"),\n\t\tDepthMeters: p.Float64(2, \"depth (m)\"),\n\t\tDepthFathoms: p.Float64(4, \"depth (fathoms)\"),\n\t}, p.Err()\n}", "func New() NopPrinter { return NopPrinter{} }", "func newNetwork(config neuralNetConfig) *neuralNet {\n\treturn &neuralNet{config: config}\n}", "func New(addr string) *BTT {\n\treturn &BTT{addr: addr, client: http.DefaultClient}\n}", "func NewProtocol(bc blockchain.Blockchain) *Protocol { return &Protocol{bc} }", "func New() *binaryTree {\n\treturn CreateDefaultTree()\n}", "func New(config config.Config) Nrdb {\n\treturn Nrdb{\n\t\tclient: http.NewClient(config),\n\t\tlogger: config.GetLogger(),\n\t}\n}", "func New(s string, base int) (v BigNum) {\n\tv.Input(s, base)\n\treturn\n}", "func New() *BTrie {\n\treturn &BTrie{}\n}", "func New(api string, h Getter) *DogNZB {\n\treturn &DogNZB{\n\t\tapi: api,\n\t\th: h,\n\t}\n}", "func (tb *TextBuf) New(nlines int) {\n\ttb.Defaults()\n\tnlines = ints.MaxInt(nlines, 1)\n\ttb.LinesMu.Lock()\n\ttb.MarkupMu.Lock()\n\ttb.Lines = make([][]rune, nlines)\n\ttb.LineBytes = make([][]byte, nlines)\n\ttb.Tags = make([]lex.Line, nlines)\n\ttb.HiTags = make([]lex.Line, nlines)\n\ttb.Markup = make([][]byte, nlines)\n\n\tif cap(tb.ByteOffs) >= nlines {\n\t\ttb.ByteOffs = tb.ByteOffs[:nlines]\n\t} else {\n\t\ttb.ByteOffs = make([]int, nlines)\n\t}\n\n\tif nlines == 1 { // this is used for a new blank doc\n\t\ttb.ByteOffs[0] = 0 // by definition\n\t\ttb.Lines[0] = []rune(\"\")\n\t\ttb.LineBytes[0] = []byte(\"\")\n\t\ttb.Markup[0] = []byte(\"\")\n\t}\n\n\ttb.NLines = nlines\n\n\ttb.PiState.SetSrc(&tb.Lines, string(tb.Filename), tb.Info.Sup)\n\ttb.Hi.Init(&tb.Info, &tb.PiState)\n\n\ttb.MarkupMu.Unlock()\n\ttb.LinesMu.Unlock()\n\ttb.Refresh()\n}", "func New(dgree int, ctx interface{}) *BTree {\n\treturn NewWithFreeList(degree, NewFreeList(DefaultFreeListSize), ctx)\n}", "func NewBulb(ip string) *Bulb {\n\tbulb := &Bulb{\n\t\tstandardCommands{},\n\t\tcommonCommands{},\n\t\tbackgroundLightCommands{},\n\t\tip,\n\t\t55443, // 55443 is a constant protocol port\n\t\tnil,\n\t\tmake(map[int]chan Response),\n\t\tsync.Mutex{},\n\t}\n\t// I know It looks badly, but \"It is working? It is working\"\n\tbulb.standardCommands.commander = bulb\n\tbulb.commonCommands.commander = bulb\n\tbulb.Bg.commander = bulb\n\tbulb.Bg.prefix = \"bg_\"\n\treturn bulb\n}", "func NewNw() *Nw {\n\tmyNet := &Nw{}\n\treturn myNet\n}", "func NewNN(g *gorgonia.ExprGraph, S NetworkStruction) *NN {\n\t// Set random seed\n\trand.Seed(time.Now().Unix())\n\tvar Ns, Bs gorgonia.Nodes\n\tfor i := 0; i < len(S.Neuron)-1; i++ {\n\t\tNs = append(Ns, gorgonia.NewMatrix(\n\t\t\tg,\n\t\t\ttensor.Float64,\n\t\t\tgorgonia.WithShape(S.Neuron[i], S.Neuron[i+1]),\n\t\t\tgorgonia.WithName(\"w\"+strconv.Itoa(i)),\n\t\t\tgorgonia.WithInit(gorgonia.GlorotN(1)),\n\t\t))\n\t}\n\tif S.Bias {\n\t\tfor i := 0; i < len(S.Neuron)-1; i++ {\n\t\t\tBs = append(Bs, gorgonia.NewMatrix(\n\t\t\t\tg,\n\t\t\t\ttensor.Float64,\n\t\t\t\tgorgonia.WithShape(1, S.Neuron[i+1]),\n\t\t\t\tgorgonia.WithName(\"b\"+strconv.Itoa(i)),\n\t\t\t\tgorgonia.WithInit(gorgonia.Zeroes()),\n\t\t\t))\n\t\t}\n\t}\n\n\treturn &NN{\n\t\tG: g,\n\t\tW: Ns,\n\t\tB: Bs,\n\t\tD: S.Dropout,\n\t\tA: S.Act,\n\t\tNormal: S.Normal,\n\t\tL1reg: S.L1reg,\n\t\tL2reg: S.L2reg,\n\t}\n}", "func NewBot() TipBot {\n\t// create sqlite databases\n\tdb, txLogger := migration()\n\treturn TipBot{\n\t\tDatabase: db,\n\t\tClient: lnbits.NewClient(internal.Configuration.Lnbits.AdminKey, internal.Configuration.Lnbits.Url),\n\t\tlogger: txLogger,\n\t\tBunt: createBunt(),\n\t\tTelegram: newTelegramBot(),\n\t}\n}", "func (b *BaseImpl) New(n Base) Base {\n\treturn n\n}", "func CreateNewNL(parentNodeID int, childNodeID int, linkType string) (NodeLink, error) {\n\tif parentNodeID == 0 || childNodeID == 0 || linkType == \"\" {\n\t\treturn NodeLink{}, errors.New(\"Not enough arguments supplied\")\n\t}\n\treturn NodeLink{\n\t\tLinkPNodeID: parentNodeID,\n\t\tLinkCNodeID: childNodeID,\n\t\tLinkType: linkType,\n\t}, nil\n}", "func New(glabel string, flags int) (*GlvlStruct, error) {\n\treturn newll(glabel, flags, LflagsDef, nil, false)\n}", "func newNGram(n int) *nGram {\n\tngram := new(nGram)\n\tngram.nValue = n\n\treturn ngram\n}", "func (t *Blockchain) New() *Blockchain {\n\tt = new(Blockchain)\n\tt.NewBlock(100, \"1\")\n\treturn t\n}", "func newParser(b []byte) parser {\n\treturn parser{\n\t\tbytes: b,\n\t\ti: 0,\n\t}\n}", "func New() *Bst {\n\treturn new(Bst).Init()\n}", "func newNLMsg(nl nl.Listener) *notifListenMsg {\n\treturn &notifListenMsg{nl: nl}\n}", "func newLabel(value string, g *Graph) *Label {\n Assert(nilGraph, g != nil)\n Assert(nilTextStore, g.textStore != nil)\n \n l := new(Label)\n t := newText(value)\n l.value = g.textStore.addText(t)\n \n return l\n}", "func NewNet() Net {\n\treturn Net{items: make(map[string]Page)}\n}", "func New(port int, id string, weight int) Node {\n\treturn &node{\n\t\tmyPort: port,\n\t\tid: id,\n\t\tring: *consistent.NewRing(),\n\t\trepCh: make(chan replicaEx),\n\t\treqCh: make(chan requestEx),\n\t\trmvCh: make(chan removeEx),\n\t\tcpyCh: make(chan copyEx),\n\t\treplaceCh: make(chan replaceEx),\n\t\tlookupCh: make(chan lookupEx),\n\t\tbulkCh: make(chan bulkEx),\n\t\tstateCh: make(chan stateEx),\n\t\tweight: weight,\n\t\tstateMap: make(map[string]rpcs.State),\n\t}\n}", "func (t *BPTree) newLeaf() *Node {\n\tleaf := t.newNode()\n\tleaf.isLeaf = true\n\treturn leaf\n}", "func (p *protoOutputNode) New(attr string) outputNode {\n\tuc := nodePool.Get().(*graph.Node)\n\tuc.Attribute = attr\n\treturn &protoOutputNode{uc}\n}", "func New() *Novis {\n\treturn &Novis{\n\t\tRoot: &Branch{\n\t\t\tbranches: make(map[string]*Branch),\n\t\t},\n\t}\n}", "func NewLLCtor() *Node {\n\t//empty node is created, small memory allocated.\n\treturn new(Node)\n}", "func New() *RBTree {\n\treturn &RBTree{\n\t\tlock: sync.RWMutex{},\n\t\tNode: nil,\n\t\tstack: newStack(nil),\n\t}\n}", "func New() *Nitro {\n\treturn NewWithConfig(DefaultConfig())\n}", "func New() Go { return Go{} }", "func New(plArgs *runtime.Unknown, _ framework.FrameworkHandle) (framework.Plugin, error) {\n\targs := &Args{}\n\tif err := framework.DecodeInto(plArgs, args); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &NodeLabel{\n\t\tpredicate: predicates.NewNodeLabelPredicate(args.Labels, args.Presence),\n\t}, nil\n}", "func New(stack consensus.Stack) consensus.Consenter {\n\t\n\tconfigLock.Lock()\n\tif config == nil{\n\t\tconfig = loadConfig()\n\t}\n\tdefer configLock.Unlock()\n\t\n\thandle, _, _ := stack.GetNetworkHandles()\n\tid, _ := getValidatorID(handle)\n\n\tswitch strings.ToLower(config.GetString(\"general.mode\")) {\n\tcase \"batch\":\n\t\treturn newObcBatch(id, config, stack)\n\tdefault:\n\t\tpanic(fmt.Errorf(\"Invalid PBFT mode: %s\", config.GetString(\"general.mode\")))\n\t}\n}", "func NewLoadBalancer() *LoadBalancer {\n\tlb := &LoadBalancer{\n\t\tnodes: make(map[string]*weightedNode),\n\t}\n\treturn lb\n}", "func newNATS(url string) (*NATS, error) {\n\tc, err := nats.Connect(url)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &NATS{conn: c}, nil\n}", "func New(opts ...Option) *HLTB {\n\th := &HLTB{}\n\n\tfor _, opt := range opts {\n\t\topt(h)\n\t}\n\n\treturn h\n}", "func New(root string) *Bag {\n\treturn &Bag{root: root, Hasher: HashSHA256}\n}", "func NewBBS() (*BBS, error) {\n\treturn &BBS{\n\t\tlines: []Line{},\n\t\tconnections: make(chan io.ReadWriter),\n\t}, nil\n}", "func New(ent []byte, lang Language) (*Mnemonic, error) {\n\tconst chunkSize = 11\n\tbits := entropy.CheckSummed(ent)\n\tlength := len(bits)\n\twords := make([]string, length/11)\n\tfor i := 0; i < length; i += chunkSize {\n\t\tstringVal := string(bits[i : chunkSize+i])\n\t\tintVal, err := strconv.ParseInt(stringVal, 2, 64)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Could not convert %s to word index\", stringVal)\n\t\t}\n\t\tword, err := GetWord(lang, intVal)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\twords[(chunkSize+i)/11-1] = word\n\t}\n\tm := Mnemonic{words, lang}\n\treturn &m, nil\n}", "func New(pg *pg.DB, index Index) *Model {\n\tInitTokenize()\n\treturn &Model{\n\t\tPG: pg,\n\t\tIndex: index,\n\t\tFiles: make(map[string]int),\n\t\tWords: make(map[string]int),\n\t}\n}", "func New() Builder {\n\treturn newBuilder()\n}", "func New(b builder.Builder, tracker, hosted string) *Builder {\n\t//create our new builder\n\tn := &Builder{\n\t\tb: b,\n\t\tbase: hosted,\n\t\trpc: gorpc.NewServer(),\n\t\ttcl: client.New(tracker, http.DefaultClient, client.JsonCodec),\n\t\tbq: rpc.NewBuilderQueue(),\n\t\tmux: http.NewServeMux(),\n\t\tdler: newDownloader(),\n\t}\n\n\t//register the build service in the rpc\n\tif err := n.rpc.RegisterService(n.bq, \"\"); err != nil {\n\t\tpanic(err)\n\t}\n\n\t//make sure we respond to pings\n\tif err := n.rpc.RegisterService(pinger.Pinger{}, \"\"); err != nil {\n\t\tpanic(err)\n\t}\n\n\t//register the codec\n\tn.rpc.RegisterCodec(json.NewCodec(), \"application/json\")\n\n\t//add the handlers to our mux\n\tn.mux.Handle(\"/\", n.rpc)\n\tn.mux.Handle(\"/download/\", http.StripPrefix(\"/download/\", n.dler))\n\n\t//start processing tasks\n\tgo n.run()\n\n\treturn n\n}", "func New() *Builder {\n\treturn &Builder{\n\t\tID: getulid(),\n\t}\n}", "func (n NetworkTypeNone) construct() NetworkTypeClass { return &n }", "func New(pN string, b uint32, p string) *Rfm12b {\n\treturn &Rfm12b{\n\t\tportName: pN, baud: b, loggerPath: p,\n\t\tchOut: make(chan []byte), ChIn: make(chan interface{}),\n\t\tlogger: logger.New(p),\n\t}\n}", "func New(data []byte) *Parser {\n\treturn &Parser{\n\t\tbook: data,\n\t}\n}", "func (t *tokenSN) NLP() *NLP {\n\treturn &t.nlp\n}", "func New(b []byte) *Buffer {\n\treturn &Buffer{b: b}\n}", "func New(length int) (nid string) {\n nid = \"\"\n\n for i := 0; i < length; i++ {\n var (\n r1 int\n )\n\n r1 = rnd.Intn(9)\n\n if i == 0 {\n for r1 == 0 {\n r1 = rnd.Intn(9)\n }\n }\n\n nid += strconv.Itoa(r1)\n }\n return\n}", "func newWordGraph(n int) wordGraph {\n\treturn wordGraph{\n\t\tn: n,\n\t\tids: make(map[string]int64),\n\t\tUndirectedGraph: simple.NewUndirectedGraph(),\n\t}\n}", "func New() *Nap {\n\treturn &Nap{\n\t\thttpClient: defaultClient,\n\t\tmethod: MethodGet,\n\t\theader: make(http.Header),\n\t\tqueryStructs: make([]interface{}, 0),\n\t\tqueryParams: make(map[string]string),\n\t\tresponseDecoder: jsonDecoder{},\n\t\tisSuccess: DecodeOnSuccess,\n\t}\n}", "func New() Tree {\n\treturn &binarySearchTree{}\n}", "func newLoadBalancer(ctx context.Context, frontend NetAddr, policy loadBalancerPolicy, backends ...NetAddr) (*LoadBalancer, error) {\n\tif ctx == nil {\n\t\treturn nil, trace.BadParameter(\"missing parameter context\")\n\t}\n\twaitCtx, waitCancel := context.WithCancel(ctx)\n\treturn &LoadBalancer{\n\t\tfrontend: frontend,\n\t\tctx: ctx,\n\t\tbackends: backends,\n\t\tpolicy: policy,\n\t\twaitCtx: waitCtx,\n\t\twaitCancel: waitCancel,\n\t\tEntry: log.WithFields(log.Fields{\n\t\t\ttrace.Component: \"loadbalancer\",\n\t\t\ttrace.ComponentFields: log.Fields{\n\t\t\t\t\"listen\": frontend.String(),\n\t\t\t},\n\t\t}),\n\t\tconnections: make(map[NetAddr]map[int64]net.Conn),\n\t}, nil\n}", "func NewBulb(address string) (*Bulb, error) {\n\tif !strings.Contains(address, \":\") {\n\t\taddress = address + \":55443\"\n\t}\n\tconn, err := net.Dial(\"tcp\", address)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not dial address: %+v\", err)\n\t}\n\treturn &Bulb{\n\t\tconn: conn,\n\t}, nil\n}", "func New() *VerbalExpression {\n\tr := new(VerbalExpression)\n\tr.flags = MULTILINE | GLOBAL\n\tr.parts = make([]string, 0)\n\treturn r\n}", "func (db *LDB) NewBatch() Batch {\n return &LDBBatch{\n db: db.db,\n batch: new(leveldb.Batch),\n }\n}", "func (t *BPTree) newNode() *Node {\n\tnode := &Node{\n\t\tKeys: make([][]byte, order-1),\n\t\tpointers: make([]interface{}, order),\n\t\tisLeaf: false,\n\t\tparent: nil,\n\t\tKeysNum: 0,\n\t\tAddress: t.LastAddress,\n\t}\n\tsize := getBinaryNodeSize()\n\tt.LastAddress += size\n\n\treturn node\n}", "func New(n int) *TopN {\n\treturn &TopN{\n\t\tn: n,\n\t\tm: make(map[string]*httpstat.HTTPStat),\n\t\tuca: urlclusteranalyzer.New(),\n\t}\n}", "func NewLob(baseAPI, apiKey, userAgent string) *lob {\n\treturn &lob{\n\t\tBaseAPI: baseAPI,\n\t\tAPIKey: apiKey,\n\t\tUserAgent: userAgent,\n\t}\n}", "func NewLoadBalance(name string) LoadBalance {\n\treturn LBS[name]()\n}", "func newMaxentClassifier(\n\tweights []float64,\n\tmapping map[string]int,\n\tlabels []string) *binaryMaxentClassifier {\n\n\tset := mapset.NewSet()\n\tfor label := range mapping {\n\t\tset.Add(strings.Split(label, \"-\")[0])\n\t}\n\n\treturn &binaryMaxentClassifier{\n\t\tset.Cardinality() + 1,\n\t\tlabels,\n\t\tmapping,\n\t\tweights}\n}", "func newBrainfog(bfSrc []byte) *brainfog {\n\tbf := &brainfog{inCh: make(chan byte), outCh: make(chan byte)}\n\n\t// Pick the instructions from the source and add them to the program\n\tinstructions := []byte(\"+-<>,.[]\")\n\tfor _, c := range bfSrc {\n\t\tif bytes.Contains(instructions, []byte{c}) {\n\t\t\tbf.program = append(bf.program, c)\n\t\t}\n\t}\n\n\t// Run the program\n\tgo bf.run()\n\treturn bf\n}", "func NewNetwork(x int, n float64, a float64, act int, b float64, h int, r int, o int) Network{\n\tnet := Network{Lrate: n, Mom: a, Bias: b, Activation: act, Inputs: x, Outputs: o, Hidden: h}\n\tfor i := 0; i < r; i++ {\n\t\tif i == 0 {\n\t\t\tl := L.NewHidden(x, n, a, act, b, h)\n\t\t\tnet.HiddenLayers = append(net.HiddenLayers, l)\t\t\t\n\t\t} else {\n\t\t\tl := L.NewHidden(h, n, a, act, b, h)\t\n\t\t\tnet.HiddenLayers = append(net.HiddenLayers, l)\t\t\t\n\t\t}\n\t}\n\tnet.OutputLayer = L.NewOutput(h, n, a, act, b, o)\n\n\treturn net\n}", "func (f *Forest) New(d *crypto.Digest, l uint32) *Tree {\n\tt := &Tree{\n\t\tleaves: l,\n\t\tdig: d,\n\t\tf: f,\n\t\tlastBlockLen: BlockSize,\n\t\tleavesComplete: make([]bool, l),\n\t}\n\tf.writeTree(t)\n\treturn t\n}", "func New(text string) (err error) {\n\treturn backendErr(errors.New(text))\n}", "func newBloblang(conf BloblangConfig) (*Bloblang, error) {\n\tvar timer *time.Ticker\n\tif len(conf.Interval) > 0 {\n\t\tduration, err := time.ParseDuration(conf.Interval)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse interval: %w\", err)\n\t\t}\n\t\ttimer = time.NewTicker(duration)\n\t}\n\texec, err := bloblang.NewMapping(\"\", conf.Mapping)\n\tif err != nil {\n\t\tif perr, ok := err.(*parser.Error); ok {\n\t\t\treturn nil, fmt.Errorf(\"failed to parse mapping: %v\", perr.ErrorAtPosition([]rune(conf.Mapping)))\n\t\t}\n\t\treturn nil, fmt.Errorf(\"failed to parse mapping: %v\", err)\n\t}\n\tremaining := int32(conf.Count)\n\tif remaining <= 0 {\n\t\tremaining = -1\n\t}\n\treturn &Bloblang{\n\t\texec: exec,\n\t\tremaining: remaining,\n\t\ttimer: timer,\n\t\tfirstIsFree: true,\n\t}, nil\n}", "func New(dir string, valueDir string) (Backend, error) {\n\topts := badger.DefaultOptions\n\topts.Dir = dir\n\topts.ValueDir = valueDir\n\n\tdb, err := badger.Open(opts)\n\treturn Backend{db}, err\n}", "func New(t testing.TB) *T {\n\treturn &T{t}\n}", "func New() *Parser {\n\treturn &Parser{\n\t\tWords: make(map[string]*wordRef),\n\t}\n}", "func NewBinaryTree(vals ...interface{}) (res *BinaryTreeNode, err error) {\n\tif len(vals) == 0 {\n\t\treturn nil, ErrEmptyInput\n\t}\n\tif res, err = createNode(vals[0]); err != nil {\n\t\treturn\n\t}\n\terr = buildTree([]*BinaryTreeNode{res}, 1, vals)\n\treturn\n}", "func newBuffer(b []byte) *buffer {\n\treturn &buffer{proto.NewBuffer(b), 0}\n}", "func New(numVisibleUnits, numHiddenUnits int) *RBM {\n\trbm := new(RBM)\n\trand.Seed(time.Now().UnixNano())\n\trbm.NumVisibleUnits = numVisibleUnits\n\trbm.NumHiddenUnits = numHiddenUnits\n\trbm.W = nnet.MakeMatrix(numHiddenUnits, numVisibleUnits)\n\trbm.B = make([]float64, numVisibleUnits)\n\trbm.C = make([]float64, numHiddenUnits)\n\trbm.GradW = nnet.MakeMatrix(numHiddenUnits, numVisibleUnits)\n\trbm.GradB = make([]float64, numVisibleUnits)\n\trbm.GradC = make([]float64, numHiddenUnits)\n\trbm.InitParam()\n\treturn rbm\n}", "func New(indexer cache.Store) *Manager {\n\tlogger := &bgplog.Logger{Entry: log}\n\tc := &metallbctl.Controller{\n\t\tClient: bgpk8s.New(logger.Logger),\n\t\tIPs: metallballoc.New(),\n\t}\n\n\tf, err := os.Open(option.Config.BGPConfigPath)\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"Failed to open BGP config file\")\n\t}\n\tdefer f.Close()\n\n\tconfig, err := bgpconfig.Parse(f)\n\tif err != nil {\n\t\tlog.WithError(err).Fatal(\"Failed to parse BGP configuration\")\n\t}\n\tc.SetConfig(logger, config)\n\n\tmgr := &Manager{\n\t\tController: c,\n\t\tlogger: logger,\n\n\t\tqueue: workqueue.New(),\n\t\tindexer: indexer,\n\t}\n\tgo mgr.run()\n\n\treturn mgr\n}", "func NewLBAgent(da DataAccess, rp pool.RunnerPool, p pool.Placer) (Agent, error) {\n\tagent := createAgent(da, false)\n\ta := &lbAgent{\n\t\tdelegatedAgent: agent,\n\t\trp: rp,\n\t\tplacer: p,\n\t}\n\treturn a, nil\n}", "func newProtobuf(typeName string) *protobuf {\n\treturn &protobuf{\n\t\tprotobufFieldSequence: newProtobufFieldSequence(false),\n\t\tTypeName: typeName,\n\t}\n}", "func (nm *nodeManager) newNode(path string, args []string, printToStdOut bool) (*nodeProcess, error) {\n\tnm.log.Debug(\"creating new node from binary at '%s'\", path)\n\tclientConfig := &plugin.ClientConfig{\n\t\tHandshakeConfig: appplugin.Handshake,\n\t\tPlugins: appplugin.PluginMap,\n\t\tCmd: subprocess.New(path, args...),\n\t\tAllowedProtocols: []plugin.Protocol{plugin.ProtocolGRPC},\n\t\tLogger: hclog.New(&hclog.LoggerOptions{Level: hclog.Error}),\n\t}\n\tif printToStdOut {\n\t\tclientConfig.SyncStdout = os.Stdout\n\t\tclientConfig.SyncStderr = os.Stderr\n\t}\n\n\tclient := plugin.NewClient(clientConfig)\n\trpcClient, err := client.Client()\n\tif err != nil {\n\t\tclient.Kill()\n\t\treturn nil, fmt.Errorf(\"couldn't get client at path %s: %w\", path, err)\n\t}\n\n\traw, err := rpcClient.Dispense(\"nodeProcess\")\n\tif err != nil {\n\t\tclient.Kill()\n\t\treturn nil, fmt.Errorf(\"couldn't dispense plugin at path %s': %w\", path, err)\n\t}\n\n\tnode, ok := raw.(*appplugin.Client)\n\tif !ok {\n\t\tclient.Kill()\n\t\treturn nil, fmt.Errorf(\"expected *node.NodeClient but got %T\", raw)\n\t}\n\n\tnp := &nodeProcess{\n\t\tlog: nm.log,\n\t\tnode: node,\n\t\trawClient: client,\n\t\tpath: path,\n\t}\n\tnm.nodes[np.path] = np\n\treturn np, nil\n}", "func NewBinaryTree(vals []Comparable) *BinaryTree {\n\treturn new(BinaryTree).Init(vals)\n}", "func New() broker.Broker {\n\treturn &natsBroker{\n\t\tsubscriptionMap: make(map[string]*natsSubscriber),\n\t}\n}", "func (n NetworkTypeOther) construct() NetworkTypeClass { return &n }", "func newBlock(t nbt.Tag) BlockState {\r\n\tblock := BlockState{}\r\n\tblock.Name = t.Compound()[\"Name\"].String()\r\n\tblock.parseProperties(t)\r\n\treturn block\r\n}", "func New(charset string) *N36 {\n\tn := &N36{\n\t\tcharset: charset,\n\t\tseedRand: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}\n\n\treturn n\n}", "func New() *SinglyLinkedList {\n return new(SinglyLinkedList).Init()\n}", "func NewNfLog(McastGroup int, IpVersion byte, Direction IpDirection, MaskBits int, a *Accounting) *NfLog {\n\th, err := C.nflog_open()\n\tif h == nil || err != nil {\n\t\tlog.Fatalf(\"Failed to open NFLOG: %s\", nflogError(err))\n\t}\n\tif *Verbose {\n\t\tlog.Println(\"Binding nfnetlink_log to AF_INET\")\n\t}\n\tif rc, err := C.nflog_bind_pf(h, C.AF_INET); rc < 0 || err != nil {\n\t\tlog.Fatalf(\"nflog_bind_pf failed: %s\", nflogError(err))\n\t}\n\n\tnflog := &NfLog{\n\t\th: h,\n\t\tfd: C.nflog_fd(h),\n\t\tMcastGroup: McastGroup,\n\t\tIpVersion: IpVersion,\n\t\tDirection: Direction,\n\t\ta: a,\n\t\tquit: make(chan struct{}),\n\t\tpackets: (*C.packets)(C.malloc(C.sizeof_packets)),\n\t}\n\tfor i := range nflogs {\n\t\tif nflogs[i] == nil {\n\t\t\tnflog.index = i\n\t\t\tnflogs[i] = nflog\n\t\t\tgoto found\n\t\t}\n\t}\n\tlog.Fatal(\"Too many filters\")\nfound:\n\tswitch IpVersion {\n\tcase 4:\n\t\tnflog.IpPacket = Ip4Packet\n\tcase 6:\n\t\tnflog.IpPacket = Ip6Packet\n\tdefault:\n\t\tlog.Fatalf(\"Bad IP version %d\", IpVersion)\n\t}\n\taddrBits := 8 * nflog.IpPacket.AddrLen\n\tnflog.UseMask = MaskBits < addrBits\n\tnflog.Mask = net.CIDRMask(MaskBits, addrBits)\n\tnflog.makeGroup(McastGroup, nflog.IpPacket.HeaderSize)\n\t// Start the background process\n\tgo nflog.Loop()\n\treturn nflog\n}", "func New(data []byte) *Index {}", "func New() Tree {\n\treturn &Node{Value: \".\"}\n}", "func New() *Baa {\n\tb := new(Baa)\n\tb.middleware = make([]HandlerFunc, 0)\n\tb.pool = sync.Pool{\n\t\tNew: func() interface{} {\n\t\t\treturn NewContext(nil, nil, b)\n\t\t},\n\t}\n\tif Env != PROD {\n\t\tb.debug = true\n\t}\n\tb.SetDIer(NewDI())\n\tb.SetDI(\"router\", NewTree(b))\n\tb.SetDI(\"logger\", log.New(os.Stderr, \"[Baa] \", log.LstdFlags))\n\tb.SetDI(\"render\", newRender())\n\tb.SetNotFound(b.DefaultNotFoundHandler)\n\treturn b\n}", "func newNode() *node {\n\tvar leafs [8]octant\n\tfor i := 0; i < 8; i++ {\n\t\tleafs[i] = newLeaf(nil)\n\t}\n\treturn &node{\n\t\tleafs: &leafs,\n\t}\n}", "func newPerson(name string,class string, nationality string ) *Person {\n\treturn &Person{name: name,job: class, nationality: nationality}\n\n}", "func New(it *msvc.ProjectIterator) *Graph {\n\tgr := &Graph{\n\t\tg: simple.NewDirectedGraph(),\n\t\tallNodes: rbtree.New(),\n\t\tnextID: 1,\n\t}\n\n\tit.Foreach(gr.newNode)\n\tait := rbtree.NewWalkInorder(gr.allNodes)\n\tait.Foreach(gr.newEdges)\n\n\treturn gr\n}", "func (tl *TextBufList) New() *TextBuf {\n\ttb := tl.AddNewChild(KiT_TextBuf, \"newbuf\").(*TextBuf)\n\treturn tb\n}", "func New(s *service.Canal) (engine *bm.Engine, err error) {\n\tcs = s\n\tengine = bm.DefaultServer(conf.Conf.BM)\n\n\tinitRouter(engine)\n\terr = engine.Start()\n\treturn\n}", "func NewProtocol(idx *indexservice.Indexer, cfg indexprotocol.HermesConfig) *Protocol {\n\treturn &Protocol{\n\t\tindexer: idx,\n\t\thermesConfig: cfg,\n\t}\n}", "func NewLoadBalancer(c config.LoadBalancerConfig) *LoadBalancer {\n\tvar lb LoadBalancer\n\tif c.Hosts != nil && len(c.Hosts) > 0 {\n\t\tlb.hosts = make([]string, len(c.Hosts))\n\t\tfor i, server := range c.Hosts {\n\t\t\tlb.hosts[i] = server\n\t\t\tgloballog.WithFields(logrus.Fields{\n\t\t\t\t\"host\": server,\n\t\t\t\t\"index\": i,\n\t\t\t}).Debug(\"adding lb host\")\n\t\t}\n\t} else {\n\t\tlb.hosts = make([]string, 10)\n\t}\n\tlb.mode = c.BalanceMode\n\tlb.hostLock = new(sync.RWMutex)\n\treturn &lb\n}", "func New(value interface{}, comparator comparator.Less) *RBTree {\n\treturn &RBTree{value: value, less: comparator, color: \"black\"}\n}" ]
[ "0.6301168", "0.62660307", "0.61914414", "0.61369634", "0.6089232", "0.6012083", "0.5968851", "0.58928835", "0.58850867", "0.58754504", "0.58220947", "0.5820667", "0.5808736", "0.5808183", "0.5801838", "0.5788649", "0.5766198", "0.57396877", "0.5721118", "0.56853133", "0.56604654", "0.564914", "0.56357443", "0.5635449", "0.560978", "0.56008554", "0.5587341", "0.55748767", "0.5564527", "0.55637753", "0.5563104", "0.55600286", "0.5552722", "0.5539268", "0.5529154", "0.5523677", "0.55211097", "0.5516548", "0.54970795", "0.54949486", "0.54946584", "0.54791015", "0.5472154", "0.54536885", "0.5442339", "0.54381627", "0.5433661", "0.5419976", "0.5414203", "0.5410814", "0.539272", "0.5392045", "0.53899837", "0.53896815", "0.5384941", "0.53819644", "0.5375944", "0.5375512", "0.5371072", "0.5369285", "0.5361803", "0.5361073", "0.5359359", "0.53557324", "0.5355704", "0.53537536", "0.535263", "0.5340964", "0.53409004", "0.53375727", "0.53340554", "0.5333666", "0.5333601", "0.5328745", "0.5325619", "0.5324726", "0.5322515", "0.53219485", "0.53183347", "0.5315561", "0.53108966", "0.53077906", "0.5304356", "0.5304347", "0.5295266", "0.52932805", "0.52868617", "0.52846307", "0.52830786", "0.52797455", "0.5275612", "0.52753043", "0.5274533", "0.52739257", "0.52644426", "0.52616775", "0.5260135", "0.52600193", "0.5258358", "0.52574235" ]
0.7635804
0
CreateBucket creates or returns a sub bucket by the specified name. TODO: implement sub buckets
func (b *Bucket) CreateBucket(name string) (*Bucket, error) { panic("n/i") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func createBucket(parentBucket *bbolt.Bucket, name []byte) (*bbolt.Bucket, error) {\n\treturn parentBucket.CreateBucket(name)\n}", "func (b *Bucket) CreateBucket( name string ) (*Bucket, error) {\n nb, err := b.bucket.CreateBucket( []byte(name) )\n if err != nil {\n return nil, err\n }\n return &Bucket{tx: b.tx, bucket: nb}, nil\n}", "func createBucket(name string) func(*bolt.Tx) error {\n\treturn func(tx *bolt.Tx) error {\n\t\tif _, err := tx.CreateBucketIfNotExists([]byte(name)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n}", "func (r *RBucket) CreateBucket(name []byte) (Bucket, error) {\n\treq := &BucketRequest{}\n\treq.Key = name\n\tresp := &BucketResponse{}\n\treq.ContextID = r.parent\n\terr := r.r.call(\"srv.CreateBucket\", req, resp)\n\tb := &RBucket{}\n\tb.tx = r.tx\n\tb.r = r.r\n\tb.id = resp.BucketID\n\tb.parent = resp.BucketContextID\n\treturn b, err\n}", "func BucketCreateChildBucket(b *bolt.Bucket, childBucketName string) (*bolt.Bucket, error) {\n\tchildBucket, err := b.CreateBucket([]byte(childBucketName))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn childBucket, nil\n}", "func CreateBucket(storageCli *storage.Client, project string, f *framework.Framework) string /*bucket Name*/ {\n\tbucketName := f.UniqueName\n\n\tbucketAttrs := &storage.BucketAttrs{\n\t\t// Force single-region by setting an explicit region.\n\t\tLocation: \"us-east1\",\n\t\tLabels: gcloud.TagsFor(f),\n\t}\n\n\tif err := storageCli.Bucket(bucketName).Create(context.Background(), project, bucketAttrs); err != nil {\n\t\tframework.FailfWithOffset(2, \"Failed to create bucket %q: %s\", bucketName, err)\n\t}\n\n\treturn bucketName\n}", "func (b *BucketService) CreateBucket(hash entities.Hash) (*entities.Bucket, error) {\n\tif err := hash.Validation(); err != nil {\n\t\treturn nil, err\n\t}\n\tvar bucket *entities.Bucket\n\tduration := time.Second * time.Duration(b.Settings.Duration)\n\n\tswitch hash.Kind {\n\tcase entities.Login:\n\t\tbucket = entities.NewBucket(b.Settings.LoginLimit, duration, hash, b.Callback)\n\tcase entities.Password:\n\t\tbucket = entities.NewBucket(b.Settings.PasswordLimit, duration, hash, b.Callback)\n\tcase entities.IP:\n\t\tbucket = entities.NewBucket(b.Settings.IPLimit, duration, hash, b.Callback)\n\tdefault:\n\t\treturn nil, exceptions.TypeNotFound\n\t}\n\terr := b.BucketStore.Add(hash, bucket)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn bucket, nil\n}", "func CreateBucket(w http.ResponseWriter, r *http.Request) {\n\n\tpageVars := PageVars{}\n\taddPageVars(r, &pageVars)\n\n\trender(w, \"createbucket\", pageVars)\n}", "func (s *Service) CreateBucket(ctx context.Context,\n\tthreadID thread.ID,\n\tthreadToken thread.Token,\n\tbuckName string,\n\tbuckPrivate bool,\n\tdataCid cid.Cid) error {\n\t_, _, _, err := s.createBucket(ctx, threadID, threadToken, buckName, buckPrivate, dataCid)\n\treturn err\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.DestinationRegion == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'DestinationRegion'\")\n\t}\n\tvar resource Bucket\n\terr := ctx.RegisterRemoteComponentResource(\"replicatedbucket:index:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func CreateBucket(provider string, bucketName string) {\n\tStep(fmt.Sprintf(\"Create bucket [%s]\", bucketName), func() {\n\t\tswitch provider {\n\t\tcase drivers.ProviderAws:\n\t\t\tCreateS3Bucket(bucketName)\n\t\tcase drivers.ProviderAzure:\n\t\t\tCreateAzureBucket(bucketName)\n\t\t}\n\t})\n}", "func Create(c *gin.Context) {\n\tlogger := correlationid.LogrusLogger(common.Log, c)\n\tlogger.Info(\"creating bucket\")\n\n\tvar request api.CreateBucketRequest\n\tif err := c.ShouldBindJSON(&request); err != nil {\n\t\terr = errors.WrapIf(err, \"could not parse request\")\n\t\tcommon.ErrorHandler.Handle(err)\n\t\tcommon.ErrorResponse(c, err)\n\t\treturn\n\t}\n\n\torg := auth.GetCurrentOrganization(c.Request)\n\n\tif len(request.Location) == 0 && (request.Cloud == providers.Amazon) {\n\t\t// location field is empty in request, get bucket location\n\t\tlocation, err := common.GetBucketLocation(request.Cloud, request.BucketName, request.SecretID, org.ID, logger)\n\t\tif err != nil {\n\t\t\terr = errors.WrapIfWithDetails(err, \"failed to get bucket region\", \"bucket\", request.BucketName)\n\t\t\tcommon.ErrorHandler.Handle(err)\n\t\t\tcommon.ErrorResponse(c, err)\n\t\t\treturn\n\t\t}\n\n\t\trequest.Location = location\n\t}\n\n\tbs := ark.BucketsServiceFactory(org, global.DB(), logger)\n\n\t_, err := bs.GetByRequest(api.FindBucketRequest{\n\t\tCloud: request.Cloud,\n\t\tBucketName: request.BucketName,\n\t\tLocation: request.Location,\n\t})\n\tif err == nil {\n\t\terr = errors.New(\"bucket already exists\")\n\t}\n\tif err != nil && err != gorm.ErrRecordNotFound {\n\t\terr = errors.WrapIf(err, \"could not create bucket\")\n\t\tcommon.ErrorHandler.Handle(err)\n\t\tcommon.ErrorResponse(c, err)\n\t\treturn\n\t}\n\n\tbucket, err := bs.FindOrCreateBucket(&api.CreateBucketRequest{\n\t\tCloud: request.Cloud,\n\t\tBucketName: request.BucketName,\n\t\tLocation: request.Location,\n\t\tSecretID: request.SecretID,\n\t})\n\tif err != nil {\n\t\terr = errors.WrapIf(err, \"could not persist bucket\")\n\t\tcommon.ErrorHandler.Handle(err)\n\t\tcommon.ErrorResponse(c, err)\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, bucket.ConvertModelToEntity())\n}", "func Bucket_create(db_name string, bucket_name string) string{\n\t// Open the my.db data_backup file in your current directory.\n\t// It will be created if it doesn't exist.\n\tdb, err := bolt.Open(db_name, 0600, nil)\n\tdefer db.Close()\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\tdb.Close()\n\t\treturn \"ERROR,DB_OPEN\"\n\t}\n\n\ttx, err := db.Begin(true)\n\tif err != nil {\n\t\tfmt.Println(\"Error : \",err)\n\t\tdb.Close()\n\t\treturn \"ERROR,DB_OPEN\"\n\t}\n\n\ttx.CreateBucket([]byte(bucket_name))\n\t// Commit the transaction and check for error.\n\tif err := tx.Commit(); err != nil {\n\t\tfmt.Println(err)\n\t\tdb.Close()\n\t\treturn \"ERROR,BUCKET_COMMIT\"\n\t}\n\n\tdb.Close()\n\treturn \"OK_BUCKET_COMMIT\"\n}", "func (s *BucketService) CreateBucket(ctx context.Context, b *influxdb.Bucket) error {\n\tspan, ctx := tracing.StartSpanFromContext(ctx)\n\tdefer span.Finish()\n\n\tp, err := influxdb.NewPermission(influxdb.WriteAction, influxdb.BucketsResourceType, b.OrgID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := IsAllowed(ctx, *p); err != nil {\n\t\treturn err\n\t}\n\n\treturn s.s.CreateBucket(ctx, b)\n}", "func (driver donutDriver) CreateBucket(bucket string) error {\n\treturn notImplemented()\n}", "func (store *Storage) CreateBucket(bucket string) error {\n\tif err := store.first.CreateBucket(bucket); err != nil {\n\t\treturn common.Error(common.WriteFailed, errors.New(\"first level fail\"), err)\n\t}\n\tif err := store.second.CreateBucket(bucket); err != nil {\n\t\treturn common.Error(common.WriteFailed, errors.New(\"second level fail\"), err)\n\t}\n\treturn nil\n}", "func (svc *S3) S3CreateBucket(bucketName string) error {\n\n\tin, err := NewCreateBucketInput(bucketName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = svc.S3.CreateBucket(in)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (gcs *StorageConnection) CreateBucket(ctx context.Context) (err error) {\n\tvar attrs *storage.BucketAttrs\n\t// Setup client bucket to work from\n\tgcs.bucket = gcs.client.Bucket(gcs.BucketName)\n\n\tbuckets := gcs.client.Buckets(ctx, gcs.ProjectID)\n\tfor {\n\t\t// TODO bucket name validation\n\t\tif gcs.BucketName == \"\" {\n\t\t\terr = fmt.Errorf(\"BucketName entered is empty %v. Re-enter.\", gcs.BucketName)\n\t\t\treturn\n\t\t}\n\t\tattrs, err = buckets.Next()\n\t\t// Assume that if Iterator end then not found and need to create bucket\n\t\tif err == iterator.Done {\n\t\t\t// Create bucket if it doesn't exist - https://cloud.google.com/storage/docs/reference/libraries\n\t\t\tif err = gcs.bucket.Create(ctx, gcs.ProjectID, &storage.BucketAttrs{\n\t\t\t\tLocation: \"US\",\n\t\t\t}); err != nil {\n\t\t\t\t// TODO - add random number to append to bucket name to resolve\n\t\t\t\treturn fmt.Errorf(\"Failed to create bucket: %v\", err)\n\n\t\t\t}\n\t\t\tlog.Printf(\"Bucket %v created.\\n\", gcs.BucketName)\n\t\t\treturn\n\t\t}\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"Issues setting up Bucket: %q due to error: %w. Double check project id.\", attrs.Name, err)\n\t\t\treturn\n\t\t}\n\t\tif attrs.Name == gcs.BucketName {\n\t\t\t//getLatestFile() // TODO set this up to check and compare what is in the bucket vs what isn't\n\t\t\tlog.Printf(\"Bucket %v exists.\\n\", gcs.BucketName)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (l *LBucket) CreateBucket(name []byte) (Bucket, error) {\n\tb, err := l.b.CreateBucket(name)\n\n\treturn &LBucket{\n\t\tb: b,\n\t}, err\n}", "func (a *FileStorageApiService) CreateBucket(Ctx _context.Context) ApiCreateBucketRequest {\n\treturn ApiCreateBucketRequest{\n\t\tApiService: a,\n\t\tCtx: Ctx,\n\t}\n}", "func (db *DB) CreateBucket(ctx context.Context, bucketName string) (newBucket storj.Bucket, err error) {\n\tdefer mon.Task()(&ctx)(&err)\n\n\tif bucketName == \"\" {\n\t\treturn storj.Bucket{}, storj.ErrNoBucket.New(\"\")\n\t}\n\n\tnewBucket, err = db.metainfo.CreateBucket(ctx, CreateBucketParams{\n\t\tName: []byte(bucketName),\n\t})\n\treturn newBucket, storj.ErrBucket.Wrap(err)\n}", "func (s *Service) createBucket(\n\tctx context.Context,\n\tdbID thread.ID,\n\tdbToken thread.Token,\n\tname string,\n\tprivate bool,\n\tbootCid cid.Cid,\n) (nctx context.Context, buck *tdb.Bucket, seed ipld.Node, err error) {\n\tvar owner thread.PubKey\n\tif dbToken.Defined() {\n\t\towner, err = dbToken.PubKey()\n\t\tif err != nil {\n\t\t\treturn ctx, nil, nil, fmt.Errorf(\"creating bucket: invalid token public key\")\n\t\t}\n\t}\n\n\t// Create bucket keys if private\n\tvar linkKey, fileKey []byte\n\tif private {\n\t\tvar err error\n\t\tlinkKey, err = dcrypto.NewKey()\n\t\tif err != nil {\n\t\t\treturn ctx, nil, nil, err\n\t\t}\n\t\tfileKey, err = dcrypto.NewKey()\n\t\tif err != nil {\n\t\t\treturn ctx, nil, nil, err\n\t\t}\n\t}\n\n\t// Make a random seed, which ensures a bucket's uniqueness\n\tseed, err = makeSeed(fileKey)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Create the bucket directory\n\tvar buckPath path.Resolved\n\tif bootCid.Defined() {\n\t\tctx, buckPath, err = s.createBootstrappedPath(ctx, \"\", seed, bootCid, linkKey, fileKey)\n\t\tif err != nil {\n\t\t\treturn ctx, nil, nil, fmt.Errorf(\"creating prepared bucket: %v\", err)\n\t\t}\n\t} else {\n\t\tctx, buckPath, err = s.createPristinePath(ctx, seed, linkKey)\n\t\tif err != nil {\n\t\t\treturn ctx, nil, nil, fmt.Errorf(\"creating pristine bucket: %v\", err)\n\t\t}\n\t}\n\n\t// Create top-level metadata\n\tnow := time.Now()\n\tmd := map[string]tdb.Metadata{\n\t\t\"\": tdb.NewDefaultMetadata(owner, fileKey, now),\n\t\tbuckets.SeedName: {\n\t\t\tRoles: make(map[string]buckets.Role),\n\t\t\tUpdatedAt: now.UnixNano(),\n\t\t},\n\t}\n\n\t// Create a new IPNS key\n\tbuckKey, err := s.IPNSManager.CreateKey(ctx, dbID, buckPath)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Create the bucket using the IPNS key as instance ID\n\tbuck, err = s.Buckets.New(\n\t\tctx,\n\t\tdbID,\n\t\tbuckKey,\n\t\tbuckPath,\n\t\tnow,\n\t\towner,\n\t\tmd,\n\t\ttdb.WithNewBucketName(name),\n\t\ttdb.WithNewBucketKey(linkKey),\n\t\ttdb.WithNewBucketToken(dbToken))\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Finally, publish the new bucket's address to the name system\n\tgo s.IPNSManager.Publish(buckPath, buck.Key)\n\treturn ctx, buck, seed, nil\n}", "func CreateBuckets(name string, buckets ...[]byte) Spec {\n\treturn CreateBucketsMigration{name, buckets}\n}", "func (service *S3Service) GetOrCreateBucket(bucketName string) (*S3Bucket,error) {\n return nil,S3Error{\"not implemented yet\"}\n}", "func NewBucket(bucketName, donutName string, nodes map[string]Node) (Bucket, error) {\n\tif bucketName == \"\" {\n\t\treturn nil, errors.New(\"invalid argument\")\n\t}\n\tb := bucket{}\n\tb.name = bucketName\n\tb.donutName = donutName\n\tb.objects = make(map[string]Object)\n\tb.nodes = nodes\n\treturn b, nil\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\targs = &BucketArgs{}\n\t}\n\targs.ApiVersion = pulumi.StringPtr(\"storage.crossplane.io/v1alpha1\")\n\targs.Kind = pulumi.StringPtr(\"Bucket\")\n\tvar resource Bucket\n\terr := ctx.RegisterResource(\"kubernetes:storage.crossplane.io/v1alpha1:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (tx *Tx) Bucket(name string) *Bucket {\n\tb, ok := tx.buckets[name]\n\tif ok {\n\t\treturn b\n\t}\n\n\tif tx.ro {\n\t\treturn nil\n\t}\n\n\ttx.db.m.Lock()\n\tif b = tx.db.buckets[name]; b == nil {\n\t\ttx.action(backend.ActionCreateBucket, name, \"\", nil)\n\t\tb = newBucket(name, tx.db.opts.DefaultMarshalFn, tx.db.opts.DefaultUnmarshalFn)\n\t\tb.m.Lock()\n\t\ttx.db.buckets[name] = b\n\t}\n\ttx.db.m.Unlock()\n\ttx.buckets[name] = b\n\n\treturn b\n}", "func createBucketRequest(w http.ResponseWriter, r *http.Request) {\n\tlog.Printf(\"received bucket creation request\")\n\n\tvars := mux.Vars(r)\n\tusername := vars[\"username\"]\n\tlog.Printf(\"username: %s\", username)\n\n\tdataType, err := getQueryField(r, \"type\")\n\tif err != nil {\n\t\trespondJSONError(w, http.StatusInternalServerError, err.Error(), dataType)\n\t\treturn\n\t}\n\n\t_, ok := validDataTypes[dataType]\n\tif !ok {\n\t\trespondJSONError(w, http.StatusInternalServerError, \"Data type %s not supported\", dataType)\n\t\treturn\n\n\t}\n\n\tbucketName, _ := getQueryField(r, \"name\")\n\n\tisPublic, _ := getQueryFieldBool(r, \"public\")\n\n\tbucketObject, err := createSageBucket(username, dataType, bucketName, isPublic)\n\tif err != nil {\n\t\trespondJSONError(w, http.StatusInternalServerError, \"bucket creation failed: %s\", err.Error())\n\t\treturn\n\t}\n\t// TODO store owner info in mysql\n\n\trespondJSON(w, http.StatusOK, bucketObject)\n\n}", "func (d *MinioDriver) createBucket() error {\n\tbucket := createName(bucketPrefix)\n\texists, err := d.c.Client.BucketExists(bucket)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !exists {\n\t\t// TODO: in the future, let the user set \"location\" so that this works with\n\t\t// aws s3.\n\t\tif err := d.c.Client.MakeBucket(bucket, \"\"); err != nil {\n\t\t\tglog.Warningf(\"Failed to create bucket %s at \", bucket)\n\t\t\treturn err\n\t\t}\n\t}\n\n\td.c.BucketName = bucket\n\treturn nil\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\targs = &BucketArgs{}\n\t}\n\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Bucket\n\terr := ctx.RegisterResource(\"alicloud:oss/bucket:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (b *Bucket) Bucket(name string) *Bucket {\n nb := b.bucket.Bucket( []byte(name) )\n if nb == nil {\n return nil\n }\n return &Bucket{ tx:b.tx, bucket: nb }\n}", "func (b fakeBucketHandle) Create(_ context.Context, _ string, attrs *storage.BucketAttrs) error {\n\tif _, ok := b.c.buckets[b.name]; ok {\n\t\treturn fmt.Errorf(\"bucket %q already exists\", b.name)\n\t}\n\tif attrs == nil {\n\t\tattrs = &storage.BucketAttrs{}\n\t}\n\tattrs.Name = b.name\n\tb.c.buckets[b.name] = &fakeBucket{attrs: attrs, objects: map[string][]byte{}}\n\treturn nil\n}", "func (b *Bucket) Bucket(name string) *Bucket { panic(\"n/i\") }", "func (s Storage) CreateBucket(bucket string) error {\n\tif !s.Opened {\n\t\treturn fmt.Errorf(\"db must be opened before creating bucket\")\n\t}\n\terr := s.DB.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists([]byte(bucket))\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Error creating bucket : %s\", err)\n\t\t}\n\t\treturn nil\n\t})\n\treturn err\n}", "func (c *defaultGcsClient) CreateBucket(ctxIn context.Context, project, bucket, location, storageClass string, lifetimeInDays uint, archiveTTM uint) error {\n\tctx, span := trace.StartSpan(ctxIn, \"(*defaultGcsClient).CreateBucket\")\n\tdefer span.End()\n\n\tvar bucketAttrs = storage.BucketAttrs{\n\t\tLocation: location,\n\t\tStorageClass: storageClass,\n\t\tBucketPolicyOnly: storage.BucketPolicyOnly{Enabled: true},\n\t\tLabels: map[string]string{\"purpose\": \"backup\"},\n\t}\n\tbucketAttrs.Lifecycle = storage.Lifecycle{}\n\n\tif archiveTTM > 0 {\n\t\truleTTM := storage.LifecycleRule{\n\t\t\tAction: storage.LifecycleAction{Type: \"SetStorageClass\", StorageClass: \"ARCHIVE\"},\n\t\t\tCondition: storage.LifecycleCondition{AgeInDays: int64(archiveTTM)},\n\t\t}\n\t\tbucketAttrs.Lifecycle.Rules = append(bucketAttrs.Lifecycle.Rules, ruleTTM)\n\t}\n\n\tif lifetimeInDays > 0 {\n\t\truleTTL := storage.LifecycleRule{\n\t\t\tAction: storage.LifecycleAction{Type: \"Delete\"},\n\t\t\tCondition: storage.LifecycleCondition{AgeInDays: int64(lifetimeInDays)},\n\t\t}\n\t\tbucketAttrs.Lifecycle.Rules = append(bucketAttrs.Lifecycle.Rules, ruleTTL)\n\t}\n\n\terr := c.client.Bucket(bucket).Create(ctx, project, &bucketAttrs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (b *Bucket) CreateBucketIfNotExists( name string ) (*Bucket, error) {\n nb, err := b.bucket.CreateBucketIfNotExists( []byte(name) )\n if err != nil {\n return nil, err\n }\n return &Bucket{tx: b.tx, bucket: nb}, nil\n}", "func (b *Bucket) CreateBucket() error {\n\t// set session\n\tsess, err := b.setSession()\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Create bucket\n\tsvc := b.newS3func(sess)\n\t_, err = svc.CreateBucket(&s3.CreateBucketInput{Bucket: aws.String(b.BucketName)})\n\treturn err\n}", "func CreateBucket(creator BucketCreator, bucketName string) error {\n\tif err := creator.MakeBucket(bucketName, ACLPublicRead, \"\"); err != nil {\n\t\tminioErr := s3.ToErrorResponse(err)\n\t\tif minioErr.Code == bucketAlreadyExistsCode {\n\t\t\treturn nil\n\t\t}\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Client) Create(ctx context.Context, bucket string) error {\n\t// Check if the bucket already exists.\n\texists, err := c.bucketExists(ctx, bucket)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif exists {\n\t\treturn fmt.Errorf(\"expected bucket %q to not be created yet)\", bucket)\n\t}\n\n\t// Create the bucket.\n\tbkt := c.gcsClient.Bucket(bucket)\n\tif err := bkt.Create(ctx, c.projectID, nil); err != nil {\n\t\treturn fmt.Errorf(\"failed to create bucket %q in project %s: %w\", bucket, c.projectID, err)\n\t}\n\tbkt.ACL().Set(ctx, gcs.AllUsers, gcs.RoleReader)\n\n\tc.bucket = bucket\n\tc.nextSeq = 0\n\treturn nil\n}", "func CreateBucketSample() {\n\t// New Client\n\tclient, err := oss.New(endpoint, accessID, accessKey)\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\tDeleteTestBucketAndObject(bucketName)\n\n\t// case 1:creates a bucket with default parameters\n\terr = client.CreateBucket(bucketName)\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\t// deletes bucket\n\terr = client.DeleteBucket(bucketName)\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\t// case 2:creates the bucket with ACL\n\terr = client.CreateBucket(bucketName, oss.ACL(oss.ACLPublicRead))\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\t// case 3:repeat the same bucket. OSS will not return error, but just no op. The ACL is not updated.\n\terr = client.CreateBucket(bucketName, oss.ACL(oss.ACLPublicReadWrite))\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\t// deletes bucket\n\terr = client.DeleteBucket(bucketName)\n\tif err != nil {\n\t\tHandleError(err)\n\t}\n\n\tfmt.Println(\"CreateBucketSample completed\")\n}", "func (db *DB) Bucket(name string) *Bucket {\n\tif err := db.Bolt.Update(func(tx *bolt.Tx) error {\n\t\tif _, err := tx.CreateBucketIfNotExists([]byte(name)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}); err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &Bucket{Name: name, db: db}\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOpt) (*Bucket, error) {\n\tinputs := make(map[string]interface{})\n\tif args == nil {\n\t\tinputs[\"accelerationStatus\"] = nil\n\t\tinputs[\"acl\"] = nil\n\t\tinputs[\"arn\"] = nil\n\t\tinputs[\"bucket\"] = nil\n\t\tinputs[\"bucketPrefix\"] = nil\n\t\tinputs[\"corsRules\"] = nil\n\t\tinputs[\"forceDestroy\"] = nil\n\t\tinputs[\"hostedZoneId\"] = nil\n\t\tinputs[\"lifecycleRules\"] = nil\n\t\tinputs[\"loggings\"] = nil\n\t\tinputs[\"objectLockConfiguration\"] = nil\n\t\tinputs[\"policy\"] = nil\n\t\tinputs[\"region\"] = nil\n\t\tinputs[\"replicationConfiguration\"] = nil\n\t\tinputs[\"requestPayer\"] = nil\n\t\tinputs[\"serverSideEncryptionConfiguration\"] = nil\n\t\tinputs[\"tags\"] = nil\n\t\tinputs[\"versioning\"] = nil\n\t\tinputs[\"website\"] = nil\n\t\tinputs[\"websiteDomain\"] = nil\n\t\tinputs[\"websiteEndpoint\"] = nil\n\t} else {\n\t\tinputs[\"accelerationStatus\"] = args.AccelerationStatus\n\t\tinputs[\"acl\"] = args.Acl\n\t\tinputs[\"arn\"] = args.Arn\n\t\tinputs[\"bucket\"] = args.Bucket\n\t\tinputs[\"bucketPrefix\"] = args.BucketPrefix\n\t\tinputs[\"corsRules\"] = args.CorsRules\n\t\tinputs[\"forceDestroy\"] = args.ForceDestroy\n\t\tinputs[\"hostedZoneId\"] = args.HostedZoneId\n\t\tinputs[\"lifecycleRules\"] = args.LifecycleRules\n\t\tinputs[\"loggings\"] = args.Loggings\n\t\tinputs[\"objectLockConfiguration\"] = args.ObjectLockConfiguration\n\t\tinputs[\"policy\"] = args.Policy\n\t\tinputs[\"region\"] = args.Region\n\t\tinputs[\"replicationConfiguration\"] = args.ReplicationConfiguration\n\t\tinputs[\"requestPayer\"] = args.RequestPayer\n\t\tinputs[\"serverSideEncryptionConfiguration\"] = args.ServerSideEncryptionConfiguration\n\t\tinputs[\"tags\"] = args.Tags\n\t\tinputs[\"versioning\"] = args.Versioning\n\t\tinputs[\"website\"] = args.Website\n\t\tinputs[\"websiteDomain\"] = args.WebsiteDomain\n\t\tinputs[\"websiteEndpoint\"] = args.WebsiteEndpoint\n\t}\n\tinputs[\"bucketDomainName\"] = nil\n\tinputs[\"bucketRegionalDomainName\"] = nil\n\ts, err := ctx.RegisterResource(\"aws:s3/bucket:Bucket\", name, true, inputs, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Bucket{s: s}, nil\n}", "func TestTransactionCreateBucket(t *testing.T) {\n\twithOpenDB(func(db *DB, path string) {\n\t\t// Create a bucket.\n\t\ttxn, _ := db.RWTransaction()\n\t\terr := txn.CreateBucket(\"widgets\")\n\t\tassert.NoError(t, err)\n\n\t\t// Commit the transaction.\n\t\terr = txn.Commit()\n\t\tassert.NoError(t, err)\n\n\t\t/*\n\t\t// Open a separate read-only transaction.\n\t\trtxn, err := db.Transaction()\n\t\tassert.NotNil(t, txn)\n\t\tassert.NoError(t, err)\n\n\t\tb, err := rtxn.Bucket(\"widgets\")\n\t\tassert.NoError(t, err)\n\t\tif assert.NotNil(t, b) {\n\t\t\tassert.Equal(t, b.Name(), \"widgets\")\n\t\t}\n\t\t*/\n\t})\n}", "func (s *S3Agent) CreateBucket(name string) error {\n\treturn s.createBucket(name, true)\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.Bucket == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Bucket'\")\n\t}\n\tif args.OutpostId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'OutpostId'\")\n\t}\n\tvar resource Bucket\n\terr := ctx.RegisterResource(\"aws:s3control/bucket:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (hp *hdfsProvider) CreateBucket(bck *meta.Bck) (errCode int, err error) {\n\treturn hp.checkDirectoryExists(bck)\n}", "func CreateUserBucket(acc string) error {\n\tbucketName := GetUserBucketName(acc)\n\n\treturn client.MakeBucket(\n\t\tcontext.Background(),\n\t\tbucketName,\n\t\tminio.MakeBucketOptions{Region: \"ap-east-1\"},\n\t)\n}", "func GetOrCreateBucket(bucket Bucket, name string) (b Bucket, err error) {\n\tif bucket == nil {\n\t\treturn nil, ErrInvalidArgument\n\t}\n\n\tif b = bucket.Bucket(name); b == nil {\n\t\tb, err = bucket.NewBucket(name)\n\t}\n\treturn\n}", "func (api *bucketAPI) Create(obj *objstore.Bucket) error {\n\tif api.ct.resolver != nil {\n\t\tapicl, err := api.ct.apiClient()\n\t\tif err != nil {\n\t\t\tapi.ct.logger.Errorf(\"Error creating API server clent. Err: %v\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = apicl.ObjstoreV1().Bucket().Create(context.Background(), obj)\n\t\tif err != nil && strings.Contains(err.Error(), \"AlreadyExists\") {\n\t\t\t_, err = apicl.ObjstoreV1().Bucket().Update(context.Background(), obj)\n\n\t\t}\n\t\treturn err\n\t}\n\n\tapi.ct.handleBucketEvent(&kvstore.WatchEvent{Object: obj, Type: kvstore.Created})\n\treturn nil\n}", "func CreateBucketAction(w http.ResponseWriter, r *http.Request) {\n\n\t// Create S3 service client\n\tsvc := s3.New(sess)\n\n\tbucket := r.FormValue(\"bucketName\")\n\n\tif len(bucket) <= 0 {\n\t\thttp.Redirect(w, r, \"/createbucket?errorM=No bucket name specified\", http.StatusSeeOther)\n\t} else {\n\t\t// Create the S3 Bucket\n\t\t_, err := svc.CreateBucket(&s3.CreateBucketInput{\n\t\t\tBucket: aws.String(bucket),\n\t\t})\n\t\tif err != nil {\n\t\t\tif awsErr, ok := err.(awserr.Error); ok {\n\t\t\t\t// process SDK error\n\t\t\t\thttp.Redirect(w, r, \"/createbucket?errorM=\"+awsErr.Message(), http.StatusSeeOther)\n\t\t\t} else {\n\t\t\t\thttp.Redirect(w, r, \"/createbucket?errorM=Failed to create bucket\", http.StatusSeeOther)\n\t\t\t}\n\t\t} else {\n\t\t\thttp.Redirect(w, r, \"/bucketlist?successM=Bucket created succcesfully\", http.StatusSeeOther)\n\t\t}\n\t}\n\n}", "func BucketGetChildBucket(b *bolt.Bucket, childBucketName string) (*bolt.Bucket, error) {\n\t// b.Bucket will return a new bucket if this childBucket is not exist.\n\tchildBucket := b.Bucket([]byte(childBucketName))\n\tvar err error\n\t// if can not find the child bucket, then create it now\n\tif childBucket == nil {\n\t\tchildBucket, err = b.CreateBucket([]byte(childBucketName))\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn childBucket, nil\n}", "func (p *minioProvisioner) createBucket(bucket string) error {\n\tcmd := exec.Command(\"/usr/bin/s3cmd\", \"mb\", \"s3://\"+ bucket, \"--no-check-certificate\")\n\tout, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"create s3 bucket %v failed with error: %v, output: %s\", bucket, err, out)\n\t}\n\n\treturn nil\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.BundleId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'BundleId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Bucket\n\terr := ctx.RegisterResource(\"aws-native:lightsail:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (store *Storage) CreateBucket(bucketID string) error {\n\treturn store.db.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists([]byte(bucketID))\n\t\tif err != nil {\n\t\t\treturn common.Error(common.WriteFailed, err)\n\t\t}\n\t\treturn nil\n\t})\n}", "func NewBucket() *Bucket {\n\treturn &Bucket{objects: map[string][]byte{}}\n}", "func (b *AmazonBucketProvider) CreateNewBucketForCluster(clusterName string, bucketKind string) (string, error) {\n\tuuid4, _ := uuid.NewV4()\n\tbucketName := fmt.Sprintf(\"%s-%s-%s\", clusterName, bucketKind, uuid4.String())\n\n\t// Max length is 63, https://docs.aws.amazon.com/AmazonS3/latest/dev/BucketRestrictions.html\n\tif len(bucketName) > 63 {\n\t\tbucketName = bucketName[:63]\n\t}\n\tbucketName = strings.TrimRight(bucketName, \"-\")\n\tbucketURL := \"s3://\" + bucketName\n\terr := b.EnsureBucketIsCreated(bucketURL)\n\tif err != nil {\n\t\treturn bucketURL, errors.Wrapf(err, \"failed to create bucket %s\", bucketURL)\n\t}\n\n\treturn bucketURL, nil\n}", "func TestTx_CreateBucket_ErrBucketNameRequired(t *testing.T) {\n\tdb := MustOpenDB()\n\tdefer MustClose(db)\n\tif err := db.Update(func(tx *pddb.Tx) error {\n\t\tif _, err := tx.CreateBucket(nil); err != pddb.ErrBucketNameRequired {\n\t\t\tt.Fatalf(\"unexpected error: %s\", err)\n\t\t}\n\t\treturn nil\n\t}); err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func (this *HttpClient) GetBucket(name string, version int) (*Bucket, error) {\n\t// fetch data\n\tresp, err := this.get(name, version, false, INITIAL_VERSION)\n\tif err != nil {\n\t\tlog.Println(\"Error fetching bucket \", err.Error())\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\terrResp := &ErrorResp{}\n\t\terr := ffjson.NewDecoder().DecodeReader(resp.Body, errResp)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error reading the response Body\")\n\t\t}\n\t\tlog.Println(\"Error fetching bucket: \", errResp)\n\t\treturn nil, errors.New(errResp.Error())\n\t}\n\n\t// create and return bucket\n\tbucket, err := this.newBucket(resp)\n\tif err != nil {\n\t\tlog.Println(\"Error creating bucket \", err.Error())\n\t\treturn nil, err\n\t}\n\n\treturn bucket, nil\n}", "func (a *FileStorageApiService) CreateBucketExecute(r ApiCreateBucketRequest) (SingleBucket, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue SingleBucket\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.Ctx, \"FileStorageApiService.CreateBucket\")\n\tif localBasePath == \"/\" {\n\t localBasePath = \"\"\n\t}\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/v2/file_storage/buckets\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\tif r.P_bucketRequest == nil {\n\t\treturn localVarReturnValue, nil, reportError(\"bucketRequest is required and must be specified\")\n\t}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{\"application/json\"}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\t// body params\n\tlocalVarPostBody = r.P_bucketRequest\n\treq, err := a.client.prepareRequest(r.Ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func NewBucket(name, accessKey, secretKey, roleArn, endpoint, region, bucketName string, insecure bool) *Bucket {\n\treturn &Bucket{\n\t\tName: name,\n\t\tAccessKey: accessKey,\n\t\tSecretKey: secretKey,\n\t\tRoleArn: roleArn,\n\t\tEndpoint: endpoint,\n\t\tRegion: region,\n\t\tBucketName: bucketName,\n\t\tinsecure: insecure,\n\t\tnewS3func: newS3,\n\t\tnewSTSfunc: newSTS,\n\t\tnewUploaderfunc: newUploader,\n\t\tnewDownloaderfunc: newDownloader,\n\t}\n}", "func (e *copyFileToNEnumerator) createBucket(ctx context.Context, destURL url.URL, metadata common.Metadata) error {\n\tswitch e.FromTo {\n\tcase common.EFromTo.FileBlob():\n\t\tif destInfo.destBlobPipeline == nil {\n\t\t\tpanic(errors.New(\"invalid state, blob type destination's pipeline is not initialized\"))\n\t\t}\n\t\ttmpContainerURL := blobURLPartsExtension{azblob.NewBlobURLParts(destURL)}.getContainerURL()\n\t\tcontainerURL := azblob.NewContainerURL(tmpContainerURL, destInfo.destBlobPipeline)\n\t\t// Create the container, in case of it doesn't exist.\n\t\t_, err := containerURL.Create(ctx, metadata.ToAzBlobMetadata(), azblob.PublicAccessNone)\n\t\tif err != nil {\n\t\t\t// Skip the error, when container already exists, or hasn't permission to create container(container might already exists).\n\t\t\tif stgErr, ok := err.(azblob.StorageError); !ok ||\n\t\t\t\t(stgErr.ServiceCode() != azblob.ServiceCodeContainerAlreadyExists &&\n\t\t\t\t\tstgErr.Response().StatusCode != http.StatusForbidden) {\n\t\t\t\treturn fmt.Errorf(\"fail to create container, %v\", err)\n\t\t\t}\n\t\t\t// the case error is container already exists\n\t\t}\n\t\t// Here could be other cases, e.g.: creating share and etc.\n\t}\n\treturn nil\n}", "func NewBucket() storage.Bucket {\n\treturn newBucket()\n}", "func (c *Client) CreateUserBucket(userID string) (bucketAddress *string, err error) {\n\t// var ctx = context.Background()\n\terr = c.C.MakeBucket(userID, \"\")\n\tbucketaddr := c.config.Endpoint + \"/\" + userID\n\tif err != nil {\n\t\t// Check to see if we already own this bucket (which happens if you run this twice)\n\t\texists, errBucketExists := c.C.BucketExists(userID)\n\t\tif errBucketExists == nil && exists {\n\t\t\tlog.Printf(\"We already own %s\\n\", userID)\n\t\t\treturn &bucketaddr, nil\n\t\t}\n\t\treturn nil, err\n\t} else if goscrappy.Debug {\n\t\tlog.Printf(\"Info(CreateUserBucket): Successfully created %s\\n\", bucketaddr)\n\t}\n\treturn &bucketaddr, nil\n}", "func (b *Buckets) NewBucket(ctx context.Context, conf Config, opts ...NewOption) (buck *Bucket, err error) {\n\targs := &newOptions{}\n\tfor _, opt := range opts {\n\t\topt(args)\n\t}\n\n\t// Ensure we're not going to overwrite an existing local config\n\tcwd, err := filepath.Abs(conf.Path)\n\tif err != nil {\n\t\treturn\n\t}\n\tbc, found, err := b.config.NewConfig(cwd, flags, false)\n\tif err != nil {\n\t\treturn\n\t}\n\tif found {\n\t\treturn nil, ErrBucketExists\n\t}\n\n\t// Check config values\n\tif !conf.Thread.Defined() {\n\t\treturn nil, ErrThreadRequired\n\t}\n\tbc.Viper.Set(\"thread\", conf.Thread.String())\n\tbc.Viper.Set(\"key\", conf.Key)\n\n\tbuck = &Bucket{\n\t\tcwd: cwd,\n\t\tconf: bc,\n\t\tclients: b.clients,\n\t\tauth: b.auth,\n\t\tpushBlock: make(chan struct{}, 1),\n\t}\n\tctx, err = buck.Context(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tinitRemote := conf.Key == \"\"\n\tif initRemote {\n\t\trep, err := b.clients.Buckets.Create(\n\t\t\tctx,\n\t\t\tclient.WithName(args.name),\n\t\t\tclient.WithPrivate(args.private),\n\t\t\tclient.WithCid(args.fromCid),\n\t\t\tclient.WithUnfreeze(args.unfreeze))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// If we're unfreezing, we simply return since the\n\t\t// bucket will get created async if the Filecoin retrieval\n\t\t// is successful. The user will `[hub] buck init -e` in the future\n\t\t// to pull the new bucket.\n\t\tif args.unfreeze {\n\t\t\tbuck.retrID = rep.RetrievalId\n\t\t\treturn buck, nil\n\t\t}\n\t\tbuck.conf.Viper.Set(\"key\", rep.Root.Key)\n\n\t\tseed := filepath.Join(cwd, buckets.SeedName)\n\t\tfile, err := os.Create(seed)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t_, err = file.Write(rep.Seed)\n\t\tif err != nil {\n\t\t\tfile.Close()\n\t\t\treturn nil, err\n\t\t}\n\t\tfile.Close()\n\n\t\tif err = buck.loadLocalRepo(ctx, cwd, b.repoName(), false); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = buck.repo.SaveFile(ctx, seed, buckets.SeedName); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsc, err := cid.Decode(rep.SeedCid)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = buck.repo.SetRemotePath(buckets.SeedName, sc); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\trp, err := util.NewResolvedPath(rep.Root.Path)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = buck.repo.SetRemotePath(\"\", rp.Cid()); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tbuck.links = &Links{URL: rep.Links.Url, WWW: rep.Links.Www, IPNS: rep.Links.Ipns}\n\t} else {\n\t\tif err := buck.loadLocalRepo(ctx, cwd, b.repoName(), true); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tr, err := buck.Roots(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = buck.repo.SetRemotePath(\"\", r.Remote); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif _, err = buck.RemoteLinks(ctx, \"\"); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Write the local config to disk\n\tdir := filepath.Join(cwd, buck.conf.Dir)\n\tif err = os.MkdirAll(dir, os.ModePerm); err != nil {\n\t\treturn\n\t}\n\tconfig := filepath.Join(dir, buck.conf.Name+\".yml\")\n\tif err = buck.conf.Viper.WriteConfigAs(config); err != nil {\n\t\treturn\n\t}\n\tcfile, err := filepath.Abs(config)\n\tif err != nil {\n\t\treturn\n\t}\n\tbuck.conf.Viper.SetConfigFile(cfile)\n\n\t// Pull remote bucket contents\n\tif !initRemote || args.fromCid.Defined() {\n\t\tif err := buck.repo.Save(ctx); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tswitch args.strategy {\n\t\tcase Soft, Hybrid:\n\t\t\tdiff, missing, remove, err := buck.diffPath(ctx, \"\", cwd, args.strategy == Hybrid)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif err = stashChanges(diff); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif _, err = buck.handleChanges(ctx, missing, remove, args.events); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif err := buck.repo.Save(ctx); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif err = applyChanges(diff); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\tcase Hard:\n\t\t\tif _, err := buck.getPath(ctx, \"\", cwd, nil, false, args.events); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif err := buck.repo.Save(ctx); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\treturn buck, nil\n}", "func (r *RBucket) CreateBucketIfNotExists(name []byte) (Bucket, error) {\n\treq := &BucketRequest{}\n\treq.Key = name\n\tresp := &BucketResponse{}\n\treq.ContextID = r.parent\n\terr := r.r.call(\"srv.CreateBucketIfNotExists\", req, resp)\n\tb := &RBucket{}\n\tb.tx = r.tx\n\tb.r = r.r\n\tb.id = resp.BucketID\n\tb.parent = resp.BucketContextID\n\treturn b, err\n}", "func (r *RBucket) Bucket(name []byte) Bucket {\n\treq := &BucketRequest{}\n\treq.Key = name\n\tresp := &BucketResponse{}\n\treq.ContextID = r.parent\n\terr := r.r.call(\"srv.Bucket\", req, resp)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tb := &RBucket{}\n\tb.tx = r.tx\n\tb.r = r.r\n\tb.id = resp.BucketID\n\tb.parent = resp.BucketContextID\n\treturn b\n}", "func CreateBucket(w http.ResponseWriter, r *http.Request) *appError {\n decoder := json.NewDecoder(r.Body)\n var ecsBucket ECSBucket\n err := decoder.Decode(&ecsBucket)\n if err != nil {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Can't decode JSON data\"}\n }\n headers := make(map[string][]string)\n if ecsBucket.ReplicationGroup != \"\" {\n headers[\"x-emc-vpool\"] = []string{ecsBucket.ReplicationGroup}\n }\n if ecsBucket.MetadataSearch != \"\" {\n headers[\"x-emc-metadata-search\"] = []string{ecsBucket.MetadataSearch}\n }\n if ecsBucket.EnableADO {\n headers[\"x-emc-is-stale-allowed\"] = []string{\"true\"}\n } else {\n headers[\"x-emc-is-stale-allowed\"] = []string{\"false\"}\n }\n if ecsBucket.EnableFS {\n headers[\"x-emc-file-system-access-enabled\"] = []string{\"true\"}\n } else {\n headers[\"x-emc-file-system-access-enabled\"] = []string{\"false\"}\n }\n if ecsBucket.EnableCompliance {\n headers[\"x-emc-compliance-enabled\"] = []string{\"true\"}\n } else {\n headers[\"x-emc-compliance-enabled\"] = []string{\"false\"}\n }\n if ecsBucket.EnableEncryption {\n headers[\"x-emc-server-side-encryption-enabled\"] = []string{\"true\"}\n } else {\n headers[\"x-emc-server-side-encryption-enabled\"] = []string{\"false\"}\n }\n var bucketCreateResponse Response\n if ecsBucket.Api == \"s3\" {\n s3, err := getS3(r)\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n bucketCreateResponse, err = s3Request(s3, ecsBucket.Name, \"PUT\", \"/\", headers, \"\")\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n if bucketCreateResponse.Code == 200 {\n rendering.JSON(w, http.StatusOK, ecsBucket.Name)\n } else {\n return &appError{err: err, status: http.StatusInternalServerError, xml: bucketCreateResponse.Body}\n }\n } else if ecsBucket.Api == \"swift\" {\n bucketCreateResponse, err = swiftRequest(ecsBucket.Endpoint, ecsBucket.User, ecsBucket.Password, ecsBucket.Name, \"PUT\", \"/\", headers, \"\")\n log.Print(bucketCreateResponse)\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n if bucketCreateResponse.Code >= 200 && bucketCreateResponse.Code < 300 {\n rendering.JSON(w, http.StatusOK, ecsBucket.Name)\n } else {\n return &appError{err: err, status: http.StatusInternalServerError, xml: bucketCreateResponse.Body}\n }\n } else if ecsBucket.Api == \"atmos\" {\n s3, err := getS3(r)\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n bucketCreateResponse, err = atmosRequest(ecsBucket.Endpoint, s3.AccessKey, s3.SecretKey, \"\", \"PUT\", \"/rest/subtenant\", headers, \"\")\n if err != nil {\n log.Print(err)\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n if bucketCreateResponse.Code >= 200 && bucketCreateResponse.Code < 300 {\n rendering.JSON(w, http.StatusOK, bucketCreateResponse.ResponseHeaders[\"Subtenantid\"][0])\n } else {\n return &appError{err: err, status: http.StatusInternalServerError, xml: bucketCreateResponse.Body}\n }\n }\n\n return nil\n}", "func (tx *Tx) Bucket(name string) (*Bucket, error) {\n\tif tx.db == nil {\n\t\treturn nil, ErrNoDatabase\n\t}\n\tif !tx.write {\n\t\treturn nil, ErrNotWriteTransaction\n\t}\n\n\tbucket, _ := tx.db.addBucket(name)\n\tb := &Bucket{\n\t\ttx: tx,\n\t\tmanaged: bucket,\n\t}\n\treturn b, nil\n}", "func CreateS3Bucket(bucketName string) {\n\tid, secret, endpoint, s3Region, disableSSLBool := GetAWSDetailsFromEnv()\n\tsess, err := session.NewSession(&aws.Config{\n\t\tEndpoint: aws.String(endpoint),\n\t\tCredentials: credentials.NewStaticCredentials(id, secret, \"\"),\n\t\tRegion: aws.String(s3Region),\n\t\tDisableSSL: aws.Bool(disableSSLBool),\n\t\tS3ForcePathStyle: aws.Bool(true),\n\t},\n\t)\n\texpect(err).NotTo(haveOccurred(),\n\t\tfmt.Sprintf(\"Failed to get S3 session to create bucket. Error: [%v]\", err))\n\n\tS3Client := s3.New(sess)\n\n\t_, err = S3Client.CreateBucket(&s3.CreateBucketInput{\n\t\tBucket: aws.String(bucketName),\n\t})\n\texpect(err).NotTo(haveOccurred(),\n\t\tfmt.Sprintf(\"Failed to create bucket [%v]. Error: [%v]\", bucketName, err))\n\n\terr = S3Client.WaitUntilBucketExists(&s3.HeadBucketInput{\n\t\tBucket: aws.String(bucketName),\n\t})\n\texpect(err).NotTo(haveOccurred(),\n\t\tfmt.Sprintf(\"Failed to wait for bucket [%v] to get created. Error: [%v]\", bucketName, err))\n}", "func (c *Common) createEmptyBucket(ctx context.Context) error {\n\tcl, done := c.Client()\n\tdefer done()\n\tx, err := cl.BucketExists(c.Bucket)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !x {\n\t\tconsole.Infof(\"Creating Bucket %q...\\n\", c.Bucket)\n\t\treturn cl.MakeBucket(c.Bucket, c.Location)\n\t}\n\tif c.Clear {\n\t\tconsole.Infof(\"Clearing Bucket %q...\\n\", c.Bucket)\n\t\tc.deleteAllInBucket(ctx)\n\t}\n\treturn nil\n}", "func NewBucket(desc metrics.Descriptor, dur time.Duration) (*Bucket, error) {\n\tvar (\n\t\tm metrics.Metric\n\t\terr error\n\t)\n\tif m, err = metrics.FromDescriptor(desc); err != nil {\n\t\treturn nil, err\n\t}\n\tshard := NewShard(m, dur)\n\treturn &Bucket{\n\t\tdescriptor: desc,\n\t\tshards: []*Shard{shard},\n\t\tshardDuration: dur,\n\t}, nil\n}", "func (blt *Bolt) CreateBuckets(buckets ...[]byte) error {\n\tvar err error\n\t// ensures that don't create bucket if is nil\n\tfor _, bkt := range buckets {\n\t\t// ignores the empty buckets\n\t\tif bytes.Equal(bkt, []byte(\"\")) {\n\t\t\tcontinue\n\t\t}\n\t\tblt.Bucket = bkt // the last is the current\n\t\terr = blt.db.Update(func(tx *b.Tx) error {\n\t\t\t_, err := tx.CreateBucketIfNotExists(blt.Bucket)\n\t\t\treturn err\n\t\t})\n\t}\n\treturn err\n}", "func (b *Buckets) New(\n\tctx context.Context,\n\tthread core.ID,\n\tkey string,\n\towner did.DID,\n\tpth path.Path,\n\tcreated time.Time,\n\tmetadata map[string]Metadata,\n\tidentity did.Token,\n\topts ...BucketOption,\n) (*Bucket, error) {\n\targs := &BucketOptions{}\n\tfor _, opt := range opts {\n\t\topt(args)\n\t}\n\tvar linkKey string\n\tif args.Key != nil {\n\t\tlinkKey = base64.StdEncoding.EncodeToString(args.Key)\n\t}\n\tif metadata == nil {\n\t\tmetadata = make(map[string]Metadata)\n\t}\n\tbucket := &Bucket{\n\t\tKey: key,\n\t\tOwner: owner,\n\t\tName: args.Name,\n\t\tVersion: int(Version1),\n\t\tLinkKey: linkKey,\n\t\tPath: pth.String(),\n\t\tMetadata: metadata,\n\t\tCreatedAt: created.UnixNano(),\n\t\tUpdatedAt: created.UnixNano(),\n\t}\n\tif _, err := b.Create(ctx, thread, bucket, WithIdentity(identity)); err != nil {\n\t\treturn nil, fmt.Errorf(\"creating bucket: %s\", err)\n\t}\n\treturn bucket, nil\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.OutpostId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'OutpostId'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Bucket\n\terr := ctx.RegisterResource(\"aws-native:s3outposts:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (rm *RsrcManager) Create(item reconciler.Object) error {\n\to := item.Obj.(*Object)\n\t_, err := rm.service.Buckets.Insert(o.ProjectID, o.Bucket).Do()\n\treturn err\n}", "func NewBucket(s *Store, key string, node storm.Node) *Bucket {\n\treturn &Bucket{\n\t\tkey: key,\n\t\tstore: s,\n\t\tnode: node,\n\t}\n}", "func NewBucket(address common.Address, backend bind.ContractBackend) (*Bucket, error) {\n\tcontract, err := bindBucket(address, backend, backend, backend)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Bucket{BucketCaller: BucketCaller{contract: contract}, BucketTransactor: BucketTransactor{contract: contract}, BucketFilterer: BucketFilterer{contract: contract}}, nil\n}", "func NewBucket(region string) (out *Bucket, err error) {\n\ta := new(Bucket)\n\tif region == \"\" {\n\t\ta.region = \"eu-central-1\"\n\t} else {\n\t\ta.region = region\n\t}\n\tsess, err := session.NewSession(&aws.Config{Region: aws.String(a.region)})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ta.session = sess\n\tsvc := s3.New(a.session)\n\ta.service = svc\n\treturn a, nil\n}", "func (g *gcs) Create(ctx context.Context, name string) (io.WriteCloser, error) {\n\treturn g.bucket.Object(name).NewWriter(ctx), nil\n}", "func CreateBucket(w http.ResponseWriter, r *http.Request) *appError {\n session, err := store.Get(r, \"session-name\")\n if err != nil {\n return &appError{err: err, status: http.StatusInternalServerError, json: http.StatusText(http.StatusInternalServerError)}\n }\n s3 := S3{\n EndPointString: session.Values[\"Endpoint\"].(string),\n AccessKey: session.Values[\"AccessKey\"].(string),\n SecretKey: session.Values[\"SecretKey\"].(string),\n Namespace: session.Values[\"Namespace\"].(string),\n }\n\n decoder := json.NewDecoder(r.Body)\n var bucket NewBucket\n err = decoder.Decode(&bucket)\n if err != nil {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Can't decode JSON data\"}\n }\n\n // Add the necessary headers for Metadata Search and Access During Outage\n createBucketHeaders := map[string][]string{}\n createBucketHeaders[\"Content-Type\"] = []string{\"application/xml\"}\n createBucketHeaders[\"x-emc-is-stale-allowed\"] = []string{\"true\"}\n createBucketHeaders[\"x-emc-metadata-search\"] = []string{\"ObjectName,x-amz-meta-image-width;Integer,x-amz-meta-image-height;Integer,x-amz-meta-gps-latitude;Decimal,x-amz-meta-gps-longitude;Decimal\"}\n\n createBucketResponse, _ := s3Request(s3, bucket.Name, \"PUT\", \"/\", createBucketHeaders, \"\")\n\n // Enable CORS after the bucket creation to allow the web browser to send requests directly to ECS\n if createBucketResponse.Code == 200 {\n enableBucketCorsHeaders := map[string][]string{}\n enableBucketCorsHeaders[\"Content-Type\"] = []string{\"application/xml\"}\n corsConfiguration := `\n <CORSConfiguration>\n <CORSRule>\n <AllowedOrigin>*</AllowedOrigin>\n <AllowedHeader>*</AllowedHeader>\n <ExposeHeader>x-amz-meta-image-width</ExposeHeader>\n <ExposeHeader>x-amz-meta-image-height</ExposeHeader>\n <ExposeHeader>x-amz-meta-gps-latitude</ExposeHeader>\n <ExposeHeader>x-amz-meta-gps-longitude</ExposeHeader>\n <AllowedMethod>HEAD</AllowedMethod>\n <AllowedMethod>GET</AllowedMethod>\n <AllowedMethod>PUT</AllowedMethod>\n <AllowedMethod>POST</AllowedMethod>\n <AllowedMethod>DELETE</AllowedMethod>\n </CORSRule>\n </CORSConfiguration>\n `\n enableBucketCorsResponse, _ := s3Request(s3, bucket.Name, \"PUT\", \"/?cors\", enableBucketCorsHeaders, corsConfiguration)\n if enableBucketCorsResponse.Code == 200 {\n rendering.JSON(w, http.StatusOK, struct {\n CorsConfiguration string `json:\"cors_configuration\"`\n Bucket string `json:\"bucket\"`\n } {\n CorsConfiguration: corsConfiguration,\n Bucket: bucket.Name,\n })\n } else {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Bucket created, but CORS can't be enabled\"}\n }\n } else {\n return &appError{err: err, status: http.StatusBadRequest, json: \"Bucket can't be created\"}\n }\n return nil\n}", "func (db *DB) Bucket(name string) (backend.Bucket, error) {\n\treturn &Bucket{\n\t\tdb: db,\n\t\tname: name,\n\t}, nil\n}", "func initializeBucket() {\n\tclient := getClient()\n\tparams := &s3.GetBucketLocationInput{\n\t\tBucket: aws.String(S3_BUCKET_NAME), // Required\n\t}\n\t_, err := client.GetBucketLocation(params)\n\tif err != nil {\n\t\t// bucket does not exist\n\t\tparams := &s3.CreateBucketInput{\n\t\t\tBucket: aws.String(S3_BUCKET_NAME), // Required\n\t\t}\n\t\t_, err := client.CreateBucket(params)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"Attempted to create bucket with name \" + S3_BUCKET_NAME + \", but failed.\")\n\t\t\tfmt.Println(\"Error was: \" + err.Error())\n\t\t\tos.Exit(2)\n\t\t}\n\t\t// fmt.Println(\"created new bucket with name: \" + S3_BUCKET_NAME)\n\t}\n}", "func NewBucket(logger log.Logger, conf []byte, component string) (*Bucket, error) {\n\tif logger == nil {\n\t\tlogger = log.NewNopLogger()\n\t}\n\n\tconfig, err := parseConfig(conf)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"parsing BOS configuration\")\n\t}\n\n\treturn NewBucketWithConfig(logger, config, component)\n}", "func NewBucket(size int) *Bucket {\n\treturn &Bucket{make(chan struct{}, size)}\n}", "func (xl xlObjects) MakeBucket(bucket string) error {\n\tnsMutex.Lock(bucket, \"\")\n\tdefer nsMutex.Unlock(bucket, \"\")\n\treturn makeBucket(xl.storage, bucket)\n}", "func (c *Client) CreateBucket(ctx context.Context, params *CreateBucketInput, optFns ...func(*Options)) (*CreateBucketOutput, error) {\n\tif params == nil {\n\t\tparams = &CreateBucketInput{}\n\t}\n\n\tresult, metadata, err := c.invokeOperation(ctx, \"CreateBucket\", params, optFns, c.addOperationCreateBucketMiddlewares)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := result.(*CreateBucketOutput)\n\tout.ResultMetadata = metadata\n\treturn out, nil\n}", "func GenerateDefaultBucket(name string, db *bolt.DB) {\n\t_ = db.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists([]byte(name))\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (this *ConfigServiceClient) initDynamicBucket(name string) (*DynamicBucket, error) {\n log.Println(\"Initializing Config bucket: \" + name)\n\n dynamicBucket := &DynamicBucket{ httpClient: this.httpClient }\n\n err := ValidateBucketName(name)\n if err != nil {\n return nil, err\n }\n\n err = dynamicBucket.init(name)\n\n if err != nil {\n log.Println(\"Error fetching bucket: \", err)\n return nil, err\n } else {\n this.dynamicBucketCache.Add(name, dynamicBucket)\n go this.httpClient.WatchBucket(name, this.dynamicBucketCache, dynamicBucket)\n return dynamicBucket, nil\n }\n}", "func NewBucket(db *mongo.Database, opts ...*options.BucketOptions) (*Bucket, error) {\n\tb := &Bucket{\n\t\tname: \"fs\",\n\t\tchunkSize: DefaultChunkSize,\n\t\tdb: db,\n\t\twc: db.WriteConcern(),\n\t\trc: db.ReadConcern(),\n\t\trp: db.ReadPreference(),\n\t}\n\n\tbo := options.MergeBucketOptions(opts...)\n\tif bo.Name != nil {\n\t\tb.name = *bo.Name\n\t}\n\tif bo.ChunkSizeBytes != nil {\n\t\tb.chunkSize = *bo.ChunkSizeBytes\n\t}\n\tif bo.WriteConcern != nil {\n\t\tb.wc = bo.WriteConcern\n\t}\n\tif bo.ReadConcern != nil {\n\t\tb.rc = bo.ReadConcern\n\t}\n\tif bo.ReadPreference != nil {\n\t\tb.rp = bo.ReadPreference\n\t}\n\n\tvar collOpts = options.Collection().SetWriteConcern(b.wc).SetReadConcern(b.rc).SetReadPreference(b.rp)\n\n\tb.chunksColl = db.Collection(b.name+\".chunks\", collOpts)\n\tb.filesColl = db.Collection(b.name+\".files\", collOpts)\n\tb.readBuf = make([]byte, b.chunkSize)\n\tb.writeBuf = make([]byte, b.chunkSize)\n\n\treturn b, nil\n}", "func InitBucket() (b map[string]Bucket, err error) {\r\n\tvar (\r\n\t\titem Bucket\r\n\t)\r\n\tb = make(map[string]Bucket)\r\n\t// bucket test\r\n\titem.BucketName = \"test\"\r\n\titem.Property = 2\r\n\titem.KeyId = \"121bce6492eba701\"\r\n\titem.KeySecret = \"1eb80603e85842542f9736eb13b7e1\"\r\n\tb[\"test\"] = item\r\n\r\n\treturn\r\n}", "func CreateAzureBucket(bucketName string) {\n\t// From the Azure portal, get your Storage account blob service URL endpoint.\n\t_, _, _, _, accountName, accountKey := GetAzureCredsFromEnv()\n\n\turlStr := fmt.Sprintf(\"https://%s.blob.core.windows.net/%s\", accountName, bucketName)\n\tlogrus.Infof(\"Create container url %s\", urlStr)\n\t// Create a ContainerURL object that wraps a soon-to-be-created container's URL and a default pipeline.\n\tu, _ := url.Parse(urlStr)\n\tcredential, err := azblob.NewSharedKeyCredential(accountName, accountKey)\n\texpect(err).NotTo(haveOccurred(),\n\t\tfmt.Sprintf(\"Failed to create shared key credential [%v]\", err))\n\n\tcontainerURL := azblob.NewContainerURL(*u, azblob.NewPipeline(credential, azblob.PipelineOptions{}))\n\tctx := context1.Background() // This example uses a never-expiring context\n\n\t_, err = containerURL.Create(ctx, azblob.Metadata{}, azblob.PublicAccessNone)\n\n\texpect(err).NotTo(haveOccurred(),\n\t\tfmt.Sprintf(\"Failed to create container. Error: [%v]\", err))\n}", "func NewBucket(boptions BucketOptions) (b *Bucket) {\n\tb = new(Bucket)\n\tb.boptions = boptions\n\tb.chs = make(map[string]*Channel, boptions.ChannelSize)\n\treturn\n}", "func (d *bucket_struct) createFile(fileName string, file []byte, ContentType string) {\n\tfmt.Fprintf(d.w, \"Creating file /%v/%v\\n\", d.bucketName, fileName)\n\n\twc := d.bucket.Object(fileName).NewWriter(d.ctx)\n\twc.ContentType = ContentType\n\tif _, err := wc.Write(file); err != nil {\n\t\td.errorf(\"createFile: unable to write data to bucket %q, file %q: %v\", d.bucketName, fileName, err)\n\t\treturn\n\t}\n\tif err := wc.Close(); err != nil {\n\t\td.errorf(\"createFile: unable to close bucket %q, file %q: %v\", d.bucketName, fileName, err)\n\t\treturn\n\t}\n}", "func NewBucket(tokens uint64) *Bucket {\n\treturn &Bucket{Added: float64(tokens)}\n}", "func (b *BitBucketClient) CreateBranch(workspace string, repositorySlug string, branchName string) (dto.BitBucketResponseBranchCreate, error) {\n\tlog.Logger().StartMessage(\"Create branch\")\n\tif err := b.beforeRequest(); err != nil {\n\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t}\n\n\tb.client.SetBaseURL(DefaultBitBucketBaseAPIUrl)\n\n\tendpoint := fmt.Sprintf(\"/repositories/%s/%s/refs/branches/%s\", workspace, repositorySlug, branchName)\n\tresponse, statusCode, err := b.client.Get(endpoint, map[string]string{})\n\tif err != nil {\n\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t}\n\n\tresponseObject := dto.BitBucketResponseBranchCreate{}\n\tif statusCode == http.StatusNotFound {\n\t\tlog.Logger().Info().Str(\"branch\", branchName).Msg(\"Release branch wasn't found. Trying to create it.\")\n\t\trequest := dto.BitBucketRequestBranchCreate{\n\t\t\tName: branchName,\n\t\t\tTarget: dto.BitBucketBranchTarget{\n\t\t\t\tHash: DefaultBitBucketMainBranch,\n\t\t\t},\n\t\t}\n\n\t\tbyteRequest, err := json.Marshal(request)\n\t\tif err != nil {\n\t\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t\t}\n\n\t\tendpoint := fmt.Sprintf(\"/repositories/%s/%s/refs/branches\", workspace, repositorySlug)\n\t\tresponse, statusCode, err := b.client.Post(endpoint, byteRequest, map[string]string{})\n\t\tif err != nil {\n\t\t\tlog.Logger().AddError(err).\n\t\t\t\tMsg(\"Failed to trigger request\")\n\t\t\tlog.Logger().FinishMessage(\"Create branch\")\n\n\t\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t\t}\n\n\t\tif err := json.Unmarshal(response, &responseObject); err != nil {\n\t\t\tlog.Logger().Info().\n\t\t\t\tStr(\"branch\", branchName).\n\t\t\t\tInt(\"status_code\", statusCode).\n\t\t\t\tStr(\"response\", string(response)).\n\t\t\t\tMsg(\"Failed to unmarshal response\")\n\t\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t\t}\n\n\t\tif statusCode == http.StatusBadRequest {\n\t\t\tif responseObject.Data.Key == ErrorBranchExists {\n\t\t\t\tlog.Logger().Info().\n\t\t\t\t\tStr(\"branch\", branchName).\n\t\t\t\t\tInt(\"status_code\", statusCode).\n\t\t\t\t\tRawJSON(\"response\", response).\n\t\t\t\t\tMsg(\"Branch already exists\")\n\t\t\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\t\t\treturn responseObject, nil\n\t\t\t}\n\t\t}\n\n\t\tif statusCode != http.StatusCreated {\n\t\t\tlog.Logger().Warn().\n\t\t\t\tStr(\"branch\", branchName).\n\t\t\t\tInt(\"status_code\", statusCode).\n\t\t\t\tInterface(\"response\", responseObject).\n\t\t\t\tMsg(\"Bad status code received\")\n\n\t\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\t\treturn dto.BitBucketResponseBranchCreate{}, errors.New(\"wrong status code received during the branch creation. See the logs for more information. \")\n\t\t}\n\n\t\tlog.Logger().Info().\n\t\t\tStr(\"branch\", branchName).\n\t\t\tInt(\"status_code\", statusCode).\n\t\t\tRawJSON(\"response\", response).Msg(\"Create branch result\")\n\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\treturn responseObject, nil\n\t}\n\n\terr = json.Unmarshal(response, &responseObject)\n\tif err != nil {\n\t\tlog.Logger().AddError(err).Msg(\"Error during response unmarshal\")\n\t\tlog.Logger().FinishMessage(\"Create branch\")\n\t\treturn dto.BitBucketResponseBranchCreate{}, err\n\t}\n\n\tlog.Logger().FinishMessage(\"Create branch\")\n\treturn responseObject, nil\n}", "func (tx *Tx) SetBucket(name string, mFn MarshalFn, uFn UnmarshalFn) *Bucket {\n\tif mFn == nil {\n\t\tmFn = tx.db.opts.DefaultMarshalFn\n\t}\n\tif uFn == nil {\n\t\tuFn = tx.db.opts.DefaultUnmarshalFn\n\t}\n\n\tb, ok := tx.buckets[name]\n\tif ok {\n\t\tgoto RET\n\t}\n\n\ttx.db.m.Lock()\n\tif b = tx.db.buckets[name]; b == nil {\n\t\ttx.action(backend.ActionCreateBucket, name, \"\", nil)\n\t\tb = newBucket(name, mFn, uFn)\n\t\ttx.db.buckets[name] = b\n\t}\n\tb.m.Lock()\n\ttx.db.m.Unlock()\n\tb.tx = tx\n\ttx.buckets[name] = b\n\nRET:\n\tb.mFn, b.uFn = mFn, uFn\n\treturn b\n}", "func CreateSearchBucketFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSearchBucket(), nil\n}", "func AddBucket(name string) {\n\tbucketsToAdd = append(bucketsToAdd, name)\n}", "func (m *MockStorage) CreateBucket(arg0 string) (storage.Bucket, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"CreateBucket\", arg0)\n\tret0, _ := ret[0].(storage.Bucket)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func New(ctx context.Context, rate, timespan int) (Bucket, error) {\n\tq := make(chan struct{}, rate)\n\tb := Bucket{ctx: ctx, queue: q, rate: rate, timespan: timespan}\n\tgo b.leak()\n\treturn b, nil // maybe return pointer?\n}" ]
[ "0.77809393", "0.7585313", "0.75450534", "0.74911594", "0.70202655", "0.6961152", "0.6842577", "0.679661", "0.6710569", "0.6691223", "0.6689105", "0.66800296", "0.66749895", "0.66719764", "0.66325283", "0.6601364", "0.6529564", "0.65010375", "0.64975137", "0.6472487", "0.64714444", "0.6444321", "0.64330477", "0.6418221", "0.6400071", "0.63848317", "0.63747174", "0.6371322", "0.63571507", "0.63432384", "0.63199526", "0.6302398", "0.62980676", "0.6296262", "0.6295976", "0.62956166", "0.62892276", "0.628455", "0.62764525", "0.6271164", "0.6270313", "0.625883", "0.62523353", "0.6239277", "0.6226169", "0.6220588", "0.62163204", "0.62134767", "0.6181833", "0.61771756", "0.615855", "0.6134471", "0.61140317", "0.6099446", "0.6095976", "0.60700977", "0.6053945", "0.6026498", "0.6026104", "0.5996336", "0.5975581", "0.5975525", "0.59430504", "0.59318036", "0.59127223", "0.5903361", "0.58913493", "0.5867046", "0.58486605", "0.5843442", "0.5824318", "0.5801955", "0.5791209", "0.5784305", "0.5780989", "0.5757855", "0.5731891", "0.57150817", "0.57007337", "0.5667916", "0.5647083", "0.56223744", "0.5600721", "0.5587939", "0.5579325", "0.5569897", "0.5568298", "0.55678016", "0.55574167", "0.55560213", "0.55395544", "0.55394125", "0.5537189", "0.5536474", "0.55343807", "0.5526627", "0.5485314", "0.5476179", "0.545577", "0.54550517" ]
0.8104922
0
Bucket returns a sub bucket by the specified name, nil if it doesn't exist. TODO: implement sub buckets
func (b *Bucket) Bucket(name string) *Bucket { panic("n/i") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (b *Bucket) Bucket(name string) *Bucket {\n nb := b.bucket.Bucket( []byte(name) )\n if nb == nil {\n return nil\n }\n return &Bucket{ tx:b.tx, bucket: nb }\n}", "func (r *RBucket) Bucket(name []byte) Bucket {\n\treq := &BucketRequest{}\n\treq.Key = name\n\tresp := &BucketResponse{}\n\treq.ContextID = r.parent\n\terr := r.r.call(\"srv.Bucket\", req, resp)\n\tif err != nil {\n\t\treturn nil\n\t}\n\tb := &RBucket{}\n\tb.tx = r.tx\n\tb.r = r.r\n\tb.id = resp.BucketID\n\tb.parent = resp.BucketContextID\n\treturn b\n}", "func (tx *Tx) Bucket(name string) *Bucket {\n\tb, ok := tx.buckets[name]\n\tif ok {\n\t\treturn b\n\t}\n\n\tif tx.ro {\n\t\treturn nil\n\t}\n\n\ttx.db.m.Lock()\n\tif b = tx.db.buckets[name]; b == nil {\n\t\ttx.action(backend.ActionCreateBucket, name, \"\", nil)\n\t\tb = newBucket(name, tx.db.opts.DefaultMarshalFn, tx.db.opts.DefaultUnmarshalFn)\n\t\tb.m.Lock()\n\t\ttx.db.buckets[name] = b\n\t}\n\ttx.db.m.Unlock()\n\ttx.buckets[name] = b\n\n\treturn b\n}", "func (this *ConfigServiceClient) GetBucket(name string, version int) (*Bucket, error) {\n if val,ok := this.staticBucketCache.Get(cacheKey(name, version)); ok {\n bucket := val.(*Bucket)\n return bucket, nil\n } else {\n //Use mutex to ensure the bucket will be fetched only once!\n this.mutex.Lock(); defer this.mutex.Unlock()\n\n //Check cache again to see if the another thread has\n //already initialized the bucket\n if val,ok := this.staticBucketCache.Get(cacheKey(name, version)); ok {\n bucket := val.(*Bucket)\n return bucket, nil;\n } else {\n // Initialize the bucket if this the first time\n return this.initStaticBucket(name, version)\n }\n }\n}", "func (service *S3Service) GetBucket(bucketName string) (*S3Bucket,error) {\n return nil,S3Error{\"not implemented yet\"}\n}", "func (db *DB) Bucket(name string) (backend.Bucket, error) {\n\treturn &Bucket{\n\t\tdb: db,\n\t\tname: name,\n\t}, nil\n}", "func (s *Server) Bucket(subKey string) *Bucket {\n\tidx := cityhash.CityHash32([]byte(subKey), uint32(len(subKey))) % s.bucketIdx\n\tif s.Debug {\n\t\ts.log.Infof(\"%s hit channel bucket index: %d use cityhash\", subKey, idx)\n\t}\n\treturn s.buckets[idx]\n}", "func (s *InMemoryRepo) GetBucket(_ context.Context, name string) (Bucket, error) {\n\ts.mu.RLock()\n\tbucket, ok := s.buckets[name]\n\ts.mu.RUnlock()\n\n\tif !ok {\n\t\tbucket.Name = name\n\t}\n\n\treturn bucket, nil\n}", "func (db *DB) Bucket(name string) *Bucket {\n\tif err := db.Bolt.Update(func(tx *bolt.Tx) error {\n\t\tif _, err := tx.CreateBucketIfNotExists([]byte(name)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}); err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn &Bucket{Name: name, db: db}\n}", "func (this *HttpClient) GetBucket(name string, version int) (*Bucket, error) {\n\t// fetch data\n\tresp, err := this.get(name, version, false, INITIAL_VERSION)\n\tif err != nil {\n\t\tlog.Println(\"Error fetching bucket \", err.Error())\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\tif resp.StatusCode != 200 {\n\t\terrResp := &ErrorResp{}\n\t\terr := ffjson.NewDecoder().DecodeReader(resp.Body, errResp)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error reading the response Body\")\n\t\t}\n\t\tlog.Println(\"Error fetching bucket: \", errResp)\n\t\treturn nil, errors.New(errResp.Error())\n\t}\n\n\t// create and return bucket\n\tbucket, err := this.newBucket(resp)\n\tif err != nil {\n\t\tlog.Println(\"Error creating bucket \", err.Error())\n\t\treturn nil, err\n\t}\n\n\treturn bucket, nil\n}", "func GetBucket(tx *bolt.Tx, path Path) (b *bolt.Bucket) {\n\tif len(path) == 0 {\n\t\tpanic(\"Path must have at least one element\")\n\t}\n\tb = tx.Bucket(path[0])\n\tif len(path) == 1 || b == nil {\n\t\treturn\n\t}\n\tfor _, bucket := range path[1:] {\n\t\tif b = b.Bucket(bucket); b == nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (tx *Tx) Bucket(name string) (*Bucket, error) {\n\tif tx.db == nil {\n\t\treturn nil, ErrNoDatabase\n\t}\n\tif !tx.write {\n\t\treturn nil, ErrNotWriteTransaction\n\t}\n\n\tbucket, _ := tx.db.addBucket(name)\n\tb := &Bucket{\n\t\ttx: tx,\n\t\tmanaged: bucket,\n\t}\n\treturn b, nil\n}", "func (service *S3Service) GetOrCreateBucket(bucketName string) (*S3Bucket,error) {\n return nil,S3Error{\"not implemented yet\"}\n}", "func (p *Pool) GetBucket(name string) (*Bucket, error) {\n\trv, ok := p.BucketMap[name]\n\tif !ok {\n\t\treturn nil, &BucketNotFoundError{bucket: name}\n\t}\n\terr := rv.Refresh()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn rv, nil\n}", "func BucketGetChildBucket(b *bolt.Bucket, childBucketName string) (*bolt.Bucket, error) {\n\t// b.Bucket will return a new bucket if this childBucket is not exist.\n\tchildBucket := b.Bucket([]byte(childBucketName))\n\tvar err error\n\t// if can not find the child bucket, then create it now\n\tif childBucket == nil {\n\t\tchildBucket, err = b.CreateBucket([]byte(childBucketName))\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn childBucket, nil\n}", "func (l *LBucket) Bucket(name []byte) Bucket {\n\treturn &LBucket{\n\t\tb: l.b.Bucket(name),\n\t}\n}", "func (s *BucketService) FindBucketByName(ctx context.Context, orgID influxdb.ID, n string) (*influxdb.Bucket, error) {\n\tspan, ctx := tracing.StartSpanFromContext(ctx)\n\tdefer span.Finish()\n\n\tb, err := s.s.FindBucketByName(ctx, orgID, n)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := authorizeReadBucket(ctx, b.OrgID, b.ID); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn b, nil\n}", "func Getbucket(name string, c *Connection) (*s3.Bucket, error) {\n\ts3Connector := s3.New(c.Auth, c.Region)\n\tbucket := s3Connector.Bucket(name)\n\t_, err := bucket.List(\"\", \"\", \"\", 1)\n\tif err != nil {\n\t\t// probably wrong name of bucket\n\t\tlog.Errorf(\"connection to s3\", name, err)\n\t} else {\n\t\tfmt.Printf(\"Connected to s3 bucket: %v\\n\", name)\n\t}\n\treturn bucket, err\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.ID, state *BucketState, opts ...pulumi.ResourceOpt) (*Bucket, error) {\n\tinputs := make(map[string]interface{})\n\tif state != nil {\n\t\tinputs[\"accelerationStatus\"] = state.AccelerationStatus\n\t\tinputs[\"acl\"] = state.Acl\n\t\tinputs[\"arn\"] = state.Arn\n\t\tinputs[\"bucket\"] = state.Bucket\n\t\tinputs[\"bucketDomainName\"] = state.BucketDomainName\n\t\tinputs[\"bucketPrefix\"] = state.BucketPrefix\n\t\tinputs[\"bucketRegionalDomainName\"] = state.BucketRegionalDomainName\n\t\tinputs[\"corsRules\"] = state.CorsRules\n\t\tinputs[\"forceDestroy\"] = state.ForceDestroy\n\t\tinputs[\"hostedZoneId\"] = state.HostedZoneId\n\t\tinputs[\"lifecycleRules\"] = state.LifecycleRules\n\t\tinputs[\"loggings\"] = state.Loggings\n\t\tinputs[\"objectLockConfiguration\"] = state.ObjectLockConfiguration\n\t\tinputs[\"policy\"] = state.Policy\n\t\tinputs[\"region\"] = state.Region\n\t\tinputs[\"replicationConfiguration\"] = state.ReplicationConfiguration\n\t\tinputs[\"requestPayer\"] = state.RequestPayer\n\t\tinputs[\"serverSideEncryptionConfiguration\"] = state.ServerSideEncryptionConfiguration\n\t\tinputs[\"tags\"] = state.Tags\n\t\tinputs[\"versioning\"] = state.Versioning\n\t\tinputs[\"website\"] = state.Website\n\t\tinputs[\"websiteDomain\"] = state.WebsiteDomain\n\t\tinputs[\"websiteEndpoint\"] = state.WebsiteEndpoint\n\t}\n\ts, err := ctx.ReadResource(\"aws:s3/bucket:Bucket\", name, id, inputs, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Bucket{s: s}, nil\n}", "func (this *ConfigServiceClient) GetDynamicBucket(name string) (*DynamicBucket, error) {\n if val,ok := this.dynamicBucketCache.Get(name); ok {\n dynamicBucket := val.(*DynamicBucket)\n return dynamicBucket, nil\n } else {\n //Use mutex to ensure the bucket will be fetched only once!\n this.mutex.Lock(); defer this.mutex.Unlock()\n\n //Check cache again to see if the another thread has\n //already initialized the bucket\n if val,ok := this.dynamicBucketCache.Get(name); ok {\n dynamicBucket := val.(*DynamicBucket)\n return dynamicBucket, nil;\n } else {\n // Initialize the bucket if this the first time\n return this.initDynamicBucket(name)\n }\n }\n}", "func GetOrCreateBucket(bucket Bucket, name string) (b Bucket, err error) {\n\tif bucket == nil {\n\t\treturn nil, ErrInvalidArgument\n\t}\n\n\tif b = bucket.Bucket(name); b == nil {\n\t\tb, err = bucket.NewBucket(name)\n\t}\n\treturn\n}", "func BucketGet(tx *bolt.Tx, bucket, key string) []byte {\n\tb := tx.Bucket([]byte(bucket))\n\tif b == nil {\n\t\treturn nil\n\t}\n\treturn b.Get([]byte(key))\n}", "func (instance *Instance) GetBucket() (bucket string) {\n\tif val := instance.GetIndexInstance(); val != nil {\n\t\treturn val.GetDefinition().GetBucket()\n\t} else {\n\t\t// TODO: should we panic ?\n\t}\n\treturn\n}", "func (m Mux) Bucket(ctx context.Context, rawurl string) (Bucket, string, error) {\n\tu, err := url.Parse(rawurl)\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\tstore, ok := m[u.Scheme]\n\tif !ok {\n\t\treturn nil, \"\", errors.E(errors.NotSupported, \"blob.Bucket\", rawurl,\n\t\t\terrors.Errorf(\"no implementation for scheme %s\", u.Scheme))\n\t}\n\tbucket, err := store.Bucket(ctx, u.Host)\n\tif err != nil {\n\t\treturn nil, \"\", err\n\t}\n\treturn bucket, strings.TrimPrefix(rawurl, bucket.Location()), err\n}", "func (p Path) Bucket() string { return p.bucket }", "func (r *Report) GetBucket(bucketname, reportPath string) *Bucket {\n\tfor _, bucket := range r.Buckets {\n\t\tif bucket.Bucketname == bucketname && bucket.ReportPath == reportPath {\n\t\t\treturn bucket\n\t\t}\n\t}\n\treturn nil\n}", "func (m *Meta) GetBucket(bucketName string, willNeed bool) (bucket Bucket, err error) {\n\tgetBucket := func() (b interface{}, err error) {\n\t\tb, err = m.Client.GetBucket(bucketName)\n\t\thelper.Logger.Println(10, \"GetBucket CacheMiss. bucket:\", bucketName)\n\t\treturn b, err\n\t}\n\tunmarshaller := func(in []byte) (interface{}, error) {\n\t\tvar bucket Bucket\n\t\terr := helper.MsgPackUnMarshal(in, &bucket)\n\t\treturn bucket, err\n\t}\n\tb, err := m.Cache.Get(redis.BucketTable, bucketName, getBucket, unmarshaller, willNeed)\n\tif err != nil {\n\t\treturn\n\t}\n\tbucket, ok := b.(Bucket)\n\tif !ok {\n\t\thelper.Debugln(\"Cast b failed:\", b)\n\t\terr = ErrInternalError\n\t\treturn\n\t}\n\treturn bucket, nil\n}", "func (obj *bucket) Bucket() buckets.Bucket {\n\treturn obj.bucket\n}", "func (this *ConfigServiceClient) initStaticBucket(name string, version int) (*Bucket, error) {\n log.Println(\"Initializing Config bucket: \" + name)\n\n err := ValidateBucketName(name)\n if err != nil {\n return nil, err\n }\n bucket, err := this.httpClient.GetBucket(name, version)\n if err != nil {\n log.Println(\"Error fetching bucket: \", err)\n return nil, err\n } else {\n this.staticBucketCache.Add(cacheKey(name, version), bucket)\n return bucket, nil\n }\n}", "func (rt RouteTable) GetBucket(id string) (bucket Bucket, ok bool) {\n\ti, err := hex.DecodeString(id)\n\tif err != nil {\n\t\treturn KBucket{}, false\n\t}\n\tb := rt.ht.GetBucket(i)\n\tif b == nil {\n\t\treturn KBucket{}, false\n\t}\n\n\treturn KBucket{\n\t\tnodes: convertNetworkNodes(b),\n\t}, true\n}", "func (b *Bucket) Get(ctx context.Context, name string) (io.ReadCloser, error) {\n\treturn b.getRange(ctx, b.name, name, 0, -1)\n}", "func (this *ConfigServiceClient) initDynamicBucket(name string) (*DynamicBucket, error) {\n log.Println(\"Initializing Config bucket: \" + name)\n\n dynamicBucket := &DynamicBucket{ httpClient: this.httpClient }\n\n err := ValidateBucketName(name)\n if err != nil {\n return nil, err\n }\n\n err = dynamicBucket.init(name)\n\n if err != nil {\n log.Println(\"Error fetching bucket: \", err)\n return nil, err\n } else {\n this.dynamicBucketCache.Add(name, dynamicBucket)\n go this.httpClient.WatchBucket(name, this.dynamicBucketCache, dynamicBucket)\n return dynamicBucket, nil\n }\n}", "func getBucket(instanceID string) string {\n\treturn fmt.Sprintf(bucketTemplate, instanceID)\n}", "func (q *Bucket) Get(k []byte) ([]byte, error) {\n\tvar v []byte\n\treturn v, q.db.boltDB.Update(func(tx *bolt.Tx) error {\n\t\tbucket, err := tx.CreateBucketIfNotExists([]byte(q.name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tv = bucket.Get(k)\n\t\tif v == nil {\n\t\t\treturn backend.ErrKeyNotFound\n\t\t}\n\t\treturn nil\n\t})\n}", "func (b *Bucket) Get(key string) (*brazier.Item, error) {\n\tvar data []byte\n\terr := b.node.Get(\"items\", key, &data)\n\tif err != nil {\n\t\tif err == storm.ErrNotFound {\n\t\t\treturn nil, store.ErrNotFound\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"boltdb.bucket.Get failed to fetch item\")\n\t}\n\n\treturn &brazier.Item{\n\t\tKey: key,\n\t\tData: data,\n\t}, nil\n}", "func (sr *scrutinyRepository) lookupBucketName(durationKey string) string {\n\tswitch durationKey {\n\tcase DURATION_KEY_WEEK:\n\t\t//data stored in the last week\n\t\treturn sr.appConfig.GetString(\"web.influxdb.bucket\")\n\tcase DURATION_KEY_MONTH:\n\t\t// data stored in the last month (after the first week)\n\t\treturn fmt.Sprintf(\"%s_weekly\", sr.appConfig.GetString(\"web.influxdb.bucket\"))\n\tcase DURATION_KEY_YEAR:\n\t\t// data stored in the last year (after the first month)\n\t\treturn fmt.Sprintf(\"%s_monthly\", sr.appConfig.GetString(\"web.influxdb.bucket\"))\n\tcase DURATION_KEY_FOREVER:\n\t\t//data stored before the last year\n\t\treturn fmt.Sprintf(\"%s_yearly\", sr.appConfig.GetString(\"web.influxdb.bucket\"))\n\t}\n\treturn sr.appConfig.GetString(\"web.influxdb.bucket\")\n}", "func (b *Bucket) Get(ctx context.Context, name string) (io.ReadCloser, error) {\n\treturn b.bkt.Object(name).NewReader(ctx)\n}", "func (b *Bucket) CreateBucket(name string) (*Bucket, error) { panic(\"n/i\") }", "func (rc RpcCmd) GetBucket() (interface{}, error) {\n\tcfg := config.GetConfigure()\n\trpcAddr := fmt.Sprintf(\"%s:%s\", cfg.LeaderRpcC.Addr, cfg.LeaderRpcC.Port)\n\topRequest := pb.OpRequest{\n\t\tOp: rc.Op,\n\t\tBucket: rc.Bucket,\n\t\tKey: rc.Key,\n\t\tValue: rc.Value,\n\t}\n\treply, err := NewRiotRPCClient().RPCRequest(rpcAddr, &opRequest)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif reply.Status != 1 {\n\t\treturn nil, fmt.Errorf(\"%s\", reply.Msg)\n\t}\n\treturn reply.Value, nil\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *BucketState, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tvar resource Bucket\n\terr := ctx.ReadResource(\"alicloud:oss/bucket:Bucket\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (l *ChannelList) Bucket(key string) *ChannelBucket {\n\tidx := int(hash.HashCrc32String(key) % uint32(Conf.ChannelBucketCount))\n\treturn l.channels[idx]\n}", "func createBucket(parentBucket *bbolt.Bucket, name []byte) (*bbolt.Bucket, error) {\n\treturn parentBucket.CreateBucket(name)\n}", "func (db *DB) GetBucket(ctx context.Context, bucketName string) (bucket storj.Bucket, err error) {\n\tdefer mon.Task()(&ctx)(&err)\n\n\tif bucketName == \"\" {\n\t\treturn storj.Bucket{}, storj.ErrNoBucket.New(\"\")\n\t}\n\n\tbucket, err = db.metainfo.GetBucket(ctx, GetBucketParams{\n\t\tName: []byte(bucketName),\n\t})\n\treturn bucket, storj.ErrBucket.Wrap(err)\n}", "func (p *bucketProvider) getObject(bucket, path string) *storage.ObjectHandle {\n\treturn p.client.Bucket(bucket).Object(path)\n}", "func (dao *Dao) GetBucketName(acc string) (string, error) {\n\tvar g mysql.Gooq\n\n\tg.SQL.Select(userpo.BucketName).From(userpo.Table).Where(c(userpo.Acc).Eq(\"?\"))\n\tg.AddValues(acc)\n\n\tvar bucketName string\n\terr := g.QueryRow(func(row *sql.Row) error {\n\t\treturn row.Scan(&bucketName)\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn bucketName, nil\n}", "func (tx *Tx) GetBucket(bucketID string) (*Bucket, error) {\n\tvar bucket Bucket\n\terr := tx.Get(&bucket, \"SELECT * FROM bucket WHERE id = $1;\", bucketID)\n\tif err != nil {\n\t\treturn nil, errors.InternalError(err)\n\t}\n\treturn &bucket, nil\n}", "func (c *Container) Bucket() *storage.BucketHandle {\n\treturn c.client.Bucket(c.name)\n}", "func (b *Bucket) Get(key string) []byte {\n return b.bucket.Get( []byte(key) )\n}", "func (jm *JobManager) GetBucket(task *api.TaskInfo) *Bucket {\n\tindex, ok := jm.podInBucket[task.Pod.UID]\n\tif !ok || index == OutOfBucket {\n\t\treturn nil\n\t}\n\n\tbucket := jm.buckets[index]\n\treturn bucket\n}", "func (r *RBucket) Get(key []byte) []byte {\n\treq := &GetReqeust{}\n\tresp := &GetResponse{}\n\treq.Key = key\n\treq.BucketID = r.id\n\treq.ContextID = r.parent\n\terr := r.r.call(\"srv.Get\", req, resp)\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn resp.Val\n}", "func CreateSearchBucketFromDiscriminatorValue(parseNode i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.ParseNode)(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, error) {\n return NewSearchBucket(), nil\n}", "func (handler *BucketWebHandler) Retreive(w http.ResponseWriter, r *http.Request) {\n\tparams := context.Get(r, CtxParamsKey).(httprouter.Params)\n\tif params.ByName(\"name\") == \"\" {\n\t\trespondWithError(w, http.StatusBadRequest, \"Invalid bucket name\", nil)\n\t\treturn\n\t}\n\n\tfs := handler.Session.DB(os.Getenv(EnvGridFSDatabase)).GridFS(os.Getenv(EnvGridFSPrefix))\n\n\tvar meta ObjectMeta\n\tbucket := &BucketMeta{}\n\tbucket.Name = params.ByName(\"name\")\n\tbucket.Objects = []ObjectMeta{}\n\n\titer := fs.Find(bson.M{\"metadata.bucket\": params.ByName(\"name\")}).Iter()\n\tfor iter.Next(&meta) {\n\t\t// drop bucket name\n\t\tif _, ok := meta.Metadata[\"bucket\"]; ok {\n\t\t\tdelete(meta.Metadata, \"bucket\")\n\t\t}\n\t\tbucket.Objects = append(bucket.Objects, meta)\n\t}\n\tif err := iter.Close(); err != nil {\n\t\trespondWithError(w, http.StatusInternalServerError, \"Operational error\", err)\n\t\treturn\n\t}\n\n\tjson.NewEncoder(w).Encode(bucket)\n}", "func GetBucket(endpoint, poolname, bucketname string) (*Bucket, error) {\n\tvar err error\n\tclient, err := Connect(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tpool, err := client.GetPool(poolname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn pool.GetBucket(bucketname)\n}", "func NewBucket() *Bucket {\n\treturn &Bucket{objects: map[string][]byte{}}\n}", "func InitBucket() (b map[string]Bucket, err error) {\r\n\tvar (\r\n\t\titem Bucket\r\n\t)\r\n\tb = make(map[string]Bucket)\r\n\t// bucket test\r\n\titem.BucketName = \"test\"\r\n\titem.Property = 2\r\n\titem.KeyId = \"121bce6492eba701\"\r\n\titem.KeySecret = \"1eb80603e85842542f9736eb13b7e1\"\r\n\tb[\"test\"] = item\r\n\r\n\treturn\r\n}", "func createBucket(name string) func(*bolt.Tx) error {\n\treturn func(tx *bolt.Tx) error {\n\t\tif _, err := tx.CreateBucketIfNotExists([]byte(name)); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n}", "func (*Template) Bucket() string {\n\treturn bucketTemplates\n}", "func (db *DB) Get(key string, bucket ...string) Value {\n\tdb.mux.RLock()\n\tdefer db.mux.RUnlock()\n\tb := &db.root\n\tfor _, bn := range bucket {\n\t\tif b = b.Buckets[bn]; b == nil {\n\t\t\treturn nil\n\t\t}\n\t}\n\treturn b.Get(key)\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *BucketState, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tvar resource Bucket\n\terr := ctx.ReadResource(\"kubernetes:storage.crossplane.io/v1alpha1:Bucket\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (r *Bucket) Bucket() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"bucket\"])\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *BucketState, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tvar resource Bucket\n\terr := ctx.ReadResource(\"aws-native:lightsail:Bucket\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *BucketState, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tvar resource Bucket\n\terr := ctx.ReadResource(\"aws:s3control/bucket:Bucket\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (o BucketOutput) Bucket() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Bucket) pulumi.StringOutput { return v.Bucket }).(pulumi.StringOutput)\n}", "func (c *FakeFirebaseStorageBuckets) Get(ctx context.Context, name string, options v1.GetOptions) (result *v1alpha1.FirebaseStorageBucket, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewGetAction(firebasestoragebucketsResource, c.ns, name), &v1alpha1.FirebaseStorageBucket{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\treturn obj.(*v1alpha1.FirebaseStorageBucket), err\n}", "func getEtcdBucket(path string) string {\n\tidx := strings.LastIndex(path, \"/\")\n\tif idx == -1 {\n\t\tpanic(\"path with no slashes \" + path)\n\t}\n\tbucket := path[:idx]\n\tif len(bucket) == 0 {\n\t\tpanic(\"invalid bucket for path \" + path)\n\t}\n\treturn bucket\n}", "func GetCephBucket(bucket string) *s3.Bucket {\n\tconn := GetCephConn()\n\treturn conn.Bucket(bucket)\n}", "func (i *Item) GetBucket() objectstorage.Bucket {\n\treturn i.folder.GetBucket()\n}", "func (db *DumbDB) Get(key []byte, bucket string) (ret_val []byte, err error) {\n\n\terr = db.dbP.View(func(tx *bolt.Tx) error {\n\t\tbkt := tx.Bucket([]byte(bucket))\n\t\tif bkt == nil {\n\t\t\tdb.err_log.Println(\"Bucket not created yet.\")\n\t\t\treturn bolt.ErrBucketNotFound\n\t\t}\n\n\t\tret_val = bkt.Get(key)\n\t\tif ret_val != nil {\n\t\t\tdb.info_log.Println(\"Found key.\")\n\t\t\treturn nil\n\t\t}\n\n\t\treturn bolt.ErrKeyRequired\n\t})\n\treturn\n}", "func (api *bucketAPI) Find(meta *api.ObjectMeta) (*Bucket, error) {\n\t// find the object\n\tobj, err := api.ct.FindObject(\"Bucket\", meta)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// asset type\n\tswitch obj.(type) {\n\tcase *Bucket:\n\t\thobj := obj.(*Bucket)\n\t\treturn hobj, nil\n\tdefault:\n\t\treturn nil, errors.New(\"incorrect object type\")\n\t}\n}", "func (h *hashMap) initializeBucket(index uint32) *bucket {\n\tparentIndex := h.getParentIndex(index)\n\n\tif h.bucketSegments.getBucket(parentIndex) == nil {\n\t\th.initializeBucket(parentIndex)\n\t}\n\n\tdummy := h.bucketSegments.getBucket(parentIndex).getDummy(index)\n\n\tif dummy != nil {\n\t\th.bucketSegments.setBucket(index, dummy)\n\t}\n\n\treturn dummy\n}", "func (km KeyValueMap) Bucket() string {\n\treturn km[kmBucket]\n}", "func (b *Bucket) Name() string {\n\treturn b.sb.BucketName\n}", "func Bucket(from string) (string, error) {\n\treturn makeRequest(\"bucket\", from)\n}", "func (c *Collection) Bucket() *Bucket {\n\treturn c.bucket\n}", "func (b *Bucket) ChkBucket() (bool, error) {\n\t// set session\n\tsess, err := b.setSession()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\t// list buckets\n\tsvc := b.newS3func(sess)\n\tresult, err := svc.ListBuckets(nil)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tklog.Info(\"Buckets:\")\n\tfound := false\n\tfor _, bu := range result.Buckets {\n\t\tklog.Infof(\"-- %s created on %s\\n\", aws.StringValue(bu.Name), aws.TimeValue(bu.CreationDate))\n\t\tif aws.StringValue(bu.Name) == b.BucketName {\n\t\t\tfound = true\n\t\t}\n\t}\n\n\treturn found, nil\n}", "func (items BlueprintsMap) Bucket() []byte {\n\treturn []byte(DBBucket)\n}", "func fetchPaymentBucket(tx *bolt.Tx) (*bolt.Bucket, error) {\n\tpbkt := tx.Bucket(poolBkt)\n\tif pbkt == nil {\n\t\tdesc := fmt.Sprintf(\"bucket %s not found\", string(poolBkt))\n\t\treturn nil, MakeError(ErrBucketNotFound, desc, nil)\n\t}\n\tbkt := pbkt.Bucket(paymentBkt)\n\tif bkt == nil {\n\t\tdesc := fmt.Sprintf(\"bucket %s not found\", string(paymentBkt))\n\t\treturn nil, MakeError(ErrBucketNotFound, desc, nil)\n\t}\n\treturn bkt, nil\n}", "func (b *Bucket) CreateBucket( name string ) (*Bucket, error) {\n nb, err := b.bucket.CreateBucket( []byte(name) )\n if err != nil {\n return nil, err\n }\n return &Bucket{tx: b.tx, bucket: nb}, nil\n}", "func GetBucketByAuthID(context *gin.Context) {\n\tresponseCode := constant.INVALID_PARAMS\n\tauthID, authErr := strconv.Atoi(context.Query(\"auth_id\"))\n\toffset := context.GetInt(\"offset\")\n\tif authErr != nil{\n\t\t//log.Println(authErr)\n\t\tutils.AppLogger.Info(authErr.Error(), zap.String(\"service\", \"GetBucketByAuthID()\"))\n\t\tcontext.AbortWithStatusJSON(http.StatusBadRequest, gin.H{\n\t\t\t\"code\": responseCode,\n\t\t\t\"data\": make(map[string]string),\n\t\t\t\"msg\": constant.GetMessage(responseCode),\n\t\t})\n\t\treturn\n\t}\n\n\tvalidCheck := validation.Validation{}\n\tvalidCheck.Required(authID, \"auth_id\").Message(\"Must have auth id\")\n\tvalidCheck.Min(authID, 1, \"auth_id\").Message(\"Auth id should be positive\")\n\tvalidCheck.Min(offset, 0, \"page_offset\").Message(\"Page offset must be >= 0\")\n\n\tdata := make(map[string]interface{})\n\tif !validCheck.HasErrors() {\n\t\tif buckets, err := models.GetBucketByAuthID(uint(authID), offset); err != nil {\n\t\t\tresponseCode = constant.INTERNAL_SERVER_ERROR\n\t\t} else {\n\t\t\tresponseCode = constant.BUCKET_GET_SUCCESS\n\t\t\tdata[\"buckets\"] = buckets\n\t\t}\n\t} else {\n\t\tfor _, err := range validCheck.Errors {\n\t\t\t//log.Println(err.Message)\n\t\t\tutils.AppLogger.Info(err.Message, zap.String(\"service\", \"GetBucketByAuthID()\"))\n\t\t}\n\t}\n\n\tcontext.JSON(http.StatusOK, gin.H{\n\t\t\"code\": responseCode,\n\t\t\"data\": data,\n\t\t\"msg\": constant.GetMessage(responseCode),\n\t})\n}", "func (g *gcsclient) Bucket() string {\n\treturn g.bucket\n}", "func NewBucket(bucketName, donutName string, nodes map[string]Node) (Bucket, error) {\n\tif bucketName == \"\" {\n\t\treturn nil, errors.New(\"invalid argument\")\n\t}\n\tb := bucket{}\n\tb.name = bucketName\n\tb.donutName = donutName\n\tb.objects = make(map[string]Object)\n\tb.nodes = nodes\n\treturn b, nil\n}", "func (rs *Restake) BucketIndex() uint64 { return rs.bucketIndex }", "func (b *Bucket) Get(_ context.Context, name string) (io.ReadCloser, error) {\n\tfile, ok := b.objects[name]\n\tif !ok {\n\t\treturn nil, errors.Errorf(\"no such file %s\", name)\n\t}\n\n\treturn ioutil.NopCloser(bytes.NewReader(file)), nil\n}", "func (c control) Bucket() *TagBucket {\n\treturn c.bucket\n}", "func (f *File) Bucket() string {\n\treturn f.bucket\n}", "func (in *Bucket) DeepCopy() *Bucket {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Bucket)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (s storageBucketNamespaceLister) Get(name string) (*v1beta1.StorageBucket, error) {\n\tobj, exists, err := s.indexer.GetByKey(s.namespace + \"/\" + name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, errors.NewNotFound(v1beta1.Resource(\"storagebucket\"), name)\n\t}\n\treturn obj.(*v1beta1.StorageBucket), nil\n}", "func NewBucket(logger log.Logger, conf []byte, component string) (*Bucket, error) {\n\tif logger == nil {\n\t\tlogger = log.NewNopLogger()\n\t}\n\n\tconfig, err := parseConfig(conf)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"parsing BOS configuration\")\n\t}\n\n\treturn NewBucketWithConfig(logger, config, component)\n}", "func getRTDBucket(token string) (string, error) {\n\ttd := common.GetTokenDetails(token)\n\tif td == nil {\n\t\treturn \"\", fmt.Errorf(\"failed to get token details for token %q\", token)\n\t}\n\n\tb := fmt.Sprintf(common.REAL_TIME_DATA_BUCKET, td.Exchange, td.Segment)\n\treturn b, nil\n\n}", "func GetBucket(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *BucketState, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tvar resource Bucket\n\terr := ctx.ReadResource(\"aws-native:s3outposts:Bucket\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func LocalBucket(localConnectStr, bucketName string) (*couchbase.Bucket, error) {\n\tlogger_utils.Debugf(\"Getting local bucket name=%v\\n\", bucketName)\n\n\tpool, err := LocalPool(localConnectStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbucket, err := pool.GetBucket(bucketName)\n\tif err != nil {\n\t\treturn nil, NewEnhancedError(fmt.Sprintf(\"Error getting bucket, %v, from pool.\", bucketName), err)\n\t}\n\n\tlogger_utils.Debugf(\"Got local bucket successfully name=%v\\n\", bucket.Name)\n\treturn bucket, err\n}", "func GetOrEnsureBucket(tx *bolt.Tx, key []byte) *bolt.Bucket {\n\tif tx.Writable() {\n\t\tb, _ := tx.CreateBucketIfNotExists(key)\n\t\treturn b\n\t}\n\treturn tx.Bucket(key)\n}", "func NewBucket(ctx *pulumi.Context,\n\tname string, args *BucketArgs, opts ...pulumi.ResourceOption) (*Bucket, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.DestinationRegion == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'DestinationRegion'\")\n\t}\n\tvar resource Bucket\n\terr := ctx.RegisterRemoteComponentResource(\"replicatedbucket:index:Bucket\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (b *Buckets) GetLocalBucket(ctx context.Context, conf Config) (*Bucket, error) {\n\tcwd, err := filepath.Abs(conf.Path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbc, found, err := b.config.NewConfig(cwd, flags, false)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif conf.Thread.Defined() {\n\t\tbc.Viper.Set(\"thread\", conf.Thread.String())\n\t}\n\tif conf.Key != \"\" {\n\t\tbc.Viper.Set(\"key\", conf.Key)\n\t}\n\tif bc.Viper.Get(\"thread\") == nil || bc.Viper.Get(\"key\") == nil {\n\t\treturn nil, ErrNotABucket\n\t}\n\tcmd.ExpandConfigVars(bc.Viper, bc.Flags)\n\tbuck := &Bucket{\n\t\tcwd: cwd,\n\t\tconf: bc,\n\t\tclients: b.clients,\n\t\tauth: b.auth,\n\t\tpushBlock: make(chan struct{}, 1),\n\t}\n\tif found {\n\t\tbp, err := buck.Path()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = buck.loadLocalRepo(ctx, bp, b.repoName(), true); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn buck, nil\n}", "func (o AccessPointOutput) Bucket() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AccessPoint) pulumi.StringOutput { return v.Bucket }).(pulumi.StringOutput)\n}", "func NewBucket(s *Store, key string, node storm.Node) *Bucket {\n\treturn &Bucket{\n\t\tkey: key,\n\t\tstore: s,\n\t\tnode: node,\n\t}\n}", "func (o BucketIntelligentTieringConfigurationOutput) Bucket() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *BucketIntelligentTieringConfiguration) pulumi.StringOutput { return v.Bucket }).(pulumi.StringOutput)\n}", "func (this *HttpClient) WatchBucket(name string, cache *lru.Cache, dynamicBucket *DynamicBucket){\n\tbackOff := &backoff.Backoff{\n\t\tMin: 1 * time.Second,\n\t\tMax: 300 * time.Second,\n\t\tJitter: true,\n\t}\n\tfor {\n\t\tlog.Println(\"Setting watch on bucket: \", name)\n\t\twatchAsync := WatchAsync{\n\t\t\tbucketName: name,\n\t\t\tdynamicBucket: dynamicBucket,\n\t\t\tasyncResp: make(chan *BucketResponse),\n\t\t\thttpClient: this,\n\t\t}\n\n\t\tselect {\n\t\tcase bucketResp := <- watchAsync.watch():\n\n\t\t\tif bucketResp.err != nil && bucketResp.statusCode == 404 {\n\t\t\t\tlog.Println(\"Stopping watch on bucket: \", name)\n\t\t\t\tdynamicBucket.DeleteBucket()\n\t\t\t\tcache.Remove(name)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif bucketResp.err != nil {\n\t\t\t\tlog.Println(\"Error fetching bucket: \", bucketResp.err)\n\t\t\t\tdynamicBucket.Disconnected(bucketResp.err)\n\t\t\t\ttime.Sleep(backOff.Duration())\n\t\t\t\tcontinue;\n\t\t\t}\n\n\t\t backOff.Reset()\n\t\t\tdynamicBucket.updateBucket(bucketResp.bucket)\n\n\t\tcase <- dynamicBucket.isShutdown():\n\t\t\tlog.Println(\"Stopping watch on bucket: \", name)\n\t\t\treturn\n\n\t\t}\n\t}\n}", "func GetPriceBucket(bid openrtb2.Bid, targetingData targetData) string {\n\tcpmStr := \"\"\n\tbucketMax := 0.0\n\tbucketMin := 0.0\n\tincrement := 0.0\n\n\tconfig := targetingData.priceGranularity //assign default price granularity\n\n\tif bidType, err := getMediaTypeForBid(bid); err == nil {\n\t\tif bidType == openrtb_ext.BidTypeBanner && targetingData.mediaTypePriceGranularity.Banner != nil {\n\t\t\tconfig = *targetingData.mediaTypePriceGranularity.Banner\n\t\t} else if bidType == openrtb_ext.BidTypeVideo && targetingData.mediaTypePriceGranularity.Video != nil {\n\t\t\tconfig = *targetingData.mediaTypePriceGranularity.Video\n\t\t} else if bidType == openrtb_ext.BidTypeNative && targetingData.mediaTypePriceGranularity.Native != nil {\n\t\t\tconfig = *targetingData.mediaTypePriceGranularity.Native\n\t\t}\n\t}\n\n\tprecision := *config.Precision\n\n\tcpm := bid.Price\n\tfor i := 0; i < len(config.Ranges); i++ {\n\t\tif config.Ranges[i].Max > bucketMax {\n\t\t\tbucketMax = config.Ranges[i].Max\n\t\t}\n\t\t// find what range cpm is in\n\t\tif cpm >= config.Ranges[i].Min && cpm <= config.Ranges[i].Max {\n\t\t\tincrement = config.Ranges[i].Increment\n\t\t\tbucketMin = config.Ranges[i].Min\n\t\t}\n\t}\n\n\tif cpm > bucketMax {\n\t\t// We are over max, just return that\n\t\tcpmStr = strconv.FormatFloat(bucketMax, 'f', precision, 64)\n\t} else if increment > 0 {\n\t\t// If increment exists, get cpm string value\n\t\tcpmStr = getCpmTarget(cpm, bucketMin, increment, precision)\n\t}\n\n\treturn cpmStr\n}", "func restGetBucketPath(w http.ResponseWriter, r *http.Request) {\n\tbucketName := r.FormValue(\"name\")\n\tif len(bucketName) < 1 {\n\t\thttp.Error(w, \"bucket name is too short or is missing\", 400)\n\t\treturn\n\t}\n\tpath, err := BucketPath(bucketName)\n\tif err != nil {\n\t\thttp.Error(w,\n\t\t\tfmt.Sprintf(\"could not compute BucketPath for name: %v, err: %v\",\n\t\t\t\tbucketName, err), 400)\n\t\treturn\n\t}\n\tw.Write([]byte(path))\n}" ]
[ "0.7497593", "0.6905284", "0.67476887", "0.6684629", "0.6612725", "0.66109204", "0.6608157", "0.6592217", "0.6591069", "0.65037", "0.63973314", "0.6395263", "0.63906217", "0.63420254", "0.62299395", "0.6223852", "0.6173834", "0.6157914", "0.61513454", "0.6138822", "0.6119121", "0.6112312", "0.6105442", "0.60886335", "0.6072091", "0.5996802", "0.5926979", "0.5905567", "0.5891682", "0.5877693", "0.5850371", "0.5845841", "0.58185565", "0.5785163", "0.5764241", "0.5759186", "0.5748859", "0.57209635", "0.571409", "0.56779796", "0.5671205", "0.56625617", "0.5655674", "0.56499606", "0.5642018", "0.5638478", "0.563196", "0.5623051", "0.5603677", "0.5602584", "0.55844074", "0.5582106", "0.55774", "0.5572042", "0.5554825", "0.55427945", "0.5540689", "0.5539822", "0.5529347", "0.5523054", "0.5520928", "0.5497853", "0.5489035", "0.5470134", "0.5469944", "0.5469704", "0.54596555", "0.5445589", "0.54357845", "0.5435539", "0.54335845", "0.5425658", "0.54253846", "0.54062754", "0.5404259", "0.5394866", "0.53901595", "0.5387494", "0.5380072", "0.5372489", "0.5365238", "0.53607655", "0.5347721", "0.53342104", "0.5329738", "0.53201354", "0.5312784", "0.53102493", "0.5309203", "0.53070176", "0.5287641", "0.5283285", "0.5279969", "0.5261651", "0.5261305", "0.5260614", "0.52519864", "0.52448803", "0.5230865", "0.52262765" ]
0.7126938
1
FindAll Find all aritcle and order created_at to desc
func (repo *ArticleRepository) FindAll(articles *[]domain.Article) error { err := repo.DB.Order("created_at desc").Limit(100).Find(articles).Error return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (qs SysDBQuerySet) OrderDescByCreatedAt() SysDBQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at DESC\"))\n}", "func GetAllAds(limit, offset int, sort, crease string) ([]*Ads, error) {\n\to := orm.NewOrm()\n\tvar ads []*Ads\n\n\tif sort == \"date\" {\n\t\tsort = \"created_at\"\n\t}\n\n\tif crease == \"decrease\" {\n\t\tsort = \"-\" + sort\n\t}\n\n\t_, err := o.QueryTable(new(Ads)).OrderBy(sort).Limit(limit).Offset(offset).All(&ads, \"name\", \"img1\", \"price\")\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\n\treturn ads, err\n}", "func GetAllChequera(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Chequera)).RelatedSel(1)\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Chequera\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAllReintegro(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Reintegro)).RelatedSel()\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Reintegro\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (qs GroupQuerySet) OrderDescByCreatedAt() GroupQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at DESC\"))\n}", "func (qs ConstraintQuerySet) OrderDescByCreatedAt() ConstraintQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at DESC\"))\n}", "func (qs SysDBQuerySet) OrderDescByUpdatedAt() SysDBQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at DESC\"))\n}", "func (qs SysDBQuerySet) OrderAscByCreatedAt() SysDBQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at ASC\"))\n}", "func (qs SysDBQuerySet) OrderAscByUpdatedAt() SysDBQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at ASC\"))\n}", "func (d *DB) All() ([]Todo, error) {\n\tdb, err := gorm.Open(\"sqlite3\", \"model/DB/test.sqlite3\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar items []Todo\n\tdb.Order(\"created_at DESC\").Find(&items)\n\tdb.Close()\n\n\treturn items, err\n}", "func GetAllLifeInsurance(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(LifeInsurance))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []LifeInsurance\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (a Author) GetAll(cfg *config.Config) ([]*Author, error) {\n\tsession := cfg.Session.Copy()\n\tvar authors []*Author\n\tif err := cfg.Database.C(AuthorCollection).Find(bson.M{}).Sort(\"-dateCreated\").All(&authors); err != nil {\n\t\treturn authors, err\n\t}\n\tdefer session.Close()\n\treturn authors, nil\n}", "func GetAllCourse(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Course))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Course\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (d *DBRepository) findAll(ctx context.Context) (*Travels, error) {\n\tc, err := d.Collection.Find(ctx, bson.D{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar travels Travels\n\n\tfor c.Next(ctx) {\n\t\tvar travel Travel\n\t\tif err := c.Decode(&travel); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttravels = append(travels, travel)\n\t}\n\tif err := c.Close(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &travels, nil\n}", "func (qs GroupQuerySet) OrderDescByUpdatedAt() GroupQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at DESC\"))\n}", "func GetAllSeguimiento(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Seguimiento))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Seguimiento\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (db *Service) FindAll(\n\tsc datatype.ServiceContainer,\n\tuser *datatype.User,\n\ttimelineFilter datatype.TimelineFilter,\n\toffset int,\n\tlimit int,\n) ([]datatype.TimelineEntry, bool, error) {\n\tresult := []datatype.TimelineEntry{}\n\tvar clause string\n\tif timelineFilter.Type == datatype.HOMETIMELINE {\n\t\tclause = \"WHERE b.status=1\"\n\t} else if timelineFilter.Type == datatype.ASSETTIMELINE {\n\t\tclause = fmt.Sprintf(\n\t\t\t\" WHERE b.status<>%d AND asset.id = %d\",\n\t\t\tdatatype.BlockRejected,\n\t\t\ttimelineFilter.AssetID,\n\t\t)\n\t} else if timelineFilter.Type == datatype.USERTIMELINE {\n\t\tclause = fmt.Sprintf(\n\t\t\t\" WHERE b.status<>%d AND (doer.id = %d OR oracle.id = %d)\",\n\t\t\tdatatype.BlockRejected,\n\t\t\ttimelineFilter.UserID,\n\t\t\ttimelineFilter.UserID,\n\t\t)\n\t}\n\trows, err := db.Query(fmt.Sprintf(`\n\t\t\tSELECT\n\t\t\t\tb.id,\n\t\t\t\tb.userId,\n\t\t\t\tdoer.username,\n\t\t\t\tdoer.profileImageUrl,\n\t\t\t\tasset.id,\n\t\t\t\tasset.name,\n\t\t\t\tasset.symbol,\n\t\t\t\toracle.id,\n\t\t\t\toracle.username,\n\t\t\t\tb.text,\n\t\t\t\tb.status,\n\t\t\t\tb.ethereumTransactionAddress,\n\t\t\t\tb.videoID,\n\t\t\t\tb.favoritesCounter,\n\t\t\t\tb.createdAt,\n\t\t\t\tIF(favorites.blockId, TRUE, FALSE),\n\t\t\t\tIF(asset_favorites.assetId, TRUE, FALSE) as following\n\t\t\tFROM asset_block b\n\t\t\tLEFT JOIN asset asset ON b.assetId=asset.Id\n\t\t\tLEFT JOIN user doer ON doer.id=b.userId\n\t\t\tLEFT JOIN user oracle ON oracle.id=asset.creatorId\n\t\t\tLEFT JOIN asset_block_favorites favorites ON b.id=favorites.blockId AND favorites.userId=?\n\t\t\tLEFT JOIN asset_favorites ON asset.id=asset_favorites.assetId AND asset_favorites.userId=?\n\t\t\t%s\n\t\t\tORDER BY b.createdAt DESC\n\t\t\tLIMIT ? OFFSET ?\n\t\t\t`, clause), user.ID, user.ID, limit+1, offset)\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\tvar c datatype.TimelineEntry\n\t\terr := rows.Scan(\n\t\t\t&c.BlockID,\n\t\t\t&c.UserID,\n\t\t\t&c.UserName,\n\t\t\t&c.UserProfileImageURL,\n\t\t\t&c.AssetID,\n\t\t\t&c.AssetName,\n\t\t\t&c.AssetSymbol,\n\t\t\t&c.OracleID,\n\t\t\t&c.OracleName,\n\t\t\t&c.Text,\n\t\t\t&c.Status,\n\t\t\t&c.EthereumTransactionAddress,\n\t\t\t&c.YtVideoID,\n\t\t\t&c.FavoritesCount,\n\t\t\t&c.CreatedAt,\n\t\t\t&c.DidUserLike,\n\t\t\t&c.DidUserLikeTopic,\n\t\t)\n\t\tif timelineFilter.Type == datatype.HOMETIMELINE && c.DidUserLikeTopic == false {\n\t\t\tcontinue\n\t\t}\n\t\tc.CreatedAtHuman = helpers.DateToHuman(c.CreatedAt)\n\t\tif err != nil {\n\t\t\tapperrors.Critical(\"timelineservice:find-all:1\", err)\n\t\t\treturn nil, false, err\n\t\t}\n\t\t// TODO optimize fetching images, bring all images for all at once,\n\t\t// not query for each entry\n\t\tc.Images, err = sc.AssetService.GetAssetBlockImages(c.BlockID)\n\t\tif err != nil {\n\t\t\tapperrors.Critical(\"timelineservice:find-all:2\", err)\n\t\t\treturn nil, false, err\n\t\t}\n\t\tc.Reactions, err = db.FindClaimReactions(c.BlockID)\n\t\tif err != nil {\n\t\t\tapperrors.Critical(\"timelineservice:find-all:3\", err)\n\t\t\treturn nil, false, err\n\t\t}\n\t\tresult = append(result, c)\n\t}\n\tif err := rows.Err(); err != nil {\n\t\treturn nil, false, err\n\t}\n\thasMore := len(result) == limit+1\n\tif hasMore {\n\t\tresult = result[:len(result)-1]\n\t}\n\treturn result, hasMore, nil\n}", "func GetAllTransactions(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []Transactions, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Transactions))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&ml, fields...); err == nil {\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAllDeck(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Deck))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Deck\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAds(c *gin.Context) {\n\tdb := models.InitDb()\n\tpage, err := strconv.Atoi(c.DefaultQuery(\"page\", \"1\"))\n\torderBy := c.DefaultQuery(\"order_by\", \"time_asc\")\n\tif err != nil || page < 1 {\n\t\tc.JSON(422, gin.H{\"error\": \"page is incorrect\"})\n\t\treturn\n\t}\n\tvar ads []models.AdShort\n\n\tswitch orderBy {\n\tcase \"time_asc\":\n\t\tdb.Table(\"advertisements\").Select(\"id, name, (CASE WHEN INSTR(pictures,',') = 0 THEN pictures ELSE substr(pictures, 0, INSTR(pictures,',')) END) as main_picture, price, created_at\").Order(\"created_at asc\").Limit(10).Offset(10 * (page - 1)).Find(&ads)\n\tcase \"time_desc\":\n\t\tdb.Table(\"advertisements\").Select(\"id, name, (CASE WHEN INSTR(pictures,',') = 0 THEN pictures ELSE substr(pictures, 0, INSTR(pictures,',')) END) as main_picture, price, created_at\").Order(\"created_at desc\").Limit(10).Offset(10 * (page - 1)).Find(&ads)\n\tcase \"price_asc\":\n\t\tdb.Table(\"advertisements\").Select(\"id, name, (CASE WHEN INSTR(pictures,',') = 0 THEN pictures ELSE substr(pictures, 0, INSTR(pictures,',')) END) as main_picture, price, created_at\").Order(\"price asc\").Limit(10).Offset(10 * (page - 1)).Find(&ads)\n\tcase \"price_desc\":\n\t\tdb.Table(\"advertisements\").Select(\"id, name, (CASE WHEN INSTR(pictures,',') = 0 THEN pictures ELSE substr(pictures, 0, INSTR(pictures,',')) END) as main_picture, price, created_at\").Order(\"price desc\").Limit(10).Offset(10 * (page - 1)).Find(&ads)\n\tdefault:\n\t\tc.JSON(422, gin.H{\"error\": \"order_by is incorrect\"})\n\t\treturn\n\t}\n\tc.JSON(200, ads)\n\n}", "func FindAllSorted(s Session, dbname string, collection string, query map[string]interface{}, sortParameters string, pageNum int, pageSize int) ([]interface{}, error) {\n\tvar object []interface{}\n\tif err := s.DB(dbname).C(collection).Find(query).Sort(sortParameters).Skip((pageNum - 1) * pageSize).Limit(pageSize).All(&object); err != nil {\n\t\treturn object, err\n\t}\n\treturn object, nil\n}", "func GetAllArticleDetail(query map[string]interface{}, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(ArticleDetail))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []ArticleDetail\n\tqs = qs.OrderBy(sortFields...).RelatedSel()\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (r Repository) All() []Quote {\n\tstmt, err := r.db.Prepare(`SELECT id_quote, content, score, uuid\n FROM quotes\n ORDER BY id_quote`)\n\tif err != nil {\n\t\tlog.Fatal(\"Malformed SQL :\" + err.Error())\n\t}\n\n\t// closes db connection\n\tdefer func() {\n\t\terr := stmt.Close()\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}()\n\treturn buildSliceFromData(stmt)\n}", "func (qs GroupQuerySet) OrderAscByUpdatedAt() GroupQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at ASC\"))\n}", "func (qs ConstraintQuerySet) OrderDescByUpdatedAt() ConstraintQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at DESC\"))\n}", "func GetAllDPSDetails(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(DPSDetail))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []DPSDetail\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAllTenant(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Tenant))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Tenant\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (qs GroupQuerySet) OrderAscByCreatedAt() GroupQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at ASC\"))\n}", "func (d *DB) All() []Article {\n\tst, err := d.db.Prepare(\"select id,feed,title,content,published,link,read,display_name from articles where deleted = false order by id\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil\n\t}\n\tdefer st.Close()\n\n\trows, err := st.Query()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil\n\t}\n\tdefer rows.Close()\n\n\tvar (\n\t\tid int\n\t\ttitle string\n\t\tcontent string\n\t\tfeed string\n\t\tlink string\n\t\tread bool\n\t\tdisplay string\n\t\tpublished time.Time\n\t)\n\n\tarticles := []Article{}\n\n\tfor rows.Next() {\n\t\terr = rows.Scan(&id, &feed, &title, &content, &published, &link, &read, &display)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\n\t\t// Check if we should higlight it\n\t\tfields := strings.Fields(title)\n\t\thighlight := false\n\t\tfor _, f := range fields {\n\t\t\tfor _, h := range d.c.conf.Highlights {\n\t\t\t\tif strings.Contains(strings.ToLower(f), strings.ToLower(h)) {\n\t\t\t\t\thighlight = true\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif highlight {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tarticles = append(articles, Article{id: id, highlight: highlight, feed: feed, title: title, content: content, published: published, link: link, read: read, feedDisplay: display})\n\t}\n\treturn articles\n}", "func GetAll(db *sql.DB) ([]models.Tag, error) {\n\tvar tags []models.Tag\n\n\trows, err := db.Query(\"select \" + tagsAllFields + \" from tags order by tag_id desc\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor rows.Next() {\n\t\tvar tag models.Tag\n\t\tif err = rows.Scan(&tag.ID, &tag.Name); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttags = append(tags, tag)\n\t}\n\n\treturn tags, nil\n}", "func GetAllArtistSong(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(ArtistSong))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []ArtistSong\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAllSoal(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Soal))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Soal\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (db *MongoDatabase) FindAllSorted(collection_name string, query interface{}, sort_fields []SortField, result interface{}) error {\n\tcurrent_session := db.GetSession()\n\tdefer current_session.Close()\n\n\tsort_fields_mgo := make([]string, len(sort_fields))\n\tfor i, field := range sort_fields {\n\t\tif field.Reversed {\n\t\t\tsort_fields_mgo[i] = fmt.Sprintf(\"-%s\", field.Name)\n\t\t} else {\n\t\t\tsort_fields_mgo[i] = field.Name\n\t\t}\n\t}\n\n\tcollection := current_session.DB(db.name).C(collection_name)\n\n\terr := collection.Find(query).Sort(sort_fields_mgo...).All(result)\n\n\treturn convertMgoError(err)\n}", "func getAll() (comics []Comic) {\n\n\tctx, client := db.Connect()\n\n\tcollection := client.Database(\"todo\").Collection(col)\n\tcursor, _ := collection.Find(ctx, bson.M{})\n\n\tdefer func() {\n\t\tcursor.Close(ctx)\n\t\tclient.Disconnect(ctx)\n\t}()\n\n\tfor cursor.Next(ctx) {\n\t\tvar comic Comic\n\t\terr := cursor.Decode(&comic)\n\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\tcomics = append(comics, comic)\n\t}\n\treturn\n}", "func (a *TodoAdapter) GetAll(ctx context.Context) ([]todo.Todo, error) {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, \"TodoAdapter-GetAll\")\n\tdefer span.Finish()\n\n\tfindOptions := options.Find()\n\n\tcur, err := a.collection.Find(ctx, bson.D{}, findOptions)\n\n\tvar todos = make([]todo.Todo, 0)\n\n\tif err != nil {\n\t\treturn todos, err\n\t}\n\n\tvar errs = make([]string, 0)\n\n\tfor cur.Next(ctx) {\n\t\tvar entity Todo\n\t\terr := cur.Decode(&entity)\n\t\tif err != nil {\n\t\t\terrs = append(errs, fmt.Sprintf(\"decode error:%s\", err.Error()))\n\t\t}\n\t\ttodos = append(todos, entity.ToModel())\n\t}\n\n\tif err := cur.Err(); err != nil {\n\t\terrs = append(errs, fmt.Sprintf(\"cursor error:%s\", err.Error()))\n\t}\n\n\t_ = cur.Close(ctx)\n\n\tif len(errs) > 0 {\n\t\treturn todos, errors.New(strings.Join(errs, \";\"))\n\t}\n\n\treturn todos, nil\n}", "func GetAllEmailTemplate(query map[string]string, fields []string, sortby []string, order []string,\noffset int64, limit int64) (ml []interface{}, meta *map[string]int64, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(EmailItem))\n\t// query k=v\n\tquery[\"type\"] = \"template\"\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []EmailItem\n\tqs = qs.OrderBy(sortFields...)\n\tobjects_count, err := qs.Count()\n\tif err != nil {\n\t\treturn\n\t}\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\tmeta := &map[string]int64{\n\t\t\t\"objects_count\": objects_count,\n\t\t\t\"limit\": limit,\n\t\t\t\"offset\": offset,\n\t\t}\n\t\treturn ml, meta, nil\n\t}\n\treturn nil, nil, err\n}", "func GetAllType(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Types))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Types\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (s *Repository) GetAll(ctx context.Context) ([]Account, error) {\n\tconst limit = 10\n\n\trows, err := s.pool.Query(\n\t\tctx,\n\t\t`select * from \"account\"\n\t\t\t order by \"createdAt\" desc\n\t\t\t limit $1`, limit)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer rows.Close()\n\n\treturn scanAccounts(limit, rows)\n}", "func GetAllByCategory(args ...string) error {\n\tif len(args) < 2 {\n\t\treturn errors.New(\"Category Get needs at least 2 arguments\")\n\t}\n\tusername, categoryName, sortBy, sortDir := strings.ToLower(args[0]), args[1], \"\", \"\"\n\n\t_, ok := category.GetCategoryIndex()[categoryName]\n\tif !ok {\n\t\treturn errors.New(\"Error - category not found \")\n\t}\n\tif !user.DoesExist(username) {\n\t\treturn errors.New(\"Error - unknown user\")\n\t}\n\n\tlistings := make([]*Listing, 0)\n\tfor _, l := range listingsByID {\n\t\tif l.categoryName == categoryName {\n\t\t\tlistings = append(listings, l)\n\t\t}\n\t}\n\tif len(args) >= 4 {\n\t\tsortBy, sortDir = strings.ToLower(args[2]), strings.ToLower(args[3])\n\t\tif sortDir != \"asc\" && sortDir != \"dsc\" {\n\t\t\tfmt.Println(\"unknown sort direction, ignoring sort\")\n\t\t} else if sortBy == \"sort_price\" {\n\t\t\tif sortDir == \"asc\" {\n\t\t\t\tsort.Slice(listings[:], func(i, j int) bool {\n\t\t\t\t\treturn listings[i].priceInCents < listings[j].priceInCents\n\t\t\t\t})\n\t\t\t} else {\n\t\t\t\tsort.Slice(listings[:], func(i, j int) bool {\n\t\t\t\t\treturn listings[i].priceInCents > listings[j].priceInCents\n\t\t\t\t})\n\t\t\t}\n\n\t\t} else if sortBy == \"sort_time\" {\n\t\t\tif sortDir == \"asc\" {\n\t\t\t\tsort.Slice(listings[:], func(i, j int) bool {\n\t\t\t\t\treturn listings[i].createdAt.UnixNano() < listings[j].createdAt.UnixNano()\n\t\t\t\t})\n\t\t\t} else {\n\t\t\t\tsort.Slice(listings[:], func(i, j int) bool {\n\t\t\t\t\treturn listings[i].createdAt.UnixNano() > listings[j].createdAt.UnixNano()\n\t\t\t\t})\n\t\t\t}\n\n\t\t}\n\t}\n\tfor _, l := range listings {\n\t\tl.print()\n\t}\n\treturn nil\n}", "func (e *ExpenseModel) ReadAll(filter interface{}) ([]Expense, error) {\n\tvar expenses []Expense\n\tcollection := e.db.Client.Database(e.db.DBName).Collection(\"expenses\")\n\tlog.Printf(\"filter: %v\\n\", filter)\n\t// sort the entries based on the `date` field\n\topts := options.FindOptions{}\n\topts.SetSort(bson.D{{\"date\", -1}})\n\tcur, err := collection.Find(context.TODO(), filter, &opts)\n\tif err != nil {\n\t\tlog.Printf(\"ERROR FINDING DATA: %v\\n\", err)\n\t\treturn expenses, err\n\t}\n\tfor cur.Next(context.TODO()) {\n\t\tvar expense Expense\n\t\terr = cur.Decode(&expense)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Error on Decoding the document: %v\\n\", err)\n\t\t}\n\t\texpenses = append(expenses, expense)\n\t}\n\tlog.Printf(\"documentReturned: %v\\n\", expenses)\n\treturn expenses, nil\n}", "func (qs ConstraintQuerySet) OrderAscByUpdatedAt() ConstraintQuerySet {\n\treturn qs.w(qs.db.Order(\"updated_at ASC\"))\n}", "func GetAllElementosMovimiento(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(ElementosMovimiento)).RelatedSel()\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []ElementosMovimiento\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (a *App) retrieveAll(c *echo.Context) error {\n\tvar tasks []*model.Task\n\ta.GetDB().Find(&tasks, struct{}{})\n\tc.JSON(http.StatusOK, tasks)\n\treturn nil\n}", "func (or *mongoOrderRepository) FetchByRange(skip int, limit int) ([]models.Order, error) {\n\tvar orders []models.Order\n\t//Find documents\n\terr := or.Conn.C(COLLECTION).Find(bson.M{}).Skip(skip).Limit(limit).All(&orders)\n\treturn orders, err\n}", "func (qs ConstraintQuerySet) OrderAscByCreatedAt() ConstraintQuerySet {\n\treturn qs.w(qs.db.Order(\"created_at ASC\"))\n}", "func (c Languages) Top() revel.Result {\n type Language struct {\n Name string\n Color string\n }\n var results []Language\n limit := 20\n if err := cache.Get(\"languages\", &results); err == nil {\n return c.RenderJson(results)\n }\n\n results = make([]Language, limit)\n languages, _ := c.Txn.Select(models.RepoStat{},\n \"select l.language from (select language from files where language != '' \" + \n \"group by language order by count(*) desc limit $1) l order by l.language\", limit)\n colors := [...]string {\"#FF0000\", \"#617C58\", \"#52D017\", \n \"#C0C0C0\", \"#0000FF\", \"#808080\", \"#0000A0\", \"#ADD8E6\",\n \"#FFA500\", \"#800080\", \"#A52A2A\", \"#FFFF00\", \"#800000\", \n \"#00FF00\", \"#008000\", \"#FF00FF\", \"#FF0000\", \"#57FEFF\", \n \"FFA62F\", \"#8E35EF\"}\n for i := 0 ; i < len(languages) ; i++ {\n results[i] = Language { \n Name: languages[i].(*models.RepoStat).Language,\n Color: colors[i],\n }\n }\n go cache.Set(\"languages\", results, 1 * time.Hour)\n return c.RenderJson(results)\n}", "func (repository *RepositoryPostCRUD) FindAll() ([]models.Post, error) {\n\tvar posts []models.Post\n\tdone := make(chan bool) //crea un canal que comunica valores boleanos\n\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\tdefer cancel()\n\t//go function\n\t//Se encarga de recuperar todos los posts de la base de datos, y transmite el resultado por medio de un canal boleano\n\tgo func(ch chan<- bool) {\n\t\tdefer cancel()\n\t\tcursor, err := repository.db.Collection(\"Posts\").Find(ctx, bson.M{})\n\t\tdefer cursor.Close(ctx)\n\t\tif err != nil {\n\t\t\tch <- false\n\t\t\treturn\n\t\t}\n\t\tfor cursor.Next(ctx) { //for each element in the database\n\t\t\tvar post models.Post\n\t\t\tcursor.Decode(&post)\n\t\t\tposts = append(posts, post) //lo acgrega al slice de posts\n\t\t}\n\t\tif len(posts) > 0 {\n\t\t\tch <- true\n\t\t}\n\t}(done)\n\n\tif channels.OK(done) {\n\t\treturn posts, nil\n\t}\n\treturn []models.Post{}, nil\n}", "func GetAllTFeedbackAction(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(TFeedbackAction))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []TFeedbackAction\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (o *Avi) FetchAll() (r []Data, err error) {\r\n\td := new(Data)\r\n\terr = o.fetch(d)\r\n\tr = append(r, *d)\r\n\treturn\r\n}", "func (ar *AutomobilesRepository) SelectAll() ([]*models.Automobiles, error) {\n\tquery := fmt.Sprintf(\"SELECT * FROM %s\", tableAutomobiles)\n\trows, err := ar.store.db.Query(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\tautomobiles := make([]*models.Automobiles, 0)\n\tfor rows.Next() {\n\t\ta := models.Automobiles{}\n\t\terr := rows.Scan(&a.ID, &a.Mark, &a.Maxspeed, &a.Distance, &a.Handler, &a.Stock)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t\tcontinue\n\t\t}\n\t\tautomobiles = append(automobiles, &a)\n\t}\n\treturn automobiles, nil\n}", "func GetAllRubroHomologado(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(RubroHomologado))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []RubroHomologado\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func Order(q *query.Query) *query.Query {\n\treturn q.Order(\"status desc\")\n}", "func GetAllGuests(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Guests))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Guests\n\tqs = qs.OrderBy(sortFields...).RelatedSel()\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (s *Service) findPaginatedAlarms(offset, limit int, orderBy string, user *accounts.User) ([]*Alarm, error) {\n\tvar alarms []*Alarm\n\n\t// Get the pagination query\n\talarmsQuery := s.alarmsQuery(user)\n\n\t// Default ordering\n\tif orderBy == \"\" {\n\t\torderBy = \"id\"\n\t}\n\n\t// Retrieve paginated results from the database\n\terr := alarmsQuery.Offset(offset).Limit(limit).Order(orderBy).\n\t\tPreload(\"User\").Preload(\"Incidents\", \"resolved_at IS NULL\").\n\t\tFind(&alarms).Error\n\tif err != nil {\n\t\treturn alarms, err\n\t}\n\n\treturn alarms, nil\n}", "func (dao *ArticleDAO) Query(rs app.RequestScope, offset, limit, categoryId int, sorting, filter string) ([]models.Article, error) {\n\tarticles := []models.Article{}\n\tq := rs.Tx().Select().OrderBy(\"id\")\n\tif categoryId != 0 {\n\t\tq.Where(dbx.HashExp{\"category_id\": categoryId})\n\t}\n\tif filter != \"\" {\n\t\tq.AndWhere(dbx.Like(\"title\", filter))\n\t}\n\tif sorting == \"asc\" {\n\t\tq.OrderBy(\"id ASC\")\n\t} else {\n\t\tq.OrderBy(\"id DESC\")\n\t}\n\terr := q.Offset(int64(offset)).Limit(int64(limit)).All(&articles)\n\treturn articles, err\n}", "func (db *MongoDBAccess) GetAll() ([]Record, error) {\n\tvar records []Record\n\tcursor, err := db.client.Database(db.database).Collection(\"days\").Find(context.Background(), bson.D{})\n\tif err != nil {\n\t\treturn []Record{}, err\n\t}\n\tdefer cursor.Close(context.Background())\n\tfor cursor.Next(context.Background()) {\n\t\tvar record Record\n\t\tif err = cursor.Decode(&record); err != nil {\n\t\t\treturn []Record{}, err\n\t\t}\n\t\trecords = append(records, record)\n\t}\n\treturn records, nil\n}", "func (hr Repository) All() ([]HistoryModel, error) {\n\trows, err := hr.db.Query(`select * from history`)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer rows.Close()\n\n\tvar historyList []HistoryModel\n\n\tfor rows.Next() {\n\t\tvar history HistoryModel\n\t\terr := rows.Scan(&history.ID, &history.Startdate, &history.Enddate, &history.Area)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\thistoryList = append(historyList, HistoryModel{\n\t\t\tID: history.ID,\n\t\t\tStartdate: history.Startdate,\n\t\t\tEnddate: history.Enddate,\n\t\t\tArea: history.Area,\n\t\t})\n\t}\n\treturn historyList, nil\n}", "func (q *Query) OrderBy(orderBys ...string) *Query {\n\tq.orderBys = append(q.orderBys, orderBys...)\n\treturn q\n}", "func (ar AlbumDbRepository) GetAll(hydrate bool) (entities domain.Albums, err error) {\n\tif !hydrate {\n\t\tquery := \"SELECT id, title, year, artist_id, cover_id, created_at FROM albums\"\n\t\t_, err = ar.AppContext.DB.Select(&entities, query)\n\n\t} else {\n\t\ttype gorpResult struct {\n\t\t\tAlbumId int\n\t\t\tAlbumTitle string\n\t\t\tAlbumYear string\n\t\t\tAlbumArtistId int\n\t\t\tAlbumCreatedAt int64\n\t\t\tdomain.Track\n\t\t\t// Cannot select domain.album.ArtistId or domain.track.AlbumId because of a Gorp error...\n\t\t\t// So we have to join on trk.album_id, but then Gorp cannot do the mapping with gorpResult, so we have\n\t\t\t// to add this property in the struct. TODO get rid of gorp.\n\t\t\tAlbum_id int\n\t\t}\n\t\tvar results []gorpResult\n\n\t\tquery := \"SELECT alb.Id AlbumId, alb.Title AlbumTitle, alb.Year AlbumYear, alb.artist_id AlbumArtistId, alb.created_at AlbumCreatedAt, trk.* \" +\n\t\t\t \"FROM albums alb, tracks trk WHERE alb.id = trk.album_id\"\n\n\t\t_, err = ar.AppContext.DB.Select(&results, query)\n\t\tif err == nil {\n\t\t\t// Deduplicate stuff.\n\t\t\tvar current domain.Album\n\t\t\tfor _, r := range results {\n\t\t\t\ttrack := domain.Track{\n\t\t\t\t\tId: r.Id,\n\t\t\t\t\tTitle: r.Title,\n\t\t\t\t\tAlbumId: r.AlbumId,\n\t\t\t\t\tArtistId: r.ArtistId,\n\t\t\t\t\tCoverId: r.CoverId,\n\t\t\t\t\tDisc: r.Disc,\n\t\t\t\t\tNumber: r.Number,\n\t\t\t\t\tDuration: r.Duration,\n\t\t\t\t\tGenre: r.Genre,\n\t\t\t\t\tPath: r.Path,\n\t\t\t\t\tDateAdded: r.DateAdded,\n\t\t\t\t}\n\n\t\t\t\tif current.Id == 0 {\n\t\t\t\t\tcurrent = domain.Album{\n\t\t\t\t\t\tId: r.AlbumId,\n\t\t\t\t\t\tTitle: r.AlbumTitle,\n\t\t\t\t\t\tYear: r.AlbumYear,\n\t\t\t\t\t\tArtistId: r.AlbumArtistId,\n\t\t\t\t\t\tDateAdded: r.AlbumCreatedAt,\n\t\t\t\t\t}\n\t\t\t\t} else if r.Id != current.Id {\n\t\t\t\t\tentities = append(entities, current)\n\t\t\t\t\t// Then change the current album\n\t\t\t\t\tcurrent = domain.Album{\n\t\t\t\t\t\tId: r.AlbumId,\n\t\t\t\t\t\tTitle: r.AlbumTitle,\n\t\t\t\t\t\tYear: r.AlbumYear,\n\t\t\t\t\t\tArtistId: r.AlbumArtistId,\n\t\t\t\t\t\tDateAdded: r.AlbumCreatedAt,\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tcurrent.Tracks = append(current.Tracks, track)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn\n}", "func FindAll() ([]entity.Hechizo, error) {\n\n\tlog.Println(\"---- Consulta FindAll ----\")\n\tdb := dbUtils.Connect()\n\tdefer dbUtils.Close(db)\n\n\tvar id int64\n\tvar mana, counter int\n\tvar nombre string\n\n\tquery := fmt.Sprintf(\"SELECT * FROM %v\", tabla)\n\tresultQuery, error := db.Query(query)\n\n\thechizos := make([]entity.Hechizo, 0)\n\tif error != nil {\n\t\treturn hechizos, error\n\t}\n\n\tfor resultQuery.Next() {\n\n\t\terrorSelect := resultQuery.Scan(&id, &nombre, &mana)\n\t\tif errorSelect != nil {\n\t\t\tlog.Println(errorSelect)\n\t\t\tcontinue\n\t\t}\n\t\tcounter++\n\t\thechizosDb := entity.Hechizo{Id: id, Nombre: nombre, Mana: mana}\n\t\thechizos = append(hechizos, hechizosDb)\n\t}\n\treturn hechizos, nil\n}", "func GetAllDevice(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, meta *map[string]int64, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Device))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Device\n\tqs = qs.RelatedSel(\"Device\", \"Node\").OrderBy(sortFields...)\n\tobjects_count, err := qs.Count()\n\tif err != nil {\n\t\treturn\n\t}\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\tmeta = &map[string]int64{\n\t\t\t\"objects_count\": objects_count,\n\t\t\t\"limit\": limit,\n\t\t\t\"offset\": offset,\n\t\t}\n\t\treturn ml, meta, nil\n\t}\n\treturn nil, nil, err\n}", "func (m *MySQL) GetAll() ([]schema.Todo, error) {\n\tquery := `\n SELECT *\n FROM Todo\n ORDER BY ID;\n `\n\n\trows, err := m.DB.Query(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar todoList []schema.Todo\n\tfor rows.Next() {\n\t\tvar t schema.Todo\n\t\tif err := rows.Scan(&t.ID, &t.Title, &t.Complete); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttodoList = append(todoList, t)\n\t}\n\n\treturn todoList, nil\n}", "func GetAllUsers(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Users))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Users\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAll(coll string) bson.D {\n\tfmt.Println(\"controllers: get all invoked\")\n\tctx, _ := context.WithTimeout(context.Background(), 50*time.Second)\n\n\tcurs, err := db.Collection(coll).Find(ctx, bson.D{})\n\tutilities.Catch(err)\n\n\tdefer curs.Close(ctx)\n\n\telements := bson.D{}\n\n\tfor curs.Next(ctx) {\n\t\terr := curs.Decode(&elements)\n\t\tutilities.Catch(err)\n\t}\n\n\treturn elements\n}", "func (b *QueryBuilder) OrderBy(nodes ...NodeI) {\n\tb.OrderBys = append(b.OrderBys, nodes...)\n}", "func GetAllMember(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Member))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\"+v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\"+v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Member\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func ByCreatedAt(opts ...sql.OrderTermOption) OrderOption {\n\treturn sql.OrderByField(FieldCreatedAt, opts...).ToFunc()\n}", "func GetAllSecUser(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(SecUser))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []SecUser\n\tqs = qs.OrderBy(sortFields...).RelatedSel()\n\tif _, err = qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (u *UserRepository) GetAll() ([]*models.User, error) {\n\n\t// Here's an array in which you can store the decoded documents\n\tvar result []*models.User\n\n\tcur, err := u.db.User.Find(context.TODO(), bson.M{})\n\t// Close the cursor once finished\n\tdefer cur.Close(context.TODO())\n\tif err != nil {\n\t\tu.log.Errorw(\"failed to get user list\",\n\t\t\t\"error\", err,\n\t\t)\n\t\treturn nil, err\n\t}\n\n\t// Finding multiple documents returns a cursor\n\t// Iterating through the cursor allows us to decode documents one at a time\n\tfor cur.Next(context.TODO()) {\n\t\t// create a value into which the single document can be decoded\n\t\tvar user models.User\n\t\terr := cur.Decode(&user)\n\t\tif err != nil {\n\t\t\tu.log.Errorw(\"failed to decode user model\",\n\t\t\t\t\"error\", err,\n\t\t\t\t\"user\", cur,\n\t\t\t)\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, &user)\n\t}\n\t//descending by Last check time\n\tsort.Slice(result, func(i, j int) bool {\n\t\treturn result[i].LastCheck.After(result[j].LastCheck)\n\t})\n\treturn result, nil\n}", "func DBFetchFeaturedAnimes() ([]projectModels.StructureAnime, error) {\n\t/*\n\t\t\tr.db(\"animedom\").table(\"animes\").filter(function(doc){\n\t\t return doc(\"id\").eq(\"199\").\n\t\t or(doc(\"id\").eq(\"31240\")).\n\t\t or(doc(\"id\").eq(\"249\")).\n\t\t or(doc(\"id\").eq(\"4722\")).\n\t\t or(doc(\"id\").eq(\"205\")).\n\t\t or(doc(\"id\").eq(\"431\")).\n\t\t or(doc(\"id\").eq(\"534\")).\n\t\t or(doc(\"id\").eq(\"20\"))\n\t\t})\n\t*/\n\tresp, err := r.Table(\"animes\").Filter(func(anime r.Term) r.Term {\n\t\treturn anime.Field(\"id\").Eq(\"199\").\n\t\t\tOr(anime.Field(\"id\").Eq(\"31240\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"249\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"4722\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"205\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"431\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"534\")).\n\t\t\tOr(anime.Field(\"id\").Eq(\"20\"))\n\t}).Run(dbSession)\n\tif err != nil {\n\t\treturn []projectModels.StructureAnime{}, err\n\t}\n\n\tif resp.IsNil() {\n\t\treturn []projectModels.StructureAnime{}, errors.New(\"Empty Result\")\n\t}\n\n\tvar featuredAnimes []projectModels.StructureAnime\n\tif err = resp.All(&featuredAnimes); err != nil {\n\t\treturn []projectModels.StructureAnime{}, err\n\t}\n\n\terr = resp.Close()\n\tcommon.CheckErrorAndPanic(err)\n\n\tif len(featuredAnimes) == 0 {\n\t\treturn []projectModels.StructureAnime{}, errors.New(\"Empty Result\")\n\t}\n\treturn featuredAnimes, nil\n}", "func findAll(findAllStruct *FindAll) ([]interface{}, error) {\n\tvar records []interface{}\n\terr := findAllStruct.Collection.Find(nil).All(&records)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn records, err\n}", "func OrderBy(field string, asc bool) Query {\n\treturn func(db *gorm.DB) *gorm.DB {\n\t\tif asc {\n\t\t\treturn db.Order(field + \" ASC\")\n\t\t}\n\n\t\treturn db.Order(field + \" DESC\")\n\t}\n}", "func (c *Command) GetAll(ctx *gin.Context) {\n\ttoken := strings.ToLower(html.EscapeString(ctx.Param(\"token\")))\n\tfilter := map[string]interface{}{\"token\": token}\n\tfromDB, err := c.Conn.GetByFilter(c.Table, filter, 0)\n\tif err != nil {\n\t\tutil.NiceError(ctx, err, http.StatusBadRequest)\n\t\treturn\n\t}\n\tif fromDB == nil {\n\t\tctx.JSON(http.StatusNotFound, make([]struct{}, 0))\n\t\treturn\n\t}\n\n\tvar respDecode ResponseSchema\n\tvar decoded = make([]map[string]interface{}, len(fromDB))\n\tfor pos, record := range fromDB {\n\t\t// If there's an issue decoding it, just log it and move on to the next record\n\t\tif err := mapstruct.Decode(record, &respDecode); err != nil {\n\t\t\tlog.Error(err.Error())\n\t\t\tcontinue\n\t\t}\n\t\tmarshalled := util.MarshalResponse(respDecode)\n\t\tdecoded[pos] = map[string]interface{}{\n\t\t\t\"id\": marshalled[\"data\"].(map[string]interface{})[\"id\"],\n\t\t\t\"attributes\": marshalled[\"data\"].(map[string]interface{})[\"attributes\"],\n\t\t\t\"meta\": marshalled[\"meta\"],\n\t\t}\n\t}\n\tvar response = make(map[string]interface{})\n\n\tresponse[\"data\"] = decoded\n\n\tctx.Header(\"x-total-count\", fmt.Sprint(len(decoded)))\n\tctx.JSON(http.StatusOK, response)\n}", "func (r *Rate) FindAll() ([]models.Rate, error) {\n\n\tdata, err := r.RateRepository.Find(&models.Rate{}) //just leave it empty object\n\tif err != nil {\n\t\treturn data, err\n\t}\n\treturn data, nil\n}", "func (r *OrderRepositoryImpl) GetAll() ([]domain.Order, error) {\n\tOrder := []domain.Order{}\n\tif err := r.Conn.Preload(\"Topic\").Find(&Order).Error; err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn Order, nil\n}", "func (qs InstantprofileQS) OrderByActiveDesc() InstantprofileQS {\n\tqs.order = append(qs.order, `\"active\" DESC`)\n\n\treturn qs\n}", "func GetAllSketchs(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Sketchs))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tif strings.Contains(k, \"isnull\") {\n\t\t\tqs = qs.Filter(k, (v == \"true\" || v == \"1\"))\n\t\t} else {\n\t\t\tqs = qs.Filter(k, v)\n\t\t}\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Sketchs\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err = qs.Limit(limit, offset).Filter(\"deleted_at__isnull\", true).RelatedSel().All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tv.loadRelations()\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tv.loadRelations()\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func GetAllUsersOld(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(UsersOld))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []UsersOld\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func getAllTask() []primitive.M {\n\tcur, err := collection.Find(context.Background(), bson.D{{}})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tvar results []primitive.M\n\tfor cur.Next(context.Background()) {\n\t\tvar result bson.M\n\t\te := cur.Decode(&result)\n\t\tif e != nil {\n\t\t\tlog.Fatal(e)\n\t\t}\n\t\t// fmt.Println(\"cur..>\", cur, \"result\", reflect.TypeOf(result), reflect.TypeOf(result[\"_id\"]))\n\t\tresults = append(results, result)\n\n\t}\n\n\tif err := cur.Err(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tcur.Close(context.Background())\n\treturn results\n}", "func getAllTask() []primitive.M {\n\tcur, err := collection.Find(context.Background(), bson.D{{}})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tvar results []primitive.M\n\tfor cur.Next(context.Background()) {\n\t\tvar result bson.M\n\t\te := cur.Decode(&result)\n\t\tif e != nil {\n\t\t\tlog.Fatal(e)\n\t\t}\n\t\t// fmt.Println(\"cur..>\", cur, \"result\", reflect.TypeOf(result), reflect.TypeOf(result[\"_id\"]))\n\t\tresults = append(results, result)\n\n\t}\n\n\tif err := cur.Err(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tcur.Close(context.Background())\n\treturn results\n}", "func FetchArchive(limit int) ([]*ArchiveEntry, error) {\n\tdb := hal.SqlDB()\n\n\t// joining reactions in here for now - might be better to let the client do it\n\t// but for now get something working\n\t// This pulls back multiple rows if there are multiple reactions. The row iteration\n\t// below uses a map to dedupe the archive rows and put reactions into a list.\n\t// This might be better written with GROUP_CONCAT later...\n\tsql := `SELECT a.id AS id,\n\t UNIX_TIMESTAMP(a.ts) AS ts,\n\t\t\t\t a.user AS user,\n\t\t\t\t a.room AS room,\n\t\t\t\t a.broker AS broker,\n\t\t\t\t a.body AS body,\n\t\t\t\t IFNULL(r.reaction,\"\") AS reaction\n\t FROM archive a\n\t\t\t LEFT OUTER JOIN reactions r ON ( r.id = a.id AND r.room = a.room )\n\t\t\t WHERE a.ts < ? AND a.ts > ?\n\t\t\t GROUP BY a.id\n\t\t\t ORDER BY a.ts DESC`\n\n\tnow := time.Now()\n\tyesterday := now.Add(-time.Hour * 24)\n\trows, err := db.Query(sql, &now, &yesterday)\n\tif err != nil {\n\t\tlog.Printf(\"archive query failed: %s\\n\", err)\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\tentries := make(map[string]*ArchiveEntry)\n\n\tfor rows.Next() {\n\t\tvar ts int64\n\t\tvar id, room, user, broker, body, reaction string\n\t\terr = rows.Scan(&id, &ts, &user, &room, &broker, &body, &reaction)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Row iteration failed: %s\\n\", err)\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif entry, exists := entries[id]; exists {\n\t\t\tif reaction != \"\" {\n\t\t\t\tentry.Reactions = append(entry.Reactions, reaction)\n\t\t\t}\n\t\t} else {\n\t\t\tae := ArchiveEntry{\n\t\t\t\tID: id,\n\t\t\t\tTimestamp: time.Unix(ts, 0),\n\t\t\t\tBroker: broker,\n\t\t\t\tBody: body,\n\t\t\t\tReactions: []string{},\n\t\t\t}\n\n\t\t\tif reaction != \"\" {\n\t\t\t\tae.Reactions = append(ae.Reactions, reaction)\n\t\t\t}\n\n\t\t\t// convert ids to names\n\t\t\tbroker := hal.Router().GetBroker(ae.Broker)\n\t\t\tae.Room = broker.RoomIdToName(room)\n\t\t\tae.User = broker.UserIdToName(user)\n\n\t\t\tentries[id] = &ae\n\t\t}\n\t}\n\n\t// hmm might want to sort these before sending...\n\taes := make([]*ArchiveEntry, len(entries))\n\tvar i int\n\tfor _, ae := range entries {\n\t\taes[i] = ae\n\t\ti++\n\t}\n\n\treturn aes, nil\n}", "func getArticles(p int) {\n\tdb, err := bolt.Open(\"../.db\", 0600, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer db.Close()\n\n\t//display 10 articles per page\n\tIdIndex := (p-1)*10 + 1\n\tvar articles ArticlesResponse\n\tvar article Article\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(\"Article\"))\n\t\tif b != nil {\n\t\t\tc := b.Cursor()\n\t\t\tk, v := c.Seek(itob(IdIndex))\n\t\t\tif k == nil {\n\t\t\t\tfmt.Println(\"Page is out of index\")\n\t\t\t\treturn errors.New(\"Page is out of index\")\n\t\t\t}\n\t\t\tkey := binary.BigEndian.Uint64(k)\n\t\t\tfmt.Print(key)\n\t\t\tif int(key) != IdIndex {\n\t\t\t\tfmt.Println(\"Page is out of index\")\n\t\t\t\treturn errors.New(\"Page is out of index\")\n\t\t\t}\n\t\t\tcount := 0\n\t\t\tvar ori_artc Article\n\t\t\tfor ; k != nil && count < 10; k, v = c.Next() {\n\t\t\t\terr = json.Unmarshal(v, &ori_artc)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tarticle.Id = ori_artc.Id\n\t\t\t\tarticle.Name = ori_artc.Name\n\t\t\t\tarticles.Articles = append(articles.Articles, article)\n\t\t\t\tcount = count + 1\n\t\t\t}\n\t\t\treturn nil\n\t\t} else {\n\t\t\treturn errors.New(\"Article Not Exists\")\n\t\t}\n\t})\n\tfor i := 0; i < len(articles.Articles); i++ {\n\t\tfmt.Println(articles.Articles[i])\n\t}\n}", "func (ql *QueueHistoryList) SortDESC() {\n\tsort.Sort(QueueHistoryByCreatedTimeDESC(ql.Items))\n}", "func (b *blogRepository) FindAll(limit int, page int) []*blog.Blog {\n\tbr := []*BlogRow{}\n\tquery := sq.Select(\"bid, blogName, blogURL, blogRSS\").\n\t\tFrom(\"blogs\").\n\t\tWhere(sq.Eq{\"active\": 1}).\n\t\tOrderBy(\"blogName ASC\").\n\t\tOffset(uint64((page - 1) * limit)).\n\t\tLimit(uint64(limit))\n\n\terr := b.db.Select(&br, query)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\tblogs := make([]*blog.Blog, 0, len(br))\n\tfor _, b := range br {\n\t\tblog := blog.NewBlog(b.BID, b.BlogName, b.BlogURL, b.BlogRSS)\n\t\tblogs = append(blogs, blog)\n\t}\n\n\treturn blogs\n}", "func (r *MongoRepository) FindAll() ([]*Definition, error) {\n\tvar result []*Definition\n\tsession, coll := r.getSession()\n\tdefer session.Close()\n\n\t// sort by name to have the same order all the time - for easier comparison\n\terr := coll.Find(nil).Sort(\"name\").All(&result)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn result, nil\n}", "func (service *Service) All() (models []ModelOperationsLog, err error) {\n\trows, err := service.pool.Query(context.Background(), `SELECT id, name, number,recipientSender,count, balanceold, balancenew, time, owner_id FROM historyoperationslog;`)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"can't get sys-test-history from db: %w\", err)\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tmodel := ModelOperationsLog{}\n\t\terr = rows.Scan(\n\t\t\t&model.Id,\n\t\t\t&model.Name,\n\t\t\t&model.Number,\n\t\t\t&model.RecipientSender,\n\t\t\t&model.Count,\n\t\t\t&model.BalanceOld,\n\t\t\t&model.BalanceNew,\n\t\t\t&model.Time,\n\t\t\t&model.OwnerID)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"can't get sys-test-history from db: %w\", err)\n\t\t}\n\t\tmodels = append(models, model)\n\t}\n\tif err = rows.Err(); err != nil {\n\t\treturn nil, fmt.Errorf(\"can't get sys-test-history from db: %w\", err)\n\t}\n\treturn models, nil\n}", "func (h *Handler) GetAll(c echo.Context) error {\n\tid := c.Param(\"id\")\n\tdb := h.DB.Clone()\n\tdefer db.Close()\n\n\tvar results []*particleio.Result\n\tif err := db.DB(\"oxylus\").C(\"metrics\").Find(bson.M{\"uuid\": id}).Sort(\"time\").All(&results); err != nil {\n\t\tlog.Println(err)\n\t}\n\treturn c.JSON(http.StatusOK, results)\n}", "func (b BladeStorage) GetAll(offset string, limit string) (count int, blades []model.Blade, err error) {\n\tif offset != \"\" && limit != \"\" {\n\t\tif err = b.db.Limit(limit).Offset(offset).Order(\"serial asc\").Find(&blades).Error; err != nil {\n\t\t\treturn count, blades, err\n\t\t}\n\t\tb.db.Model(&model.Blade{}).Order(\"serial asc\").Count(&count)\n\t} else {\n\t\tif err = b.db.Order(\"serial\").Find(&blades).Error; err != nil {\n\t\t\treturn count, blades, err\n\t\t}\n\t}\n\treturn count, blades, err\n}", "func (repo *reminderRepository) FindAll(ctx context.Context) ([]*Reminder, error) {\n\tkind := repo.Kind(ctx, &Reminder{})\n\tq := repo.NewQuery(kind).Filter(\"Enabled =\", true)\n\n\tvar dst []*Reminder\n\treturn dst, repo.GetAll(ctx, q, &dst)\n}", "func FetchAll() history.CollectionYearHistory {\n\treturn history.FetchCollectionYearHistory()\n\n}", "func (q ReleaseQuery) OrderBy(cmd string) ReleaseQuery {\n\tq.builder = q.builder.OrderBy(cmd)\n\treturn q\n}", "func Timeline() []dto.Article {\n\tlogfile, er := os.OpenFile(utils.LogFile, os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0666)\n\tif er != nil {\n\t\tpanic(er.Error())\n\t}\n\tdefer logfile.Close()\n\t// Initalize DB Connection\n\tsql, sqlErr := utils.DBInit()\n\tif sqlErr != nil {\n\t\tlog.SetOutput(io.MultiWriter(logfile, os.Stdout))\n\t\tlog.SetFlags(log.Ldate | log.Ltime)\n\t\tlog.Fatal(sqlErr)\n\t}\n\t// Close DB connection at the end.\n\tdefer sql.Close()\n\t// SQL syntax\n\tgetPosts := `SELECT \n\t\t\t\t\tarticle_table.user_id,\n\t\t\t\t\tarticle_table.article_id, \n\t\t\t\tCOALESCE(COUNT(liked_table.user_id), 0) AS liked_sum, \n\t\t\t\t\tuser_table.user_name, \n\t\t\t\t\tarticle_table.title, \n\t\t\t\t\tarticle_table.content, \n\t\t\t\t\tarticle_table.created_time, \n\t\t\t\t\tarticle_table.modified_time \n\t\t\t\t\tFROM(\n\t\t\t\t\t\t\tarticle_table \n\t\t\t\t\t\tINNER JOIN \n\t\t\t\t\t\t\tuser_table \n\t\t\t\t\t\tON \n\t\t\t\t\t\t\tarticle_table.user_id = user_table.user_id\n\t\t\t\t\t\t) \n\t\t\t\t\tLEFT JOIN \n\t\t\t\t\t\tliked_table \n\t\t\t\t\tON \n\t\t\t\t\t\tarticle_table.article_id = liked_table.article_id \n\t\t\t\t\tGROUP BY \n\t\t\t\t\t\tarticle_table.article_id \n\t\t\t\t\tORDER BY \n\t\t\t\t\t\tarticle_table.created_time DESC`\n\n\trow, err := sql.Query(getPosts)\n\n\tif err != nil {\n\t\tlog.SetOutput(io.MultiWriter(logfile, os.Stdout))\n\t\tlog.SetFlags(log.Ldate | log.Ltime)\n\t\tlog.Fatal(err)\n\t}\n\n\t// Prepare an array which save JSON results.\n\tvar postArray []dto.Article\n\n\tfor row.Next() {\n\t\tposts := dto.Article{}\n\t\tif err := row.Scan(&posts.UserID, &posts.ArticleID, &posts.LikedSum, &posts.UserName, &posts.Title, &posts.Content, &posts.CreatedTime, &posts.ModifiedTime); err != nil {\n\t\t\tlog.SetOutput(io.MultiWriter(logfile, os.Stdout))\n\t\t\tlog.SetFlags(log.Ldate | log.Ltime)\n\t\t\tlog.Fatal(err)\n\t\t}\n\n\t\t// Appending JSON in array.\n\t\tpostArray = append(postArray, posts)\n\t}\n\n\treturn postArray\n}", "func (snapshots EBSSnapshots) SortByLatest() {\n\tsort.Sort(sort.Reverse(byStartTime{snapshots}))\n}", "func (q *queryImpl) OrderBy(orderBys ...string) Query {\n\tq.orderBys = append(q.orderBys, orderBys...)\n\treturn q\n}", "func GetAllDepartment(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(Department))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []Department\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func SortByUpdatedAt() SortEnum {\n\treturn SortEnum{value: \"updated_at\"}\n}", "func GetAllSunStoreWatermark(query map[string]string, fields []string, sortby []string, order []string,\n\toffset int64, limit int64) (ml []interface{}, err error) {\n\to := orm.NewOrm()\n\tqs := o.QueryTable(new(SunStoreWatermark))\n\t// query k=v\n\tfor k, v := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, v)\n\t}\n\t// order by:\n\tvar sortFields []string\n\tif len(sortby) != 0 {\n\t\tif len(sortby) == len(order) {\n\t\t\t// 1) for each sort field, there is an associated order\n\t\t\tfor i, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[i] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[i] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t\tqs = qs.OrderBy(sortFields...)\n\t\t} else if len(sortby) != len(order) && len(order) == 1 {\n\t\t\t// 2) there is exactly one order, all the sorted fields will be sorted by this order\n\t\t\tfor _, v := range sortby {\n\t\t\t\torderby := \"\"\n\t\t\t\tif order[0] == \"desc\" {\n\t\t\t\t\torderby = \"-\" + v\n\t\t\t\t} else if order[0] == \"asc\" {\n\t\t\t\t\torderby = v\n\t\t\t\t} else {\n\t\t\t\t\treturn nil, errors.New(\"Error: Invalid order. Must be either [asc|desc]\")\n\t\t\t\t}\n\t\t\t\tsortFields = append(sortFields, orderby)\n\t\t\t}\n\t\t} else if len(sortby) != len(order) && len(order) != 1 {\n\t\t\treturn nil, errors.New(\"Error: 'sortby', 'order' sizes mismatch or 'order' size is not 1\")\n\t\t}\n\t} else {\n\t\tif len(order) != 0 {\n\t\t\treturn nil, errors.New(\"Error: unused 'order' fields\")\n\t\t}\n\t}\n\n\tvar l []SunStoreWatermark\n\tqs = qs.OrderBy(sortFields...)\n\tif _, err := qs.Limit(limit, offset).All(&l, fields...); err == nil {\n\t\tif len(fields) == 0 {\n\t\t\tfor _, v := range l {\n\t\t\t\tml = append(ml, v)\n\t\t\t}\n\t\t} else {\n\t\t\t// trim unused fields\n\t\t\tfor _, v := range l {\n\t\t\t\tm := make(map[string]interface{})\n\t\t\t\tval := reflect.ValueOf(v)\n\t\t\t\tfor _, fname := range fields {\n\t\t\t\t\tm[fname] = val.FieldByName(fname).Interface()\n\t\t\t\t}\n\t\t\t\tml = append(ml, m)\n\t\t\t}\n\t\t}\n\t\treturn ml, nil\n\t}\n\treturn nil, err\n}", "func (overlapRepo *OverlapRepo) Fetch() ([]*models.Overlap, error) {\n\toverlaps := []*models.Overlap{}\n\n\t// TODO pagination support\n\terr := overlapRepo.Conn.Model(models.Overlap{}).\n\t\tOrder(\"created_at asc\").\n\t\tFind(&overlaps).\n\t\tError\n\n\treturn overlaps, err\n}", "func (a *AvatarDAO) FindAll() ([]AvatarDAO, error) {\n\tvar avatars []AvatarDAO\n\terr := a.db.C(a.collection).Find(bson.M{}).All(&avatars)\n\treturn avatars, err\n}", "func FetchAllTodo(c *gin.Context) {\n\tvar todos []TodoModel\n\n\tdb.Find(&todos)\n\n\tif len(todos) <= 0 {\n\t\terrs := []commons.ErrorMsgs{}\n\t\terrs = append(errs, commons.ErrorMsgs{\n\t\t\tField: \"Todo\",\n\t\t\tMotive: \"Not Found\"})\n\t\tfmt.Println(errs)\n\t\tc.JSON(http.StatusNotFound, errs)\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, todos)\n}" ]
[ "0.61049104", "0.57657576", "0.57463163", "0.5727657", "0.56536424", "0.55227464", "0.54856706", "0.5468018", "0.53491807", "0.5293025", "0.5285763", "0.52729464", "0.5260392", "0.5237611", "0.5224621", "0.522297", "0.52228534", "0.52193165", "0.52149683", "0.5201512", "0.515416", "0.5106453", "0.51024854", "0.5093654", "0.50747067", "0.5064133", "0.5060681", "0.5028022", "0.50056374", "0.50044197", "0.4972034", "0.49672636", "0.4937976", "0.49226764", "0.49175698", "0.49075484", "0.49043638", "0.490067", "0.489048", "0.4844606", "0.4841148", "0.48355663", "0.48343754", "0.48288453", "0.48268235", "0.4809922", "0.48080447", "0.48031753", "0.48028868", "0.47956693", "0.47921062", "0.47900304", "0.47859615", "0.478346", "0.47781533", "0.4775647", "0.47622275", "0.4747041", "0.47456315", "0.47455555", "0.47340304", "0.4722998", "0.47194892", "0.4718648", "0.47103226", "0.470115", "0.46930197", "0.46756718", "0.4670529", "0.46598703", "0.46579927", "0.46494335", "0.46293086", "0.4626528", "0.46248293", "0.4623618", "0.46129662", "0.46118203", "0.45931184", "0.45931184", "0.4589556", "0.45827886", "0.45814273", "0.45782536", "0.4575827", "0.45707226", "0.45663777", "0.45638964", "0.45611107", "0.45577663", "0.45552108", "0.45497194", "0.45439738", "0.45414865", "0.45406657", "0.45365578", "0.4533297", "0.4532777", "0.45236275", "0.45217252" ]
0.5033738
27
Find Find only one article
func (repo *ArticleRepository) Find(article *domain.Article) error { err := repo.DB.Where(article).First(article).Error return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func FindArticle(id int64) (*Article, error) {\n\tif id == 0 {\n\t\treturn nil, errors.New(\"Invalid ID: it can't be zero\")\n\t}\n\t_article := Article{}\n\terr := DB.Get(&_article, DB.Rebind(`SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles WHERE articles.id = ? LIMIT 1`), id)\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn &_article, nil\n}", "func SearchArticle(db *sql.DB, words string) ([]blogmodel.Summary, error) {\n\tql := fmt.Sprintf(\"SELECT id, title, abstract, created_time FROM summary WHERE id in (SELECT id FROM content WHERE MATCH(substance) AGAINST('%s' IN BOOLEAN MODE));\", words)\n\trows, err := db.Query(ql)\n\tdefer func() {\n\t\tif rows != nil {\n\t\t\t// 可以关闭掉未scan连接一直占用\n\t\t\trows.Close()\n\t\t}\n\t}()\n\tif err != nil {\n\t\tlog.Error(\"query all err \", err)\n\t\treturn []blogmodel.Summary{}, err\n\t}\n\tvar res []blogmodel.Summary\n\tvar summary blogmodel.Summary\n\tfor rows.Next() {\n\t\terr = rows.Scan(&summary.SId, &summary.Title, &summary.Abstract, &summary.CreatedTime)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"search Article error\", err)\n\t\t}\n\t\ttimeParse, terr := time.Parse(\"2006-01-02 15:04:05\", summary.CreatedTime)\n\t\tif terr != nil {\n\t\t\tlog.Error(\"Parser time error \")\n\t\t}\n\t\tsummary.CreatedTime = fmt.Sprint(timeParse.Format(\"02 Jan 06\"))\n\t\tres = append(res, summary)\n\t}\n\ttimeParse, terr := time.Parse(\"2006-01-02 15:04:05\", summary.CreatedTime)\n\tif terr != nil {\n\t\tlog.Error(\"Parser time error \")\n\t}\n\tsummary.CreatedTime = fmt.Sprint(timeParse.Format(\"02 Jan 06\"))\n\tlog.Print(\"Search article by words:\", words)\n\treturn res, nil\n}", "func (*articleDAO) FirstArtilce() (Article, error) {\n\tvar article Article\n\terr := DBInstance.First(&article).Error\n\treturn article, err\n}", "func FirstArticle() (*Article, error) {\n\t_article := Article{}\n\terr := DB.Get(&_article, DB.Rebind(`SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles ORDER BY articles.id ASC LIMIT 1`))\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn &_article, nil\n}", "func getAnArticle(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\tfor _, article := range articles {\n\t\tif article.ID == params[\"id\"] {\n\t\t\tjson.NewEncoder(w).Encode(article)\n\t\t\treturn\t\n\t\t}\n\t}\n\tjson.NewEncoder(w).Encode(&article{})\n\tw.WriteHeader(http.StatusOK)\n}", "func ArticleRetrieve(c *gin.Context) {\n\tslug := c.Param(\"slug\")\n\t// if slug == \"feed\" {\n\t// \tArticleFeed(c)\n\t// \treturn\n\t// }\n\tarticleModel, _ := FindOneArticle(&ArticleModel{Slug: slug})\n\t// if err != nil {\n\t// \tc.JSON(http.StatusNotFound, common.NewError(\"articles\", errors.New(\"Invalid slug\")))\n\t// \treturn\n\t// }\n\tserializer := ArticleSerializer{c, articleModel}\n\tc.JSON(http.StatusOK, gin.H{\"article\": serializer.Response()})\n}", "func getArticle(w http.ResponseWriter, r *http.Request) {\r\n\tw.Header().Set(\"Content-Type\", \"application/json\")\r\n\tparams := mux.Vars(r) // Gets params\r\n\t// Looping through articles and find one with the id from the params\r\n\tfor _, item := range Articles {\r\n\t\tif item.Name == params[\"Id\"] {\r\n\t\t\tjson.NewEncoder(w).Encode(item)\r\n\t\t\treturn\r\n\t\t}\r\n\t}\r\n\tjson.NewEncoder(w).Encode(&Article{})\r\n}", "func getArticleByID(id int) (*article, error) {\n\tarticleList = nil\n\tprintarticle()\n\tfor _, a := range articleList {\n\t\tif a.ID == id {\n\t\t\treturn &a, nil\n\t\t}\n\t}\n\treturn nil, errors.New(\"Article not found\")\n}", "func (s *Service) FindByArticleID(ctx context.Context, articleID string) (*Article, error) {\n\ta, err := s.repo.FindByArticleID(ctx, articleID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn a, nil\n}", "func (service *RssService) GetArticle(id uint) *models.Articles {\n\t// get article\n\tvar article models.Articles\n\tservice.dbp().First(&article, id)\n\n\t// update state\n\tarticle.IsRead = true\n\tservice.dbp().Save(&article)\n\n\tif service.AppSettings.MarkSameRead {\n\t\tgo service.markSameArticles(article.Link, article.FeedId)\n\t}\n\n\treturn &article\n}", "func (s *Service) FindByID(ctx context.Context, id string) (*Article, error) {\n\ta, err := s.repo.FindByID(ctx, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn a, nil\n}", "func (s *Service) Article(id alarmquote.ArticleID) (*alarmquote.Article, error) {\n\treturn s.repo.Retrieve(id)\n}", "func (s *Service) article(c context.Context, mid, zoneid int64, highlight int, keyword, mobiApp, device, platform, buvid, filtered, order, sType string, plat int8, categoryID, build, pn, ps int, now time.Time) (res *search.TypeSearch, err error) {\n\tif res, err = s.srchDao.ArticleByType(c, mid, zoneid, keyword, mobiApp, device, platform, buvid, filtered, order, sType, plat, categoryID, build, highlight, pn, ps, now); err != nil {\n\t\tlog.Error(\"%+v\", err)\n\t\treturn\n\t}\n\tif res != nil && len(res.Items) > 0 {\n\t\tvar mids []int64\n\t\tfor _, v := range res.Items {\n\t\t\tmids = append(mids, v.Mid)\n\t\t}\n\t\tvar infom map[int64]*account.Info\n\t\tif infom, err = s.accDao.Infos3(c, mids); err != nil {\n\t\t\tlog.Error(\"%+v\", err)\n\t\t\terr = nil\n\t\t\treturn\n\t\t}\n\t\tfor _, item := range res.Items {\n\t\t\tif info, ok := infom[item.Mid]; ok {\n\t\t\t\titem.Name = info.Name\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func (d daoStruct) FindByArticleID(articleId string) (*Rule, error) {\n\trule := &Rule{}\n\tfilter := bson.NewDocument(bson.EC.String(\"articleId\", articleId))\n\tif err := d.collection.FindOne(context.Background(), filter).Decode(rule); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn rule, nil\n}", "func GetArticle(w http.ResponseWriter, r *http.Request) {\n params := mux.Vars(r)\n for _, item := range article {\n if item.ID == params[\"id\"] {\n item.ShortText = \"\"\n json.NewEncoder(w).Encode(item)\n return\n }\n }\n json.NewEncoder(w).Encode(&Article{})\n}", "func grabArticle(doc *goquery.Document, articleTitle string) (*goquery.Selection, string) {\n\t// Create initial variable\n\tauthor := \"\"\n\telementsToScore := []*goquery.Selection{}\n\n\t// First, node prepping. Trash nodes that look cruddy (like ones with the\n\t// class name \"comment\", etc), and turn divs into P tags where they have been\n\t// used inappropriately (as in, where they contain no other block level elements.)\n\tdoc.Find(\"*\").Each(func(i int, s *goquery.Selection) {\n\t\tmatchString := s.AttrOr(\"class\", \"\") + \" \" + s.AttrOr(\"id\", \"\")\n\n\t\t// If byline, remove this element\n\t\tif rel := s.AttrOr(\"rel\", \"\"); rel == \"author\" || rxByline.MatchString(matchString) {\n\t\t\ttext := s.Text()\n\t\t\ttext = strings.TrimSpace(text)\n\t\t\tif isValidByline(text) {\n\t\t\t\tauthor = text\n\t\t\t\ts.Remove()\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\n\t\t// Remove unlikely candidates\n\t\tif rxUnlikelyCandidates.MatchString(matchString) &&\n\t\t\t!rxOkMaybeItsACandidate.MatchString(matchString) &&\n\t\t\t!s.Is(\"body\") && !s.Is(\"a\") {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\tif rxUnlikelyElements.MatchString(goquery.NodeName(s)) {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\t// Remove DIV, SECTION, and HEADER nodes without any content(e.g. text, image, video, or iframe).\n\t\tif s.Is(\"div,section,header,h1,h2,h3,h4,h5,h6\") && isElementWithoutContent(s) {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\tif s.Is(\"section,h2,h3,h4,h5,h6,p,td,pre\") {\n\t\t\telementsToScore = append(elementsToScore, s)\n\t\t}\n\n\t\t// Turn all divs that don't have children block level elements into p's\n\t\tif s.Is(\"div\") {\n\t\t\t// Sites like http://mobile.slate.com encloses each paragraph with a DIV\n\t\t\t// element. DIVs with only a P element inside and no text content can be\n\t\t\t// safely converted into plain P elements to avoid confusing the scoring\n\t\t\t// algorithm with DIVs with are, in practice, paragraphs.\n\t\t\tif hasSinglePInsideElement(s) {\n\t\t\t\tnewNode := s.Children().First()\n\t\t\t\ts.ReplaceWithSelection(newNode)\n\t\t\t\telementsToScore = append(elementsToScore, s)\n\t\t\t} else if !hasChildBlockElement(s) {\n\t\t\t\tsetNodeTag(s, \"p\")\n\t\t\t\telementsToScore = append(elementsToScore, s)\n\t\t\t}\n\t\t}\n\t})\n\n\t// Loop through all paragraphs, and assign a score to them based on how content-y they look.\n\t// Then add their score to their parent node.\n\t// A score is determined by things like number of commas, class names, etc. Maybe eventually link density.\n\tcandidates := make(map[string]candidateItem)\n\tfor _, s := range elementsToScore {\n\t\t// If this paragraph is less than 25 characters, don't even count it.\n\t\tinnerText := normalizeText(s.Text())\n\t\tif strLen(innerText) < 25 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Exclude nodes with no ancestor.\n\t\tancestors := getNodeAncestors(s, 3)\n\t\tif len(ancestors) == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Calculate content score\n\t\t// Add a point for the paragraph itself as a base.\n\t\tcontentScore := 1.0\n\n\t\t// Add points for any commas within this paragraph.\n\t\tcontentScore += float64(strings.Count(innerText, \",\"))\n\t\tcontentScore += float64(strings.Count(innerText, \",\"))\n\n\t\t// For every 100 characters in this paragraph, add another point. Up to 3 points.\n\t\tcontentScore += math.Min(math.Floor(float64(strLen(innerText)/100)), 3)\n\n\t\t// Initialize and score ancestors.\n\t\tfor level, ancestor := range ancestors {\n\t\t\t// Node score divider:\n\t\t\t// - parent: 1 (no division)\n\t\t\t// - grandparent: 2\n\t\t\t// - great grandparent+: ancestor level * 3\n\t\t\tscoreDivider := 0\n\t\t\tif level == 0 {\n\t\t\t\tscoreDivider = 1\n\t\t\t} else if level == 1 {\n\t\t\t\tscoreDivider = 2\n\t\t\t} else {\n\t\t\t\tscoreDivider = level * 3\n\t\t\t}\n\n\t\t\tancestorHash := hashNode(ancestor)\n\t\t\tif _, ok := candidates[ancestorHash]; !ok {\n\t\t\t\tcandidates[ancestorHash] = initializeNodeScore(ancestor)\n\t\t\t}\n\n\t\t\tcandidate := candidates[ancestorHash]\n\t\t\tcandidate.score += contentScore / float64(scoreDivider)\n\t\t\tcandidates[ancestorHash] = candidate\n\t\t}\n\t}\n\n\t// Scale the final candidates score based on link density. Good content\n\t// should have a relatively small link density (5% or less) and be mostly\n\t// unaffected by this operation.\n\ttopCandidate := candidateItem{}\n\tfor hash, candidate := range candidates {\n\t\tcandidate.score = candidate.score * (1 - getLinkDensity(candidate.node))\n\t\tcandidates[hash] = candidate\n\n\t\tif topCandidate.node == nil || candidate.score > topCandidate.score {\n\t\t\ttopCandidate = candidate\n\t\t}\n\t}\n\n\t// If we still have no top candidate, use the body as a last resort.\n\tif topCandidate.node == nil {\n\t\tbody := doc.Find(\"body\").First()\n\n\t\tbodyHTML, _ := body.Html()\n\t\tnewHTML := fmt.Sprintf(`<div id=\"xxx-readability-body\">%s<div>`, bodyHTML)\n\t\tbody.AppendHtml(newHTML)\n\n\t\ttempReadabilityBody := body.Find(\"div#xxx-readability-body\").First()\n\t\ttempReadabilityBody.RemoveAttr(\"id\")\n\n\t\ttempHash := hashNode(tempReadabilityBody)\n\t\tif _, ok := candidates[tempHash]; !ok {\n\t\t\tcandidates[tempHash] = initializeNodeScore(tempReadabilityBody)\n\t\t}\n\n\t\ttopCandidate = candidates[tempHash]\n\t}\n\n\t// Create new document to save the final article content.\n\treader := strings.NewReader(`<div id=\"readability-content\"></div>`)\n\tnewDoc, _ := goquery.NewDocumentFromReader(reader)\n\tarticleContent := newDoc.Find(\"div#readability-content\").First()\n\n\t// Now that we have the top candidate, look through its siblings for content\n\t// that might also be related. Things like preambles, content split by ads\n\t// that we removed, etc.\n\ttopCandidateClass, _ := topCandidate.node.Attr(\"class\")\n\tsiblingScoreThreshold := math.Max(10.0, topCandidate.score*0.2)\n\ttopCandidate.node.Parent().Children().Each(func(_ int, sibling *goquery.Selection) {\n\t\tappendSibling := false\n\n\t\tif sibling.IsSelection(topCandidate.node) {\n\t\t\tappendSibling = true\n\t\t} else {\n\t\t\tcontentBonus := 0.0\n\t\t\tsiblingClass, _ := sibling.Attr(\"class\")\n\t\t\tif siblingClass == topCandidateClass && topCandidateClass != \"\" {\n\t\t\t\tcontentBonus += topCandidate.score * 0.2\n\t\t\t}\n\n\t\t\tsiblingHash := hashNode(sibling)\n\t\t\tif item, ok := candidates[siblingHash]; ok && item.score > siblingScoreThreshold {\n\t\t\t\tappendSibling = true\n\t\t\t} else if sibling.Is(\"p\") {\n\t\t\t\tlinkDensity := getLinkDensity(sibling)\n\t\t\t\tnodeContent := normalizeText(sibling.Text())\n\t\t\t\tnodeLength := strLen(nodeContent)\n\n\t\t\t\tif nodeLength > 80 && linkDensity < 0.25 {\n\t\t\t\t\tappendSibling = true\n\t\t\t\t} else if nodeLength < 80 && nodeLength > 0 &&\n\t\t\t\t\tlinkDensity == 0 && rxPIsSentence.MatchString(nodeContent) {\n\t\t\t\t\tappendSibling = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\tif appendSibling {\n\t\t\tarticleContent.AppendSelection(sibling)\n\t\t}\n\t})\n\n\t// So we have all of the content that we need.\n\t// Now we clean it up for presentation.\n\tprepArticle(articleContent, articleTitle)\n\n\treturn articleContent, author\n}", "func (idearepo *IdeaRepo) SearchIdeaByTitle(title string ) ([]*entity.Idea , error){\n\tideas := []*entity.Idea{}\n\tcursor , era := idearepo.DB.Collection(entity.IDEA).Find(context.TODO(), bson.M{ \"$text\": bson.D{{\"$search\" , title } } })\n\tif era != nil {\n\t\tfmt.Println(era.Error())\n\t\treturn ideas , era \n\t}\n\ttempoIdeas := map[string]*entity.Idea{} \n\tfor cursor.Next(context.TODO()) {\n\t\tusr := &entity.Idea{}\n\t\tcursor.Decode(usr)\n\t\tif usr.ID != \"\" {\n\t\t\ttempoIdeas[usr.ID] = usr\n\t\t}\n\t}\n\tif len(tempoIdeas) > 0 {\n\t\tfor _ , val := range tempoIdeas {\n\t\t\tideas = append(ideas , val )\n\t\t}\n\t}\n\tif len( ideas) ==0 {\n\t\treturn ideas , errors.New(\"No Record Found \")\n\t}\n\treturn ideas , nil \n}", "func (service *RssService) GetArticle(id int64, feedID int64, userID int64) *models.Articles {\n\trss := service.GetRss(userID)\n\n\tif len(rss) == 0 {\n\t\treturn nil\n\t}\n\n\t// get article\n\tvar article models.Articles\n\tservice.db.Where(&models.Articles{Id: id, FeedId: feedID}).First(&article)\n\n\tvar settings models.Settings // todo: to func\n\tservice.db.Where(models.Settings{UserId: userID}).Find(&settings)\n\n\t// update state\n\tarticle.IsRead = true\n\tservice.db.Save(&article)\n\n\tif settings.MarkSameRead {\n\t\tgo service.markSameArticles(article.Link, article.FeedId)\n\t}\n\n\treturn &article\n}", "func getArticleByID(id int) (*article, error) {\n\tclient, clierr := mongo.NewClient(options.Client().ApplyURI(dbConfig.dburi))\n\tif clierr != nil {\n\t\tlog.Fatal(clierr)\n\t}\n\tctx, _ := context.WithTimeout(context.Background(), 10*time.Second)\n\tconnerr := client.Connect(ctx)\n\tif connerr != nil {\n\t\tlog.Fatal(connerr)\n\t}\n\tdefer client.Disconnect(ctx)\n\tdb := client.Database(dbConfig.dbName).Collection(dbConfig.usersColName)\n\n\tcursor, err := db.Find(ctx, bson.D{})\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfor cursor.Next(ctx) {\n\t\tvar articleOne article\n\t\tif err := cursor.Decode(&articleOne); err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tif articleOne.ID == id {\n\t\t\treturn &articleOne, nil\n\t\t}\n\t}\n\treturn nil, errors.New(\"Article Can't be found\")\n}", "func FindArticlesWhere(where string, args ...interface{}) (articles []Article, err error) {\n\tsql := \"SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles\"\n\tif len(where) > 0 {\n\t\tsql = sql + \" WHERE \" + where\n\t}\n\tstmt, err := DB.Preparex(DB.Rebind(sql))\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\terr = stmt.Select(&articles, args...)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\treturn articles, nil\n}", "func Find(rawurl string) (idvideoteca string, err error) {\n\n\tif strings.Contains(rawurl, \"%\") {\n\t// if isURLEncoded.MatchString(rawurl) {\n\t\trawurl, err = url.QueryUnescape(rawurl)\n\t\tif err != nil {\n\t\t\t// log.Println(err.Error())\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\n\t// Cerca la prima corrispondenza\n\tidvideoteca = idv.FindString(rawurl)\n\n\t// idvideoteca = onlyNum.ReplaceAllString(element, \"\")\n\n\tidvideoteca = strings.Replace(idvideoteca, \"/\", \"\", 2)\n\n\treturn idvideoteca, err\n}", "func (s NounResource) FindOne(ID string, r api2go.Request) (api2go.Responder, error) {\n\tnoun, err := s.NounStorage.GetOne(ID)\n\tif err != nil {\n\t\treturn &Response{}, api2go.NewHTTPError(err, err.Error(), http.StatusNotFound)\n\t}\n\treturn &Response{Res: noun}, nil\n}", "func FindArticleBy(field string, val interface{}) (*Article, error) {\n\t_article := Article{}\n\tsqlFmt := `SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles WHERE %s = ? LIMIT 1`\n\tsqlStr := fmt.Sprintf(sqlFmt, field)\n\terr := DB.Get(&_article, DB.Rebind(sqlStr), val)\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn &_article, nil\n}", "func findExistingPermanode(qs search.QueryDescriber, wholeRef blob.Ref) (pn blob.Ref, err error) {\n\tres, err := qs.Query(&search.SearchQuery{\n\t\tConstraint: &search.Constraint{\n\t\t\tPermanode: &search.PermanodeConstraint{\n\t\t\t\tAttr: \"camliContent\",\n\t\t\t\tValueInSet: &search.Constraint{\n\t\t\t\t\tFile: &search.FileConstraint{\n\t\t\t\t\t\tWholeRef: wholeRef,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tDescribe: &search.DescribeRequest{\n\t\t\tDepth: 1,\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn\n\t}\n\tif res.Describe == nil {\n\t\treturn pn, os.ErrNotExist\n\t}\nRes:\n\tfor _, resBlob := range res.Blobs {\n\t\tbr := resBlob.Blob\n\t\tdesBlob, ok := res.Describe.Meta[br.String()]\n\t\tif !ok || desBlob.Permanode == nil {\n\t\t\tcontinue\n\t\t}\n\t\tattrs := desBlob.Permanode.Attr\n\t\tfor _, attr := range sensitiveAttrs {\n\t\t\tif attrs.Get(attr) != \"\" {\n\t\t\t\tcontinue Res\n\t\t\t}\n\t\t}\n\t\treturn br, nil\n\t}\n\treturn pn, os.ErrNotExist\n}", "func (d *DB) NextArticle(categoryID uint16, count int, user *User) (*Article, error) {\n\tarticles, err := d.Articles(categoryID, count, user)\n\tif err != nil {\n\t\tglog.Error(err)\n\t\treturn nil, err\n\t}\n\tfor _, article := range articles {\n\t\tscore, alreadySeen := user.Items[article.ID.Hex()]\n\t\tif !alreadySeen {\n\t\t\treturn article, nil\n\t\t}\n\t\t// maybe show again, say after some time....\n\t\tif score == 0 {\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"none\")\n}", "func (s *Searcher) FindOne() (interface{}, error) {\n\tcontext := s.parseContext()\n\tresult := s.Resource.NewStruct()\n\terr := s.Resource.CallFindOne(result, nil, context)\n\treturn result, err\n}", "func FindArticleBySql(sql string, args ...interface{}) (*Article, error) {\n\tstmt, err := DB.Preparex(DB.Rebind(sql))\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\t_article := &Article{}\n\terr = stmt.Get(_article, args...)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil, err\n\t}\n\treturn _article, nil\n}", "func (a ArticleDB) QueryOne(sess *sqlabble.Session, st stmt.Statement) (Article, error) {\n\tquery, values := sess.Builder.Build(st)\n\trows, err := sess.Query(query, values...)\n\tif err != nil {\n\t\treturn Article{}, err\n\t}\n\tms, err := a.Map(rows)\n\tif err != nil {\n\t\treturn Article{}, err\n\t}\n\tif len(ms) == 0 {\n\t\treturn Article{}, sqlabble.NewErrRecordNotFound(a.Table.Name)\n\t}\n\tif len(ms) > 1 {\n\t\treturn Article{}, sqlabble.NewErrFoundMultipleRecords(a.Table.Name)\n\t}\n\treturn ms[0], nil\n}", "func userArticleIsFavarite(userId string, articleId string) bool {\n user := bson.ObjectIdHex(userId)\n article := bson.ObjectIdHex(articleId)\n count, err := db.favarite.Find(bson.M{\"user\": user, \"article\": article}).Count()\n if err != nil || count == 0 {\n return false\n }\n return true\n}", "func createNewArticle(title, content string) (*article, error) {\n\tclient, clierr := mongo.NewClient(options.Client().ApplyURI(dbConfig.dburi))\n\tif clierr != nil {\n\t\tlog.Fatal(clierr)\n\t}\n\tctx, _ := context.WithTimeout(context.Background(), 10*time.Second)\n\tconnerr := client.Connect(ctx)\n\tif connerr != nil {\n\t\tlog.Fatal(connerr)\n\t}\n\tdefer client.Disconnect(ctx)\n\tdb := client.Database(dbConfig.dbName).Collection(dbConfig.articleColName)\n\t// Set the ID of a new article to one more than the number of articles\n\ta := article{ID: len(getAllArticles()) + 1, Title: title, Content: content}\n\n\t// Add the article to the list of articles\n\t_, err := db.InsertOne(ctx, a)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treturn &a, nil\n}", "func findOne(m Model, wh Where, args ...interface{}) error {\n\tvar err error\n\t// trim spaces just in case (injection)\n\tfld, crit, val := wh.trimSpace()\n\tq := \"SELECT * FROM \" + m.TableName() + \" WHERE \" + fld + crit + \"? LIMIT 1\"\n\t// write data to dest\n\terr = DB.QueryRow(q, val).Scan(args...)\n\treturn err\n}", "func (r *readability) getArticleContent(doc *goquery.Document) *goquery.Selection {\n\t// First, node prepping. Trash nodes that look cruddy (like ones with the\n\t// class name \"comment\", etc), and turn divs into P tags where they have been\n\t// used inappropriately (as in, where they contain no other block level elements.)\n\tdoc.Find(\"*\").Each(func(i int, s *goquery.Selection) {\n\t\tmatchString := s.AttrOr(\"class\", \"\") + \" \" + s.AttrOr(\"id\", \"\")\n\n\t\t// If byline, remove this element\n\t\tif rel := s.AttrOr(\"rel\", \"\"); rel == \"author\" || byline.MatchString(matchString) {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\t// Remove unlikely candidates\n\t\tif unlikelyCandidates.MatchString(matchString) &&\n\t\t\t!okMaybeItsACandidate.MatchString(matchString) &&\n\t\t\t!s.Is(\"body\") && !s.Is(\"a\") {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\tif unlikelyElements.MatchString(r.getTagName(s)) {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\t// Remove DIV, SECTION, and HEADER nodes without any content(e.g. text, image, video, or iframe).\n\t\tif s.Is(\"div,section,header,h1,h2,h3,h4,h5,h6\") && r.isElementEmpty(s) {\n\t\t\ts.Remove()\n\t\t\treturn\n\t\t}\n\n\t\t// Turn all divs that don't have children block level elements into p's\n\t\tif s.Is(\"div\") {\n\t\t\tsHTML, _ := s.Html()\n\t\t\tif !divToPElements.MatchString(sHTML) {\n\t\t\t\ts.Nodes[0].Data = \"p\"\n\t\t\t}\n\t\t}\n\t})\n\n\t// Loop through all paragraphs, and assign a score to them based on how content-y they look.\n\t// Then add their score to their parent node.\n\t// A score is determined by things like number of commas, class names, etc. Maybe eventually link density.\n\tr.candidates = make(map[string]candidateItem)\n\tdoc.Find(\"p\").Each(func(i int, s *goquery.Selection) {\n\t\t// If this paragraph is less than 25 characters, don't even count it.\n\t\tinnerText := normalizeText(s.Text())\n\t\tif strLen(innerText) < 25 {\n\t\t\treturn\n\t\t}\n\n\t\t// Exclude nodes with no ancestor.\n\t\tancestors := r.getNodeAncestors(s, 3)\n\t\tif len(ancestors) == 0 {\n\t\t\treturn\n\t\t}\n\n\t\t// Calculate content score\n\t\t// Add a point for the paragraph itself as a base.\n\t\tcontentScore := 1.0\n\n\t\t// Add points for any commas within this paragraph.\n\t\tcontentScore += float64(strings.Count(innerText, \",\"))\n\t\tcontentScore += float64(strings.Count(innerText, \",\"))\n\n\t\t// For every 100 characters in this paragraph, add another point. Up to 3 points.\n\t\tcontentScore += math.Min(math.Floor(float64(strLen(innerText)/100)), 3)\n\n\t\t// Initialize and score ancestors.\n\t\tfor level, ancestor := range ancestors {\n\t\t\t// Node score divider:\n\t\t\t// - parent: 1 (no division)\n\t\t\t// - grandparent: 2\n\t\t\t// - great grandparent+: ancestor level * 3\n\t\t\tscoreDivider := 0\n\t\t\tif level == 0 {\n\t\t\t\tscoreDivider = 1\n\t\t\t} else if level == 1 {\n\t\t\t\tscoreDivider = 2\n\t\t\t} else {\n\t\t\t\tscoreDivider = level * 3\n\t\t\t}\n\n\t\t\tancestorHash := hashStr(ancestor)\n\t\t\tif _, ok := r.candidates[ancestorHash]; !ok {\n\t\t\t\tr.candidates[ancestorHash] = r.initializeNodeScore(ancestor)\n\t\t\t}\n\n\t\t\tcandidate := r.candidates[ancestorHash]\n\t\t\tcandidate.score += contentScore / float64(scoreDivider)\n\t\t\tr.candidates[ancestorHash] = candidate\n\t\t}\n\t})\n\n\t// After we've calculated scores, loop through all of the possible\n\t// candidate nodes we found and find the one with the highest score.\n\tvar topCandidate *candidateItem\n\tfor hash, candidate := range r.candidates {\n\t\tcandidate.score = candidate.score * (1 - r.getLinkDensity(candidate.node))\n\t\tr.candidates[hash] = candidate\n\n\t\tif topCandidate == nil || candidate.score > topCandidate.score {\n\t\t\tif topCandidate == nil {\n\t\t\t\ttopCandidate = new(candidateItem)\n\t\t\t}\n\n\t\t\ttopCandidate.score = candidate.score\n\t\t\ttopCandidate.node = candidate.node\n\t\t}\n\t}\n\n\t// If top candidate not found, stop\n\tif topCandidate == nil {\n\t\treturn nil\n\t}\n\n\tr.prepArticle(topCandidate.node)\n\treturn topCandidate.node\n}", "func (h *handler) FindOne(ctx context.Context, param db.Params) error {\n\tbsonFilter := bson.M{}\n\tfor key, val := range param.Filter {\n\t\tbsonFilter[key] = val\n\t}\n\tif err := h.getDatabase(param.Database).C(param.Collection).Find(bsonFilter).One(param.Result); err != nil {\n\t\tif err == mgo.ErrNotFound {\n\t\t\terr = db.ErrNotFound\n\t\t}\n\t\treturn err\n\t}\n\treturn nil\n}", "func (db *Client) FindOne(collection string, filters *bson.M) *mongo.SingleResult {\n\treturn db.Database.Collection(collection).FindOne(context.TODO(), filters)\n}", "func GetArticle(c *gin.Context) {\n\tvar loginUser models.User\n\tclaims, err := middlewares.Authenticate(c)\n\n\tclient, ctx, cancel := utils.GetConnection()\n\tdefer cancel()\n\tuserCollection := client.Database(\"conduit\").Collection(\"users\")\n\tarticleCollection := client.Database(\"conduit\").Collection(\"articles\")\n\n\tid, err := primitive.ObjectIDFromHex(claims.UserID)\n\tif err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\n\t\t\t\"error\": err.Error(),\n\t\t})\n\t\treturn\n\t}\n\terr = userCollection.FindOne(ctx, bson.M{\n\t\t\"_id\": id,\n\t}).Decode(&loginUser)\n\n\tif err != nil {\n\t\tc.JSON(http.StatusNotFound, gin.H{\n\t\t\t\"error\": err.Error(),\n\t\t})\n\t\treturn\n\t}\n\n\tslug := c.Param(\"slug\")\n\n\tvar article models.ArticleWithAuthor\n\tmatchStage := bson.D{{Key: \"$match\", Value: bson.D{{Key: \"slug\", Value: slug}}}}\n\tlookupStage := bson.D{{Key: \"$lookup\", Value: bson.D{{Key: \"from\", Value: \"users\"}, {Key: \"localField\", Value: \"author\"}, {Key: \"foreignField\", Value: \"_id\"}, {Key: \"as\", Value: \"author\"}}}}\n\tcursor, err := articleCollection.Aggregate(ctx, mongo.Pipeline{matchStage, lookupStage})\n\tif err != nil {\n\t\tc.JSON(http.StatusNotFound, gin.H{\n\t\t\t\"error\": err.Error(),\n\t\t})\n\t\treturn\n\t}\n\tfor cursor.Next(ctx) {\n\t\terr := cursor.Decode(&article)\n\t\tif err != nil {\n\t\t\tc.JSON(http.StatusInternalServerError, gin.H{\n\t\t\t\t\"error\": err.Error(),\n\t\t\t})\n\t\t\treturn\n\t\t}\n\t\tbreak\n\t}\n\tvar articleJSON = models.ArticleJSON{ArticleBase: article.ArticleBase, Author: article.Author.ToProfile(&loginUser)}\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"article\": articleJSON,\n\t})\n}", "func Deletearticles(title string) string {\n\tvar db, err = sql.Open(\"mysql\", \"saggarwal98:shubham@tcp(127.0.0.1:3306)/NewMysqlApi\")\n\tdefer db.Close()\n\tflag := false\n\tresults, err := db.Query(\"Select * from Articles\")\n\tfor results.Next() {\n\t\tvar a Article\n\t\terr = results.Scan(&a.ID, &a.Title, &a.Description, &a.Price)\n\t\tif err != nil {\n\t\t\tlog.Print(err.Error())\n\t\t}\n\t\tif a.Title == title {\n\t\t\tflag = true\n\t\t\t_, err := db.Query(\"Delete from Articles where Title='\" + a.Title + \"'\")\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(err.Error())\n\t\t\t\treturn \"error received\"\n\t\t\t}\n\t\t}\n\t}\n\tif flag == false {\n\t\treturn \"No article found with that title\"\n\t}\n\treturn \"Article Deleted\"\n}", "func (c WrapperCollection) FindOne(ctx context.Context, filter interface{}, opts ...option.FindOneOptioner) DocumentResultLayer {\n\tvar docResult *WrapperDocumentResult\n\n\t// Recover on panic() from mongo driver.\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tdocResult = nil\n\t\t}\n\t}()\n\n\tdocResult = &WrapperDocumentResult{c.Collection.FindOne(ctx, filter, opts...)}\n\treturn docResult\n}", "func (e *Element) findOne(name string, attrs Attributes) *Element {\n\tif e.Type == TextType {\n\t\treturn nil\n\t}\n\tif e.Content == name && e.containsAttrs(attrs) {\n\t\treturn e\n\t}\n\tfor _, child := range e.Children {\n\t\tfounds := child.findOne(name, attrs)\n\t\tif founds != nil {\n\t\t\treturn founds\n\t\t}\n\t}\n\treturn nil\n}", "func getArticleById(id int) {\n\tdb, err := bolt.Open(\"../my.db\", 0600, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer db.Close()\n\n\tvar article Article\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(\"Article\"))\n\t\tfmt.Println(b)\n\t\tif b != nil {\n\t\t\tv := b.Get(itob(id))\n\t\t\tif v == nil {\n\t\t\t\tfmt.Println(id, \" Article Not Exists\")\n\t\t\t\treturn errors.New(\"Article Not Exists\")\n\t\t\t} else {\n\t\t\t\t_ = json.Unmarshal(v, &article)\n\t\t\t\treturn nil\n\t\t\t}\n\t\t} else {\n\t\t\tfmt.Println(\"Article Not Exists\")\n\t\t\treturn errors.New(\"Article Not Exists\")\n\t\t}\n\t})\n\n}", "func FindTitle(where string, args ...any) (*Title, error) {\n\tvar op = dbi.DB.Operation()\n\top.Dbg = dbi.Debug\n\tvar t = &Title{}\n\top.Select(\"titles\", &Title{}).Where(where, args...).First(t)\n\treturn t, op.Err()\n}", "func (s TensePresIndResource) FindOne(ID string, r api2go.Request) (api2go.Responder, error) {\n\tqueryConstraints, err := constraints.ApplySingleConstraints(r)\n\tif err != nil {\n\t\treturn &Response{}, err\n\t}\n\n\ttense, err := s.TensePresIndStorage.GetOne(ID, queryConstraints)\n\tif err != nil {\n\t\treturn &Response{}, api2go.NewHTTPError(err, err.Error(), http.StatusNotFound)\n\t}\n\treturn &Response{Res: tense}, nil\n}", "func return_one_article(w http.ResponseWriter, r *http.Request) {\n\n var Urls = strings.Split(r.URL.Path, \"/\")\n var Temp_Article = Articles[0]\n //Calling Id's\n Temp_Article.ID = \"NOTFOUND\"\n for i := 0; i < len(Articles); i++{\n //loop for calling Ids\n if Urls[2]==Articles[i].ID {\n Temp_Article = Articles[i]\n }\n\n }\n if Temp_Article.ID == \"NOTFOUND\" {\n\n w.WriteHeader(http.StatusNotFound)\n return\n }\n\n if len(Urls) != 3 {\n\n w.WriteHeader(http.StatusNotFound)\n return\n }\n json.NewEncoder(w).Encode(Temp_Article)\n\n}", "func (d *DB) NextUnseenArticle(user *User) (*Article, error) {\n\tarticle := Article{}\n\tusersSeen := []bson.ObjectId{}\n\tfor id, _ := range user.Items {\n\t\tusersSeen = append(usersSeen, bson.ObjectIdHex(id))\n\t}\n\terr := d.Items.Find(bson.M{\"_id\": bson.M{\"$nin\": usersSeen}}).One(&article)\n\treturn &article, err\n}", "func (c *cache) find(id int) (item, bool) {\n\tfor i := 0; i < c.curSize; i++ {\n\t\tif c.stories[i].Item.ID == id {\n\t\t\treturn c.stories[i], true\n\t\t}\n\t}\n\treturn item{}, false\n}", "func (d *Document) FindOne(ctx context.Context, id string) (*flare.Document, error) {\n\td.mutex.RLock()\n\tdefer d.mutex.RUnlock()\n\n\tdocument, ok := d.documents[id]\n\tif !ok {\n\t\treturn nil, &errMemory{message: fmt.Sprintf(\"document '%s' not found\", id), notFound: true}\n\t}\n\treturn &document, nil\n}", "func (repo *mongoBaseRepo) FindOne(filter interface{}, result interface{}, args ...interface{}) error {\n\t// Default values\n\ttimeout := DefaultTimeout\n\topts := &options.FindOneOptions{}\n\n\tfor i := 0; i < len(args); i++ {\n\t\tswitch val := args[i].(type) {\n\t\tcase time.Duration:\n\t\t\ttimeout = val\n\t\tcase *options.FindOneOptions:\n\t\t\topts = val\n\t\t}\n\t}\n\n\tif repo.locale != nil && opts.Collation == nil {\n\t\topts.SetCollation(&options.Collation{\n\t\t\tLocale: *repo.locale,\n\t\t})\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), timeout)\n\tdefer cancel()\n\n\t// Find and convert no documents error\n\terr := repo.collection.FindOne(ctx, filter, opts).Decode(result)\n\tif err != nil {\n\t\tif errors.Is(err, mongo.ErrNoDocuments) {\n\t\t\treturn ErrNotFound\n\t\t}\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (clnt *Client) FindOne(coll string, query interface{}) *mongo.SingleResult {\n\tcollection := clnt.DB.Collection(coll)\n\tsingleResult := collection.FindOne(context.TODO(), query)\n\treturn singleResult\n}", "func (as *ArticleService) GetArticle(id string) (*entities.Article, []error) {\n\tarticle, errs := as.articleRepo.GetArticle(id) // here article is pointer btw, it is accesing from the one we did in the repository layer\n\n\tif len(errs) > 0 {\n\t\treturn nil, errs\n\t}\n\treturn article, nil\n}", "func (r *Resolver) ArticleFindList() runtime.ArticleFindListResolver {\n\treturn &articleFindListResolver{r}\n}", "func (s *PetStore) FindOne(q *PetQuery) (*Pet, error) {\n\tq.Limit(1)\n\tq.Offset(0)\n\trs, err := s.Find(q)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !rs.Next() {\n\t\treturn nil, kallax.ErrNotFound\n\t}\n\n\trecord, err := rs.Get()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := rs.Close(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn record, nil\n}", "func (s *PersonStore) FindOne(q *PersonQuery) (*Person, error) {\n\tq.Limit(1)\n\tq.Offset(0)\n\trs, err := s.Find(q)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !rs.Next() {\n\t\treturn nil, kallax.ErrNotFound\n\t}\n\n\trecord, err := rs.Get()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := rs.Close(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn record, nil\n}", "func (p *APIView) FindOne(result interface{}, context *Context) error {\n\tprimaryQuerySQL, primaryParams := p.toPrimaryQueryParams(result, context.ResourceID, context)\n\tdb := context.GetDB()\n\tif db == nil {\n\t\treturn errors.New(\"db is nil\")\n\t}\n\tif primaryQuerySQL != \"\" {\n\t\treturn db.First(result, append([]interface{}{primaryQuerySQL}, primaryParams...)...).Error\n\t}\n\n\treturn errors.New(\"failed to find\")\n}", "func FindOne(q *mgo.Query, d interface{}) (bool, error) {\n\terr := q.One(d)\n\tif err == nil {\n\t\treturn true, nil\n\t}\n\n\tif err.Error() == \"not found\" {\n\t\treturn false, nil\n\t}\n\n\treturn false, err\n}", "func (s *Server) Article(ctx context.Context, in *pb.ArticleRequest) (*pb.ArticleReply, error) {\n\ta, err := s.db.Get(ctx, in.Id)\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Internal, fmt.Sprintf(\"failed to get article: %v\", err))\n\t}\n\treturn &pb.ArticleReply{Article: a}, nil\n}", "func (e *ExpenseModel) ReadOne(filter interface{}) (Expense, error) {\n\tvar expense Expense\n\tcollection := e.db.Client.Database(e.db.DBName).Collection(\"expenses\")\n\tdocumentReturned := collection.FindOne(context.TODO(), filter)\n\tdocumentReturned.Decode(&expense)\n\treturn expense, nil\n}", "func findExistingPermanode(qs search.QueryDescriber, wholeRef blob.Ref) (pn blob.Ref, picasaAttrs url.Values, err error) {\n\tres, err := qs.Query(&search.SearchQuery{\n\t\tConstraint: &search.Constraint{\n\t\t\tPermanode: &search.PermanodeConstraint{\n\t\t\t\tAttr: \"camliContent\",\n\t\t\t\tValueInSet: &search.Constraint{\n\t\t\t\t\tFile: &search.FileConstraint{\n\t\t\t\t\t\tWholeRef: wholeRef,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tDescribe: &search.DescribeRequest{\n\t\t\tDepth: 1,\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn\n\t}\n\tif res.Describe == nil {\n\t\treturn pn, nil, os.ErrNotExist\n\t}\nRes:\n\tfor _, resBlob := range res.Blobs {\n\t\tbr := resBlob.Blob\n\t\tdesBlob, ok := res.Describe.Meta[br.String()]\n\t\tif !ok || desBlob.Permanode == nil {\n\t\t\tcontinue\n\t\t}\n\t\tattrs := desBlob.Permanode.Attr\n\t\tif attrs.Get(picasa.AttrMediaURL) != \"\" {\n\t\t\t// If we found a picasa permanode, we're going to reuse it, in order to avoid\n\t\t\t// creating what would look like duplicates. We let the caller deal with merging\n\t\t\t// properly on the node the existing (Picasa) attributes, with the new (Google\n\t\t\t// Photos) attributes.\n\t\t\treturn br, attrs, nil\n\t\t}\n\t\t// otherwise, only keep it if attributes are not conflicting.\n\t\tfor _, attr := range sensitiveAttrs {\n\t\t\tif attrs.Get(attr) != \"\" {\n\t\t\t\tcontinue Res\n\t\t\t}\n\t\t}\n\t\treturn br, nil, nil\n\t}\n\treturn pn, nil, os.ErrNotExist\n}", "func FindOne(s Session, dbname string, collection string, query map[string]interface{}) (interface{}, error) {\n\tvar object interface{}\n\tif err := s.DB(dbname).C(collection).Find(query).One(&object); err != nil {\n\t\treturn object, err\n\t}\n\treturn object, nil\n}", "func FindOne(condition Equipment) (Equipment, error) {\n\tdb := common.GetDB()\n\tvar eq Equipment\n\terr := db.Where(condition).First(&eq).Error\n\treturn eq, err\n}", "func (l *Link) FindOne(database, collection string, filter interface{}, dest interface{}) error {\n\tif err := l.linkCheck(\"link.FindOne\"); err != nil {\n\t\treturn err\n\t}\n\n\tif dest == nil {\n\t\treturn fmt.Errorf(`given \"dest\" is null`)\n\t}\n\n\tctx, cancel := context.WithTimeout(context.Background(), l.execTimeout())\n\n\tdefer cancel()\n\n\tif filter == nil {\n\t\tfilter = bson.M{}\n\t}\n\n\trs := l.client.Database(database).Collection(collection).FindOne(ctx, filter, options.FindOne())\n\n\tif err := rs.Err(); err != nil {\n\t\t// If not connected, try once again, reconnecting. otherwise, just return/leave\n\t\tif !errors.Is(err, mongo.ErrClientDisconnected) {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := l.connect(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tctx2, cancel2 := context.WithTimeout(context.Background(), l.execTimeout())\n\n\t\tdefer cancel2()\n\n\t\trs = l.client.Database(database).Collection(collection).FindOne(ctx2, filter, options.FindOne())\n\n\t\tif err := rs.Err(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err := rs.Decode(dest); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (repo *ArticleRepository) FindAll(articles *[]domain.Article) error {\n\terr := repo.DB.Order(\"created_at desc\").Limit(100).Find(articles).Error\n\treturn err\n}", "func (p *Parser) GetArticle(id int) (*model.Article, int, error) {\n\tdoc, code, err := p.getDoc(utils.BuildURL(malURL, \"featured\", id), \".content-left\")\n\tif err != nil {\n\t\treturn nil, code, err\n\t}\n\treturn p.article.GetDetails(doc), http.StatusOK, nil\n}", "func (i *Invoice) FindOne(id string, client *Client) (*Invoice, error) {\n\tresp, err := client.Get(i.Endpoint() + \"/\" + id)\n\trespData, err := SendResp(resp, err, i)\n\tif err != nil {\n\t\treturn i, err\n\t}\n\treturn &respData.Invoice, err\n}", "func (reajuste *Reajuste) FindReajusteByID(db *gorm.DB, anoRef, codLote uint32) (*Reajuste, error) {\n\n\t//\tBusca um elemento no banco de dados a partir de sua chave primaria\n\terr := db.Debug().Model(Reajuste{}).Where(\"ano_ref = ? AND cod_lote = ?\", anoRef, codLote).Take(&reajuste).Error\n\tif err != nil {\n\t\treturn &Reajuste{}, err\n\t}\n\n\treturn reajuste, err\n}", "func (m *mongoDriver) FindOne(ctx context.Context, query bson.M, opts *options.FindOneOptions) (*mongo.SingleResult, error) {\n\tcoll := m.client.Database(database).Collection(collection)\n\tres := coll.FindOne(ctx, query, opts)\n\n\tif err := res.Err(); err != nil {\n\t\treturn nil, fmt.Errorf(\"FindOne: unable to find book: %s\", err)\n\t}\n\n\treturn res, nil\n}", "func (d *DBRepository) findOne(ctx context.Context, id string) (*Travel, error) {\n\tobjectId, err := primitive.ObjectIDFromHex(id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres := d.Collection.FindOne(ctx, bson.M{\"_id\": objectId})\n\tvar travel Travel\n\tif err := res.Decode(&travel); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &travel, nil\n}", "func (ads *ArticleDS) Get(idS string) (*articles.Article, error) {\n\tid, err := strconv.Atoi(idS)\n\tif err != nil {\n\t\treturn nil, status.ErrBadRequest.WithMessage(err.Error())\n\t}\n\tdb := ads.db\n\ta := articles.Article{}\n\tfindStmt := `SELECT id, title, publish_date, body, tags FROM ARTICLES.ARTICLE WHERE id = $1`\n\terr = db.QueryRow(\n\t\tfindStmt,\n\t\tid).Scan(&a.ID, &a.Title, &a.Date, &a.Body, pq.Array(&a.Tags))\n\tif err == sql.ErrNoRows {\n\t\treturn nil, status.ErrNotFound.WithMessage(err.Error())\n\t}\n\treturn &a, nil\n}", "func (c *ArticlesController) Article(ctx *app.ArticleArticlesContext) error {\n\t// ArticlesController_Article: start_implement\n\n\t// Put your logic here\n\tarticleTable := models.NewArticleDB(c.db)\n\tarticle, err := articleTable.OneGoaNewsAPIArticle(ctx.Context, ctx.ID)\n\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\n\t// ArticlesController_Article: end_implement\n\tres := &app.GoaNewsAPIArticle{}\n\tres = article\n\treturn ctx.OK(res)\n}", "func (sp *SessionProvider) FindOne(db, collection string, skip int, query interface{}, sort []string, into interface{}, flags int) error {\n\tsession, err := sp.GetSession()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer session.Close()\n\n\tq := session.DB(db).C(collection).Find(query).Sort(sort...).Skip(skip)\n\tq = ApplyFlags(q, session, flags)\n\treturn q.One(into)\n}", "func (service *RssService) Search(searchString string, isBookmark bool, feedID uint) *models.ArticlesJSON {\n\t// fixme\n\tbm := 0\n\n\tif isBookmark {\n\t\tbm = 1\n\t}\n\n\tvar articles []models.Articles\n\tquery := service.dbp().\n\t\tSelect(\"Id, Title, IsBookmark, IsRead\").\n\t\tWhere(\"IsBookmark = ? AND (Title LIKE ? OR Body LIKE ?)\", bm, \"%\"+searchString+\"%\", \"%\"+searchString+\"%\")\n\n\tif feedID != 0 {\n\t\tquery = query.Where(&models.Articles{Id: feedID})\n\t}\n\n\tquery.Find(&articles)\n\n\treturn &models.ArticlesJSON{Articles: articles}\n}", "func (d *Document) FindOneWithRevision(\n\tcontext.Context, string, int64,\n) (*flare.Document, error) {\n\treturn nil, errors.New(\"not implemented\")\n}", "func (r *AnswerRepo) Find(ID int) models.Answer {\n\tdb := database.Connect()\n\tdefer db.Close() // nolint: gas,errcheck\n\n\tvar answer models.Answer\n\tdb.First(&answer, ID)\n\treturn answer\n}", "func (d *Dao) FindOne(query interface{}) (*Item, error) {\n\titem := d.Create()\n\n\tcollection := d.Collection.Name\n\n\tdb := d.Database.Clone()\n\tdefer db.Close()\n\n\terr := db.C(collection).Find(query).One(item.Value)\n\n\tif mgo.ErrNotFound == err {\n\t\treturn nil, nil\n\t}\n\n\tif nil != err {\n\t\treturn nil, err\n\t}\n\n\titem.saved = true\n\titem.updated = true\n\treturn item, nil\n}", "func prepArticle(articleContent *goquery.Selection, articleTitle string) {\n\tif articleContent == nil {\n\t\treturn\n\t}\n\n\t// Check for data tables before we continue, to avoid removing items in\n\t// those tables, which will often be isolated even though they're\n\t// visually linked to other content-ful elements (text, images, etc.).\n\tmarkDataTables(articleContent)\n\n\t// Remove style attribute\n\tcleanStyle(articleContent)\n\n\t// Clean out junk from the article content\n\tcleanConditionally(articleContent, \"form\")\n\tcleanConditionally(articleContent, \"fieldset\")\n\tclean(articleContent, \"h1\")\n\tclean(articleContent, \"object\")\n\tclean(articleContent, \"embed\")\n\tclean(articleContent, \"footer\")\n\tclean(articleContent, \"link\")\n\n\t// Clean out elements have \"share\" in their id/class combinations from final top candidates,\n\t// which means we don't remove the top candidates even they have \"share\".\n\tarticleContent.Find(\"*\").Each(func(_ int, s *goquery.Selection) {\n\t\tid, _ := s.Attr(\"id\")\n\t\tclass, _ := s.Attr(\"class\")\n\t\tmatchString := class + \" \" + id\n\t\tif strings.Contains(matchString, \"share\") {\n\t\t\ts.Remove()\n\t\t}\n\t})\n\n\t// If there is only one h2 and its text content substantially equals article title,\n\t// they are probably using it as a header and not a subheader,\n\t// so remove it since we already extract the title separately.\n\th2s := articleContent.Find(\"h2\")\n\tif h2s.Length() == 1 {\n\t\th2 := h2s.First()\n\t\th2Text := normalizeText(h2.Text())\n\t\tlengthSimilarRate := float64(strLen(h2Text)-strLen(articleTitle)) /\n\t\t\tfloat64(strLen(articleTitle))\n\n\t\tif math.Abs(lengthSimilarRate) < 0.5 {\n\t\t\ttitlesMatch := false\n\t\t\tif lengthSimilarRate > 0 {\n\t\t\t\ttitlesMatch = strings.Contains(h2Text, articleTitle)\n\t\t\t} else {\n\t\t\t\ttitlesMatch = strings.Contains(articleTitle, h2Text)\n\t\t\t}\n\n\t\t\tif titlesMatch {\n\t\t\t\th2.Remove()\n\t\t\t}\n\t\t}\n\t}\n\n\tclean(articleContent, \"iframe\")\n\tclean(articleContent, \"input\")\n\tclean(articleContent, \"textarea\")\n\tclean(articleContent, \"select\")\n\tclean(articleContent, \"button\")\n\tcleanHeaders(articleContent)\n\n\t// Do these last as the previous stuff may have removed junk\n\t// that will affect these\n\tcleanConditionally(articleContent, \"table\")\n\tcleanConditionally(articleContent, \"ul\")\n\tcleanConditionally(articleContent, \"div\")\n\n\t// Remove extra paragraphs\n\t// At this point, nasty iframes have been removed, only remain embedded video ones.\n\tarticleContent.Find(\"p\").Each(func(_ int, p *goquery.Selection) {\n\t\timgCount := p.Find(\"img\").Length()\n\t\tembedCount := p.Find(\"embed\").Length()\n\t\tobjectCount := p.Find(\"object\").Length()\n\t\tiframeCount := p.Find(\"iframe\").Length()\n\t\ttotalCount := imgCount + embedCount + objectCount + iframeCount\n\n\t\tpText := normalizeText(p.Text())\n\t\tif totalCount == 0 && strLen(pText) == 0 {\n\t\t\tp.Remove()\n\t\t}\n\t})\n\n\tarticleContent.Find(\"br\").Each(func(_ int, br *goquery.Selection) {\n\t\tif br.Next().Is(\"p\") {\n\t\t\tbr.Remove()\n\t\t}\n\t})\n}", "func findOneObj(obj interface{}, find *bson.M, selection *bson.M, sort string, customProvider PriceRuleCustomProvider) (interface{}, error) {\n\tvar p *persistence.Persistor\n\tp = GetPersistorForObject(obj)\n\n\tif find == nil {\n\t\tfind = &bson.M{}\n\t}\n\tif selection == nil {\n\t\tselection = &bson.M{}\n\t}\n\n\tsession, collection := p.GetCollection()\n\tdefer session.Close()\n\n\tif sort != \"\" {\n\t\terr := collection.Find(find).Select(selection).Sort(sort).One(obj)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\terr := collection.Find(find).Select(selection).One(obj)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif customProvider != nil {\n\t\tvar err error\n\t\ttypedObject, err := mapDecodeObj(obj, customProvider)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tobj = typedObject\n\t}\n\n\tif obj == nil {\n\t\treturn nil, errors.New(string(shop_error.ErrorNotFound))\n\t}\n\treturn obj, nil\n}", "func FindArticles(ids ...int64) ([]Article, error) {\n\tif len(ids) == 0 {\n\t\tmsg := \"At least one or more ids needed\"\n\t\tlog.Println(msg)\n\t\treturn nil, errors.New(msg)\n\t}\n\t_articles := []Article{}\n\tidsHolder := strings.Repeat(\",?\", len(ids)-1)\n\tsql := DB.Rebind(fmt.Sprintf(`SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles WHERE articles.id IN (?%s)`, idsHolder))\n\tidsT := []interface{}{}\n\tfor _,id := range ids {\n\t\tidsT = append(idsT, interface{}(id))\n\t}\n\terr := DB.Select(&_articles, sql, idsT...)\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn _articles, nil\n}", "func (h *HeroRepo) FindOne(heroID int64, hero *models.Hero) error {\n\n\tquery := func(c *mgo.Collection) error {\n\t\terr := c.Find(bson.M{\"id\": heroID}).One(&hero)\n\t\tif err != nil {\n\t\t\tfmt.Println(\"query broke\", err)\n\n\t\t}\n\t\treturn err\n\t}\n\n\terr := mongo.WithCollection(\"heroes\", query)\n\tif err != nil {\n\t\tfmt.Println(\"can't find one -repo layer\")\n\t}\n\treturn err\n\n}", "func LastArticle() (*Article, error) {\n\t_article := Article{}\n\terr := DB.Get(&_article, DB.Rebind(`SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles ORDER BY articles.id DESC LIMIT 1`))\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn &_article, nil\n}", "func GetArticleDetail(query map[string]interface{}) (v *ArticleDetail, err error) {\n\to := orm.NewOrm()\n\t// query k=v\n\tqs := o.QueryTable(new(ArticleDetail))\n\tfor k, item := range query {\n\t\t// rewrite dot-notation to Object__Attribute\n\t\tk = strings.Replace(k, \".\", \"__\", -1)\n\t\tqs = qs.Filter(k, item)\n\t}\n\tv = &ArticleDetail{}\n\tif err = qs.RelatedSel().One(v); err == nil {\n\t\treturn v, nil\n\t}\n\treturn nil, err\n}", "func FindOne(name string) (string, string) {\n\n\tsvc := utils.OpenDynamoDb()\n\n\tume := models.Dog{Name: \"Ume\", Kind: \"Mix\"}\n\tselctUme := selectItem(ume)\n\n\tresult, err := selctUme(svc)\n\n\tif err != nil {\n\t\tfmt.Println(\"GetItem Error\", err)\n\t\treturn \"\", \"\"\n\t}\n\n\titem := formatToDog(result)\n\treturn item.Name, item.Kind\n}", "func (c WrapperCollection) FindOneAndDelete(ctx context.Context, filter interface{}, opts ...option.FindOneAndDeleteOptioner) DocumentResultLayer {\n\tvar docResult *WrapperDocumentResult\n\n\t// Recover on panic() from mongo driver.\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tdocResult = nil\n\t\t}\n\t}()\n\n\tdocResult = &WrapperDocumentResult{c.Collection.FindOneAndDelete(ctx, filter, opts...)}\n\treturn docResult\n}", "func findSubject(c *Container) string {\n\tif article1 := c.Article; article1 == nil {\n\t\tif c.Child == nil || c.Child.Article == nil || c.Child.Article.Subject == \"\" {\n\t\t\treturn \"\"\n\t\t}\n\t\treturn c.Child.Article.Subject\n\t} else {\n\t\treturn article1.Subject\n\t}\n\n\treturn \"\"\n}", "func (updateservice UpdateService) FindOneAndReplace(filter interface{}, update interface{}, coll string) (interface{}, error, bool) {\r\n\t//check Collection is exist\r\n\tif !checkCollectionExist(coll) {\r\n\t\treturn nil, nil, false\r\n\t}\r\n\r\n\t//create the context\r\n\texp := 5 * time.Second\r\n\tctx, cancel := context.WithTimeout(context.Background(), exp)\r\n\tdefer cancel()\r\n\r\n\t//select the collection\r\n\tcollection := Database.Collection(coll)\r\n\r\n\t//create an instance of an options and set the desired options\r\n\tupsert := true\r\n\tafter := options.After\r\n\topt := options.FindOneAndReplaceOptions{\r\n\t\tReturnDocument: &after,\r\n\t\tUpsert: &upsert,\r\n\t}\r\n\r\n\tlog.Printf(\"filter : %v\", filter)\r\n\tlog.Printf(\"update : %v\", update)\r\n\r\n\tresult := collection.FindOneAndReplace(ctx, filter, update, &opt)\r\n\tif result.Err() != nil {\r\n\t\treturn nil, result.Err(), true\r\n\t}\r\n\r\n\tid := filter.(map[string]interface{})[\"id\"]\r\n\r\n\tdoc := bson.M{}\r\n\tdecodeErr := result.Decode(&doc)\r\n\tlog.Printf(\"result : %v\", decodeErr)\r\n\r\n\treturn id, decodeErr, true\r\n}", "func FindSingle(single_id uuid.UUID) (single *Single, err error) {\n\tsingle = &Single{}\n\terr = DB.Find(single, single_id)\n\treturn\n}", "func findOneObj(obj interface{}, find *bson.M, selection *bson.M, sort string, customProvider PriceRuleCustomProvider) (interface{}, error) {\n\tvar p *persistence.Persistor\n\tp = GetPersistorForObject(obj)\n\n\tif find == nil {\n\t\tfind = &bson.M{}\n\t}\n\tif selection == nil {\n\t\tselection = &bson.M{}\n\t}\n\tif sort != \"\" {\n\t\terr := p.GetCollection().Find(find).Select(selection).Sort(sort).One(obj)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\terr := p.GetCollection().Find(find).Select(selection).One(obj)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif customProvider != nil {\n\t\tvar err error\n\t\ttypedObject, err := mapDecodeObj(obj, customProvider)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tobj = typedObject\n\t}\n\n\tif obj == nil {\n\t\treturn nil, errors.New(string(shop_error.ErrorNotFound))\n\t}\n\treturn obj, nil\n}", "func FindOne(collection string, query interface{},\n\tprojection interface{}, sort []string, out interface{}) error {\n\n\tsession, db, err := GetGlobalSessionFactory().GetSession()\n\tif err != nil {\n\t\tgrip.Errorf(\"error establishing db connection: %+v\", err)\n\t\treturn err\n\t}\n\tdefer session.Close()\n\n\tq := db.C(collection).Find(query).Select(projection)\n\tif len(sort) != 0 {\n\t\tq = q.Sort(sort...)\n\t}\n\treturn q.One(out)\n}", "func (s *Service) FindOne(identifier Identifier) (IdentifiableLimit, error) {\n\tresult, err := s.Find(identifier)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(result) == 0 {\n\t\treturn nil, ErrNotFound\n\t}\n\treturn result[0], nil\n}", "func NSFindOne(el *etree.Element, namespace, tag string) (*etree.Element, error) {\n\treturn NSFindOneCtx(NewDefaultNSContext(), el, namespace, tag)\n}", "func (r *ShowsService) One(traktID string) (show *Show, result *Result) {\n\turl, _ := showURL.Expand(M{\"traktID\":traktID})\n\tresult = r.client.get(url, &show)\n\treturn\n}", "func getArticles(p int) {\n\tdb, err := bolt.Open(\"../.db\", 0600, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdefer db.Close()\n\n\t//display 10 articles per page\n\tIdIndex := (p-1)*10 + 1\n\tvar articles ArticlesResponse\n\tvar article Article\n\terr = db.View(func(tx *bolt.Tx) error {\n\t\tb := tx.Bucket([]byte(\"Article\"))\n\t\tif b != nil {\n\t\t\tc := b.Cursor()\n\t\t\tk, v := c.Seek(itob(IdIndex))\n\t\t\tif k == nil {\n\t\t\t\tfmt.Println(\"Page is out of index\")\n\t\t\t\treturn errors.New(\"Page is out of index\")\n\t\t\t}\n\t\t\tkey := binary.BigEndian.Uint64(k)\n\t\t\tfmt.Print(key)\n\t\t\tif int(key) != IdIndex {\n\t\t\t\tfmt.Println(\"Page is out of index\")\n\t\t\t\treturn errors.New(\"Page is out of index\")\n\t\t\t}\n\t\t\tcount := 0\n\t\t\tvar ori_artc Article\n\t\t\tfor ; k != nil && count < 10; k, v = c.Next() {\n\t\t\t\terr = json.Unmarshal(v, &ori_artc)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tarticle.Id = ori_artc.Id\n\t\t\t\tarticle.Name = ori_artc.Name\n\t\t\t\tarticles.Articles = append(articles.Articles, article)\n\t\t\t\tcount = count + 1\n\t\t\t}\n\t\t\treturn nil\n\t\t} else {\n\t\t\treturn errors.New(\"Article Not Exists\")\n\t\t}\n\t})\n\tfor i := 0; i < len(articles.Articles); i++ {\n\t\tfmt.Println(articles.Articles[i])\n\t}\n}", "func (db *DB) FindOne(collection string, query bson.M, obj interface{}) error {\n\tsession := db.session.Copy()\n\tdefer session.Close()\n\tif err := session.DB(db.conf.Mongo.DatabaseName).C(collection).Find(query).One(obj); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func FirstArticles(n uint32) ([]Article, error) {\n\t_articles := []Article{}\n\tsql := fmt.Sprintf(\"SELECT COALESCE(articles.text, '') AS text, articles.id, articles.title, articles.created_at, articles.updated_at FROM articles ORDER BY articles.id ASC LIMIT %v\", n)\n\terr := DB.Select(&_articles, DB.Rebind(sql))\n\tif err != nil {\n\t\tlog.Printf(\"Error: %v\\n\", err)\n\t\treturn nil, err\n\t}\n\treturn _articles, nil\n}", "func (m MongoClient) FindOne(ctx context.Context, filter interface{}, data interface{}) error {\n\treturn m.Conn.FindOne(ctx, filter).Decode(data)\n}", "func (a *Artist) Find() error {\n\treturn DB.Where(&a).First(&a).Error\n}", "func (qf *QuickFind) Find(p int) (int, error) {\n\tif p >= len(qf.id) {\n\t\treturn 0, fmt.Errorf(\"Find(p) failed, p out of bounds: p=%d qf.count=%d\", p, qf.count)\n\t}\n\treturn qf.id[p], nil\n}", "func findMovie(c *gin.Context) {\n\t// Get model if exist\n\tvar movie Movie\n\tif err := DB.Where(\"id = ?\", c.Param(\"id\")).First(&movie).Error; err != nil {\n\t\tc.JSON(http.StatusBadRequest, gin.H{\"error\": \"Record not found!\"})\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"data\": movie})\n}", "func (model *Log) SingleFindFilter(filter interface{}) error {\n\treturn basemodel.SingleFindFilter(&model, filter)\n}", "func (d *DB) All() []Article {\n\tst, err := d.db.Prepare(\"select id,feed,title,content,published,link,read,display_name from articles where deleted = false order by id\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil\n\t}\n\tdefer st.Close()\n\n\trows, err := st.Query()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn nil\n\t}\n\tdefer rows.Close()\n\n\tvar (\n\t\tid int\n\t\ttitle string\n\t\tcontent string\n\t\tfeed string\n\t\tlink string\n\t\tread bool\n\t\tdisplay string\n\t\tpublished time.Time\n\t)\n\n\tarticles := []Article{}\n\n\tfor rows.Next() {\n\t\terr = rows.Scan(&id, &feed, &title, &content, &published, &link, &read, &display)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\n\t\t// Check if we should higlight it\n\t\tfields := strings.Fields(title)\n\t\thighlight := false\n\t\tfor _, f := range fields {\n\t\t\tfor _, h := range d.c.conf.Highlights {\n\t\t\t\tif strings.Contains(strings.ToLower(f), strings.ToLower(h)) {\n\t\t\t\t\thighlight = true\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif highlight {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tarticles = append(articles, Article{id: id, highlight: highlight, feed: feed, title: title, content: content, published: published, link: link, read: read, feedDisplay: display})\n\t}\n\treturn articles\n}", "func (r *readability) getArticleMetadata(doc *goquery.Document) Metadata {\n\tmetadata := Metadata{}\n\tmapAttribute := make(map[string]string)\n\n\tdoc.Find(\"meta\").Each(func(_ int, meta *goquery.Selection) {\n\t\tmetaName, _ := meta.Attr(\"name\")\n\t\tmetaProperty, _ := meta.Attr(\"property\")\n\t\tmetaContent, _ := meta.Attr(\"content\")\n\n\t\tmetaName = strings.TrimSpace(metaName)\n\t\tmetaProperty = strings.TrimSpace(metaProperty)\n\t\tmetaContent = strings.TrimSpace(metaContent)\n\n\t\t// Fetch author name\n\t\tif strings.Contains(metaName+metaProperty, \"author\") {\n\t\t\tmetadata.Author = metaContent\n\t\t\treturn\n\t\t}\n\n\t\t// Fetch description and title\n\t\tif metaName == \"title\" ||\n\t\t\tmetaName == \"description\" ||\n\t\t\tmetaName == \"twitter:title\" ||\n\t\t\tmetaName == \"twitter:image\" ||\n\t\t\tmetaName == \"twitter:description\" {\n\t\t\tif _, exist := mapAttribute[metaName]; !exist {\n\t\t\t\tmapAttribute[metaName] = metaContent\n\t\t\t}\n\t\t\treturn\n\t\t}\n\n\t\tif metaProperty == \"og:description\" ||\n\t\t\tmetaProperty == \"og:image\" ||\n\t\t\tmetaProperty == \"og:title\" {\n\t\t\tif _, exist := mapAttribute[metaProperty]; !exist {\n\t\t\t\tmapAttribute[metaProperty] = metaContent\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t})\n\n\t// Set final image\n\tif _, exist := mapAttribute[\"og:image\"]; exist {\n\t\tmetadata.Image = mapAttribute[\"og:image\"]\n\t} else if _, exist := mapAttribute[\"twitter:image\"]; exist {\n\t\tmetadata.Image = mapAttribute[\"twitter:image\"]\n\t}\n\n\tif metadata.Image != \"\" && strings.HasPrefix(metadata.Image, \"//\") {\n\t\tmetadata.Image = \"http:\" + metadata.Image\n\t}\n\n\t// Set final description\n\tif _, exist := mapAttribute[\"description\"]; exist {\n\t\tmetadata.Excerpt = mapAttribute[\"description\"]\n\t} else if _, exist := mapAttribute[\"og:description\"]; exist {\n\t\tmetadata.Excerpt = mapAttribute[\"og:description\"]\n\t} else if _, exist := mapAttribute[\"twitter:description\"]; exist {\n\t\tmetadata.Excerpt = mapAttribute[\"twitter:description\"]\n\t}\n\n\t// Set final title\n\tmetadata.Title = r.getArticleTitle(doc)\n\tif metadata.Title == \"\" {\n\t\tif _, exist := mapAttribute[\"og:title\"]; exist {\n\t\t\tmetadata.Title = mapAttribute[\"og:title\"]\n\t\t} else if _, exist := mapAttribute[\"twitter:title\"]; exist {\n\t\t\tmetadata.Title = mapAttribute[\"twitter:title\"]\n\t\t}\n\t}\n\n\treturn metadata\n}", "func (m *MongoDB) FindOne(\n\tname string,\n\tfilter interface{},\n\tdoc interface{},\n) error {\n\tcoll, err := m.selCollection(name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn coll.FindOne(m.ctx, filter).Decode(doc)\n}", "func createNewArticle(title, content string) (*article, error) {\n\t// Set the ID of a new article to one more than the number of articles\n\ta := article{Title: title, Content: content}\n\taddarticle(a)\n\treturn &a, nil\n}" ]
[ "0.6177107", "0.6014521", "0.60014904", "0.5923695", "0.5902255", "0.57037485", "0.56767607", "0.5588242", "0.55838203", "0.5573647", "0.5543387", "0.5540758", "0.5499482", "0.54978687", "0.5468035", "0.54162866", "0.5413777", "0.5383615", "0.53822386", "0.5378711", "0.5344943", "0.5340329", "0.53358537", "0.5325252", "0.53116953", "0.53001136", "0.52775174", "0.5251422", "0.5236752", "0.52355367", "0.52175295", "0.5203725", "0.51953197", "0.51799685", "0.5166625", "0.5163181", "0.51336974", "0.5126014", "0.5119769", "0.5112479", "0.5106749", "0.5096529", "0.5094087", "0.5084233", "0.5083453", "0.5083105", "0.50762147", "0.50517863", "0.50446874", "0.50181824", "0.50172913", "0.50144607", "0.5012992", "0.5012042", "0.50061756", "0.49996436", "0.49977252", "0.49959064", "0.49937043", "0.4982795", "0.49789608", "0.4973718", "0.4972748", "0.4968453", "0.4964825", "0.49604404", "0.49593434", "0.49500296", "0.4949026", "0.49479744", "0.49390852", "0.49271765", "0.4924131", "0.49233806", "0.49117523", "0.49100643", "0.49068898", "0.49043944", "0.49039122", "0.49031973", "0.49025106", "0.48931667", "0.4890871", "0.48795536", "0.48720822", "0.48690623", "0.486308", "0.48474032", "0.48436335", "0.4839975", "0.4838488", "0.48374176", "0.4830919", "0.4828016", "0.48249808", "0.482288", "0.4817338", "0.4814846", "0.48139617", "0.480852" ]
0.6327357
0
NewFormatter returns a new instance of formatter
func NewFormatter(t Type, fileName string, logger *log.Logger) Formatter { switch t { case StdoutType: return newStdoutFormatter(logger) case FileType: return newFileFormatter(fileName, logger) case JSONType: return newJSONFormatter(fileName, logger) case YamlType: return newYamlFormatter(fileName, logger) case SkopeoType: return newSkopeoFormatter(fileName, logger) default: return newStdoutFormatter(logger) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New() *Formatter {\n\treturn &Formatter{\n\t\tformats: map[string]Format{},\n\t}\n}", "func NewFormatter() *formatter {\n\treturn &formatter{}\n}", "func NewFormatter(providers map[string]Provider) Formatter {\n\treturn Formatter{\n\t\tproviders: providers,\n\t}\n}", "func NewFormatter() (*Formatter, error) {\n\tlexer, err := NewLexer()\n\n\tif err != nil {\n\t\treturn &Formatter{}, err\n\t}\n\n\treturn &Formatter{&lexer}, nil\n}", "func New() *Formatter {\n\tf := &Formatter{\n\t\tLabels: map[string]string{},\n\t}\n\treturn f\n}", "func NewFormatter(t string) Formatter {\n\tswitch t {\n\tcase \"stdout\":\n\t\treturn newSTDOUTFormatter()\n\tcase \"plain\":\n\t\treturn newPlainFormatter()\n\tcase \"json\":\n\t\treturn newJSONFormatter()\n\tcase \"yaml\":\n\t\treturn newYamlFormatter()\n\tdefault:\n\t\treturn newSTDOUTFormatter()\n\t}\n}", "func New(options ...Option) *Formatter {\n\tf := &Formatter{\n\t\ttabStr: []byte(\" \"),\n\t\tnewline: []byte(\"\\n\"),\n\t\ttextFormatters: func(tag Tag) TextFormatter {\n\t\t\treturn nil\n\t\t},\n\t}\n\tfor _, option := range options {\n\t\toption(f)\n\t}\n\treturn f\n}", "func NewFormatter(format string) Formatter {\n\tif fun, ok := formatMap[format]; ok {\n\t\treturn fun()\n\t}\n\treturn nil\n}", "func NewFormatter() *Formatter {\n\treturn &Formatter{\n\t\tPassColor: color.New(color.FgGreen),\n\t\tFailColor: color.New(color.FgRed),\n\t\tStringMaxLength: 0,\n\t\tDisabledColor: false,\n\t\tIndent: 2,\n\t}\n}", "func newLogFormatter(globals server.Globals) handlers.LogFormatter {\n\treturn logFormatter{\n\t\tlogger: globals.Logger,\n\t}.format\n}", "func NewFormatter(withoutUnit bool, duration time.Duration) *Formatter {\n\treturn &Formatter{withoutUnit, duration}\n}", "func NewFormatter(w io.Writer, fmt string) (*Formatter, error) {\n\tif fmt == \"\" {\n\t\tfmt = DefaultTemplate\n\t}\n\ttmpl, err := template.New(\"out\").Parse(fmt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Formatter{\n\t\toutput: w,\n\t\ttemplate: tmpl,\n\t\tColorize: false,\n\t\tShowFields: true,\n\t\tMaxFieldLength: 30,\n\t\tShowPrefix: true,\n\t\tShowSuffix: true,\n\t\tExcludeFields: defaultExcludes,\n\t}, nil\n}", "func NewFormatter(format string, colored bool) (*Formatter, error) {\n\tfm := new(Formatter)\n\tfm.colored = colored\n\tif err := fm.SetFormat(format); err != nil {\n\t\treturn nil, err\n\t}\n\treturn fm, nil\n}", "func NewFormatter(printID bool) *Formatter {\n\tf := &Formatter{\n\t\tprintID: printID,\n\t\tw: uilive.New(),\n\t\ts: NewSpinner(),\n\t}\n\treturn f\n}", "func NewMockFormatter(ctrl *gomock.Controller) *MockFormatter {\n\tmock := &MockFormatter{ctrl: ctrl}\n\tmock.recorder = &MockFormatterMockRecorder{mock}\n\treturn mock\n}", "func NewFormatter() *Formatter {\n\treturn &Formatter{\n\t\tKeyColor: ansi.NewColor(ansi.FgBlue, ansi.Bold),\n\t\tStringColor: nil,\n\t\tBoolColor: ansi.NewColor(ansi.FgYellow, ansi.Bold),\n\t\tNumberColor: ansi.NewColor(ansi.FgMagenta),\n\t\tNullColor: ansi.NewColor(ansi.Bold),\n\t\tDisableColor: false,\n\t\tIndent: 2,\n\t}\n}", "func Formatter() *formatter {\n\treturn &formatter{}\n}", "func newInfluxFormatter() influxFormatter {\n\treturn influxFormatter{}\n}", "func NewFormatter() *PlainTextFormatter {\n\treturn &PlainTextFormatter{}\n}", "func NewFormat(ctx context.Context, client *github.Client, debug bool) *Format {\n\treturn &Format{ctx: ctx, client: client, debug: debug}\n}", "func NewFormat(source string, quiet bool) formatter.Format {\n\tswitch source {\n\tcase formatter.TableFormatKey:\n\t\tif quiet {\n\t\t\treturn formatter.DefaultQuietFormat\n\t\t}\n\t\treturn defaultNetworkTableFormat\n\tcase formatter.RawFormatKey:\n\t\tif quiet {\n\t\t\treturn `network_id: {{.ID}}`\n\t\t}\n\t\treturn `network_id: {{.ID}}\\nname: {{.Name}}\\ndriver: {{.Driver}}\\nscope: {{.Scope}}\\n`\n\t}\n\treturn formatter.Format(source)\n}", "func newFormat(format string) eval.Format {\n\treturn parseFormat(format, NO_STRING, NO_STRING, nil)\n}", "func NewFormat(prefix, indent, quote, lineEnd string) Format {\n\tif quote == \"\" {\n\t\tquote = `\"`\n\t}\n\tif lineEnd == \"\" {\n\t\tlineEnd = \"\\n\"\n\t}\n\treturn Format{\n\t\tPrefix: prefix,\n\t\tIndent: indent,\n\t\tIsBreaking: prefix != \"\" || indent != \"\",\n\t\tQuote: quote,\n\t\tLineEnd: lineEnd,\n\t}\n}", "func NewDiffFormat(source string) formatter.Format {\n\tswitch source {\n\tcase formatter.TableFormatKey:\n\t\treturn defaultDiffTableFormat\n\t}\n\treturn formatter.Format(source)\n}", "func NewJSON() Formatter {\n\treturn &jsonFormat{\n\t\tOutput: os.Stdout,\n\t}\n}", "func NewStandardFormatter() (standardFormatter *StandardFormatter) {\n\treturn &StandardFormatter{\n\t\tappendNewLine: true,\n\t\tdateTimeLayout: \"2006-01-02 15:04:05\",\n\t\tlayout: \"%(dateTime) [%(logLevel)] (%(pid)) %(program) %(loggerName) %(fileName) %(lineNum) %(message)\",\n\t\tmutex: new(sync.RWMutex),\n\t}\n}", "func NewFormatterFilter() Filter {\n\treturn &FormatterFilter{}\n}", "func NewParser(f string) (*LogHandler, error) {\n\tif !ValidFormat(f) {\n\t\treturn nil, fmt.Errorf(\"%s is not a valid parser format\", f)\n\t}\n\n\tvar p = &LogHandler{}\n\n\tif f == RFC5424Name || f == RFC5424Standard {\n\t\tp.Parser = &parser.RFC5424{}\n\t\tp.Fmt = RFC5424Standard\n\t} else if f == WatchguardName || f == WatchguardFirebox {\n\t\tp.Parser = &parser.Watchguard{}\n\t\tp.Fmt = WatchguardFirebox\n\t}\n\n\tlog.Printf(\"input format parser created for %s\", f)\n\tp.Stats = stats.Add\n\tp.Parser.Init()\n\treturn p, nil\n}", "func NewFormatterUtils() *FormatterUtils {\n return &FormatterUtils{}\n}", "func wrapFormatter(module string, formatter *logrus.TextFormatter) *moduleFormatter {\n\treturn &moduleFormatter{\n\t\tmodule: module,\n\t\tTextFormatter: formatter,\n\t}\n}", "func NewFormat(chans int, freq freq.T, sc sample.Codec) *Format {\n\treturn &Format{\n\t\tchannels: chans,\n\t\tfreq: freq,\n\t\tCodec: sc}\n}", "func (b *Basic) Formatter() Formatter {\n\treturn b.formatter\n}", "func NewConsoleFormatter(color bool, opts ...OptionFormatter) *ConsoleFormatter {\n\tf := &ConsoleFormatter{\n\t\tConsoleTimestampFormat: \"2018-01-02/15:04:05\",\n\t\tConsoleLevelFormat: \"%.1s\",\n\t\tConsoleLogFormat: \"%localtime% %LEVEL% %msg% %fields%\",\n\t\tConsoleCallerFormat: \"%file%:%line% %fun%()\",\n\t\tConsoleFieldSep: \", \",\n\t\tConsoleFieldKVSep: \":\",\n\t\tConsoleFieldsWrap: \" [%s]\", // \"「%s」\"\n\t\tConsoleLogColor: color,\n\t}\n\t// call option functions on instance to set options on it\n\tfor _, opt := range opts {\n\t\topt(f)\n\t}\n\treturn f\n}", "func NewFLATE() *FLATEFormat { return &FLATEFormat{} }", "func New(name string) (Format, error) {\n\tif name == \"\" {\n\t\treturn nil, NewErrFormatNameIsEmpty()\n\t}\n\n\tif synonym, ok := synonyms[name]; ok {\n\t\tname = synonym\n\t}\n\n\tswitch name {\n\tcase JSON:\n\t\treturn NewJSON(), nil\n\tcase YAML:\n\t\treturn NewYAML(), nil\n\tcase TOML:\n\t\treturn NewTOML(), nil\n\tcase HEX:\n\t\treturn NewHEX(), nil\n\tcase BASE64:\n\t\treturn NewBASE64(), nil\n\tcase QUERY:\n\t\treturn NewQUERY(), nil\n\tcase FLATE:\n\t\treturn NewFLATE(), nil\n\tdefault:\n\t\treturn nil, NewErrNotSupported(name)\n\t}\n}", "func NewMockIFormatter(ctrl *gomock.Controller) *MockIFormatter {\n\tmock := &MockIFormatter{ctrl: ctrl}\n\tmock.recorder = &MockIFormatterMockRecorder{mock}\n\treturn mock\n}", "func CreateCryptoFormatter(format string) *AcraCryptoFormatter {\n\tvar formatter *AcraCryptoFormatter\n\tswitch strings.ToLower(format) {\n\tcase JSONFormatString:\n\t\tformatter = NewCryptoFormatter(JSONFormatter())\n\tcase CefFormatString:\n\t\tformatter = NewCryptoFormatter(CEFFormatter())\n\tdefault:\n\t\tformatter = NewCryptoFormatter(TextFormatter())\n\t}\n\treturn formatter\n}", "func NewMySQLFormatter(template bool) *MySQLFormatter {\n\treturn &MySQLFormatter{Template: template}\n}", "func Newf(format string, args ...interface{}) error {\n\terr := &wrappedError{\n\t\ttraceMessage: fmt.Sprintf(format, args...),\n\t}\n\tif IncludeCaller {\n\t\terr.addCaller(1)\n\t}\n\treturn err\n}", "func Newf(msgFormat string, v ...any) Entry {\n\treturn makeEntry(msgFormat, v...)\n}", "func NewJSONFormatter(keys map[string]string, full bool) (Formatter, error) {\n\tif len(keys) > 0 {\n\t\tstructure := true\n\t\tmapping := map[string]string{\n\t\t\t\"name\": \"name\", \"time\": \"time\", \"level\": \"level\", \"message\": \"message\",\n\t\t\t\"fields\": \"fields\", \"caller\": \"caller\", \"stack\": \"stack\",\n\t\t}\n\t\tfor key, value := range keys {\n\t\t\tif mapping[key] == \"\" {\n\t\t\t\t// We require that the key-name map must be pure.\n\t\t\t\treturn nil, fmt.Errorf(\"invalid json formatter key %q\", key)\n\t\t\t}\n\t\t\t// We ignore the case where all fields are mapped as empty, which is more practical.\n\t\t\tif value != \"\" && mapping[key] != value {\n\t\t\t\tstructure = false\n\t\t\t\tmapping[key] = value\n\t\t\t}\n\t\t}\n\t\t// when the json field cannot be predicted in advance, we use map to package the log data.\n\t\t// is there a better solution to improve the efficiency of json serialization?\n\t\tif !structure {\n\t\t\treturn NewJSONFormatterFromPool(newJSONFormatterMapPool(full, mapping)), nil\n\t\t}\n\t}\n\t// In most cases, the performance of json serialization of structure is higher than\n\t// that of json serialization of map. When the json field name has not changed, we\n\t// try to use structure for json serialization.\n\treturn NewJSONFormatterFromPool(newJSONFormatterObjectPool(full)), nil\n}", "func NewTimeFormatter(timestamps bool) TimeFormatter {\n\ttimeFormatter := timeFormatter(timestamps)\n\treturn &timeFormatter\n}", "func New(target io.Writer, prefix string) Structured {\n\t// wrapedy wrap!\n\treturn &basicStructured{&defaultLogger{log.New(target, prefix, log.LstdFlags)}}\n}", "func Newf(classFormat string, args ...interface{}) *Ex {\n\treturn &Ex{class: fmt.Sprintf(classFormat, args...), stack: callers()}\n}", "func NewConsoleFormatter() Formatter {\n\treturn new(consoleFormatter)\n}", "func WithFormatter(v logrus.Formatter) Option {\n\treturn formatterOption{f: v}\n}", "func NewHumanizeFormatter() *Formatter {\n\treturn &Formatter{\n\t\tCompact: false,\n\t\tDateTimeFormat: \"2006-01-02T15:04:05\",\n\t\tFallback: &logrus.JSONFormatter{},\n\t}\n}", "func (f *Formattable) Formatter() Formatter {\n\tif f.formatter == nil {\n\t\tf.formatter = NewTextFormatter()\n\t}\n\treturn f.formatter\n}", "func NewCliOpFormatter(workingDirPath, dataDirPath string) CliOpFormatter {\n\treturn CliOpFormatter{workingDirPath, dataDirPath}\n}", "func NewJsonFormatter() Formatter {\n\treturn JsonFormatter{}\n}", "func Newf(tag Tag, format string, insert ...interface{}) Gerror {\n\treturn New(tag, fmt.Sprintf(format, insert...))\n}", "func (f *FakeFormatter) Format(b *bytes.Buffer, lvl golog.Level, ctx golog.Context, msg string, trace []byte) *bytes.Buffer {\n\tb.WriteString(msg)\n\treturn b\n}", "func newPrinter(format OutputFormat, writerType WriterType) *printer {\n\treturn &printer{Formatter: NewFormatter(format, writerType), Encoder: NewEncoder()}\n}", "func New(out io.Writer, format string) Printer {\n\tfn := printTable\n\tif format == \"csv\" {\n\t\tfn = printCSV\n\t}\n\n\treturn &printer{\n\t\tout: out,\n\t\tfn: fn,\n\t}\n}", "func New(f *os.File, level gol.Level) logging.Logger {\n\t// Leveled formatted file backend.\n\tbackend := gol.AddModuleLevel(\n\t\tgol.NewBackendFormatter(\n\t\t\tgol.NewLogBackend(f, \"\", 0),\n\t\t\tgol.MustStringFormatter(fmt)))\n\tbackend.SetLevel(level, \"\")\n\tlogger := gol.MustGetLogger(\"\")\n\tlogger.SetBackend(backend)\n\treturn Wrap(logger)\n}", "func (f *Formatter) Format(dst io.Writer, src io.Reader) error {\n\tp := newParser(src, f.tabStr)\n\n\ttokens, err := p.parse()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\titer := &tokenIterator{\n\t\ttokens: tokens,\n\t\tpos: -1,\n\t}\n\n\tw := &writer{\n\t\tdst: dst,\n\t\tf: f,\n\t\titer: iter,\n\t\tenableDebug: false,\n\t}\n\n\tvar formatText TextFormatter = nil\n\tvar inPre bool\n\n\tfor {\n\t\tcurr := iter.Next()\n\t\tif curr == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif inPre && !curr.inPre {\n\t\t\tw.write(curr.raw)\n\t\t\tcontinue\n\t\t}\n\n\t\tprev := iter.Prev()\n\t\tnext := iter.Peek()\n\n\t\tif curr.text.isWhitespaceOnly {\n\t\t\tif prev == nil && leadingNewlineRe.Match(curr.raw) {\n\t\t\t\t// Preserve one leading newline.\n\t\t\t\tw.newline()\n\t\t\t}\n\n\t\t\tif next == nil && trailingNewlineRe.Match(curr.raw) {\n\t\t\t\t// Preserve one trailing newline.\n\t\t\t\tw.newline()\n\t\t\t}\n\n\t\t\tif prev == nil || !prev.inPre {\n\t\t\t\t// Nothing more to do.\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\tvar newlineAttribute bool\n\t\tif curr.typ != html.TextToken && f.newlineAttributePlaceholder != \"\" {\n\t\t\tnewlineAttribute = !curr.tag.Attributes.ByKey(f.newlineAttributePlaceholder).IsZero()\n\t\t\tif newlineAttribute && !curr.isVoid() {\n\t\t\t\treturn errors.New(\"newline attributes is for void attributes only\")\n\t\t\t}\n\t\t}\n\n\t\tif newlineAttribute {\n\t\t\t// Insert newline only.\n\t\t\tw.newlineForced()\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch curr.typ {\n\t\tcase html.StartTagToken:\n\t\t\tif curr.inPre {\n\t\t\t\tinPre = true\n\t\t\t\tw.write(curr.raw)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// A text formatter for e.g. JavaScript script tags currently assumes\n\t\t\t// a single wrapped text element and any whitespace handling is\n\t\t\t// delegated to the custom text formatter.\n\t\t\tformatText = f.textFormatters(curr.tag)\n\n\t\t\tvar needsNewlineAppended bool\n\n\t\t\tif formatText == nil {\n\t\t\t\tneedsNewlineAppended = curr.needsNewlineAppended()\n\t\t\t\tif needsNewlineAppended {\n\t\t\t\t\tcurr.indented = true\n\t\t\t\t\tw.depth++\n\t\t\t\t} else if prev != nil && next != nil && curr.isVoid() {\n\t\t\t\t\tif w.newline() {\n\t\t\t\t\t\tw.tab()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tw.write(curr.raw)\n\n\t\t\tif formatText == nil {\n\t\t\t\tif needsNewlineAppended || (prev != nil && next != nil && curr.isVoid()) {\n\t\t\t\t\tif w.newline() {\n\t\t\t\t\t\tw.tab()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase html.SelfClosingTagToken, html.CommentToken, html.DoctypeToken:\n\t\t\tw.write(curr.raw)\n\t\t\tif prev == nil && next != nil {\n\t\t\t\tw.newline()\n\t\t\t}\n\t\tcase html.EndTagToken:\n\t\t\tif curr.inPre {\n\t\t\t\tinPre = false\n\t\t\t\tw.write(curr.raw)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif formatText == nil {\n\t\t\t\tif curr.isStartIndented() {\n\t\t\t\t\tn := w.newline()\n\t\t\t\t\tw.depth--\n\t\t\t\t\tif w.depth < 0 {\n\t\t\t\t\t\tw.depth = 0\n\t\t\t\t\t}\n\t\t\t\t\tif n {\n\t\t\t\t\t\tw.tab()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tw.write(curr.raw)\n\n\t\t\tif next != nil && !next.isInline() {\n\t\t\t\tnextStart := iter.PeekStart()\n\t\t\t\tif nextStart != nil && curr.depth == nextStart.depth {\n\t\t\t\t\tif w.newline() {\n\t\t\t\t\t\tw.tab()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase html.TextToken:\n\t\t\tif prev != nil && (prev.typ == html.EndTagToken || prev.isVoid()) && prev.isBlock() {\n\t\t\t\tif w.newline() {\n\t\t\t\t\tw.tab()\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// Preserve one leading newline.\n\t\t\tif prev == nil && curr.text.hadLeadingNewline {\n\t\t\t\tif w.newline() {\n\t\t\t\t\tw.tab()\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif formatText != nil {\n\t\t\t\tw.write(formatText(curr.raw, w.depth))\n\t\t\t} else {\n\t\t\t\tw.handleTextToken(prev, curr, next)\n\t\t\t}\n\n\t\t\t// Preserve one trailing newline.\n\t\t\tif next != nil && !next.isInline() && curr.text.hadTrailingNewline && next.depth == curr.depth {\n\t\t\t\tif w.newline() {\n\t\t\t\t\tw.tab()\n\t\t\t\t}\n\t\t\t}\n\t\tdefault:\n\t\t\tpanic(\"Unhandled token\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func DefaultFormatter() logrus.Formatter {\n\treturn textFormatterInstance\n}", "func newParser(br string) (*parser, []string, error) {\n\tloc, err := bugreportutils.TimeZone(br)\n\tif err != nil {\n\t\treturn nil, []string{}, err\n\t}\n\tpm, warnings := bugreportutils.ExtractPIDMappings(br)\n\t// Extract the year and month from the bugreport dumpstate line.\n\td, err := bugreportutils.DumpState(br)\n\tif err != nil {\n\t\treturn nil, warnings, fmt.Errorf(\"could not find dumpstate information in the bugreport: %v\", err)\n\t}\n\tbuf := new(bytes.Buffer)\n\treturn &parser{\n\t\treferenceYear: d.Year(),\n\t\treferenceMonth: d.Month(),\n\t\tloc: loc,\n\t\tbuf: buf,\n\t\tcsvState: csv.NewState(buf, true),\n\t\tpidMappings: pm,\n\t}, warnings, nil\n}", "func Newf(format string, args ...interface{}) error {\n\terr := &wrap{\n\t\tprevious: fmt.Errorf(format, args...),\n\t}\n\terr.setLocation(1)\n\treturn err\n}", "func BuildFormatterFromConfig(ca *config.Accessor) (logging.StringFormatter, error) {\n\n\tvar mode string\n\tvar err error\n\n\tentryPath := \"LogWriting.Format.Entry\"\n\n\tif mode, err = ca.StringVal(entryPath); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif mode == textEntryMode {\n\n\t\tlmf := new(logging.LogMessageFormatter)\n\n\t\tif err := ca.Populate(\"LogWriting.Format\", lmf); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif lmf.PrefixFormat == \"\" && lmf.PrefixPreset == \"\" {\n\t\t\tlmf.PrefixPreset = logging.FrameworkPresetPrefix\n\t\t}\n\n\t\treturn lmf, lmf.Init()\n\t} else if mode == jsonEntryMode {\n\n\t\tjmf := new(logging.JSONLogFormatter)\n\n\t\tcfg := new(logging.JSONConfig)\n\n\t\tca.Populate(\"LogWriting.Format.JSON\", cfg)\n\t\tjmf.Config = cfg\n\n\t\tcfg.UTC, _ = ca.BoolVal(\"LogWriting.Format.UtcTimes\")\n\n\t\tcfg.ParsedFields = logging.ConvertFields(cfg.Fields)\n\n\t\tif err := logging.ValidateJSONFields(cfg.ParsedFields); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tif mb, err := logging.CreateMapBuilder(cfg); err == nil {\n\t\t\tjmf.MapBuilder = mb\n\t\t} else {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn jmf, nil\n\t}\n\n\treturn nil, fmt.Errorf(\"%s is a not a supported value for %s. Should be %s or %s\", mode, entryPath, textEntryMode, jsonEntryMode)\n\n}", "func New(output *os.File, NoColors ...bool) *Tiny {\n\tnocolors := false\n\tif len(NoColors) > 0 {\n\t\tnocolors = NoColors[0]\n\t}\n\tl := logrus.New()\n\tl.SetLevel(logrus.DebugLevel)\n\tlog.SetOutput(output)\n\tl.SetFormatter(&f.Formatter{\n\t\tNoColors: nocolors,\n\t\tHideKeys: true,\n\t\tFieldsOrder: []string{\"component\", \"category\"},\n\t})\n\treturn &Tiny{l}\n}", "func newPrettyStdout(w io.Writer) *prettyStdout {\n\treturn &prettyStdout{\n\t\twriter: w,\n\t\tbuffer: bytes.NewBuffer([]byte{}),\n\t}\n}", "func Format(name, description string, f FormatterFunc) {\n\tformatters = append(formatters, &registeredFormatter{\n\t\tname: name,\n\t\tfmt: f,\n\t\tdescription: description,\n\t})\n}", "func GetFormatter(format LogFormat) logrus.Formatter {\n\tswitch format {\n\tcase LogFormatText:\n\t\treturn &logrus.TextFormatter{\n\t\t\tDisableTimestamp: true,\n\t\t\tDisableColors: true,\n\t\t}\n\tcase LogFormatJSON:\n\t\treturn &logrus.JSONFormatter{\n\t\t\tDisableTimestamp: true,\n\t\t}\n\tcase LogFormatJSONTimestamp:\n\t\treturn &logrus.JSONFormatter{\n\t\t\tDisableTimestamp: false,\n\t\t\tTimestampFormat: time.RFC3339Nano,\n\t\t}\n\t}\n\n\treturn nil\n}", "func NewCSVFormatter(w io.Writer, fields []string, headers []string) Formatter {\n\tif headers == nil {\n\t\theaders = fields\n\t}\n\treturn &csvFormatter{w: csv.NewWriter(w), fields: fields, headers: headers}\n}", "func New(uf undo.Factory) TextStorer {\n\treturn &Store{undoFac: uf, delim: \"\\n\"}\n}", "func NewCryptoFormatter(formatter Formatter) *AcraCryptoFormatter {\n\treturn &AcraCryptoFormatter{Formatter: formatter}\n}", "func New(format string, args ...interface{}) error {\n\treturn Error(fmt.Sprintf(format, args...))\n}", "func newParser(invocation string, verbose bool) *parser {\n\tp := &parser{\n\t\tScanner: &scanner.Scanner{},\n\t\toriginal: invocation,\n\t\tverbose: verbose,\n\t\terrors: make([]error, 0),\n\t}\n\tp.Init(strings.NewReader(invocation))\n\tp.Error = func(s *scanner.Scanner, msg string) {\n\t\tp.errors = append(p.errors, p.error(s.Position, msg))\n\t}\n\treturn p\n}", "func New(w io.Writer, template string) (*Logger, error) {\n\tformatters, isTimeRequired, err := compileFormat(template)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// Create a dummy event to see how long the log line is with the provided\n\t// template.\n\tbuf := make([]byte, 0, 64)\n\tvar e event\n\tfor _, formatter := range formatters {\n\t\tformatter(&e, &buf)\n\t}\n\tmin := len(buf) + 64\n\tif min < 128 {\n\t\tmin = 128\n\t}\n\tparent := &base{\n\t\tc: min,\n\t\tformatters: formatters,\n\t\tisTimeRequired: isTimeRequired,\n\t\tw: w,\n\t}\n\treturn &Logger{parent: parent, level: Warning}, nil\n}", "func newLogrus(level string, formatter string, output io.Writer) *logrus.Logger {\n\tl, err := logrus.ParseLevel(level)\n\tif err != nil {\n\t\tfmt.Printf(\"Bad level: %v, set it to 'debug'\", level)\n\t\tl = logrus.DebugLevel\n\t}\n\tlogger := &logrus.Logger{\n\t\tOut: output,\n\t\tHooks: make(logrus.LevelHooks),\n\t\tLevel: l,\n\t}\n\tswitch formatter {\n\tcase \"json\":\n\t\tlogger.Formatter = &logrus.JSONFormatter{TimestampFormat: TimestampFormat}\n\tcase \"text\":\n\t\tfallthrough\n\tdefault:\n\t\tlogger.Formatter = &logrus.TextFormatter{DisableColors: true,\n\t\t\tDisableSorting: false, TimestampFormat: TimestampFormat}\n\t}\n\treturn logger\n}", "func New(filename string, prefix string, flag int) *Logger {\n\treturn NewEx(filename, prefix, flag, SPLIT_FILE_SIZE, TOTAL_ROTATE_SPLIT)\n}", "func New(level string, writer string, prettyprint string) Logger {\n\tvar lg Logger\n\tlg.level = stringToLevel()[level]\n\tlg.logger = json.NewEncoder(stringToWriter(writer))\n\tif prettyprint == \"true\" {\n\t\tlg.logger.SetIndent(\"\", \" \")\n\t}\n\n\tvar process = strings.Split(os.Args[0], \"/\")\n\tlg.json.Process = process[len(process)-1]\n\n\treturn lg\n}", "func New(f io.RuneReader, init StateFn) Interface {\n\t// add line 1 to file\n\t// f.AddLine(0, 1)\n\treturn &Lexer{\n\t\tf: f,\n\t\ti: init,\n\t\t// initial q size must be an exponent of 2\n\t\tq: &queue{items: make([]Item, 2)},\n\t}\n}", "func New(message string) (*MessageFormat, error) {\n\tp := &parser{\n\t\tinput: []rune(message),\n\t}\n\tif err := p.parse(); err != nil {\n\t\treturn nil, fmt.Errorf(\"messageformat: cannot parse message: %s\", err)\n\t}\n\n\treturn &MessageFormat{p.blocks}, nil\n}", "func WithFormatter(formatter Formatter) Option {\n\treturn option{\n\t\ttable: func(enc *TableEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\texpanded: func(enc *ExpandedEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tjson: func(enc *JSONEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tunaligned: func(enc *UnalignedEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\ttemplate: func(enc *TemplateEncoder) error {\n\t\t\tenc.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t\tcrosstab: func(view *CrosstabView) error {\n\t\t\tview.formatter = formatter\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func newExporter(w io.Writer) (trace.SpanExporter, error) {\n\treturn stdouttrace.New(\n\t\tstdouttrace.WithWriter(w),\n\t\t// Use human-readable output.\n\t\tstdouttrace.WithPrettyPrint(),\n\t\t// Do not print timestamps for the demo.\n\t\tstdouttrace.WithoutTimestamps(),\n\t)\n}", "func New() *Parser {\n\treturn &Parser{\n\t\tdelimiter: \".\",\n\t}\n}", "func New(f string, x ...interface{}) error {\n\treturn &prefixError{s: format(f, x...)}\n}", "func NewPrinter(src io.Reader, format Formatter, c transport.Codec) Printer {\n\tif c == nil {\n\t\tc = codec.Intermediate{}\n\t}\n\tif format == nil {\n\t\tformat = formats(\"go\")\n\t}\n\treturn Printer{\n\t\tsrc: src,\n\t\tcodec: c,\n\t\tformat: format,\n\t}\n}", "func NewFloat64Formatter(p *float64) *Float64Formatter {\n\treturn &Float64Formatter{p}\n}", "func NewLine(format string) *Line {\n\treturn &Line{format}\n}", "func newLog(prefix string) *logging.Logger {\n\tfdFmt := logging.MustStringFormatter(\n\t\t`%{level:.4s}[%{id:03x}]%{time:2006-01-02 15:04:05.000}: %{message}`,\n\t)\n\t// nolint\n\tfd, err := os.OpenFile(logFile, os.O_CREATE|os.O_WRONLY|os.O_APPEND, 0o644)\n\tif err != nil {\n\t\tfmt.Fprint(os.Stderr, err.Error())\n\t}\n\n\tfdLog := logging.NewLogBackend(fd, \"\", 0)\n\tfdFmttr = logging.NewBackendFormatter(fdLog, fdFmt)\n\n\tsysFmttr, err := logging.NewSyslogBackend(prog + \": \")\n\tif err != nil {\n\t\tfmt.Fprint(os.Stderr, err.Error())\n\t}\n\n\tlogging.SetBackend(fdFmttr, sysFmttr)\n\n\treturn logging.MustGetLogger(prog)\n}", "func New(sink io.Writer) Debug {\n\tflags := log.Lshortfile\n\treturn log.New(sink, \"\", flags)\n}", "func Newf(format string, args ...interface{}) error {\n\treturn &Error{\n\t\tstack: callers(),\n\t\ttext: fmt.Sprintf(format, args...),\n\t\tcode: gcode.CodeNil,\n\t}\n}", "func Newf(c int, format string, a ...interface{}) *Status {\n\treturn New(c, fmt.Sprintf(format, a...))\n}", "func NewIlpFormatter(measurement string) (ilp *IlpFormatter) {\n\tilp = &IlpFormatter{\n\t\tname: ilpFormatterName,\n\t\tmeasurement: measurement,\n\t}\n\treturn ilp\n}", "func Newf(format string, a ...interface{}) (err error) {\n\treturn backendErr(fmt.Errorf(format, a...))\n}", "func NewJSONFormatterFromPool(p JSONFormatterObjectPool) Formatter {\n\treturn &jsonFormatter{pool: p}\n}", "func NewConnectivityFormat(source string, quiet bool) Format {\n\tswitch source {\n\tcase TableFormatKey:\n\t\tif quiet {\n\t\t\treturn connectivityTableQuietFormat\n\t\t}\n\t\treturn connectivityTableFormat\n\tcase RawFormatKey:\n\t\tif quiet {\n\t\t\treturn connectivityRawQuietFormat\n\t\t}\n\t\treturn connectivityRawFormat\n\tcase SummaryFormatKey:\n\t\treturn connectivitySummaryFormat\n\t}\n\treturn Format(source)\n}", "func newFSFormatV1() (format *formatConfigV1) {\n\treturn &formatConfigV1{\n\t\tVersion: \"1\",\n\t\tFormat: \"fs\",\n\t\tFS: &fsFormat{\n\t\t\tVersion: \"1\",\n\t\t},\n\t}\n}", "func (c *context) New(prefixes ...interface{}) Logger {\n\treturn newContext(c.logger, c.prefix, prefixes...)\n}", "func newCompressor(format CompressEncodingType) (compressor, error) {\n\tvar (\n\t\twriter encoder\n\t\terr error\n\t)\n\n\tswitch format {\n\tcase GZIPCompression:\n\t\twriter = gzip.NewWriter(io.Discard)\n\tcase DeflateCompression:\n\t\twriter, err = flate.NewWriter(io.Discard, flate.BestSpeed)\n\t\tif err != nil {\n\t\t\treturn compressor{}, err\n\t\t}\n\tcase NoCompression:\n\t\twriter = nil\n\tdefault:\n\t\treturn compressor{}, fmt.Errorf(\"invalid format: %s\", format)\n\t}\n\n\treturn compressor{\n\t\tformat: format,\n\t\twriter: writer,\n\t}, nil\n}", "func New(tfProvider *schema.Provider, outputParam string, resourceFormat ResourceFormat) (OutputSink, error) {\n\tswitch resourceFormat {\n\tcase KRMResourceFormat:\n\t\treturn newKRM(tfProvider, outputParam)\n\tcase HCLResourceFormat:\n\t\treturn newHCL(tfProvider, outputParam)\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unknown resource format '%v'\", resourceFormat)\n\t}\n}", "func New() *Tree {\n\treturn &Tree{\n\t\tDelimiter: DefaultDelimiter,\n\t\tFormatter: SimpleFormatter,\n\t\tErrors: make(map[string]error),\n\t}\n}", "func New() NopPrinter { return NopPrinter{} }", "func MustNewJSONFormatter(keys map[string]string, full bool) Formatter {\n\tf, err := NewJSONFormatter(keys, full)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn f\n}", "func Formatter(k string, verb string, v interface{}) Field {\n\treturn Field{Key: k, Value: valf.Formatter(verb, v)}\n}", "func newSpanParser(src []byte) *spanParser {\n\tp := &parser{\n\t\tsrc: src,\n\t}\n\tsp := &spanParser{parser: p, ref: make(map[string]*reference), spanChan: make(chan Span)}\n\tgo sp.run()\n\treturn sp\n}", "func New(f string, args ...interface{}) error {\n\treturn fmt.Errorf(f, args...)\n}" ]
[ "0.8013783", "0.78027093", "0.7800939", "0.75259215", "0.7440931", "0.7186287", "0.71804726", "0.7076727", "0.70714617", "0.7042193", "0.7010463", "0.6986868", "0.6966276", "0.6866522", "0.6745573", "0.67250544", "0.6631822", "0.6554914", "0.6272679", "0.62581867", "0.6230031", "0.62052846", "0.6106857", "0.6008597", "0.59064233", "0.58747345", "0.5815067", "0.58150226", "0.5806014", "0.57786", "0.5721937", "0.5673563", "0.5665305", "0.5655425", "0.5613802", "0.5594178", "0.55718976", "0.55573463", "0.54941434", "0.54919505", "0.5449529", "0.5410787", "0.5389873", "0.5365359", "0.5349459", "0.5341154", "0.5336356", "0.53290695", "0.53235525", "0.5317677", "0.53090316", "0.53078085", "0.5296739", "0.5241671", "0.5240106", "0.52355504", "0.5225389", "0.52141297", "0.5201789", "0.5200597", "0.5172352", "0.51676965", "0.51510805", "0.5151018", "0.5142349", "0.51400965", "0.513835", "0.51323503", "0.5126558", "0.51243687", "0.5120611", "0.5117749", "0.5098062", "0.50658554", "0.50623035", "0.50578624", "0.5042957", "0.5041266", "0.5032862", "0.50153255", "0.5005434", "0.4999307", "0.4985417", "0.49819154", "0.49808758", "0.497951", "0.49722514", "0.49679863", "0.4967963", "0.49634778", "0.49509898", "0.49465585", "0.49461526", "0.4944553", "0.49280873", "0.49217874", "0.49191445", "0.49107334", "0.49076003", "0.4901273" ]
0.66417533
16
NewHostMetrics returns a new instance of input host level load metrics. Should be one instance per input host
func NewHostMetrics() *HostMetrics { return &HostMetrics{ CounterBank: common.NewCounterBank(numHostMetrics), } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewHostMetrics() *HostMetrics {\n\tthis := HostMetrics{}\n\treturn &this\n}", "func newMetrics(hostAndPort string) *metrics {\n\tm := metrics{\n\t\tmetricsCh: make(chan metricType),\n\t\thostAndPort: hostAndPort,\n\t}\n\n\treturn &m\n}", "func newHostMetricsReceiver(\n\tctx context.Context,\n\tlogger *zap.Logger,\n\tconfig *Config,\n\tfactories map[string]internal.Factory,\n\tconsumer consumer.MetricsConsumer,\n) (*receiver, error) {\n\n\tscrapers := make([]internal.Scraper, 0)\n\tfor key, cfg := range config.Scrapers {\n\t\tfactory := factories[key]\n\t\tif factory == nil {\n\t\t\treturn nil, fmt.Errorf(\"host metrics scraper factory not found for key: %s\", key)\n\t\t}\n\n\t\tscraper, err := factory.CreateMetricsScraper(ctx, logger, cfg)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"cannot create scraper: %s\", err.Error())\n\t\t}\n\t\tscrapers = append(scrapers, scraper)\n\t}\n\n\thmr := &receiver{\n\t\tconfig: config,\n\t\tscrapers: scrapers,\n\t\tconsumer: consumer,\n\t}\n\n\treturn hmr, nil\n}", "func NewHostMetricCollector() (*HostMetricCollector, error) {\n\tcpuTimes, err := cpu.Times(false)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cpu.Times() failed: %s\", err)\n\t} else if len(cpuTimes) == 0 {\n\t\treturn nil, fmt.Errorf(\"cpu.Times() returns no cpus\")\n\t}\n\tt := cpuTimes[0]\n\treturn &HostMetricCollector{\n\t\tlastJiffy: t.Total(),\n\t\tlastTimes: t,\n\t}, nil\n}", "func NewInputHost(serviceName string, sVice common.SCommon, mClient metadata.TChanMetadataService, opts *InOptions) (*InputHost, []thrift.TChanServer) {\n\n\t// Get the deployment name for logger field\n\tdeploymentName := sVice.GetConfig().GetDeploymentName()\n\tbs := InputHost{\n\t\tlogger: (sVice.GetConfig().GetLogger()).WithFields(bark.Fields{common.TagIn: common.FmtIn(sVice.GetHostUUID()), common.TagDplName: common.FmtDplName(deploymentName)}),\n\t\tSCommon: sVice,\n\t\tpathCache: make(map[string]*inPathCache),\n\t\tpathCacheByDestPath: make(map[string]string), // simple map which just resolves the path to uuid\n\t\tcacheTimeout: defaultIdleTimeout,\n\t\tshutdown: make(chan struct{}),\n\t\thostMetrics: load.NewHostMetrics(),\n\t\tlastLoadReportedTime: time.Now().UnixNano(),\n\t}\n\n\t// Set the host limits from the common package\n\t// TODO: once limits are moved behind an interface we can change this\n\t// this is mainly exposed via getters and setters for testing\n\tbs.SetHostConnLimit(int32(common.HostOverallConnLimit))\n\tbs.SetHostConnLimitPerSecond(int32(common.HostPerSecondConnLimit))\n\tbs.SetMaxConnPerDest(int32(common.HostMaxConnPerDestination))\n\n\t// create the token bucket for this host\n\tbs.SetTokenBucketValue(int32(bs.GetHostConnLimitPerSecond()))\n\n\tbs.m3Client = metrics.NewClient(sVice.GetMetricsReporter(), metrics.Inputhost)\n\tif opts != nil {\n\t\tbs.cacheTimeout = opts.CacheIdleTimeout\n\t}\n\n\tbs.mClient = mm.NewMetadataMetricsMgr(mClient, bs.m3Client, bs.logger)\n\n\t// manage uconfig, regiester handerFunc and verifyFunc for uConfig values\n\tbs.dConfigClient = sVice.GetDConfigClient()\n\tbs.dynamicConfigManage()\n\tbs.SetNodeStatus(controller.NodeStatus_UP)\n\treturn &bs, []thrift.TChanServer{cherami.NewTChanBInServer(&bs), admin.NewTChanInputHostAdminServer(&bs)}\n}", "func NewMetrics(component string, sampleRate float64, client metrics.Client) BaseMetrics {\n\treturn BaseMetrics{\n\t\tcomponent: component,\n\t\trate: sampleRate,\n\t\tmetrics: client,\n\t\tmetMap: map[string]string{\n\t\t\t\"latency\": \"comp.\" + component + \".requests.latency\",\n\t\t\t\"request\": \"comp.\" + component + \".requests.%d\",\n\t\t\t\"mLatency\": \"comp.\" + component + \".requests.%s.latency\",\n\t\t\t\"mRequest\": \"comp.\" + component + \".requests.%s.%d\",\n\t\t},\n\t}\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tInputBytesTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_input_bytes_total\",\n\t\t\t\tHelp: \"Total number of bytes received\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tOutputBytesTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_output_bytes_total\",\n\t\t\t\tHelp: \"Total number of bytes sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tInputPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_input_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets received\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tOutputPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_output_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tDroppedPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_dropped_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets dropped by the router. This metric reports \" +\n\t\t\t\t\t\"the number of packets that were dropped because of errors.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tInterfaceUp: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_interface_up\",\n\t\t\t\tHelp: \"Either zero or one depending on whether the interface is up.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDInterfaceStateChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_state_changes_total\",\n\t\t\t\tHelp: \"Total number of BFD state changes.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDPacketsSent: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sent_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDPacketsReceived: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_received_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets received.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tServiceInstanceCount: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_service_instance_count\",\n\t\t\t\tHelp: \"Number of service instances known by the data plane.\",\n\t\t\t},\n\t\t\t[]string{\"service\", \"isd_as\"},\n\t\t),\n\t\tServiceInstanceChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_service_instance_changes_total\",\n\t\t\t\tHelp: \"Number of total service instance changes. Both addition and removal of a \" +\n\t\t\t\t\t\"service instance is accumulated.\",\n\t\t\t},\n\t\t\t[]string{\"service\", \"isd_as\"},\n\t\t),\n\t\tSiblingReachable: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_sibling_reachable\",\n\t\t\t\tHelp: \"Either zero or one depending on whether a sibling router \" +\n\t\t\t\t\t\"instance is reachable.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDPacketsSent: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sent_sibling_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets sent to sibling router instance.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDPacketsReceived: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_received_sibling_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets received from sibling router instance.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDStateChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sibling_state_changes_total\",\n\t\t\t\tHelp: \"Total number of BFD state changes for sibling router instances\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t}\n}", "func NewMetrics(conf *probepb.HermesProbeDef, target *probepb.Target) (*Metrics, error) {\n\tm := &Metrics{\n\t\tProbeOpLatency: make(map[ProbeOperation]map[ExitStatus]*metrics.EventMetrics, len(ProbeOpName)),\n\t\tAPICallLatency: make(map[APICall]map[ExitStatus]*metrics.EventMetrics, len(APICallName)),\n\t}\n\n\tprobeOpLatDist, err := metrics.NewDistributionFromProto(conf.GetProbeLatencyDistribution())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"invalid argument: error creating probe latency distribution from the specification (%v): %w\", conf.GetProbeLatencyDistribution(), err)\n\t}\n\n\tfor op := range ProbeOpName {\n\t\tm.ProbeOpLatency[op] = make(map[ExitStatus]*metrics.EventMetrics, len(ExitStatusName))\n\t\tfor e := range ExitStatusName {\n\t\t\tm.ProbeOpLatency[op][e] = metrics.NewEventMetrics(time.Now()).\n\t\t\t\tAddMetric(\"hermes_probe_latency_seconds\", probeOpLatDist.Clone()).\n\t\t\t\tAddLabel(\"storage_system\", target.GetTargetSystem().String()).\n\t\t\t\tAddLabel(\"target\", fmt.Sprintf(\"%s:%s\", target.GetName(), target.GetBucketName())).\n\t\t\t\tAddLabel(\"probe_operation\", ProbeOpName[op]).\n\t\t\t\tAddLabel(\"exit_status\", ExitStatusName[e])\n\t\t}\n\t}\n\n\tapiCallLatDist, err := metrics.NewDistributionFromProto(conf.GetApiCallLatencyDistribution())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"invalid argument: error creating probe latency distribution from the specification (%v): %v\", conf.GetApiCallLatencyDistribution(), err)\n\t}\n\n\tfor call := range APICallName {\n\t\tm.APICallLatency[call] = make(map[ExitStatus]*metrics.EventMetrics, len(ExitStatusName))\n\t\tfor e := range ExitStatusName {\n\t\t\tm.APICallLatency[call][e] = metrics.NewEventMetrics(time.Now()).\n\t\t\t\tAddMetric(\"hermes_api_latency_seconds\", apiCallLatDist.Clone()).\n\t\t\t\tAddLabel(\"storage_system\", target.GetTargetSystem().String()).\n\t\t\t\tAddLabel(\"target\", fmt.Sprintf(\"%s:%s\", target.GetName(), target.GetBucketName())).\n\t\t\t\tAddLabel(\"api_call\", APICallName[call]).\n\t\t\t\tAddLabel(\"exit_status\", ExitStatusName[e])\n\t\t}\n\t}\n\n\treturn m, nil\n}", "func NewMetrics(p fabricmetrics.Provider) *Metrics {\n\treturn &Metrics{\n\t\tRefreshTimer: p.NewHistogram(refreshTimer),\n\t}\n}", "func newMetrics() *metrics {\n\treturn new(metrics)\n}", "func NewMetrics() *MetricsHolder {\n\tm := &MetricsHolder{\n\t\tlines: make(map[string]*Reading),\n\t\tchannel: make(chan interface{}),\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\tw, ok := <-m.channel\n\t\t\treading := w.(*Reading)\n\t\t\tif !ok {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif val, ok := m.lines[reading.Key]; ok {\n\t\t\t\tm.lines[reading.Key] = val.Accept(reading)\n\t\t\t} else {\n\t\t\t\tm.lines[reading.Key] = reading\n\t\t\t}\n\t\t}\n\t}()\n\treturn m\n}", "func NewMetrics(period time.Duration, maxQueueSize int) (*Metrics, error) {\n\tmetrics := &Metrics{\n\t\tmaxQueueSize: maxQueueSize,\n\t\tperiod: period,\n\t\tinitialized: true,\n\t\tqueue: make([]Measurement, 0),\n\t\tlastSendingDate: -1,\n\t}\n\n\tif UseGlobalEngine {\n\t\tmetrics.Engine = Engine\n\t} else {\n\t\tmetrics.Engine = &req.Engine{}\n\t}\n\n\terr := validateMetrics(metrics)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif sources == nil {\n\t\tsources = make([]DataSource, 0)\n\t\tgo sendingLoop()\n\t}\n\n\tsources = append(sources, metrics)\n\n\treturn metrics, nil\n}", "func NewMetrics(factory promutil.Factory) *Metrics {\n\treturn &Metrics{\n\t\tImporterEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"importer_engine\",\n\t\t\t\tHelp: \"counting open and closed importer engines\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tIdleWorkersGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"idle_workers\",\n\t\t\t\tHelp: \"counting idle workers\",\n\t\t\t}, []string{\"name\"}),\n\n\t\tKvEncoderCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"kv_encoder\",\n\t\t\t\tHelp: \"counting kv open and closed kv encoder\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tTableCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"tables\",\n\t\t\t\tHelp: \"count number of tables processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tProcessedEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"engines\",\n\t\t\t\tHelp: \"count number of engines processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tChunkCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunks\",\n\t\t\t\tHelp: \"count number of chunks processed\",\n\t\t\t}, []string{\"state\"}),\n\n\t\tBytesCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"bytes\",\n\t\t\t\tHelp: \"count of total bytes\",\n\t\t\t}, []string{\"state\"}),\n\t\t// state can be one of:\n\t\t// - estimated (an estimation derived from the file size)\n\t\t// - pending\n\t\t// - running\n\t\t// - finished\n\t\t// - failed\n\n\t\tRowsCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"rows\",\n\t\t\t\tHelp: \"count of total rows\",\n\t\t\t}, []string{\"state\", \"table\"}),\n\n\t\tImportSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"import_seconds\",\n\t\t\t\tHelp: \"time needed to import a table\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.125, 2, 6),\n\t\t\t}),\n\n\t\tChunkParserReadBlockSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunk_parser_read_block_seconds\",\n\t\t\t\tHelp: \"time needed for chunk parser read a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tApplyWorkerSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"apply_worker_seconds\",\n\t\t\t\tHelp: \"time needed to apply a worker\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}, []string{\"name\"}),\n\n\t\tRowReadSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_seconds\",\n\t\t\t\tHelp: \"time needed to parse a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 7),\n\t\t\t}),\n\n\t\tRowReadBytesHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_bytes\",\n\t\t\t\tHelp: \"number of bytes being read out from data source\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1024, 2, 8),\n\t\t\t}),\n\n\t\tRowEncodeSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_encode_seconds\",\n\t\t\t\tHelp: \"time needed to encode a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tRowKVDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_kv_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver kvs of a single row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tBlockDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tBlockDeliverBytesHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_bytes\",\n\t\t\t\tHelp: \"number of bytes being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(512, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tBlockDeliverKVPairsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_kv_pairs\",\n\t\t\t\tHelp: \"number of KV pairs being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tChecksumSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"checksum_seconds\",\n\t\t\t\tHelp: \"time needed to complete the checksum stage\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}),\n\t\tSSTSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"sst_seconds\",\n\t\t\t\tHelp: \"time needed to complete the sst operations\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}, []string{\"kind\"}),\n\n\t\tLocalStorageUsageBytesGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"local_storage_usage_bytes\",\n\t\t\t\tHelp: \"disk/memory size currently occupied by intermediate files in local backend\",\n\t\t\t}, []string{\"medium\"}),\n\n\t\tProgressGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"progress\",\n\t\t\t\tHelp: \"progress of lightning phase\",\n\t\t\t}, []string{\"phase\"}),\n\t}\n}", "func newMetrics() *Metrics {\n\treturn newMetricsFrom(DefaultMetricsOpts)\n}", "func NewHost(clusterName string, mainCfg conf.Main, hostCfg conf.Host, lg *zap.Logger, ms *metrics.Prom) *Host {\n\ttargetPort := mainCfg.TargetPort\n\tif hostCfg.Port != 0 {\n\t\ttargetPort = hostCfg.Port\n\t}\n\n\tpromLabels := prometheus.Labels{\n\t\t\"cluster\": clusterName,\n\t\t\"upstream_host\": hostCfg.Name,\n\t}\n\treturn &Host{\n\t\tName: hostCfg.Name,\n\t\tPort: targetPort,\n\t\tCh: make(chan *rec.Rec, mainCfg.HostQueueSize),\n\t\tLg: lg,\n\t\tstop: make(chan int),\n\n\t\tSendTimeoutSec: mainCfg.SendTimeoutSec,\n\t\tConnTimeoutSec: mainCfg.OutConnTimeoutSec,\n\t\tTCPOutBufFlushPeriodSec: mainCfg.TCPOutBufFlushPeriodSec,\n\t\toutRecs: ms.OutRecs.With(promLabels),\n\t\tthrottled: ms.ThrottledHosts.With(promLabels),\n\t\tprocessingDuration: ms.ProcessingDuration,\n\t\tbufSize: mainCfg.TCPOutBufSize,\n\t\tMaxReconnectPeriodMs: mainCfg.MaxHostReconnectPeriodMs,\n\t\tReconnectPeriodDeltaMs: mainCfg.MaxHostReconnectPeriodMs,\n\t}\n}", "func NewMetrics(reg prometheus.Registerer) *Metrics {\n\tvar m Metrics\n\tm.reg = reg\n\n\tm.dockerEntries = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_entries_total\",\n\t\tHelp: \"Total number of successful entries sent to the Docker target\",\n\t})\n\tm.dockerErrors = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_parsing_errors_total\",\n\t\tHelp: \"Total number of parsing errors while receiving Docker messages\",\n\t})\n\n\tif reg != nil {\n\t\treg.MustRegister(\n\t\t\tm.dockerEntries,\n\t\t\tm.dockerErrors,\n\t\t)\n\t}\n\n\treturn &m\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tcfgwarn.Beta(\"The vsphere host metricset is beta\")\n\n\tconfig := struct {\n\t\tUsername string `config:\"username\"`\n\t\tPassword string `config:\"password\"`\n\t\tInsecure bool `config:\"insecure\"`\n\t}{}\n\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\tu, err := url.Parse(base.HostData().URI)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu.User = url.UserPassword(config.Username, config.Password)\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tHostURL: u,\n\t\tInsecure: config.Insecure,\n\t}, nil\n}", "func newHttpMetrics() *httpMetrics {\n\treturn &httpMetrics{\n\t\tRequestsTotal: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_requests_total\",\n\t\t\tHelp: \"Total number of HTTP requests.\",\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t\tRequestDurationHistogram: promauto.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_request_duration_seconds\",\n\t\t\tHelp: \"Seconds spent serving HTTP requests.\",\n\t\t\tBuckets: prometheus.DefBuckets,\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t}\n}", "func NewHost(hostLine string) (*Host, error) {\n\tptn := regexp.MustCompile(HostLineFmtRegexp)\n\tdata := ptn.FindStringSubmatch(hostLine)\n\tif len(data) != 3 {\n\t\treturn nil, errors.New(\"Parse failed\")\n\t}\n\n\treturn &Host{data[1], data[2]}, nil\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{items: make(map[string]*metric), rm: &sync.RWMutex{}}\n}", "func NewMetrics(ctx context.Context, output string, tenant string, refreshRate time.Duration) Metrics {\n\treturn Metrics{\n\t\tDaemonSupport: utils.NewDaemonSupport(ctx, \"metrics\"),\n\t\tstorage: localfs.NewPlaintextStorage(output),\n\t\ttenant: tenant,\n\t\trefreshRate: refreshRate,\n\t\tpromisesAccepted: metrics.NewCounter(),\n\t\tcommitsAccepted: metrics.NewCounter(),\n\t\trollbacksAccepted: metrics.NewCounter(),\n\t\tcreatedAccounts: metrics.NewCounter(),\n\t\tupdatedSnapshots: metrics.NewMeter(),\n\t\tsnapshotCronLatency: metrics.NewTimer(),\n\t}\n}", "func NewHostMonitor(duration time.Duration) (pkg.HostMonitor, error) {\n\tif duration == 0 {\n\t\tduration = 2 * time.Second\n\t}\n\twatcher, err := fsnotify.NewWatcher()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to initialize fs watcher\")\n\t}\n\n\t// the file can not exist if the system was booted from overlay\n\tif _, err := os.Stat(upgrade.FlistInfoFile); err == nil {\n\t\tif err := watcher.Add(upgrade.FlistInfoFile); err != nil {\n\t\t\treturn nil, errors.Wrapf(err, \"failed to watch '%s'\", upgrade.FlistInfoFile)\n\t\t}\n\t}\n\n\treturn &hostMonitor{\n\t\tduration: duration,\n\t}, nil\n}", "func NewMetrics(app, metricsPrefix, version, hash, date string) *Metrics {\n\tlabels := map[string]string{\n\t\t\"app\": app,\n\t\t\"version\": version,\n\t\t\"hash\": hash,\n\t\t\"buildTime\": date,\n\t}\n\n\tif metricsPrefix != \"\" {\n\t\tmetricsPrefix += \"_\"\n\t}\n\n\tpm := &Metrics{\n\t\tresponseTime: prometheus.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tName: metricsPrefix + \"response_time_seconds\",\n\t\t\t\tHelp: \"Description\",\n\t\t\t\tConstLabels: labels,\n\t\t\t},\n\t\t\t[]string{\"endpoint\"},\n\t\t),\n\t\ttotalRequests: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_total\",\n\t\t\tHelp: \"number of requests\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tduration: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_duration_seconds\",\n\t\t\tHelp: \"duration of a requests in seconds\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tresponseSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"response_size_bytes\",\n\t\t\tHelp: \"size of the responses in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\trequestSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_size_bytes\",\n\t\t\tHelp: \"size of the requests in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\thandlerStatuses: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_statuses_total\",\n\t\t\tHelp: \"count number of responses per status\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"method\", \"status_bucket\"}),\n\t}\n\n\terr := prometheus.Register(pm)\n\tif e := new(prometheus.AlreadyRegisteredError); errors.As(err, e) {\n\t\treturn pm\n\t} else if err != nil {\n\t\tpanic(err)\n\t}\n\n\tgrpcPrometheus.EnableHandlingTimeHistogram()\n\n\treturn pm\n}", "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseImpl: newBaseImpl(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func NewMetrics(namespace string, logger Logger) Metrics {\n\tlog := logger.GetLogger()\n\n\treturn &metricsImpl{\n\t\tinternalMetrics: metrics.NewMetrics(\"\", log),\n\t\texternalMetrics: metrics.NewMetrics(strings.ToLower(namespace), log),\n\t}\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tPath: defaultPath,\n\t\tAddr: defaultAddr,\n\t\textraLabels: []extraLabel{},\n\t}\n}", "func New() *Metrics {\n\treturn &Metrics{\n\t\tSectionCounts: make(map[string]int),\n\t}\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{\n\t\tTimeMetrics: make(map[string]*TimeStats),\n\t\tNumberMetrics: make(map[string]*NumberStats),\n\t\tBoolMetrics: make(map[string]*BoolStats),\n\t}\n\treturn m\n}", "func New(hosts ...string) *P2C {\n\tp := &P2C{\n\t\thosts: []*host{},\n\t\tloadMap: map[string]*host{},\n\t\trndm: rand.New(rand.NewSource(time.Now().UnixNano())),\n\t}\n\n\tfor _, h := range hosts {\n\t\tp.Add(h)\n\t}\n\n\treturn p\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{}\n\tm.Reset()\n\treturn m\n}", "func newMetricsFrom(opts *MetricsOpts) *Metrics {\n\tmetrics := &Metrics{\n\t\tcounters: make(map[string]prometheus.Counter, 512),\n\t\tgauges: make(map[string]prometheus.Gauge, 512),\n\t\thistorams: make(map[string]prometheus.Histogram, 512),\n\t\tsummaries: make(map[string]prometheus.Summary, 512),\n\t\tdefBuckets: opts.DefBuckets,\n\t\tdefQuantile: opts.DefQuantile,\n\t\tregistry: prometheus.NewRegistry(),\n\t}\n\treturn metrics\n}", "func NewHostMetricsWithDefaults() *HostMetrics {\n\tthis := HostMetrics{}\n\treturn &this\n}", "func New() *SystemMetrics {\n\treturn &SystemMetrics{}\n}", "func NewHost(host string) Host {\n\treturn Host(host)\n}", "func newProcessMetrics(id string) *processMetrics {\n\tcommonTags := tags{TAG_INGESTER_ID: id, TAG_INGESTER_SOURCE: \"poll\"}\n\treturn &processMetrics{\n\t\tignoredByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"ignored\"}),\n\t\tprocessedByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"processed\"}),\n\t\tliveness: metrics2.NewLiveness(id, tags{TAG_INGESTER_SOURCE: \"poll\", TAG_INGESTION_METRIC: \"since-last-run\"}),\n\t}\n}", "func NewMetrics() *Metrics {\n\tmtrcs := &Metrics{\n\t\tcounters: make(map[MetricName]int),\n\t\tSidecarSyncErrors: SidecarSyncErrors,\n\t\tSidecarVaultTokenErrors: SidecarVaultTokenErrors,\n\t\tSidecarSecretErrors: SidecarSecretErrors,\n\t}\n\n\treturn mtrcs\n}", "func CreateMetrics(protocol string, host string, port int, tag string) (*Metrics, error) {\n\tvar m *Metrics\n\tswitch protocol {\n\tcase \"tcp\":\n\t\tm = &Metrics{Host: host, Port: port, Protocol: \"tcp\", Tag: tag}\n\tcase \"udp\":\n\t\tm = &Metrics{Host: host, Port: port, Protocol: \"udp\", Tag: tag}\n\t}\n\t// Initialize values\n\tm.Timeout = 0\n\tm.conn = nil\n\tm.Counters = make(map[string]Counter)\n\t// Connect\n\tif err := m.Connect(); err != nil {\n\t\treturn m, err\n\t}\n\tm.Ready = true\n\treturn m, nil\n}", "func NewMetrics(reg *prometheus.Registry, namespace, subsystem string, methodsFrom interface{}) (metric Metrics) {\n\tmetric.callErrTotal = prometheus.NewCounterVec(\n\t\tprometheus.CounterOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"errors_total\",\n\t\t\tHelp: \"Amount of DAL errors.\",\n\t\t},\n\t\t[]string{methodLabel},\n\t)\n\treg.MustRegister(metric.callErrTotal)\n\tmetric.callDuration = prometheus.NewHistogramVec(\n\t\tprometheus.HistogramOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"call_duration_seconds\",\n\t\t\tHelp: \"DAL call latency.\",\n\t\t},\n\t\t[]string{methodLabel},\n\t)\n\treg.MustRegister(metric.callDuration)\n\n\tfor _, methodName := range reflectx.MethodsOf(methodsFrom) {\n\t\tl := prometheus.Labels{\n\t\t\tmethodLabel: methodName,\n\t\t}\n\t\tmetric.callErrTotal.With(l)\n\t\tmetric.callDuration.With(l)\n\t}\n\n\treturn metric\n}", "func New(metrics ...interface{}) Master {\n\tvar sentries []Sentry\n\tvar entries []Metrics\n\n\tfor _, item := range metrics {\n\t\tswitch rItem := item.(type) {\n\t\tcase Metrics:\n\t\t\tentries = append(entries, rItem)\n\t\tcase Sentry:\n\t\t\tsentries = append(sentries, rItem)\n\t\t}\n\t}\n\n\treturn Master{\n\t\tmetrics: append(entries, Sentries(sentries...)),\n\t}\n}", "func NewHostFilter(host string) *HostFilter {\n\tctx, cancel := context.WithCancel(context.Background())\n\tf := &HostFilter{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\n\t\thost: host,\n\n\t\toutputCh: make(chan map[string][]*targetgroup.Group),\n\t}\n\treturn f\n}", "func New(provider PayloadEncoderDecoderProvider, cluster Cluster) messageprocessors.PayloadEncoderDecoder {\n\treturn &host{\n\t\tcluster: cluster,\n\t\tprovider: provider,\n\n\t\tcache: gcache.New(cacheSize).LFU().Build(),\n\t}\n}", "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (consumer.Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseConsumer: newBaseConsumer(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func init() {\n\n\t// cpu\n\tRegistryMetricCreateInput(\"cpu\", \"CPU usage\", monitor.METRIC_RES_TYPE_HOST, monitor.METRIC_DATABASE_TELE, 1,\n\t\t[]monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_active\", \"CPU active state utilization rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"usage_idle\", \"CPU idle state utilization rate\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t\tnewMetricFieldCreateInput(\"usage_system\", \"CPU system state utilization rate\", monitor.METRIC_UNIT_PERCENT, 3),\n\t\t\tnewMetricFieldCreateInput(\"usage_user\", \"CPU user mode utilization rate\", monitor.METRIC_UNIT_PERCENT, 4),\n\t\t\tnewMetricFieldCreateInput(\"usage_iowait\", \"CPU IO usage\", monitor.METRIC_UNIT_PERCENT, 5),\n\t\t\tnewMetricFieldCreateInput(\"usage_irq\", \"CPU IRQ usage\", monitor.METRIC_UNIT_PERCENT, 6),\n\t\t\tnewMetricFieldCreateInput(\"usage_guest\", \"CPU guest usage\", monitor.METRIC_UNIT_PERCENT, 7),\n\t\t\tnewMetricFieldCreateInput(\"usage_nice\", \"CPU priority switch utilization\", monitor.METRIC_UNIT_PERCENT, 8),\n\t\t\tnewMetricFieldCreateInput(\"usage_softirq\", \"CPU softirq usage\", monitor.METRIC_UNIT_PERCENT, 9),\n\t\t})\n\n\t// disk\n\tRegistryMetricCreateInput(\"disk\", \"Disk usage\", monitor.METRIC_RES_TYPE_HOST,\n\t\tmonitor.METRIC_DATABASE_TELE, 3,\n\t\t[]monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Percentage of used disks\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"free\", \"Free space size\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"used\", \"Used disk size\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t\tnewMetricFieldCreateInput(\"total\", \"Total disk size\", monitor.METRIC_UNIT_BYTE, 4),\n\t\t\tnewMetricFieldCreateInput(\"inodes_free\", \"Available inode\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"inodes_used\", \"Number of inodes used\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"inodes_total\", \"Total inodes\", monitor.METRIC_UNIT_COUNT, 7),\n\t\t})\n\n\t// diskio\n\tRegistryMetricCreateInput(\"diskio\", \"Disk traffic and timing\",\n\t\tmonitor.METRIC_RES_TYPE_HOST, monitor.METRIC_DATABASE_TELE, 4, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"read_bps\", \"Disk read rate\", monitor.METRIC_UNIT_BPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"write_bps\", \"Disk write rate\", monitor.METRIC_UNIT_BPS, 2),\n\t\t\tnewMetricFieldCreateInput(\"read_iops\", \"Disk read operate rate\", monitor.METRIC_UNIT_COUNT, 3),\n\t\t\tnewMetricFieldCreateInput(\"write_iops\", \"Disk write operate rate\", monitor.METRIC_UNIT_COUNT, 4),\n\t\t\tnewMetricFieldCreateInput(\"reads\", \"Number of reads\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"writes\", \"Number of writes\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"read_bytes\", \"Bytes read\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"write_bytes\", \"Bytes write\", monitor.METRIC_UNIT_BYTE, 8),\n\t\t\tnewMetricFieldCreateInput(\"write_time\", \"Time to wait for write\", monitor.METRIC_UNIT_MS, 9),\n\t\t\tnewMetricFieldCreateInput(\"io_time\", \"I / O request queuing time\", monitor.METRIC_UNIT_MS, 10),\n\t\t\tnewMetricFieldCreateInput(\"weighted_io_time\", \"I / O request waiting time\", monitor.METRIC_UNIT_MS, 11),\n\t\t\tnewMetricFieldCreateInput(\"iops_in_progress\", \"Number of I / O requests issued but not yet completed\", monitor.METRIC_UNIT_COUNT, 12),\n\t\t})\n\n\t// mem\n\tRegistryMetricCreateInput(\"mem\", \"Memory\", monitor.METRIC_RES_TYPE_HOST,\n\t\tmonitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Used memory rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"available_percent\", \"Available memory rate\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t\tnewMetricFieldCreateInput(\"used\", \"Used memory\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t\tnewMetricFieldCreateInput(\"free\", \"Free memory\", monitor.METRIC_UNIT_BYTE, 4),\n\t\t\tnewMetricFieldCreateInput(\"active\", \"The amount of active memory\", monitor.METRIC_UNIT_BYTE, 5),\n\t\t\tnewMetricFieldCreateInput(\"inactive\", \"The amount of inactive memory\", monitor.METRIC_UNIT_BYTE, 6),\n\t\t\tnewMetricFieldCreateInput(\"cached\", \"Cache memory\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"buffered\", \"Buffer memory\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"slab\", \"Number of kernel caches\", monitor.METRIC_UNIT_BYTE, 8),\n\t\t\tnewMetricFieldCreateInput(\"available\", \"Available memory\", monitor.METRIC_UNIT_BYTE, 9),\n\t\t\tnewMetricFieldCreateInput(\"total\", \"Total memory\", monitor.METRIC_UNIT_BYTE, 10),\n\t\t})\n\n\t// net\n\tRegistryMetricCreateInput(\"net\", \"Network interface and protocol usage\",\n\t\tmonitor.METRIC_RES_TYPE_HOST, monitor.METRIC_DATABASE_TELE, 5, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bytes_sent\", \"The total number of bytes sent by the network interface\", monitor.METRIC_UNIT_BYTE, 1),\n\t\t\tnewMetricFieldCreateInput(\"bytes_recv\", \"The total number of bytes received by the network interface\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"packets_sent\", \"The total number of packets sent by the network interface\", monitor.METRIC_UNIT_COUNT, 3),\n\t\t\tnewMetricFieldCreateInput(\"packets_recv\", \"The total number of packets received by the network interface\", monitor.METRIC_UNIT_COUNT, 4),\n\t\t\tnewMetricFieldCreateInput(\"err_in\", \"The total number of receive errors detected by the network interface\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"err_out\", \"The total number of transmission errors detected by the network interface\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"drop_in\", \"The total number of received packets dropped by the network interface\", monitor.METRIC_UNIT_COUNT, 7),\n\t\t\tnewMetricFieldCreateInput(\"drop_out\", \"The total number of transmission packets dropped by the network interface\", monitor.METRIC_UNIT_COUNT, 8),\n\t\t})\n\n\t// vm_cpu\n\tRegistryMetricCreateInput(\"vm_cpu\", \"Guest CPU usage\", monitor.METRIC_RES_TYPE_GUEST,\n\t\tmonitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_active\", \"CPU active state utilization rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"cpu_usage_pcore\", \"CPU utilization rate per core\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t\tnewMetricFieldCreateInput(\"cpu_usage_idle_pcore\", \"CPU idle rate per core\", monitor.METRIC_UNIT_PERCENT, 3),\n\t\t\tnewMetricFieldCreateInput(\"cpu_time_system\", \"CPU system state time\", monitor.METRIC_UNIT_MS, 4),\n\t\t\tnewMetricFieldCreateInput(\"cpu_time_user\", \"CPU user state time\", monitor.METRIC_UNIT_MS, 5),\n\t\t\tnewMetricFieldCreateInput(\"thread_count\", \"The number of threads used by the process\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t})\n\n\t// vm_diskio\n\tRegistryMetricCreateInput(\"vm_diskio\", \"Guest disk traffic\", monitor.METRIC_RES_TYPE_GUEST,\n\t\tmonitor.METRIC_DATABASE_TELE, 3, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"read_bps\", \"Disk read rate\", monitor.METRIC_UNIT_BYTEPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"write_bps\", \"Disk write rate\", monitor.METRIC_UNIT_BYTEPS, 2),\n\t\t\tnewMetricFieldCreateInput(\"read_iops\", \"Disk read operate rate\", monitor.METRIC_UNIT_COUNT, 3),\n\t\t\tnewMetricFieldCreateInput(\"write_iops\", \"Disk write operate rate\", monitor.METRIC_UNIT_COUNT, 4),\n\t\t\tnewMetricFieldCreateInput(\"read_bytes\", \"Bytes read\", monitor.METRIC_UNIT_BYTE, 5),\n\t\t\tnewMetricFieldCreateInput(\"write_bytes\", \"Bytes write\", monitor.METRIC_UNIT_BYTE, 6),\n\t\t})\n\n\t// vm_mem\n\tRegistryMetricCreateInput(\"vm_mem\", \"Guest memory\", monitor.METRIC_RES_TYPE_GUEST,\n\t\tmonitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Used memory rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"vms\", \"Virtual memory consumption\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"rss\", \"Actual use of physical memory\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t})\n\n\t// vm_netio\n\tRegistryMetricCreateInput(\"vm_netio\", \"Guest network traffic\", monitor.METRIC_RES_TYPE_GUEST,\n\t\tmonitor.METRIC_DATABASE_TELE, 4, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bps_recv\", \"Received traffic per second\", monitor.METRIC_UNIT_BPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"bps_sent\", \"Send traffic per second\", monitor.METRIC_UNIT_BPS, 2),\n\t\t})\n\n\t// oss_latency\n\tRegistryMetricCreateInput(\"oss_latency\", \"Object storage latency\",\n\t\tmonitor.METRIC_RES_TYPE_OSS, monitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"req_late\", \"Request average E2E delay\", monitor.METRIC_UNIT_MS, 1),\n\t\t})\n\n\t// oss_netio\n\tRegistryMetricCreateInput(\"oss_netio\", \"Object storage network traffic\",\n\t\tmonitor.METRIC_RES_TYPE_OSS, monitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bps_recv\", \"Receive byte\", monitor.METRIC_UNIT_BYTE, 1),\n\t\t\tnewMetricFieldCreateInput(\"bps_sent\", \"Send byte\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t})\n\n\t// oss_req\n\tRegistryMetricCreateInput(\"oss_req\", \"Object store request\", monitor.METRIC_RES_TYPE_OSS,\n\t\tmonitor.METRIC_DATABASE_TELE, 3, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"req_count\", \"request count\", monitor.METRIC_UNIT_COUNT, 1),\n\t\t})\n\n\t// rds_conn\n\tRegistryMetricCreateInput(\"rds_conn\", \"Rds connect\", monitor.METRIC_RES_TYPE_RDS,\n\t\tmonitor.METRIC_DATABASE_TELE, 5, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Connection usage\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// rds_cpu\n\tRegistryMetricCreateInput(\"rds_cpu\", \"Rds CPU usage\", monitor.METRIC_RES_TYPE_RDS,\n\t\tmonitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_active\", \"CPU active state utilization rate\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t})\n\n\t// rds_mem\n\tRegistryMetricCreateInput(\"rds_mem\", \"Rds memory\", monitor.METRIC_RES_TYPE_RDS,\n\t\tmonitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Used memory rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// rds_netio\n\tRegistryMetricCreateInput(\"rds_netio\", \"Rds network traffic\", monitor.METRIC_RES_TYPE_RDS,\n\t\tmonitor.METRIC_DATABASE_TELE, 4, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bps_recv\", \"Received traffic per second\", monitor.METRIC_UNIT_BPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"bps_sent\", \"Send traffic per second\", monitor.METRIC_UNIT_BPS, 2),\n\t\t})\n\n\t// rds_disk\n\tRegistryMetricCreateInput(\"rds_disk\", \"Rds disk usage\", monitor.METRIC_RES_TYPE_RDS,\n\t\tmonitor.METRIC_DATABASE_TELE, 3, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Percentage of used disks\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// dcs_cpu\n\tRegistryMetricCreateInput(\"dcs_cpu\", \"Redis CPU usage\", monitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_percent\", \"CPU active state utilization rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// dcs_mem\n\tRegistryMetricCreateInput(\"dcs_mem\", \"Redis memory\", monitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Used memory rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// dcs_netio\n\tRegistryMetricCreateInput(\"dcs_netio\", \"Redis network traffic\",\n\t\tmonitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 4, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bps_recv\", \"Received traffic per second\", monitor.METRIC_UNIT_BPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"bps_sent\", \"Send traffic per second\", monitor.METRIC_UNIT_BPS, 2),\n\t\t})\n\n\t// dcs_conn\n\tRegistryMetricCreateInput(\"dcs_conn\", \"Redis connect\", monitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 5, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Connection usage\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t})\n\n\t// dcs_instantopt\n\tRegistryMetricCreateInput(\"dcs_instantopt\", \"Redis operator\",\n\t\tmonitor.METRIC_RES_TYPE_REDIS, monitor.METRIC_DATABASE_TELE, 5, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"opt_sec\", \"Number of commands processed per second\", monitor.METRIC_UNIT_COUNT, 1),\n\t\t})\n\n\t// dcs_cachekeys\n\tRegistryMetricCreateInput(\"dcs_cachekeys\", \"Redis keys\", monitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 6, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"key_count\", \"Number of cache keys\", monitor.METRIC_UNIT_COUNT, 1),\n\t\t})\n\n\t// dcs_datamem\n\tRegistryMetricCreateInput(\"dcs_datamem\", \"Redis data memory\", monitor.METRIC_RES_TYPE_REDIS,\n\t\tmonitor.METRIC_DATABASE_TELE, 3, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_byte\", \"Data node memory usage\", monitor.METRIC_UNIT_BYTE, 1),\n\t\t})\n\n\t// cloudaccount_balance\n\tRegistryMetricCreateInput(\"cloudaccount_balance\", \"Cloud account balance\",\n\t\tmonitor.METRIC_RES_TYPE_CLOUDACCOUNT,\n\t\tmonitor.METRIC_DATABASE_METER, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"balance\", \"balance\", monitor.METRIC_UNIT_RMB, 1),\n\t\t})\n\n\t// cpu\n\tRegistryMetricCreateInput(\"agent_cpu\", \"CPU usage\", monitor.METRIC_RES_TYPE_AGENT, monitor.METRIC_DATABASE_TELE, 1,\n\t\t[]monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_active\", \"CPU active state utilization rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"usage_idle\", \"CPU idle state utilization rate\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t\tnewMetricFieldCreateInput(\"usage_system\", \"CPU system state utilization rate\", monitor.METRIC_UNIT_PERCENT, 3),\n\t\t\tnewMetricFieldCreateInput(\"usage_user\", \"CPU user mode utilization rate\", monitor.METRIC_UNIT_PERCENT, 4),\n\t\t\tnewMetricFieldCreateInput(\"usage_iowait\", \"CPU IO usage\", monitor.METRIC_UNIT_PERCENT, 5),\n\t\t\tnewMetricFieldCreateInput(\"usage_irq\", \"CPU IRQ usage\", monitor.METRIC_UNIT_PERCENT, 6),\n\t\t\tnewMetricFieldCreateInput(\"usage_guest\", \"CPU guest usage\", monitor.METRIC_UNIT_PERCENT, 7),\n\t\t\tnewMetricFieldCreateInput(\"usage_nice\", \"CPU priority switch utilization\", monitor.METRIC_UNIT_PERCENT, 8),\n\t\t\tnewMetricFieldCreateInput(\"usage_softirq\", \"CPU softirq usage\", monitor.METRIC_UNIT_PERCENT, 9),\n\t\t})\n\n\t// disk\n\tRegistryMetricCreateInput(\"agent_disk\", \"Disk usage\", monitor.METRIC_RES_TYPE_AGENT,\n\t\tmonitor.METRIC_DATABASE_TELE, 3,\n\t\t[]monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Percentage of used disks\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"free\", \"Free space size\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"used\", \"Used disk size\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t\tnewMetricFieldCreateInput(\"total\", \"Total disk size\", monitor.METRIC_UNIT_BYTE, 4),\n\t\t\tnewMetricFieldCreateInput(\"inodes_free\", \"Available inode\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"inodes_used\", \"Number of inodes used\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"inodes_total\", \"Total inodes\", monitor.METRIC_UNIT_COUNT, 7),\n\t\t})\n\n\t// diskio\n\tRegistryMetricCreateInput(\"agent_diskio\", \"Disk traffic and timing\",\n\t\tmonitor.METRIC_RES_TYPE_AGENT, monitor.METRIC_DATABASE_TELE, 4, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"read_bps\", \"Disk read rate\", monitor.METRIC_UNIT_BPS, 1),\n\t\t\tnewMetricFieldCreateInput(\"write_bps\", \"Disk write rate\", monitor.METRIC_UNIT_BPS, 2),\n\t\t\tnewMetricFieldCreateInput(\"read_iops\", \"Disk read operate rate\", monitor.METRIC_UNIT_COUNT, 3),\n\t\t\tnewMetricFieldCreateInput(\"write_iops\", \"Disk write operate rate\", monitor.METRIC_UNIT_COUNT, 4),\n\t\t\tnewMetricFieldCreateInput(\"reads\", \"Number of reads\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"writes\", \"Number of writes\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"read_bytes\", \"Bytes read\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"write_bytes\", \"Bytes write\", monitor.METRIC_UNIT_BYTE, 8),\n\t\t\tnewMetricFieldCreateInput(\"write_time\", \"Time to wait for write\", monitor.METRIC_UNIT_MS, 9),\n\t\t\tnewMetricFieldCreateInput(\"io_time\", \"I / O request queuing time\", monitor.METRIC_UNIT_MS, 10),\n\t\t\tnewMetricFieldCreateInput(\"weighted_io_time\", \"I / O request waiting time\", monitor.METRIC_UNIT_MS, 11),\n\t\t\tnewMetricFieldCreateInput(\"iops_in_progress\", \"Number of I / O requests issued but not yet completed\", monitor.METRIC_UNIT_COUNT, 12),\n\t\t})\n\n\t// mem\n\tRegistryMetricCreateInput(\"agent_mem\", \"Memory\", monitor.METRIC_RES_TYPE_AGENT,\n\t\tmonitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"used_percent\", \"Used memory rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"available_percent\", \"Available memory rate\", monitor.METRIC_UNIT_PERCENT, 2),\n\t\t\tnewMetricFieldCreateInput(\"used\", \"Used memory\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t\tnewMetricFieldCreateInput(\"free\", \"Free memory\", monitor.METRIC_UNIT_BYTE, 4),\n\t\t\tnewMetricFieldCreateInput(\"active\", \"The amount of active memory\", monitor.METRIC_UNIT_BYTE, 5),\n\t\t\tnewMetricFieldCreateInput(\"inactive\", \"The amount of inactive memory\", monitor.METRIC_UNIT_BYTE, 6),\n\t\t\tnewMetricFieldCreateInput(\"cached\", \"Cache memory\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"buffered\", \"Buffer memory\", monitor.METRIC_UNIT_BYTE, 7),\n\t\t\tnewMetricFieldCreateInput(\"slab\", \"Number of kernel caches\", monitor.METRIC_UNIT_BYTE, 8),\n\t\t\tnewMetricFieldCreateInput(\"available\", \"Available memory\", monitor.METRIC_UNIT_BYTE, 9),\n\t\t\tnewMetricFieldCreateInput(\"total\", \"Total memory\", monitor.METRIC_UNIT_BYTE, 10),\n\t\t})\n\n\t// net\n\tRegistryMetricCreateInput(\"agent_net\", \"Network interface and protocol usage\",\n\t\tmonitor.METRIC_RES_TYPE_AGENT, monitor.METRIC_DATABASE_TELE, 5, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"bytes_sent\", \"The total number of bytes sent by the network interface\", monitor.METRIC_UNIT_BYTE, 1),\n\t\t\tnewMetricFieldCreateInput(\"bytes_recv\", \"The total number of bytes received by the network interface\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"packets_sent\", \"The total number of packets sent by the network interface\", monitor.METRIC_UNIT_COUNT, 3),\n\t\t\tnewMetricFieldCreateInput(\"packets_recv\", \"The total number of packets received by the network interface\", monitor.METRIC_UNIT_COUNT, 4),\n\t\t\tnewMetricFieldCreateInput(\"err_in\", \"The total number of receive errors detected by the network interface\", monitor.METRIC_UNIT_COUNT, 5),\n\t\t\tnewMetricFieldCreateInput(\"err_out\", \"The total number of transmission errors detected by the network interface\", monitor.METRIC_UNIT_COUNT, 6),\n\t\t\tnewMetricFieldCreateInput(\"drop_in\", \"The total number of received packets dropped by the network interface\", monitor.METRIC_UNIT_COUNT, 7),\n\t\t\tnewMetricFieldCreateInput(\"drop_out\", \"The total number of transmission packets dropped by the network interface\", monitor.METRIC_UNIT_COUNT, 8),\n\t\t})\n\n\tRegistryMetricCreateInput(\"storage\", \"Storage usage\",\n\t\tmonitor.METRIC_RES_TYPE_STORAGE, monitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"usage_active\", \"Storage utilization rate\", monitor.METRIC_UNIT_PERCENT, 1),\n\t\t\tnewMetricFieldCreateInput(\"free\", \"Free storage\", monitor.METRIC_UNIT_MB, 2),\n\t\t})\n\n\t//jenkins\n\tRegistryMetricCreateInput(\"jenkins_node\", \"jenkins node\",\n\t\tmonitor.METRIC_RES_TYPE_JENKINS, monitor.METRIC_DATABASE_TELE, 1, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"disk_available\", \"disk_available\", monitor.METRIC_UNIT_BYTE, 1),\n\t\t\tnewMetricFieldCreateInput(\"temp_available\", \"temp_available\", monitor.METRIC_UNIT_BYTE, 2),\n\t\t\tnewMetricFieldCreateInput(\"memory_available\", \"memory_available\", monitor.METRIC_UNIT_BYTE, 3),\n\t\t\tnewMetricFieldCreateInput(\"memory_total\", \"memory_total\", monitor.METRIC_UNIT_BYTE, 4),\n\t\t\tnewMetricFieldCreateInput(\"swap_available\", \"swap_available\", monitor.METRIC_UNIT_BYTE, 5),\n\t\t\tnewMetricFieldCreateInput(\"swap_total\", \"swap_total\", monitor.METRIC_UNIT_BYTE, 6),\n\t\t})\n\tRegistryMetricCreateInput(\"jenkins_job\", \"jenkins job\",\n\t\tmonitor.METRIC_RES_TYPE_JENKINS, monitor.METRIC_DATABASE_TELE, 2, []monitor.MetricFieldCreateInput{\n\t\t\tnewMetricFieldCreateInput(\"duration\", \"duration\", monitor.METRIC_UNIT_MS, 1),\n\t\t\tnewMetricFieldCreateInput(\"number\", \"number\", monitor.METRIC_UNIT_COUNT, 2),\n\t\t})\n\n}", "func NewHosts(hosts ...Host) *Hosts {\n\treturn &Hosts{\n\t\thosts: hosts,\n\t\tstopped: make(chan struct{}),\n\t}\n}", "func LoadHost() *loadbalance.HostInfo {\n\thostInfo := new(loadbalance.HostInfo)\n\thostInfo.IsMultiTarget = true\n\tfor _, proxy := range FloodDataConfig.HttpProxy {\n\t\thostInfo.MultiTarget = append(hostInfo.MultiTarget, proxy.Proxypass)\n\t}\n\thostInfo.MultiTargetMode = 1\n\thostInfo.Length = len(hostInfo.MultiTarget)\n\treturn hostInfo\n}", "func NewHost(conf Config, middlewares ...Middleware) (host *Host) {\n\thost = &Host{\n\t\thandlers: map[string]*endpoint{},\n\t\tconf: conf,\n\n\t\tbasepath: \"\",\n\t\tmstack: middlewares,\n\t}\n\tif !conf.DisableAutoReport {\n\t\tos.Stdout.WriteString(\"Registration Info:\\r\\n\")\n\t}\n\thost.initCheck()\n\treturn\n}", "func NewHost(ip net.IP, hostname string, aliases ...string) Host {\n\treturn Host{\n\t\tIP: ip,\n\t\tHostname: hostname,\n\t\tAliases: aliases,\n\t}\n}", "func New() *CloudMetrics {\n\treturn &CloudMetrics{\n\t\tAPIRequestsCounter: promauto.NewCounter(prometheus.CounterOpts{\n\t\t\tNamespace: provisionerNamespace,\n\t\t\tSubsystem: provisionerSubsystemAPI,\n\t\t\tName: \"requests_total\",\n\t\t\tHelp: \"The total number of http API requests\",\n\t\t}),\n\n\t\tAPITimesHistograms: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemAPI,\n\t\t\t\tName: \"requests_duration\",\n\t\t\t\tHelp: \"The duration of http API requests\",\n\t\t\t},\n\t\t\t[]string{\"handler\", \"method\", \"status_code\"},\n\t\t),\n\n\t\tInstallationCreationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_creation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation creation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationUpdateDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_update_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation update tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationHibernationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_hibernation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation hibernation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationWakeUpDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_wakeup_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation wake up tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tClusterInstallationReconcilingDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_installation_reconciling_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster installation reconciliation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"cluster\"},\n\t\t),\n\n\t\tClusterInstallationDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_installation_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster installation deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"cluster\"},\n\t\t),\n\t\tClusterCreationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_creation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster creation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterUpgradeDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_upgrade_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster upgrade tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterProvisioningDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_provisioning_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster provisioning tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterResizeDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_resize_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster resize tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t}\n}", "func newpcpInstanceMetric(vals Instances, indom *PCPInstanceDomain, desc *pcpMetricDesc) (*pcpInstanceMetric, error) {\n\tif len(vals) != indom.InstanceCount() {\n\t\treturn nil, errors.New(\"values for all instances in the instance domain only should be passed\")\n\t}\n\n\tmvals := make(map[string]*instanceValue)\n\n\tfor name := range indom.instances {\n\t\tval, present := vals[name]\n\t\tif !present {\n\t\t\treturn nil, errors.Errorf(\"Instance %v not initialized\", name)\n\t\t}\n\n\t\tif !desc.t.IsCompatible(val) {\n\t\t\treturn nil, errors.Errorf(\"value %v is incompatible with type %v for Instance %v\", val, desc.t, name)\n\t\t}\n\n\t\tval = desc.t.resolve(val)\n\t\tmvals[name] = newinstanceValue(val)\n\t}\n\n\treturn &pcpInstanceMetric{desc, indom, mvals}, nil\n}", "func (o *HostMetrics) UnmarshalJSON(bytes []byte) (err error) {\n\traw := map[string]interface{}{}\n\tall := struct {\n\t\tCpu *float64 `json:\"cpu,omitempty\"`\n\t\tIowait *float64 `json:\"iowait,omitempty\"`\n\t\tLoad *float64 `json:\"load,omitempty\"`\n\t}{}\n\terr = json.Unmarshal(bytes, &all)\n\tif err != nil {\n\t\terr = json.Unmarshal(bytes, &raw)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\to.UnparsedObject = raw\n\t\treturn nil\n\t}\n\to.Cpu = all.Cpu\n\to.Iowait = all.Iowait\n\to.Load = all.Load\n\treturn nil\n}", "func NewProcMetrics() *ProcMetrics {\n\treturn NewProcMetricsWith(stats.DefaultEngine, os.Getpid())\n}", "func (c *Collector) LoadFromRuntimeMetrics(h *metrics.Float64Histogram) {\n\tc.Lock()\n\tdefer c.Unlock()\n\n\tc.Buckets = make([]Bucket, len(h.Buckets)-1)\n\tc.BucketsLimit = len(h.Buckets)\n\tc.Bucket = Bucket{\n\t\tMin: h.Buckets[0],\n\t\tMax: h.Buckets[0],\n\t}\n\n\tfor i, b := range h.Buckets[1:] {\n\t\tbb := Bucket{\n\t\t\tMin: c.Bucket.Max,\n\t\t\tMax: b,\n\t\t\tCount: int(h.Counts[i]),\n\t\t}\n\n\t\tif bb.Count != 0 && !math.IsInf(b, 0) {\n\t\t\tbb.Sum = float64(bb.Count) * b\n\t\t\tc.Bucket.Sum += bb.Sum\n\t\t}\n\n\t\tc.Bucket.Count += bb.Count\n\t\tc.Bucket.Max = b\n\n\t\tc.Buckets[i] = bb\n\t}\n}", "func (d *loadReporterImpl) ReportHostMetric(metrics controller.NodeMetrics) error {\n\tcontrollerClient, err := d.getControllerClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trequest := controller.NewReportNodeMetricRequest()\n\trequest.HostId = StringPtr(d.hostUUID)\n\trequest.Sku = SKUPtr(d.sku)\n\trequest.Role = RolePtr(d.role)\n\trequest.Metrics = NodeMetricsPtr(metrics)\n\trequest.Timestamp = Int64Ptr(time.Now().UnixNano())\n\n\tctx, cancel := thrift.NewContext(reportLoadMetricThriftTimeout)\n\tdefer cancel()\n\n\treturn controllerClient.ReportNodeMetric(ctx, request)\n}", "func New(resolution int) HMSketch {\n\tvar x HMSketch\n\tx.Resolution = resolution\n\tx.Registers = make([]hist.Histogram, 0)\n\tx.Index = make(map[int64]int)\n\tx = x.insert(map[string]string{\"__global__\": \"__global__\"}, 0, 0)\n\treturn x\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tcfgwarn.Beta(\"The mssql performance metricset is beta.\")\n\n\tlogger := logp.NewLogger(\"mssql.performance\").With(\"host\", base.HostData().SanitizedURI)\n\n\tdb, err := mssql.NewConnection(base.HostData().URI)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not create connection to db\")\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tlog: logger,\n\t\tdb: db,\n\t}, nil\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tsuccessScope := scope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := scope.Tagged(map[string]string{\"result\": \"fail\"})\n\ttimeoutScope := scope.Tagged(map[string]string{\"result\": \"timeout\"})\n\tapiScope := scope.SubScope(\"api\")\n\tserverScope := scope.SubScope(\"server\")\n\tplacement := scope.SubScope(\"placement\")\n\trecovery := scope.SubScope(\"recovery\")\n\n\treturn &Metrics{\n\t\tAPIEnqueueGangs: apiScope.Counter(\"enqueue_gangs\"),\n\t\tEnqueueGangSuccess: successScope.Counter(\"enqueue_gang\"),\n\t\tEnqueueGangFail: failScope.Counter(\"enqueue_gang\"),\n\n\t\tAPIDequeueGangs: apiScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangSuccess: successScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangTimeout: timeoutScope.Counter(\"dequeue_gangs\"),\n\n\t\tAPIGetPreemptibleTasks: apiScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksSuccess: successScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksTimeout: timeoutScope.Counter(\"get_preemptible_tasks\"),\n\n\t\tAPISetPlacements: apiScope.Counter(\"set_placements\"),\n\t\tSetPlacementSuccess: successScope.Counter(\"set_placements\"),\n\t\tSetPlacementFail: failScope.Counter(\"set_placements\"),\n\n\t\tAPIGetPlacements: apiScope.Counter(\"get_placements\"),\n\t\tGetPlacementSuccess: successScope.Counter(\"get_placements\"),\n\t\tGetPlacementFail: failScope.Counter(\"get_placements\"),\n\n\t\tAPILaunchedTasks: apiScope.Counter(\"launched_tasks\"),\n\n\t\tRecoverySuccess: successScope.Counter(\"recovery\"),\n\t\tRecoveryFail: failScope.Counter(\"recovery\"),\n\t\tRecoveryRunningSuccessCount: successScope.Counter(\"task_count\"),\n\t\tRecoveryRunningFailCount: failScope.Counter(\"task_count\"),\n\t\tRecoveryEnqueueFailedCount: failScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryEnqueueSuccessCount: successScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryTimer: recovery.Timer(\"running_tasks\"),\n\n\t\tPlacementQueueLen: placement.Gauge(\"placement_queue_length\"),\n\t\tPlacementFailed: placement.Counter(\"fail\"),\n\n\t\tElected: serverScope.Gauge(\"elected\"),\n\t}\n}", "func New(r *chi.Mux, log *logging.Logger, m servermetrics.Metrics) *API {\n\tapi := &API{\n\t\tmetrics: m,\n\t\tstartedAt: time.Now(),\n\t\tminuteDecValues: make(map[*dmsg.SessionCommon]uint64),\n\t\tminuteEncValues: make(map[*dmsg.SessionCommon]uint64),\n\t\tsecondDecValues: make(map[*dmsg.SessionCommon]uint64),\n\t\tsecondEncValues: make(map[*dmsg.SessionCommon]uint64),\n\t\trouter: r,\n\t}\n\tr.Use(httputil.SetLoggerMiddleware(log))\n\tr.Get(\"/health\", api.health)\n\treturn api\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tm := &Metrics{\n\t\tProcedures: map[string]*PerProcedureMetrics{},\n\t}\n\tfor _, procedure := range _procedures {\n\t\tresponseCodes := make(map[api.ResponseCode]*PerResponseCodeMetrics)\n\t\tfor _, responseCode := range api.ResponseCode_Values() {\n\t\t\tresponseCodeText, exists := _responseCodeToText[responseCode]\n\t\t\tif !exists {\n\t\t\t\tresponseCodeText = \"unknown-error\"\n\t\t\t}\n\t\t\ttag := map[string]string{\n\t\t\t\tTagProcedure: procedure,\n\t\t\t\tTagResponseCode: responseCodeText,\n\t\t\t\t// Fill empty string here so that prometheus won't panic\n\t\t\t\t// when the number of tags is changed inside subscope\n\t\t\t\tTagService: \"\",\n\t\t\t}\n\t\t\tsubscope := scope.Tagged(tag)\n\t\t\tresponseCodes[responseCode] = &PerResponseCodeMetrics{\n\t\t\t\tScope: subscope,\n\t\t\t\tCalls: subscope.Counter(MetricNameCalls),\n\t\t\t\tCallLatency: subscope.Timer(MetricNameCallLatency),\n\t\t\t}\n\t\t}\n\t\tm.Procedures[procedure] = &PerProcedureMetrics{\n\t\t\tResponseCodes: responseCodes,\n\t\t}\n\t}\n\treturn m\n}", "func NewMetrics(registry metrics.Registry) Metrics {\n\treturn &defaultMetrics{registry: registry}\n}", "func (c *HostMetricCollector) Run() (HostMetrics, error) {\n\tcpuTimes, err := cpu.Times(false)\n\tif err != nil {\n\t\t// note: can't happen on Linux. gopsutil doesn't\n\t\t// return an error\n\t\treturn HostMetrics{}, fmt.Errorf(\"cpu.Times() failed: %s\", err)\n\t}\n\tif len(cpuTimes) == 0 {\n\t\t// possible with hardware failure\n\t\treturn HostMetrics{}, fmt.Errorf(\"cpu.Times() returns no cpus\")\n\t}\n\tt := cpuTimes[0]\n\tjiffy := t.Total()\n\ttoPercent := 100 / (jiffy - c.lastJiffy)\n\n\tlastTimes := c.lastTimes\n\tc.lastJiffy = jiffy\n\tc.lastTimes = t\n\n\tconst mbSize float64 = 1024 * 1024\n\tvmem, err := mem.VirtualMemory()\n\tif err != nil {\n\t\t// only possible if can't parse numbers in /proc/meminfo\n\t\t// that would be massive failure\n\t\treturn HostMetrics{}, fmt.Errorf(\"mem.VirtualMemory() failed: %s:\", err)\n\t}\n\n\treturn HostMetrics{\n\t\tCPUUser: ((t.User + t.Nice) - (lastTimes.User + lastTimes.Nice)) * toPercent,\n\t\tCPUSystem: ((t.System + t.Irq + t.Softirq) - (lastTimes.System + lastTimes.Irq + lastTimes.Softirq)) * toPercent,\n\t\tCPUIowait: (t.Iowait - lastTimes.Iowait) * toPercent,\n\t\tCPUIdle: (t.Idle - lastTimes.Idle) * toPercent,\n\t\tCPUStolen: (t.Steal - lastTimes.Steal) * toPercent,\n\t\tCPUGuest: (t.Guest - lastTimes.Guest) * toPercent,\n\t\tMemTotal: float64(vmem.Total) / mbSize,\n\t\tMemFree: float64(vmem.Free) / mbSize,\n\t\tMemUsed: float64(vmem.Total-vmem.Free) / mbSize,\n\t\tMemUsable: float64(vmem.Available) / mbSize,\n\t\tMemPctUsable: float64(100-vmem.UsedPercent) / 100,\n\t}, nil\n}", "func New(d *docker.Docker, containersInterval, metricsInterval time.Duration) *Metrics {\n\treturn &Metrics{\n\t\tdocker: d,\n\t\tmetricsMap: &metricsMap{\n\t\t\tmetrics: make(map[string]*docker.ContainerStats),\n\t\t},\n\t\tcInterval: containersInterval,\n\t\tmInterval: metricsInterval,\n\t}\n}", "func NewMetrics(healthyChan chan bool, config *openapi.AlgoRunnerConfig) Metrics {\n\n\tgo func() {\n\t\tfor h := range healthyChan {\n\t\t\thealthy = h\n\t\t}\n\t}()\n\n\tregisterMetrics(config)\n\n\treturn Metrics{\n\t\tRunnerRuntimeHistogram: runnerRuntimeHistogram,\n\t\tAlgoRuntimeHistogram: algoRuntimeHistogram,\n\t\tMsgBytesInputCounter: msgBytesInputCounter,\n\t\tMsgBytesOutputCounter: msgBytesOutputCounter,\n\t\tDataBytesInputCounter: dataBytesInputCounter,\n\t\tDataBytesOutputCounter: dataBytesOutputCounter,\n\t\tRetryCounter: retryCounter,\n\t\tDlqCounter: dlqCounter,\n\t\tAlgoErrorCounter: algoErrorCounter,\n\t\tRunnerErrorCounter: runnerErrorCounter,\n\t\tMsgOK: msgOK,\n\t\tMsgNOK: msgNOK,\n\t\tMsgDropped: msgDropped,\n\t\tProducerQueueLen: &producerQueueLen,\n\t\tEventIgnored: eventIgnored,\n\t\tMsgInTransit: msgInTransit,\n\t\tLibRdKafkaVersion: libRdKafkaVersion,\n\t\tLastProducerStartTime: &lastProducerStartTime,\n\t\tMetricCertExpirationTime: &metricCertExpirationTime,\n\t\tMetricCaExpirationTime: &metricCaExpirationTime,\n\t\tMetricKafkaEventsQueueLen: &metricKafkaEventsQueueLen,\n\t\tMetricRDKafkaGlobal: metricRDKafkaGlobal,\n\t\tMetricRDKafkaBroker: metricRDKafkaBroker,\n\t\tMetricRDKafkaTopic: metricRDKafkaTopic,\n\t\tMetricRDKafkaPartition: metricRDKafkaPartition,\n\n\t\tDeploymentLabel: deploymentLabel,\n\t\tPipelineLabel: pipelineLabel,\n\t\tComponentLabel: componentLabel,\n\t\tAlgoLabel: algoLabel,\n\t\tAlgoVersionLabel: algoVersionLabel,\n\t\tAlgoIndexLabel: algoIndexLabel,\n\t}\n}", "func newMaglevLoadBalancer(info types.ClusterInfo, set types.HostSet) types.LoadBalancer {\n\tnames := []string{}\n\tfor _, host := range set.Hosts() {\n\t\tnames = append(names, host.AddressString())\n\t}\n\tmgv := &maglevLoadBalancer{\n\t\thosts: set,\n\t}\n\n\tnameCount := len(names)\n\t// if host count > BigM, maglev table building will cross array boundary\n\t// maglev lb will not work in this scenario\n\tif nameCount >= maglev.BigM {\n\t\tlog.DefaultLogger.Errorf(\"[lb][maglev] host count too large, expect <= %d, get %d\",\n\t\t\tmaglev.BigM, nameCount)\n\t\treturn mgv\n\t}\n\tif nameCount == 0 {\n\t\treturn mgv\n\t}\n\n\tmaglevM := maglev.SmallM\n\t// according to test, 30000 host with testing 1e8 times, hash distribution begins to go wrong,\n\t// max=4855, mean=3333.3333333333335, peak-to-mean=1.4565\n\t// so use BigM when host >= 30000\n\tlimit := 30000\n\tif nameCount >= limit {\n\t\tlog.DefaultLogger.Infof(\"[lb][maglev] host count %d >= %d, using maglev.BigM\", nameCount, limit)\n\t\tmaglevM = maglev.BigM\n\t}\n\n\tmgv.maglev = maglev.New(names, uint64(maglevM))\n\treturn mgv\n}", "func NewHost(conf config.Config) CurrentHost {\n\treturn &currentHostInfo{\n\t\thostName: conf.HostName,\n\t\tnextHostAddress: conf.NextHostAddress,\n\t\tnextHostPort: conf.NextHostPort,\n\t}\n}", "func (h *singleLhcNodeHarness) getMetrics() *metrics {\n\treturn &metrics{\n\t\ttimeSinceLastCommitMillis: h.metricRegistry.Get(\"ConsensusAlgo.LeanHelix.TimeSinceLastCommit.Millis\").(*metric.Histogram),\n\t\ttimeSinceLastElectionMillis: h.metricRegistry.Get(\"ConsensusAlgo.LeanHelix.TimeSinceLastElection.Millis\").(*metric.Histogram),\n\t\tcurrentElectionCount: h.metricRegistry.Get(\"ConsensusAlgo.LeanHelix.CurrentElection.Number\").(*metric.Gauge),\n\t\tcurrentLeaderMemberId: h.metricRegistry.Get(\"ConsensusAlgo.LeanHelix.CurrentLeaderMemberId.Number\").(*metric.Text),\n\t\tlastCommittedTime: h.metricRegistry.Get(\"ConsensusAlgo.LeanHelix.LastCommitted.TimeNano\").(*metric.Gauge),\n\t}\n}", "func NewMetrics(scope tally.Scope) Metrics {\n\tscope = scope.SubScope(\"aggregation\")\n\treturn Metrics{\n\t\tCounter: newCounterMetrics(scope.SubScope(\"counters\")),\n\t\tGauge: newGaugeMetrics(scope.SubScope(\"gauges\")),\n\t}\n}", "func NewHost(addr Address, peerCount, channelLimit uint64, incomingBandwidth, outgoingBandwidth uint32) (Host, error) {\n\tvar cAddr *C.struct__ENetAddress\n\tif addr != nil {\n\t\tcAddr = &(addr.(*enetAddress)).cAddr\n\t}\n\n\thost := C.enet_host_create(\n\t\tcAddr,\n\t\t(C.size_t)(peerCount),\n\t\t(C.size_t)(channelLimit),\n\t\t(C.enet_uint32)(incomingBandwidth),\n\t\t(C.enet_uint32)(outgoingBandwidth),\n\t)\n\n\tif host == nil {\n\t\treturn nil, errors.New(\"unable to create host\")\n\t}\n\n\treturn &enetHost{\n\t\tcHost: host,\n\t}, nil\n}", "func newMetrics() metrics {\n\treturn metrics{\n\t\tsize: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"pool_size\",\n\t\t\t\tHelp: \"Size of pool\",\n\t\t\t},\n\t\t),\n\n\t\tstatus: prometheus.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"pool_status\",\n\t\t\t\tHelp: `Status of pool (0, 1, 2, 3, 4, 5, 6)= {\"Offline\", \"Online\", \"Degraded\", \"Faulted\", \"Removed\", \"Unavail\", \"NoPoolsAvailable\"}`,\n\t\t\t},\n\t\t\t[]string{\"pool\"},\n\t\t),\n\n\t\tusedCapacity: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"used_pool_capacity\",\n\t\t\t\tHelp: \"Capacity used by pool\",\n\t\t\t},\n\t\t),\n\n\t\tfreeCapacity: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"free_pool_capacity\",\n\t\t\t\tHelp: \"Free capacity in pool\",\n\t\t\t},\n\t\t),\n\n\t\tusedCapacityPercent: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"used_pool_capacity_percent\",\n\t\t\t\tHelp: \"Capacity used by pool in percent\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolListparseErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_list_parse_error_count\",\n\t\t\t\tHelp: \"Total no of parsing errors\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolRejectRequestCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_reject_request_count\",\n\t\t\t\tHelp: \"Total no of rejected requests of zpool command\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolCommandErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_command_error\",\n\t\t\t\tHelp: \"Total no of zpool command errors\",\n\t\t\t},\n\t\t),\n\n\t\tnoPoolAvailableErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"no_pool_available_error\",\n\t\t\t\tHelp: \"Total no of no pool available errors\",\n\t\t\t},\n\t\t),\n\n\t\tincompleteOutputErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_list_incomplete_stdout_error\",\n\t\t\t\tHelp: \"Total no of incomplete stdout of zpool list command errors\",\n\t\t\t},\n\t\t),\n\t}\n}", "func newMonitoringMetrics() monitoringMetrics {\n\treturn monitoringMetrics{\n\t\tCSR: csrCounts.With(prometheus.Labels{}),\n\t\tAuthnError: authnErrorCounts.With(prometheus.Labels{}),\n\t\tSuccess: successCounts.With(prometheus.Labels{}),\n\t\tCSRError: csrParsingErrorCounts.With(prometheus.Labels{}),\n\t\tIDExtractionError: idExtractionErrorCounts.With(prometheus.Labels{}),\n\t\tcertSignErrors: certSignErrorCounts,\n\t}\n}", "func (p *Proxy) NewHost(c *exec.Cmd) (*Host, error) {\n\th := &Host{\n\t\tcmd: c,\n\t\tproxy: p,\n\t}\n\tvar err error\n\th.httpTransfer, h.httpsTransfer, err = h.setupCmd(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn h, nil\n}", "func newCustomMetrics() ICustomMetrics {\n\n\tcounters := make(map[string]prometheus.Counter)\n\tgauges := make(map[string]prometheus.Gauge)\n\tsummaries := make(map[string]prometheus.Summary)\n\thistograms := make(map[string]prometheus.Histogram)\n\n\treturn &customMetrics{\n\t\tcounters: counters,\n\t\tgauges: gauges,\n\t\tsummaries: summaries,\n\t\thistograms: histograms,\n\t}\n}", "func NewHost(uri string) Host {\n\t// no need to decompose uri using net/url package\n\treturn Host{uri: uri, client: http.Client{}}\n}", "func NewHost(config v2.Host, clusterInfo types.ClusterInfo) types.Host {\n\taddr, _ := net.ResolveTCPAddr(\"tcp\", config.Address)\n\n\treturn &host{\n\t\thostInfo: newHostInfo(addr, config, clusterInfo),\n\t\tweight: config.Weight,\n\t}\n}", "func newMetricsMetadata(cluster *string, containerInstance *string) *ecstcs.MetricsMetadata {\n\treturn &ecstcs.MetricsMetadata{\n\t\tCluster: cluster,\n\t\tContainerInstance: containerInstance,\n\t}\n}", "func NewHost(ctx *pulumi.Context,\n\tname string, args *HostArgs, opts ...pulumi.ResourceOption) (*Host, error) {\n\tif args == nil || args.Hostname == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Hostname'\")\n\t}\n\tif args == nil || args.Password == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Password'\")\n\t}\n\tif args == nil || args.Username == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Username'\")\n\t}\n\tif args == nil {\n\t\targs = &HostArgs{}\n\t}\n\tvar resource Host\n\terr := ctx.RegisterResource(\"vsphere:index/host:Host\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tlogp.Warn(\"EXPERIMENTAL: The haproxy info metricset is experimental\")\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tstatsAddr: base.Host(),\n\t}, nil\n}", "func newMetricsCollector(devices func() ([]*wgtypes.Device, error)) prometheus.Collector {\n\t// common labels for all metrics\n\tlabels := []string{\"device\", \"public_key\"}\n\n\treturn &collector{\n\t\tDeviceInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_device_info\",\n\t\t\t\"Metadata about a device.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_info\",\n\t\t\t\"Metadata about a peer. The public_key label on peer metrics refers to the peer's public key; not the device's public key.\",\n\t\t\tappend(labels, []string{\"endpoint\"}...),\n\t\t\tnil,\n\t\t),\n\t\tPeerAllowedIPsInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_allowed_ips_info\",\n\t\t\t\"Metadata about each of a peer's allowed IP subnets for a given device.\",\n\t\t\tappend(labels, []string{\"allowed_ips\"}...),\n\t\t\tnil,\n\t\t),\n\t\tPeerReceiveBytes: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_receive_bytes_total\",\n\t\t\t\"Number of bytes received from a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerTransmitBytes: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_transmit_bytes_total\",\n\t\t\t\"Number of bytes transmitted to a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerLastHandshake: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_last_handshake_seconds\",\n\t\t\t\"UNIX timestamp for the last handshake with a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tdevices: devices,\n\t}\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tRangeFeedCatchupScanNanos: metric.NewCounter(metaRangeFeedCatchupScanNanos),\n\t\tRangeFeedSlowClosedTimestampLogN: log.Every(5 * time.Second),\n\t\tRangeFeedSlowClosedTimestampNudgeSem: make(chan struct{}, 1024),\n\t}\n}", "func newListMetrics() *listMetrics {\n\treturn new(listMetrics)\n}", "func NewPrometheusMetrics(namespace string, registry metrics.RegisterGatherer) *prometheusMetrics {\n\tm := &prometheusMetrics{\n\t\tregistry: registry,\n\t}\n\n\tm.AvailableIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_ips\",\n\t\tHelp: \"Total available IPs on Node for IPAM allocation\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.UsedIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"used_ips\",\n\t\tHelp: \"Total used IPs on Node for IPAM allocation\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.NeededIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"needed_ips\",\n\t\tHelp: \"Number of IPs that are needed on the Node to satisfy IPAM allocation requests\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.IPsAllocated = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ips\",\n\t\tHelp: \"Number of IPs allocated\",\n\t}, []string{\"type\"})\n\n\tm.AllocateIpOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ip_allocation_ops\",\n\t\tHelp: \"Number of IP allocation operations\",\n\t}, []string{\"subnet_id\"})\n\n\tm.ReleaseIpOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ip_release_ops\",\n\t\tHelp: \"Number of IP release operations\",\n\t}, []string{\"subnet_id\"})\n\n\tm.AllocateInterfaceOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"interface_creation_ops\",\n\t\tHelp: \"Number of interfaces allocated\",\n\t}, []string{\"subnet_id\"})\n\n\tm.AvailableInterfaces = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_interfaces\",\n\t\tHelp: \"Number of interfaces with addresses available\",\n\t})\n\n\tm.InterfaceCandidates = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"interface_candidates\",\n\t\tHelp: \"Number of attached interfaces with IPs available for allocation\",\n\t})\n\n\tm.EmptyInterfaceSlots = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"empty_interface_slots\",\n\t\tHelp: \"Number of empty interface slots available for interfaces to be attached\",\n\t})\n\n\tm.AvailableIPsPerSubnet = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_ips_per_subnet\",\n\t\tHelp: \"Number of available IPs per subnet ID\",\n\t}, []string{\"subnet_id\", \"availability_zone\"})\n\n\tm.Nodes = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"nodes\",\n\t\tHelp: \"Number of nodes by category { total | in-deficit | at-capacity }\",\n\t}, []string{\"category\"})\n\n\tm.Resync = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"resync_total\",\n\t\tHelp: \"Number of resync operations to synchronize and resolve IP deficit of nodes\",\n\t})\n\n\tm.Allocation = prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"allocation_duration_seconds\",\n\t\tHelp: \"Allocation ip or interface latency in seconds\",\n\t\tBuckets: merge(\n\t\t\tprometheus.LinearBuckets(0.25, 0.25, 2), // 0.25s, 0.50s\n\t\t\tprometheus.LinearBuckets(1, 1, 60), // 1s, 2s, 3s, ... 60s,\n\t\t),\n\t}, []string{\"type\", \"status\", \"subnet_id\"})\n\n\tm.Release = prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"release_duration_seconds\",\n\t\tHelp: \"Release ip or interface latency in seconds\",\n\t\tBuckets: merge(\n\t\t\tprometheus.LinearBuckets(0.25, 0.25, 2), // 0.25s, 0.50s\n\t\t\tprometheus.LinearBuckets(1, 1, 60), // 1s, 2s, 3s, ... 60s,\n\t\t),\n\t}, []string{\"type\", \"status\", \"subnet_id\"})\n\n\t// pool_maintainer is a more generic name, but for backward compatibility\n\t// of dashboard, keep the metric name deficit_resolver unchanged\n\tm.poolMaintainer = NewTriggerMetrics(namespace, \"deficit_resolver\")\n\tm.k8sSync = NewTriggerMetrics(namespace, \"k8s_sync\")\n\tm.resync = NewTriggerMetrics(namespace, \"resync\")\n\n\tregistry.MustRegister(m.AvailableIPs)\n\tregistry.MustRegister(m.UsedIPs)\n\tregistry.MustRegister(m.NeededIPs)\n\n\tregistry.MustRegister(m.IPsAllocated)\n\tregistry.MustRegister(m.AllocateIpOps)\n\tregistry.MustRegister(m.ReleaseIpOps)\n\tregistry.MustRegister(m.AllocateInterfaceOps)\n\tregistry.MustRegister(m.AvailableInterfaces)\n\tregistry.MustRegister(m.InterfaceCandidates)\n\tregistry.MustRegister(m.EmptyInterfaceSlots)\n\tregistry.MustRegister(m.AvailableIPsPerSubnet)\n\tregistry.MustRegister(m.Nodes)\n\tregistry.MustRegister(m.Resync)\n\tregistry.MustRegister(m.Allocation)\n\tregistry.MustRegister(m.Release)\n\tm.poolMaintainer.Register(registry)\n\tm.k8sSync.Register(registry)\n\tm.resync.Register(registry)\n\n\treturn m\n}", "func NewProcMetricsWith(eng *stats.Engine, pid int) *ProcMetrics {\n\tp := &ProcMetrics{engine: eng, pid: pid}\n\n\tp.cpu.user.typ = \"user\"\n\tp.cpu.system.typ = \"system\"\n\n\tp.memory.resident.typ = \"resident\"\n\tp.memory.shared.typ = \"shared\"\n\tp.memory.text.typ = \"text\"\n\tp.memory.data.typ = \"data\"\n\n\tp.memory.pagefault.major.typ = \"major\"\n\tp.memory.pagefault.minor.typ = \"minor\"\n\n\tp.threads.switches.voluntary.typ = \"voluntary\"\n\tp.threads.switches.involuntary.typ = \"involuntary\"\n\n\treturn p\n}", "func (t *OpenconfigSystem_System_Dns_HostEntries) NewHostEntry(Hostname string) (*OpenconfigSystem_System_Dns_HostEntries_HostEntry, error){\n\n\t// Initialise the list within the receiver struct if it has not already been\n\t// created.\n\tif t.HostEntry == nil {\n\t\tt.HostEntry = make(map[string]*OpenconfigSystem_System_Dns_HostEntries_HostEntry)\n\t}\n\n\tkey := Hostname\n\n\t// Ensure that this key has not already been used in the\n\t// list. Keyed YANG lists do not allow duplicate keys to\n\t// be created.\n\tif _, ok := t.HostEntry[key]; ok {\n\t\treturn nil, fmt.Errorf(\"duplicate key %v for list HostEntry\", key)\n\t}\n\n\tt.HostEntry[key] = &OpenconfigSystem_System_Dns_HostEntries_HostEntry{\n\t\tHostname: &Hostname,\n\t}\n\n\treturn t.HostEntry[key], nil\n}", "func NewHostSet(hkr renter.HostKeyResolver, currentHeight types.BlockHeight) *HostSet {\n\treturn &HostSet{\n\t\thkr: hkr,\n\t\tcurrentHeight: currentHeight,\n\t\tsessions: make(map[hostdb.HostPublicKey]*lockedHost),\n\t\tlockTimeout: 10 * time.Second,\n\t\tonConnect: func(*proto.Session) {},\n\t}\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tcfgwarn.Beta(\"The kvm dommemstat metricset is beta.\")\n\tu, err := url.Parse(base.HostData().URI)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tTimeout: base.Module().Config().Timeout,\n\t\tHostURL: u,\n\t}, nil\n}", "func NewHostHandler() *HostHandler {\n\th := &HostHandler{\n\t\teligibleHosts: make(map[string]*http.ServeMux),\n\t}\n\n\treturn h\n}", "func NewHostsManager(path string) (*HostsManager, error) {\n\th := &HostsManager{\n\t\thosts: make(map[string]string),\n\t\tpath: path,\n\t}\n\t// if hosts file doesnt exist, return\n\tif _, err := os.Stat(path); os.IsNotExist(err) {\n\t\tlog.Printf(\"[WARN] writing a new hosts file at: %s\", path)\n\t\treturn h, nil\n\t}\n\t// parse hosts file\n\tinput, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlines := strings.Split(string(input), \"\\n\")\n\tfor _, l := range lines {\n\t\tx := strings.Split(l, \" \")\n\t\tif len(x) < 2 {\n\t\t\tcontinue\n\t\t}\n\t\tip := strings.TrimSpace(x[0])\n\t\thostname := strings.TrimSpace(x[1])\n\t\tif len(x) == 3 {\n\t\t\thostname = x[2]\n\t\t}\n\t\th.hosts[ip] = hostname\n\t}\n\treturn h, nil\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\treadyScope := scope.SubScope(\"ready\")\n\ttrackerScope := scope.SubScope(\"tracker\")\n\ttaskStateScope := scope.SubScope(\"tasks_state\")\n\n\treconcilerScope := scope.SubScope(\"reconciler\")\n\tleakScope := reconcilerScope.SubScope(\"leaks\")\n\tsuccessScope := reconcilerScope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := reconcilerScope.Tagged(map[string]string{\"result\": \"fail\"})\n\n\treturn &Metrics{\n\t\tReadyQueueLen: readyScope.Gauge(\"ready_queue_length\"),\n\t\tTasksCountInTracker: trackerScope.Gauge(\"task_len_tracker\"),\n\t\tTaskStatesGauge: map[task.TaskState]tally.Gauge{\n\t\t\ttask.TaskState_PENDING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_pending\"),\n\t\t\ttask.TaskState_READY: taskStateScope.Gauge(\n\t\t\t\t\"task_state_ready\"),\n\t\t\ttask.TaskState_PLACING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placing\"),\n\t\t\ttask.TaskState_PLACED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placed\"),\n\t\t\ttask.TaskState_LAUNCHING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launching\"),\n\t\t\ttask.TaskState_LAUNCHED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launched\"),\n\t\t\ttask.TaskState_RUNNING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_running\"),\n\t\t\ttask.TaskState_SUCCEEDED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_succeeded\"),\n\t\t\ttask.TaskState_FAILED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_failed\"),\n\t\t\ttask.TaskState_KILLED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_killed\"),\n\t\t\ttask.TaskState_LOST: taskStateScope.Gauge(\n\t\t\t\t\"task_state_lost\"),\n\t\t\ttask.TaskState_PREEMPTING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_preempting\"),\n\t\t},\n\t\tResourcesHeldByTaskState: map[task.TaskState]scalar.GaugeMaps{\n\t\t\ttask.TaskState_READY: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_ready\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placing\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placed\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launching\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launched\"),\n\t\t\t),\n\t\t\ttask.TaskState_RUNNING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_running\"),\n\t\t\t),\n\t\t\ttask.TaskState_STARTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_starting\"),\n\t\t\t),\n\t\t\ttask.TaskState_PREEMPTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_preempting\"),\n\t\t\t),\n\t\t\ttask.TaskState_KILLING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_killing\"),\n\t\t\t),\n\t\t},\n\t\tLeakedResources: scalar.NewGaugeMaps(leakScope),\n\t\tReconciliationSuccess: successScope.Counter(\"run\"),\n\t\tReconciliationFail: failScope.Counter(\"run\"),\n\t\tOrphanTasks: scope.Gauge(\"orphan_tasks\"),\n\t}\n}", "func NewBaseMetrics() (*BaseMetrics, error) {\n\tm, err := metrics.NewBaseMemoryStats(nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &BaseMetrics{innerMetrics: m}, nil\n}", "func New(config *Config) (*Metrics, error) {\n\tm := &Metrics{\n\t\tconfig: config,\n\t\tcounters: make(map[string]prometheus.Counter),\n\t\tcounterVecs: make(map[string]*prometheus.CounterVec),\n\t\tgauges: make(map[string]prometheus.Gauge),\n\t\tgaugeVecs: make(map[string]*prometheus.GaugeVec),\n\t}\n\n\tif config.Enable {\n\t\tgo func() {\n\t\t\thttp.Handle(\"/metrics\", promhttp.Handler())\n\t\t\terr := http.ListenAndServe(config.Addr, http.DefaultServeMux)\n\t\t\tlog.Error().Err(err).Msg(\"could not start metrics HTTP server\")\n\t\t}()\n\t}\n\n\treturn m, nil\n}", "func NewMetricsWrapper(ctx context.Context, base api.Backend) api.Backend {\n\tmetricsOnce.Do(func() {\n\t\tprometheus.MustRegister(registeryCollectors...)\n\t})\n\n\t// XXX: When the registry backends support node deregistration,\n\t// handle this on the metrics side.\n\n\twrapper := &metricsWrapper{\n\t\tBackend: base,\n\t\tcloseCh: make(chan struct{}),\n\t\tclosedCh: make(chan struct{}),\n\t}\n\n\twrapper.updatePeriodicMetrics(ctx)\n\tgo wrapper.worker(ctx)\n\n\treturn wrapper\n}", "func NewMetricsClient(credential azcore.TokenCredential, options *MetricsClientOptions) (*MetricsClient, error) {\n\tif options == nil {\n\t\toptions = &MetricsClientOptions{}\n\t}\n\tif reflect.ValueOf(options.Cloud).IsZero() {\n\t\toptions.Cloud = cloud.AzurePublic\n\t}\n\tc, ok := options.Cloud.Services[ServiceNameMetrics]\n\tif !ok || c.Audience == \"\" || c.Endpoint == \"\" {\n\t\treturn nil, errors.New(\"provided Cloud field is missing Azure Monitor Metrics configuration\")\n\t}\n\n\tauthPolicy := runtime.NewBearerTokenPolicy(credential, []string{c.Audience + \"/.default\"}, nil)\n\tpl := runtime.NewPipeline(moduleName, version, runtime.PipelineOptions{PerRetry: []policy.Policy{authPolicy}}, &options.ClientOptions)\n\treturn &MetricsClient{host: c.Endpoint, pl: pl}, nil\n}", "func NewMetrics(name string, r prometheus.Registerer) *Metrics {\n\treg := prometheus.WrapRegistererWith(prometheus.Labels{\"controller\": name}, r)\n\tm := Metrics{\n\t\treg: reg,\n\t\treconcileErrorsCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_reconcile_errors_total\",\n\t\t\tHelp: \"Number of errors that occurred while reconciling the statefulset\",\n\t\t}),\n\t\ttriggerByCounter: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_triggered_total\",\n\t\t\tHelp: \"Number of times a Kubernetes object add, delete or update event\" +\n\t\t\t\t\" triggered the Prometheus Operator to reconcile an object\",\n\t\t}, []string{\"triggered_by\", \"action\"}),\n\t\tstsDeleteCreateCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_reconcile_sts_delete_create_total\",\n\t\t\tHelp: \"Number of times that reconciling a statefulset required deleting and re-creating it\",\n\t\t}),\n\t\tlistCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_list_operations_total\",\n\t\t\tHelp: \"Total number of list operations\",\n\t\t}),\n\t\tlistFailedCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_list_operations_failed_total\",\n\t\t\tHelp: \"Total number of list operations that failed\",\n\t\t}),\n\t\twatchCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_watch_operations_total\",\n\t\t\tHelp: \"Total number of watch operations\",\n\t\t}),\n\t\twatchFailedCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_watch_operations_failed_total\",\n\t\t\tHelp: \"Total number of watch operations that failed\",\n\t\t}),\n\t}\n\tm.reg.MustRegister(\n\t\tm.reconcileErrorsCounter,\n\t\tm.triggerByCounter,\n\t\tm.stsDeleteCreateCounter,\n\t\tm.listCounter,\n\t\tm.listFailedCounter,\n\t\tm.watchCounter,\n\t\tm.watchFailedCounter,\n\t)\n\treturn &m\n}", "func New(issueser getIssueser, metricser metricser, queries map[string]string) *Monitoring {\n\tlastActiveIssues := make(map[string]map[string]model.Issue)\n\tfor queryName := range queries {\n\t\tlastActiveIssues[queryName] = make(map[string]model.Issue)\n\t}\n\n\treturn &Monitoring{\n\t\tissueser: issueser,\n\t\tmetricser: metricser,\n\t\tlastActiveIssues: lastActiveIssues,\n\t\tqueries: queries,\n\t}\n}", "func New() *Metrics {\n\tm := &Metrics{\n\t\tBuildInfo: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\tNamespace: Namespace,\n\t\t\tSubsystem: Subsystem,\n\t\t\tName: \"build_info\",\n\t\t\tHelp: \"Build information\",\n\t\t}, []string{\"version\"}),\n\t}\n\n\t_ = prometheus.Register(m.BuildInfo)\n\t// TODO: implement metrics\n\treturn m\n}", "func newDatabaseMetrics(db *database, opts *pluginOpts) (*databaseMetrics, error) {\n\tgauges, err := newDatabaseGauges(opts.prometheusNamespace)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not create database gauges\")\n\t}\n\n\treturn &databaseMetrics{\n\t\tgauges: gauges,\n\t\tdb: db,\n\t}, nil\n}", "func New(agents []config.Agent) *SamplerateCache {\n\tc := &SamplerateCache{\n\t\tcache: make(map[string]uint64),\n\t}\n\n\t// Initialize cache with configured samplerates\n\tfor _, a := range agents {\n\t\tc.Set(net.ParseIP(a.IPAddress), a.SampleRate)\n\t}\n\n\treturn c\n}", "func (m *podMetrics) New() runtime.Object {\n\treturn &metrics.PodMetrics{}\n}", "func NewMetric(name string, prog string, kind Kind, keys ...string) *Metric {\n\tm := &Metric{Name: name, Program: prog, Kind: kind,\n\t\tKeys: make([]string, len(keys), len(keys)),\n\t\tLabelValues: make([]*LabelValue, 0)}\n\tcopy(m.Keys, keys)\n\treturn m\n}", "func New(opts ...Option) *Metric {\n\tvar options Options\n\tfor _, opt := range opts {\n\t\topt(&options)\n\t}\n\tm := &Metric{\n\t\tOptions: options,\n\t\thistograms: make(map[string]metrics.Histogram),\n\t\tkeyLabels: make(map[string]map[string]string),\n\t}\n\tgo m.watch()\n\treturn m\n}" ]
[ "0.76299536", "0.6583562", "0.6368021", "0.61275035", "0.56682813", "0.5607558", "0.560199", "0.5576542", "0.5574288", "0.5560153", "0.5492024", "0.5471501", "0.5465395", "0.5453144", "0.54210395", "0.54023623", "0.5395166", "0.5385599", "0.5382379", "0.53610957", "0.53390855", "0.53325385", "0.5317617", "0.5303308", "0.52897185", "0.52768016", "0.52371687", "0.5214042", "0.5213923", "0.52124524", "0.52036554", "0.5199035", "0.51946104", "0.51815075", "0.5151227", "0.5106754", "0.50944483", "0.50898", "0.50804466", "0.50703", "0.5068172", "0.505557", "0.50413334", "0.50393707", "0.50378495", "0.5023074", "0.50094235", "0.50039274", "0.49813506", "0.49795344", "0.49792725", "0.4949894", "0.49413985", "0.49305", "0.4926857", "0.49249586", "0.49186024", "0.49098414", "0.49058768", "0.48888955", "0.4886473", "0.48856276", "0.4872623", "0.4869975", "0.4864273", "0.4862562", "0.4862041", "0.48512885", "0.4848975", "0.4840396", "0.48251125", "0.48197985", "0.48183823", "0.4813721", "0.4810861", "0.480386", "0.47976995", "0.47847676", "0.4769453", "0.4759886", "0.47577572", "0.4747057", "0.47316375", "0.4728699", "0.47275737", "0.4727436", "0.47140616", "0.4712569", "0.47089505", "0.47004253", "0.4698582", "0.46869105", "0.46852532", "0.46840855", "0.46803564", "0.46711436", "0.4668075", "0.46659538", "0.46633795" ]
0.75794894
1
NewCGMetrics returns a new instance of CGMetrics. Should be one instance per consumer group UUID
func NewCGMetrics() *CGMetrics { return &CGMetrics{ CounterBank: common.NewCounterBank(numCGMetrics), } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (consumer.Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseConsumer: newBaseConsumer(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseImpl: newBaseImpl(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func NewMetrics(ctx context.Context, output string, tenant string, refreshRate time.Duration) Metrics {\n\treturn Metrics{\n\t\tDaemonSupport: utils.NewDaemonSupport(ctx, \"metrics\"),\n\t\tstorage: localfs.NewPlaintextStorage(output),\n\t\ttenant: tenant,\n\t\trefreshRate: refreshRate,\n\t\tpromisesAccepted: metrics.NewCounter(),\n\t\tcommitsAccepted: metrics.NewCounter(),\n\t\trollbacksAccepted: metrics.NewCounter(),\n\t\tcreatedAccounts: metrics.NewCounter(),\n\t\tupdatedSnapshots: metrics.NewMeter(),\n\t\tsnapshotCronLatency: metrics.NewTimer(),\n\t}\n}", "func newMetrics() *metrics {\n\treturn new(metrics)\n}", "func newMetrics() *Metrics {\n\treturn newMetricsFrom(DefaultMetricsOpts)\n}", "func NewMetrics(period time.Duration, maxQueueSize int) (*Metrics, error) {\n\tmetrics := &Metrics{\n\t\tmaxQueueSize: maxQueueSize,\n\t\tperiod: period,\n\t\tinitialized: true,\n\t\tqueue: make([]Measurement, 0),\n\t\tlastSendingDate: -1,\n\t}\n\n\tif UseGlobalEngine {\n\t\tmetrics.Engine = Engine\n\t} else {\n\t\tmetrics.Engine = &req.Engine{}\n\t}\n\n\terr := validateMetrics(metrics)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif sources == nil {\n\t\tsources = make([]DataSource, 0)\n\t\tgo sendingLoop()\n\t}\n\n\tsources = append(sources, metrics)\n\n\treturn metrics, nil\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{items: make(map[string]*metric), rm: &sync.RWMutex{}}\n}", "func newMetrics(hostAndPort string) *metrics {\n\tm := metrics{\n\t\tmetricsCh: make(chan metricType),\n\t\thostAndPort: hostAndPort,\n\t}\n\n\treturn &m\n}", "func New(d *docker.Docker, containersInterval, metricsInterval time.Duration) *Metrics {\n\treturn &Metrics{\n\t\tdocker: d,\n\t\tmetricsMap: &metricsMap{\n\t\t\tmetrics: make(map[string]*docker.ContainerStats),\n\t\t},\n\t\tcInterval: containersInterval,\n\t\tmInterval: metricsInterval,\n\t}\n}", "func NewMetrics(subsystem string) *Metrics {\n\tbase := metrics.NewBase(subsystem, \"\")\n\treturn &Metrics{\n\t\tBase: base,\n\t\tFailedDatabaseMethods: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"failed_database_operations\",\n\t\t\tHelp: \"Tracks the number of database failures\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}, []string{methodLabel}),\n\t\tDepositIDMismatch: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"deposit_id_mismatch\",\n\t\t\tHelp: \"Set to 1 when the postgres and the disrburser contract \" +\n\t\t\t\t\"disagree on the next deposit id, and 0 otherwise\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tMissingDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"missing_disbursements\",\n\t\t\tHelp: \"Number of deposits that are missing disbursements in \" +\n\t\t\t\t\"postgres below our supposed next deposit id\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tSuccessfulDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"successful_disbursements\",\n\t\t\tHelp: \"Number of disbursements that emit a success event \" +\n\t\t\t\t\"from a given tx\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tFailedDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"failed_disbursements\",\n\t\t\tHelp: \"Number of disbursements that emit a failed event \" +\n\t\t\t\t\"from a given tx\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tPostgresLastDisbursedID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"postgres_last_disbursed_id\",\n\t\t\tHelp: \"Latest recorded disbursement id in postgres\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tContractNextDisbursementID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"contract_next_disbursement_id\",\n\t\t\tHelp: \"Next disbursement id expected by the disburser contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tContractNextDepositID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"contract_next_deposit_id\",\n\t\t\tHelp: \"next deposit id expected by the deposit contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tDisburserBalance: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"disburser_balance\",\n\t\t\tHelp: \"Balance in Wei of Teleportr's disburser wallet\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tDepositContractBalance: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"deposit_contract_balance\",\n\t\t\tHelp: \"Balance in Wei of Teleportr's deposit contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tFailedTXSubmissions: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"failed_tx_submissions\",\n\t\t\tHelp: \"Number of failed transaction submissions\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}, []string{\n\t\t\t\"type\",\n\t\t}),\n\t}\n}", "func NewMetrics(component string, sampleRate float64, client metrics.Client) BaseMetrics {\n\treturn BaseMetrics{\n\t\tcomponent: component,\n\t\trate: sampleRate,\n\t\tmetrics: client,\n\t\tmetMap: map[string]string{\n\t\t\t\"latency\": \"comp.\" + component + \".requests.latency\",\n\t\t\t\"request\": \"comp.\" + component + \".requests.%d\",\n\t\t\t\"mLatency\": \"comp.\" + component + \".requests.%s.latency\",\n\t\t\t\"mRequest\": \"comp.\" + component + \".requests.%s.%d\",\n\t\t},\n\t}\n}", "func NewMetrics(scope tally.Scope) Metrics {\n\tscope = scope.SubScope(\"aggregation\")\n\treturn Metrics{\n\t\tCounter: newCounterMetrics(scope.SubScope(\"counters\")),\n\t\tGauge: newGaugeMetrics(scope.SubScope(\"gauges\")),\n\t}\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{}\n\tm.Reset()\n\treturn m\n}", "func New() *Metrics {\n\treturn &Metrics{\n\t\tSectionCounts: make(map[string]int),\n\t}\n}", "func NewMetrics(p fabricmetrics.Provider) *Metrics {\n\treturn &Metrics{\n\t\tRefreshTimer: p.NewHistogram(refreshTimer),\n\t}\n}", "func newMetricsCollector(devices func() ([]*wgtypes.Device, error)) prometheus.Collector {\n\t// common labels for all metrics\n\tlabels := []string{\"device\", \"public_key\"}\n\n\treturn &collector{\n\t\tDeviceInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_device_info\",\n\t\t\t\"Metadata about a device.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_info\",\n\t\t\t\"Metadata about a peer. The public_key label on peer metrics refers to the peer's public key; not the device's public key.\",\n\t\t\tappend(labels, []string{\"endpoint\"}...),\n\t\t\tnil,\n\t\t),\n\t\tPeerAllowedIPsInfo: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_allowed_ips_info\",\n\t\t\t\"Metadata about each of a peer's allowed IP subnets for a given device.\",\n\t\t\tappend(labels, []string{\"allowed_ips\"}...),\n\t\t\tnil,\n\t\t),\n\t\tPeerReceiveBytes: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_receive_bytes_total\",\n\t\t\t\"Number of bytes received from a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerTransmitBytes: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_transmit_bytes_total\",\n\t\t\t\"Number of bytes transmitted to a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tPeerLastHandshake: prometheus.NewDesc(\n\t\t\t\"semaphore_wg_peer_last_handshake_seconds\",\n\t\t\t\"UNIX timestamp for the last handshake with a given peer.\",\n\t\t\tlabels,\n\t\t\tnil,\n\t\t),\n\t\tdevices: devices,\n\t}\n}", "func New() (*Metrics, error) {\n\treturn NewCapa(DefaultBufferSize, DefaultSamplingFactor)\n}", "func NewMetricsMetricsSystem(a string, b SparkConfInterface, c SecurityManagerInterface) (*MetricsMetricsSystem) {\n\tconv_a := javabind.NewGoToJavaString()\n\tconv_b := javabind.NewGoToJavaCallable()\n\tconv_c := javabind.NewGoToJavaCallable()\n\tif err := conv_a.Convert(a); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := conv_b.Convert(b); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := conv_c.Convert(c); err != nil {\n\t\tpanic(err)\n\t}\n\n\tobj, err := javabind.GetEnv().NewObject(\"org/apache/spark/metrics/MetricsSystem\", conv_a.Value().Cast(\"java/lang/String\"), conv_b.Value().Cast(\"org/apache/spark/SparkConf\"), conv_c.Value().Cast(\"org/apache/spark/SecurityManager\"))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tconv_a.CleanUp()\n\tconv_b.CleanUp()\n\tconv_c.CleanUp()\n\tx := &MetricsMetricsSystem{}\n\tx.Callable = &javabind.Callable{obj}\n\treturn x\n}", "func NewMetrics() *MetricsHolder {\n\tm := &MetricsHolder{\n\t\tlines: make(map[string]*Reading),\n\t\tchannel: make(chan interface{}),\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\tw, ok := <-m.channel\n\t\t\treading := w.(*Reading)\n\t\t\tif !ok {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif val, ok := m.lines[reading.Key]; ok {\n\t\t\t\tm.lines[reading.Key] = val.Accept(reading)\n\t\t\t} else {\n\t\t\t\tm.lines[reading.Key] = reading\n\t\t\t}\n\t\t}\n\t}()\n\treturn m\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tsuccessScope := scope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := scope.Tagged(map[string]string{\"result\": \"fail\"})\n\ttimeoutScope := scope.Tagged(map[string]string{\"result\": \"timeout\"})\n\tapiScope := scope.SubScope(\"api\")\n\tserverScope := scope.SubScope(\"server\")\n\tplacement := scope.SubScope(\"placement\")\n\trecovery := scope.SubScope(\"recovery\")\n\n\treturn &Metrics{\n\t\tAPIEnqueueGangs: apiScope.Counter(\"enqueue_gangs\"),\n\t\tEnqueueGangSuccess: successScope.Counter(\"enqueue_gang\"),\n\t\tEnqueueGangFail: failScope.Counter(\"enqueue_gang\"),\n\n\t\tAPIDequeueGangs: apiScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangSuccess: successScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangTimeout: timeoutScope.Counter(\"dequeue_gangs\"),\n\n\t\tAPIGetPreemptibleTasks: apiScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksSuccess: successScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksTimeout: timeoutScope.Counter(\"get_preemptible_tasks\"),\n\n\t\tAPISetPlacements: apiScope.Counter(\"set_placements\"),\n\t\tSetPlacementSuccess: successScope.Counter(\"set_placements\"),\n\t\tSetPlacementFail: failScope.Counter(\"set_placements\"),\n\n\t\tAPIGetPlacements: apiScope.Counter(\"get_placements\"),\n\t\tGetPlacementSuccess: successScope.Counter(\"get_placements\"),\n\t\tGetPlacementFail: failScope.Counter(\"get_placements\"),\n\n\t\tAPILaunchedTasks: apiScope.Counter(\"launched_tasks\"),\n\n\t\tRecoverySuccess: successScope.Counter(\"recovery\"),\n\t\tRecoveryFail: failScope.Counter(\"recovery\"),\n\t\tRecoveryRunningSuccessCount: successScope.Counter(\"task_count\"),\n\t\tRecoveryRunningFailCount: failScope.Counter(\"task_count\"),\n\t\tRecoveryEnqueueFailedCount: failScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryEnqueueSuccessCount: successScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryTimer: recovery.Timer(\"running_tasks\"),\n\n\t\tPlacementQueueLen: placement.Gauge(\"placement_queue_length\"),\n\t\tPlacementFailed: placement.Counter(\"fail\"),\n\n\t\tElected: serverScope.Gauge(\"elected\"),\n\t}\n}", "func New() *CloudMetrics {\n\treturn &CloudMetrics{\n\t\tAPIRequestsCounter: promauto.NewCounter(prometheus.CounterOpts{\n\t\t\tNamespace: provisionerNamespace,\n\t\t\tSubsystem: provisionerSubsystemAPI,\n\t\t\tName: \"requests_total\",\n\t\t\tHelp: \"The total number of http API requests\",\n\t\t}),\n\n\t\tAPITimesHistograms: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemAPI,\n\t\t\t\tName: \"requests_duration\",\n\t\t\t\tHelp: \"The duration of http API requests\",\n\t\t\t},\n\t\t\t[]string{\"handler\", \"method\", \"status_code\"},\n\t\t),\n\n\t\tInstallationCreationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_creation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation creation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationUpdateDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_update_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation update tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationHibernationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_hibernation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation hibernation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationWakeUpDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_wakeup_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation wake up tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tInstallationDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"installation_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of installation deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"group\"},\n\t\t),\n\n\t\tClusterInstallationReconcilingDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_installation_reconciling_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster installation reconciliation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"cluster\"},\n\t\t),\n\n\t\tClusterInstallationDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_installation_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster installation deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{\"cluster\"},\n\t\t),\n\t\tClusterCreationDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_creation_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster creation tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterUpgradeDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_upgrade_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster upgrade tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterProvisioningDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_provisioning_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster provisioning tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterResizeDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_resize_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster resize tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t\tClusterDeletionDurationHist: promauto.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: provisionerNamespace,\n\t\t\t\tSubsystem: provisionerSubsystemApp,\n\t\t\t\tName: \"cluster_deletion_duration_seconds\",\n\t\t\t\tHelp: \"The duration of cluster deletion tasks\",\n\t\t\t\tBuckets: standardDurationBuckets(),\n\t\t\t},\n\t\t\t[]string{},\n\t\t),\n\t}\n}", "func newCustomMetrics() ICustomMetrics {\n\n\tcounters := make(map[string]prometheus.Counter)\n\tgauges := make(map[string]prometheus.Gauge)\n\tsummaries := make(map[string]prometheus.Summary)\n\thistograms := make(map[string]prometheus.Histogram)\n\n\treturn &customMetrics{\n\t\tcounters: counters,\n\t\tgauges: gauges,\n\t\tsummaries: summaries,\n\t\thistograms: histograms,\n\t}\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{\n\t\tTimeMetrics: make(map[string]*TimeStats),\n\t\tNumberMetrics: make(map[string]*NumberStats),\n\t\tBoolMetrics: make(map[string]*BoolStats),\n\t}\n\treturn m\n}", "func (it *Mcmc4intmcMetricsIterator) Create(key uint64) (*Mcmc4intmcMetrics, error) {\n\ttmtr := &Mcmc4intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc4intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func newProcessMetrics(id string) *processMetrics {\n\tcommonTags := tags{TAG_INGESTER_ID: id, TAG_INGESTER_SOURCE: \"poll\"}\n\treturn &processMetrics{\n\t\tignoredByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"ignored\"}),\n\t\tprocessedByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"processed\"}),\n\t\tliveness: metrics2.NewLiveness(id, tags{TAG_INGESTER_SOURCE: \"poll\", TAG_INGESTION_METRIC: \"since-last-run\"}),\n\t}\n}", "func (it *Mcmc6intmcMetricsIterator) Create(key uint64) (*Mcmc6intmcMetrics, error) {\n\ttmtr := &Mcmc6intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc6intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewGoMetrics(reporter *Reporter, dims map[string]string) *GoMetrics {\n\tstart := time.Now()\n\tmstat := runtime.MemStats{}\n\tret := &GoMetrics{\n\t\treporter: reporter,\n\t}\n\n\tret.metrics = []Metric{\n\t\tWrapGauge(\"go-metric-alloc\", dims, Value(&mstat.Alloc)),\n\t\tWrapCumulativeCounter(\n\t\t\t\"go-metric-total-alloc\",\n\t\t\tdims,\n\t\t\tValue(&mstat.TotalAlloc),\n\t\t),\n\t\tWrapGauge(\"go-metric-sys\", dims, Value(&mstat.Sys)),\n\t\tWrapCumulativeCounter(\"go-metric-lookups\", dims, Value(&mstat.Lookups)),\n\t\tWrapCumulativeCounter(\"go-metric-mallocs\", dims, Value(&mstat.Mallocs)),\n\t\tWrapCumulativeCounter(\"go-metric-frees\", dims, Value(&mstat.Frees)),\n\t\tWrapGauge(\"go-metric-heap-alloc\", dims, Value(&mstat.HeapAlloc)),\n\t\tWrapGauge(\"go-metric-heap-sys\", dims, Value(&mstat.HeapSys)),\n\t\tWrapGauge(\"go-metric-heap-idle\", dims, Value(&mstat.HeapIdle)),\n\t\tWrapGauge(\"go-metric-heap-in-use\", dims, Value(&mstat.HeapInuse)),\n\t\tWrapGauge(\"go-metric-heap-released\", dims, Value(&mstat.HeapReleased)),\n\t\tWrapGauge(\"go-metric-heap-objects\", dims, Value(&mstat.HeapObjects)),\n\t\tWrapGauge(\"go-metric-stack-in-use\", dims, Value(&mstat.StackInuse)),\n\t\tWrapGauge(\"go-metric-stack-sys\", dims, Value(&mstat.StackSys)),\n\t\tWrapGauge(\"go-metric-mspan-in-use\", dims, Value(&mstat.MSpanInuse)),\n\t\tWrapGauge(\"go-metric-mspan-sys\", dims, Value(&mstat.MSpanSys)),\n\t\tWrapGauge(\"go-metric-mcache-in-use\", dims, Value(&mstat.MCacheInuse)),\n\t\tWrapGauge(\"go-metric-mcache-sys\", dims, Value(&mstat.MCacheSys)),\n\t\tWrapGauge(\"go-metric-buck-hash-sys\", dims, Value(&mstat.BuckHashSys)),\n\t\tWrapGauge(\"go-metric-gc-sys\", dims, Value(&mstat.GCSys)),\n\t\tWrapGauge(\"go-metric-other-sys\", dims, Value(&mstat.OtherSys)),\n\t\tWrapGauge(\"go-metric-next-gc\", dims, Value(&mstat.NextGC)),\n\t\tWrapGauge(\"go-metric-last-gc\", dims, Value(&mstat.LastGC)),\n\t\tWrapCumulativeCounter(\n\t\t\t\"go-metric-pause-total-ns\",\n\t\t\tdims,\n\t\t\tValue(&mstat.PauseTotalNs),\n\t\t),\n\t\tWrapGauge(\"go-metric-num-gc\", dims, Value(&mstat.NumGC)),\n\n\t\tWrapGauge(\n\t\t\t\"go-metric-gomaxprocs\",\n\t\t\tdims,\n\t\t\tGetterFunc(func() (interface{}, error) {\n\t\t\t\treturn runtime.GOMAXPROCS(0), nil\n\t\t\t}),\n\t\t),\n\t\tWrapGauge(\n\t\t\t\"go-metric-uptime-ns\",\n\t\t\tdims,\n\t\t\tGetterFunc(func() (interface{}, error) {\n\t\t\t\treturn time.Since(start).Nanoseconds(), nil\n\t\t\t}),\n\t\t),\n\t\tWrapGauge(\n\t\t\t\"go-metric-num-cpu\",\n\t\t\tdims,\n\t\t\tGetterFunc(func() (interface{}, error) {\n\t\t\t\treturn runtime.NumCPU(), nil\n\t\t\t}),\n\t\t),\n\t\tWrapCumulativeCounter(\n\t\t\t\"go-metric-num-cgo-call\",\n\t\t\tdims,\n\t\t\tGetterFunc(func() (interface{}, error) {\n\t\t\t\treturn runtime.NumCgoCall(), nil\n\t\t\t}),\n\t\t),\n\t\tWrapGauge(\n\t\t\t\"go-metric-num-goroutine\",\n\t\t\tdims,\n\t\t\tGetterFunc(func() (interface{}, error) {\n\t\t\t\treturn runtime.NumGoroutine(), nil\n\t\t\t}),\n\t\t),\n\t}\n\treporter.Track(ret.metrics...)\n\n\treporter.AddPreReportCallback(func() {\n\t\truntime.ReadMemStats(&mstat)\n\t})\n\n\treturn ret\n}", "func (it *Mcmc4mchintmcMetricsIterator) Create(key uint64) (*Mcmc4mchintmcMetrics, error) {\n\ttmtr := &Mcmc4mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc4mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc6mchintmcMetricsIterator) Create(key uint64) (*Mcmc6mchintmcMetrics, error) {\n\ttmtr := &Mcmc6mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc6mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tRangeFeedCatchupScanNanos: metric.NewCounter(metaRangeFeedCatchupScanNanos),\n\t\tRangeFeedSlowClosedTimestampLogN: log.Every(5 * time.Second),\n\t\tRangeFeedSlowClosedTimestampNudgeSem: make(chan struct{}, 1024),\n\t}\n}", "func (it *Mcmc5mchintmcMetricsIterator) Create(key uint64) (*Mcmc5mchintmcMetrics, error) {\n\ttmtr := &Mcmc5mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc5mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func New() *SystemMetrics {\n\treturn &SystemMetrics{}\n}", "func NewMetrics() *Metrics {\n\tmtrcs := &Metrics{\n\t\tcounters: make(map[MetricName]int),\n\t\tSidecarSyncErrors: SidecarSyncErrors,\n\t\tSidecarVaultTokenErrors: SidecarVaultTokenErrors,\n\t\tSidecarSecretErrors: SidecarSecretErrors,\n\t}\n\n\treturn mtrcs\n}", "func New(ctx context.Context) *Group {\n\t// Monitor goroutine context and cancelation.\n\tmctx, cancel := context.WithCancel(ctx)\n\n\tg := &Group{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\n\t\taddC: make(chan struct{}),\n\t\tlenC: make(chan int),\n\t}\n\n\tg.wg.Add(1)\n\tgo func() {\n\t\tdefer g.wg.Done()\n\t\tg.monitor(mctx)\n\t}()\n\n\treturn g\n}", "func NewMetrics(factory metrics.Factory, globalTags map[string]string) *Metrics {\n\tm := &Metrics{}\n\tmetrics.Init(m, factory.Namespace(\"jaeger\", nil), globalTags)\n\treturn m\n}", "func (it *Mcmc5intmcMetricsIterator) Create(key uint64) (*Mcmc5intmcMetrics, error) {\n\ttmtr := &Mcmc5intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc5intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewMetrics(app, metricsPrefix, version, hash, date string) *Metrics {\n\tlabels := map[string]string{\n\t\t\"app\": app,\n\t\t\"version\": version,\n\t\t\"hash\": hash,\n\t\t\"buildTime\": date,\n\t}\n\n\tif metricsPrefix != \"\" {\n\t\tmetricsPrefix += \"_\"\n\t}\n\n\tpm := &Metrics{\n\t\tresponseTime: prometheus.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tName: metricsPrefix + \"response_time_seconds\",\n\t\t\t\tHelp: \"Description\",\n\t\t\t\tConstLabels: labels,\n\t\t\t},\n\t\t\t[]string{\"endpoint\"},\n\t\t),\n\t\ttotalRequests: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_total\",\n\t\t\tHelp: \"number of requests\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tduration: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_duration_seconds\",\n\t\t\tHelp: \"duration of a requests in seconds\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tresponseSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"response_size_bytes\",\n\t\t\tHelp: \"size of the responses in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\trequestSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_size_bytes\",\n\t\t\tHelp: \"size of the requests in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\thandlerStatuses: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_statuses_total\",\n\t\t\tHelp: \"count number of responses per status\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"method\", \"status_bucket\"}),\n\t}\n\n\terr := prometheus.Register(pm)\n\tif e := new(prometheus.AlreadyRegisteredError); errors.As(err, e) {\n\t\treturn pm\n\t} else if err != nil {\n\t\tpanic(err)\n\t}\n\n\tgrpcPrometheus.EnableHandlingTimeHistogram()\n\n\treturn pm\n}", "func NewMetrics(namespace string, logger Logger) Metrics {\n\tlog := logger.GetLogger()\n\n\treturn &metricsImpl{\n\t\tinternalMetrics: metrics.NewMetrics(\"\", log),\n\t\texternalMetrics: metrics.NewMetrics(strings.ToLower(namespace), log),\n\t}\n}", "func New() *Metrics {\n\tm := &Metrics{\n\t\tBuildInfo: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\tNamespace: Namespace,\n\t\t\tSubsystem: Subsystem,\n\t\t\tName: \"build_info\",\n\t\t\tHelp: \"Build information\",\n\t\t}, []string{\"version\"}),\n\t}\n\n\t_ = prometheus.Register(m.BuildInfo)\n\t// TODO: implement metrics\n\treturn m\n}", "func New(cfg *Config) (*CirconusMetrics, error) {\n\n\tif cfg == nil {\n\t\treturn nil, errors.New(\"invalid configuration (nil)\")\n\t}\n\n\tcm := &CirconusMetrics{\n\t\tcounters: make(map[string]uint64),\n\t\tcounterFuncs: make(map[string]func() uint64),\n\t\tgauges: make(map[string]interface{}),\n\t\tgaugeFuncs: make(map[string]func() int64),\n\t\thistograms: make(map[string]*Histogram),\n\t\ttext: make(map[string]string),\n\t\ttextFuncs: make(map[string]func() string),\n\t\tcustom: make(map[string]Metric),\n\t\tlastMetrics: &prevMetrics{},\n\t}\n\n\t// Logging\n\t{\n\t\tcm.Debug = cfg.Debug\n\t\tcm.DumpMetrics = cfg.DumpMetrics\n\t\tcm.Log = cfg.Log\n\n\t\tif (cm.Debug || cm.DumpMetrics) && cm.Log == nil {\n\t\t\tcm.Log = log.New(os.Stderr, \"\", log.LstdFlags)\n\t\t}\n\t\tif cm.Log == nil {\n\t\t\tcm.Log = log.New(ioutil.Discard, \"\", log.LstdFlags)\n\t\t}\n\t}\n\n\t// Flush Interval\n\t{\n\t\tfi := defaultFlushInterval\n\t\tif cfg.Interval != \"\" {\n\t\t\tfi = cfg.Interval\n\t\t}\n\n\t\tdur, err := time.ParseDuration(fi)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"parsing flush interval\")\n\t\t}\n\t\tcm.flushInterval = dur\n\t}\n\n\t// metric resets\n\n\tcm.resetCounters = true\n\tif cfg.ResetCounters != \"\" {\n\t\tsetting, err := strconv.ParseBool(cfg.ResetCounters)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"parsing reset counters\")\n\t\t}\n\t\tcm.resetCounters = setting\n\t}\n\n\tcm.resetGauges = true\n\tif cfg.ResetGauges != \"\" {\n\t\tsetting, err := strconv.ParseBool(cfg.ResetGauges)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"parsing reset gauges\")\n\t\t}\n\t\tcm.resetGauges = setting\n\t}\n\n\tcm.resetHistograms = true\n\tif cfg.ResetHistograms != \"\" {\n\t\tsetting, err := strconv.ParseBool(cfg.ResetHistograms)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"parsing reset histograms\")\n\t\t}\n\t\tcm.resetHistograms = setting\n\t}\n\n\tcm.resetText = true\n\tif cfg.ResetText != \"\" {\n\t\tsetting, err := strconv.ParseBool(cfg.ResetText)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"parsing reset text\")\n\t\t}\n\t\tcm.resetText = setting\n\t}\n\n\t// check manager\n\t{\n\t\tcfg.CheckManager.Debug = cm.Debug\n\t\tcfg.CheckManager.Log = cm.Log\n\n\t\tcheck, err := checkmgr.New(&cfg.CheckManager)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"creating new check manager\")\n\t\t}\n\t\tcm.check = check\n\t}\n\n\t// start initialization (serialized or background)\n\tif err := cm.check.Initialize(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// if automatic flush is enabled, start it.\n\t// NOTE: submit will jettison metrics until initialization has completed.\n\tif cm.flushInterval > time.Duration(0) {\n\t\tgo func() {\n\t\t\tfor range time.NewTicker(cm.flushInterval).C {\n\t\t\t\tcm.Flush()\n\t\t\t}\n\t\t}()\n\t}\n\n\treturn cm, nil\n}", "func (it *Mcmc0mchintmcMetricsIterator) Create(key uint64) (*Mcmc0mchintmcMetrics, error) {\n\ttmtr := &Mcmc0mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc0mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func newMetricsMetadata(cluster *string, containerInstance *string) *ecstcs.MetricsMetadata {\n\treturn &ecstcs.MetricsMetadata{\n\t\tCluster: cluster,\n\t\tContainerInstance: containerInstance,\n\t}\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tm := &Metrics{\n\t\tProcedures: map[string]*PerProcedureMetrics{},\n\t}\n\tfor _, procedure := range _procedures {\n\t\tresponseCodes := make(map[api.ResponseCode]*PerResponseCodeMetrics)\n\t\tfor _, responseCode := range api.ResponseCode_Values() {\n\t\t\tresponseCodeText, exists := _responseCodeToText[responseCode]\n\t\t\tif !exists {\n\t\t\t\tresponseCodeText = \"unknown-error\"\n\t\t\t}\n\t\t\ttag := map[string]string{\n\t\t\t\tTagProcedure: procedure,\n\t\t\t\tTagResponseCode: responseCodeText,\n\t\t\t\t// Fill empty string here so that prometheus won't panic\n\t\t\t\t// when the number of tags is changed inside subscope\n\t\t\t\tTagService: \"\",\n\t\t\t}\n\t\t\tsubscope := scope.Tagged(tag)\n\t\t\tresponseCodes[responseCode] = &PerResponseCodeMetrics{\n\t\t\t\tScope: subscope,\n\t\t\t\tCalls: subscope.Counter(MetricNameCalls),\n\t\t\t\tCallLatency: subscope.Timer(MetricNameCallLatency),\n\t\t\t}\n\t\t}\n\t\tm.Procedures[procedure] = &PerProcedureMetrics{\n\t\t\tResponseCodes: responseCodes,\n\t\t}\n\t}\n\treturn m\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tPath: defaultPath,\n\t\tAddr: defaultAddr,\n\t\textraLabels: []extraLabel{},\n\t}\n}", "func NewPrometheusMetrics(namespace string, registry metrics.RegisterGatherer) *prometheusMetrics {\n\tm := &prometheusMetrics{\n\t\tregistry: registry,\n\t}\n\n\tm.AvailableIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_ips\",\n\t\tHelp: \"Total available IPs on Node for IPAM allocation\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.UsedIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"used_ips\",\n\t\tHelp: \"Total used IPs on Node for IPAM allocation\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.NeededIPs = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"needed_ips\",\n\t\tHelp: \"Number of IPs that are needed on the Node to satisfy IPAM allocation requests\",\n\t}, []string{LabelTargetNodeName})\n\n\tm.IPsAllocated = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ips\",\n\t\tHelp: \"Number of IPs allocated\",\n\t}, []string{\"type\"})\n\n\tm.AllocateIpOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ip_allocation_ops\",\n\t\tHelp: \"Number of IP allocation operations\",\n\t}, []string{\"subnet_id\"})\n\n\tm.ReleaseIpOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"ip_release_ops\",\n\t\tHelp: \"Number of IP release operations\",\n\t}, []string{\"subnet_id\"})\n\n\tm.AllocateInterfaceOps = prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"interface_creation_ops\",\n\t\tHelp: \"Number of interfaces allocated\",\n\t}, []string{\"subnet_id\"})\n\n\tm.AvailableInterfaces = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_interfaces\",\n\t\tHelp: \"Number of interfaces with addresses available\",\n\t})\n\n\tm.InterfaceCandidates = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"interface_candidates\",\n\t\tHelp: \"Number of attached interfaces with IPs available for allocation\",\n\t})\n\n\tm.EmptyInterfaceSlots = prometheus.NewGauge(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"empty_interface_slots\",\n\t\tHelp: \"Number of empty interface slots available for interfaces to be attached\",\n\t})\n\n\tm.AvailableIPsPerSubnet = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"available_ips_per_subnet\",\n\t\tHelp: \"Number of available IPs per subnet ID\",\n\t}, []string{\"subnet_id\", \"availability_zone\"})\n\n\tm.Nodes = prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"nodes\",\n\t\tHelp: \"Number of nodes by category { total | in-deficit | at-capacity }\",\n\t}, []string{\"category\"})\n\n\tm.Resync = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"resync_total\",\n\t\tHelp: \"Number of resync operations to synchronize and resolve IP deficit of nodes\",\n\t})\n\n\tm.Allocation = prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"allocation_duration_seconds\",\n\t\tHelp: \"Allocation ip or interface latency in seconds\",\n\t\tBuckets: merge(\n\t\t\tprometheus.LinearBuckets(0.25, 0.25, 2), // 0.25s, 0.50s\n\t\t\tprometheus.LinearBuckets(1, 1, 60), // 1s, 2s, 3s, ... 60s,\n\t\t),\n\t}, []string{\"type\", \"status\", \"subnet_id\"})\n\n\tm.Release = prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: ipamSubsystem,\n\t\tName: \"release_duration_seconds\",\n\t\tHelp: \"Release ip or interface latency in seconds\",\n\t\tBuckets: merge(\n\t\t\tprometheus.LinearBuckets(0.25, 0.25, 2), // 0.25s, 0.50s\n\t\t\tprometheus.LinearBuckets(1, 1, 60), // 1s, 2s, 3s, ... 60s,\n\t\t),\n\t}, []string{\"type\", \"status\", \"subnet_id\"})\n\n\t// pool_maintainer is a more generic name, but for backward compatibility\n\t// of dashboard, keep the metric name deficit_resolver unchanged\n\tm.poolMaintainer = NewTriggerMetrics(namespace, \"deficit_resolver\")\n\tm.k8sSync = NewTriggerMetrics(namespace, \"k8s_sync\")\n\tm.resync = NewTriggerMetrics(namespace, \"resync\")\n\n\tregistry.MustRegister(m.AvailableIPs)\n\tregistry.MustRegister(m.UsedIPs)\n\tregistry.MustRegister(m.NeededIPs)\n\n\tregistry.MustRegister(m.IPsAllocated)\n\tregistry.MustRegister(m.AllocateIpOps)\n\tregistry.MustRegister(m.ReleaseIpOps)\n\tregistry.MustRegister(m.AllocateInterfaceOps)\n\tregistry.MustRegister(m.AvailableInterfaces)\n\tregistry.MustRegister(m.InterfaceCandidates)\n\tregistry.MustRegister(m.EmptyInterfaceSlots)\n\tregistry.MustRegister(m.AvailableIPsPerSubnet)\n\tregistry.MustRegister(m.Nodes)\n\tregistry.MustRegister(m.Resync)\n\tregistry.MustRegister(m.Allocation)\n\tregistry.MustRegister(m.Release)\n\tm.poolMaintainer.Register(registry)\n\tm.k8sSync.Register(registry)\n\tm.resync.Register(registry)\n\n\treturn m\n}", "func newMetricsFrom(opts *MetricsOpts) *Metrics {\n\tmetrics := &Metrics{\n\t\tcounters: make(map[string]prometheus.Counter, 512),\n\t\tgauges: make(map[string]prometheus.Gauge, 512),\n\t\thistorams: make(map[string]prometheus.Histogram, 512),\n\t\tsummaries: make(map[string]prometheus.Summary, 512),\n\t\tdefBuckets: opts.DefBuckets,\n\t\tdefQuantile: opts.DefQuantile,\n\t\tregistry: prometheus.NewRegistry(),\n\t}\n\treturn metrics\n}", "func newMetrics() metrics {\n\treturn metrics{\n\t\tsize: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"pool_size\",\n\t\t\t\tHelp: \"Size of pool\",\n\t\t\t},\n\t\t),\n\n\t\tstatus: prometheus.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"pool_status\",\n\t\t\t\tHelp: `Status of pool (0, 1, 2, 3, 4, 5, 6)= {\"Offline\", \"Online\", \"Degraded\", \"Faulted\", \"Removed\", \"Unavail\", \"NoPoolsAvailable\"}`,\n\t\t\t},\n\t\t\t[]string{\"pool\"},\n\t\t),\n\n\t\tusedCapacity: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"used_pool_capacity\",\n\t\t\t\tHelp: \"Capacity used by pool\",\n\t\t\t},\n\t\t),\n\n\t\tfreeCapacity: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"free_pool_capacity\",\n\t\t\t\tHelp: \"Free capacity in pool\",\n\t\t\t},\n\t\t),\n\n\t\tusedCapacityPercent: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"used_pool_capacity_percent\",\n\t\t\t\tHelp: \"Capacity used by pool in percent\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolListparseErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_list_parse_error_count\",\n\t\t\t\tHelp: \"Total no of parsing errors\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolRejectRequestCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_reject_request_count\",\n\t\t\t\tHelp: \"Total no of rejected requests of zpool command\",\n\t\t\t},\n\t\t),\n\n\t\tzpoolCommandErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_command_error\",\n\t\t\t\tHelp: \"Total no of zpool command errors\",\n\t\t\t},\n\t\t),\n\n\t\tnoPoolAvailableErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"no_pool_available_error\",\n\t\t\t\tHelp: \"Total no of no pool available errors\",\n\t\t\t},\n\t\t),\n\n\t\tincompleteOutputErrorCounter: prometheus.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"openebs\",\n\t\t\t\tName: \"zpool_list_incomplete_stdout_error\",\n\t\t\t\tHelp: \"Total no of incomplete stdout of zpool list command errors\",\n\t\t\t},\n\t\t),\n\t}\n}", "func NewMetrics(reg prometheus.Registerer) *Metrics {\n\tvar m Metrics\n\tm.reg = reg\n\n\tm.dockerEntries = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_entries_total\",\n\t\tHelp: \"Total number of successful entries sent to the Docker target\",\n\t})\n\tm.dockerErrors = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_parsing_errors_total\",\n\t\tHelp: \"Total number of parsing errors while receiving Docker messages\",\n\t})\n\n\tif reg != nil {\n\t\treg.MustRegister(\n\t\t\tm.dockerEntries,\n\t\t\tm.dockerErrors,\n\t\t)\n\t}\n\n\treturn &m\n}", "func (it *Mcmc2mchintmcMetricsIterator) Create(key uint64) (*Mcmc2mchintmcMetrics, error) {\n\ttmtr := &Mcmc2mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc2mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc3mchintmcMetricsIterator) Create(key uint64) (*Mcmc3mchintmcMetrics, error) {\n\ttmtr := &Mcmc3mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc3mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc7mchintmcMetricsIterator) Create(key uint64) (*Mcmc7mchintmcMetrics, error) {\n\ttmtr := &Mcmc7mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc7mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewMetrics(healthyChan chan bool, config *openapi.AlgoRunnerConfig) Metrics {\n\n\tgo func() {\n\t\tfor h := range healthyChan {\n\t\t\thealthy = h\n\t\t}\n\t}()\n\n\tregisterMetrics(config)\n\n\treturn Metrics{\n\t\tRunnerRuntimeHistogram: runnerRuntimeHistogram,\n\t\tAlgoRuntimeHistogram: algoRuntimeHistogram,\n\t\tMsgBytesInputCounter: msgBytesInputCounter,\n\t\tMsgBytesOutputCounter: msgBytesOutputCounter,\n\t\tDataBytesInputCounter: dataBytesInputCounter,\n\t\tDataBytesOutputCounter: dataBytesOutputCounter,\n\t\tRetryCounter: retryCounter,\n\t\tDlqCounter: dlqCounter,\n\t\tAlgoErrorCounter: algoErrorCounter,\n\t\tRunnerErrorCounter: runnerErrorCounter,\n\t\tMsgOK: msgOK,\n\t\tMsgNOK: msgNOK,\n\t\tMsgDropped: msgDropped,\n\t\tProducerQueueLen: &producerQueueLen,\n\t\tEventIgnored: eventIgnored,\n\t\tMsgInTransit: msgInTransit,\n\t\tLibRdKafkaVersion: libRdKafkaVersion,\n\t\tLastProducerStartTime: &lastProducerStartTime,\n\t\tMetricCertExpirationTime: &metricCertExpirationTime,\n\t\tMetricCaExpirationTime: &metricCaExpirationTime,\n\t\tMetricKafkaEventsQueueLen: &metricKafkaEventsQueueLen,\n\t\tMetricRDKafkaGlobal: metricRDKafkaGlobal,\n\t\tMetricRDKafkaBroker: metricRDKafkaBroker,\n\t\tMetricRDKafkaTopic: metricRDKafkaTopic,\n\t\tMetricRDKafkaPartition: metricRDKafkaPartition,\n\n\t\tDeploymentLabel: deploymentLabel,\n\t\tPipelineLabel: pipelineLabel,\n\t\tComponentLabel: componentLabel,\n\t\tAlgoLabel: algoLabel,\n\t\tAlgoVersionLabel: algoVersionLabel,\n\t\tAlgoIndexLabel: algoIndexLabel,\n\t}\n}", "func (it *Mcmc1mchintmcMetricsIterator) Create(key uint64) (*Mcmc1mchintmcMetrics, error) {\n\ttmtr := &Mcmc1mchintmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc1mchintmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewMetrics(kubeclientset kubernetes.Interface,\n\tkubeInformerFactory kubeinformers.SharedInformerFactory,\n\tcerebralclientset cerebral.Interface,\n\tcInformerFactory cinformers.SharedInformerFactory,\n\tscaleRequestCh chan<- ScaleRequest) *MetricsController {\n\trateLimiter := workqueue.NewItemExponentialFailureRateLimiter(metricsDelayBetweenRequeues, metricsMaxRequeues)\n\n\tc := &MetricsController{\n\t\tkubeclientset: kubeclientset,\n\t\tcerebralclientset: cerebralclientset,\n\t\tworkqueue: workqueue.NewNamedRateLimitingQueue(rateLimiter, metricsControllerName),\n\t\tpollManagers: make(map[string]pollManager),\n\t\tscaleRequestCh: scaleRequestCh,\n\t}\n\n\teventBroadcaster := record.NewBroadcaster()\n\teventBroadcaster.StartLogging(log.Infof)\n\teventBroadcaster.StartRecordingToSink(&typedcorev1.EventSinkImpl{\n\t\tInterface: kubeclientset.CoreV1().Events(\"\"),\n\t})\n\tc.recorder = eventBroadcaster.NewRecorder(scheme.Scheme, corev1.EventSource{\n\t\tComponent: controllerName,\n\t})\n\n\tasgInformer := cInformerFactory.Cerebral().V1alpha1().AutoscalingGroups()\n\taspInformer := cInformerFactory.Cerebral().V1alpha1().AutoscalingPolicies()\n\taseInformer := cInformerFactory.Cerebral().V1alpha1().AutoscalingEngines()\n\n\tlog.Infof(\"%s: setting up event handlers\", metricsControllerName)\n\n\tasgInformer.Informer().AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: c.enqueueAutoscalingGroup,\n\t\tUpdateFunc: func(old, new interface{}) {\n\t\t\tnewASG := new.(*cerebralv1alpha1.AutoscalingGroup)\n\t\t\toldASG := old.(*cerebralv1alpha1.AutoscalingGroup)\n\t\t\t// Generation need to be checked so that the ASG only gets enqueued if the\n\t\t\t// spec changes and ignores status update changes, as well as sync events\n\t\t\tif newASG.ResourceVersion == oldASG.ResourceVersion ||\n\t\t\t\tnewASG.Generation == oldASG.Generation {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tc.enqueueAutoscalingGroup(new)\n\t\t},\n\t\tDeleteFunc: c.enqueueAutoscalingGroup,\n\t})\n\n\taspInformer.Informer().AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: c.enqueueASGsForAutoscalingPolicy,\n\t\tUpdateFunc: func(old, new interface{}) {\n\t\t\tnewASP := new.(*cerebralv1alpha1.AutoscalingPolicy)\n\t\t\toldASP := old.(*cerebralv1alpha1.AutoscalingPolicy)\n\t\t\tif newASP.ResourceVersion == oldASP.ResourceVersion {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tc.enqueueASGsForAutoscalingPolicy(new)\n\t\t},\n\t\tDeleteFunc: c.enqueueASGsForAutoscalingPolicy,\n\t})\n\n\t// Explicitly ignore AutoscalingEngine updates since we only care about\n\t// starting poll managers when the AutoscalingEngine is added, and removing\n\t// them when the AutoscalingEngine is deleted. Miscellaneous updates to engine\n\t// configuration can be ignored, even when changing engine \"type\" since we can\n\t// assume schema is valid, and have no reason to believe the scaling action will fail\n\taseInformer.Informer().AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: c.enqueueASGsForAutoscalingEngine,\n\t\tDeleteFunc: c.enqueueASGsForAutoscalingEngine,\n\t})\n\n\tc.asgLister = asgInformer.Lister()\n\tc.asgSynced = asgInformer.Informer().HasSynced\n\n\tc.aspLister = aspInformer.Lister()\n\tc.aspSynced = aspInformer.Informer().HasSynced\n\n\tc.aseLister = aseInformer.Lister()\n\tc.aseSynced = aseInformer.Informer().HasSynced\n\n\treturn c\n}", "func (it *Mcmc0intmcMetricsIterator) Create(key uint64) (*Mcmc0intmcMetrics, error) {\n\ttmtr := &Mcmc0intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc0intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc3intmcMetricsIterator) Create(key uint64) (*Mcmc3intmcMetrics, error) {\n\ttmtr := &Mcmc3intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc3intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc2intmcMetricsIterator) Create(key uint64) (*Mcmc2intmcMetrics, error) {\n\ttmtr := &Mcmc2intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc2intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func New(\n\tname string,\n\ttags map[string]string,\n\tmeta map[string]string,\n\tfields map[string]interface{},\n\ttm time.Time,\n) (CCMetric, error) {\n\tm := &ccMetric{\n\t\tname: name,\n\t\ttags: make(map[string]string, len(tags)),\n\t\tmeta: make(map[string]string, len(meta)),\n\t\tfields: make(map[string]interface{}, len(fields)),\n\t\ttm: tm,\n\t}\n\n\t// deep copy tags, meta data tags and fields\n\tfor k, v := range tags {\n\t\tm.tags[k] = v\n\t}\n\tfor k, v := range meta {\n\t\tm.meta[k] = v\n\t}\n\tfor k, v := range fields {\n\t\tv := convertField(v)\n\t\tif v == nil {\n\t\t\tcontinue\n\t\t}\n\t\tm.fields[k] = v\n\t}\n\n\treturn m, nil\n}", "func (m *podMetrics) New() runtime.Object {\n\treturn &metrics.PodMetrics{}\n}", "func New() *AllocGrp {\n\tvar m AllocGrp\n\treturn &m\n}", "func NewMetrics(registry metrics.Registry, exchanges []openrtb_ext.BidderName, disableAccountMetrics config.DisabledMetrics, syncerKeys []string, moduleStageNames map[string][]string) *Metrics {\n\tnewMetrics := NewBlankMetrics(registry, exchanges, disableAccountMetrics, moduleStageNames)\n\tnewMetrics.ConnectionCounter = metrics.GetOrRegisterCounter(\"active_connections\", registry)\n\tnewMetrics.TMaxTimeoutCounter = metrics.GetOrRegisterCounter(\"tmax_timeout\", registry)\n\tnewMetrics.ConnectionAcceptErrorMeter = metrics.GetOrRegisterMeter(\"connection_accept_errors\", registry)\n\tnewMetrics.ConnectionCloseErrorMeter = metrics.GetOrRegisterMeter(\"connection_close_errors\", registry)\n\tnewMetrics.ImpMeter = metrics.GetOrRegisterMeter(\"imps_requested\", registry)\n\n\tnewMetrics.ImpsTypeBanner = metrics.GetOrRegisterMeter(\"imp_banner\", registry)\n\tnewMetrics.ImpsTypeVideo = metrics.GetOrRegisterMeter(\"imp_video\", registry)\n\tnewMetrics.ImpsTypeAudio = metrics.GetOrRegisterMeter(\"imp_audio\", registry)\n\tnewMetrics.ImpsTypeNative = metrics.GetOrRegisterMeter(\"imp_native\", registry)\n\n\tnewMetrics.NoCookieMeter = metrics.GetOrRegisterMeter(\"no_cookie_requests\", registry)\n\tnewMetrics.AppRequestMeter = metrics.GetOrRegisterMeter(\"app_requests\", registry)\n\tnewMetrics.DebugRequestMeter = metrics.GetOrRegisterMeter(\"debug_requests\", registry)\n\tnewMetrics.RequestTimer = metrics.GetOrRegisterTimer(\"request_time\", registry)\n\tnewMetrics.DNSLookupTimer = metrics.GetOrRegisterTimer(\"dns_lookup_time\", registry)\n\tnewMetrics.TLSHandshakeTimer = metrics.GetOrRegisterTimer(\"tls_handshake_time\", registry)\n\tnewMetrics.PrebidCacheRequestTimerSuccess = metrics.GetOrRegisterTimer(\"prebid_cache_request_time.ok\", registry)\n\tnewMetrics.PrebidCacheRequestTimerError = metrics.GetOrRegisterTimer(\"prebid_cache_request_time.err\", registry)\n\tnewMetrics.StoredResponsesMeter = metrics.GetOrRegisterMeter(\"stored_responses\", registry)\n\tnewMetrics.OverheadTimer = makeOverheadTimerMetrics(registry)\n\tnewMetrics.BidderServerResponseTimer = metrics.GetOrRegisterTimer(\"bidder_server_response_time_seconds\", registry)\n\n\tfor _, dt := range StoredDataTypes() {\n\t\tfor _, ft := range StoredDataFetchTypes() {\n\t\t\ttimerName := fmt.Sprintf(\"stored_%s_fetch_time.%s\", string(dt), string(ft))\n\t\t\tnewMetrics.StoredDataFetchTimer[dt][ft] = metrics.GetOrRegisterTimer(timerName, registry)\n\t\t}\n\t\tfor _, e := range StoredDataErrors() {\n\t\t\tmeterName := fmt.Sprintf(\"stored_%s_error.%s\", string(dt), string(e))\n\t\t\tnewMetrics.StoredDataErrorMeter[dt][e] = metrics.GetOrRegisterMeter(meterName, registry)\n\t\t}\n\t}\n\n\tnewMetrics.AmpNoCookieMeter = metrics.GetOrRegisterMeter(\"amp_no_cookie_requests\", registry)\n\n\tnewMetrics.CookieSyncMeter = metrics.GetOrRegisterMeter(\"cookie_sync_requests\", registry)\n\tfor _, s := range CookieSyncStatuses() {\n\t\tnewMetrics.CookieSyncStatusMeter[s] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"cookie_sync_requests.%s\", s), registry)\n\t}\n\n\tnewMetrics.SetUidMeter = metrics.GetOrRegisterMeter(\"setuid_requests\", registry)\n\tfor _, s := range SetUidStatuses() {\n\t\tnewMetrics.SetUidStatusMeter[s] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"setuid_requests.%s\", s), registry)\n\t}\n\n\tfor _, syncerKey := range syncerKeys {\n\t\tnewMetrics.SyncerRequestsMeter[syncerKey] = make(map[SyncerCookieSyncStatus]metrics.Meter)\n\t\tfor _, status := range SyncerRequestStatuses() {\n\t\t\tnewMetrics.SyncerRequestsMeter[syncerKey][status] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"syncer.%s.request.%s\", syncerKey, status), registry)\n\t\t}\n\n\t\tnewMetrics.SyncerSetsMeter[syncerKey] = make(map[SyncerSetUidStatus]metrics.Meter)\n\t\tfor _, status := range SyncerSetUidStatuses() {\n\t\t\tnewMetrics.SyncerSetsMeter[syncerKey][status] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"syncer.%s.set.%s\", syncerKey, status), registry)\n\t\t}\n\t}\n\n\tfor _, a := range exchanges {\n\t\tregisterAdapterMetrics(registry, \"adapter\", string(a), newMetrics.AdapterMetrics[a])\n\t}\n\n\tfor typ, statusMap := range newMetrics.RequestStatuses {\n\t\tfor stat := range statusMap {\n\t\t\tstatusMap[stat] = metrics.GetOrRegisterMeter(\"requests.\"+string(stat)+\".\"+string(typ), registry)\n\t\t}\n\t}\n\n\tfor _, cacheRes := range CacheResults() {\n\t\tnewMetrics.StoredReqCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"stored_request_cache_%s\", string(cacheRes)), registry)\n\t\tnewMetrics.StoredImpCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"stored_imp_cache_%s\", string(cacheRes)), registry)\n\t\tnewMetrics.AccountCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"account_cache_%s\", string(cacheRes)), registry)\n\t}\n\n\tnewMetrics.RequestsQueueTimer[\"video\"][true] = metrics.GetOrRegisterTimer(\"queued_requests.video.accepted\", registry)\n\tnewMetrics.RequestsQueueTimer[\"video\"][false] = metrics.GetOrRegisterTimer(\"queued_requests.video.rejected\", registry)\n\n\tnewMetrics.TimeoutNotificationSuccess = metrics.GetOrRegisterMeter(\"timeout_notification.ok\", registry)\n\tnewMetrics.TimeoutNotificationFailure = metrics.GetOrRegisterMeter(\"timeout_notification.failed\", registry)\n\n\tnewMetrics.PrivacyCCPARequest = metrics.GetOrRegisterMeter(\"privacy.request.ccpa.specified\", registry)\n\tnewMetrics.PrivacyCCPARequestOptOut = metrics.GetOrRegisterMeter(\"privacy.request.ccpa.opt-out\", registry)\n\tnewMetrics.PrivacyCOPPARequest = metrics.GetOrRegisterMeter(\"privacy.request.coppa\", registry)\n\tnewMetrics.PrivacyLMTRequest = metrics.GetOrRegisterMeter(\"privacy.request.lmt\", registry)\n\tfor _, version := range TCFVersions() {\n\t\tnewMetrics.PrivacyTCFRequestVersion[version] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"privacy.request.tcf.%s\", string(version)), registry)\n\t}\n\n\tnewMetrics.AdsCertRequestsSuccess = metrics.GetOrRegisterMeter(\"ads_cert_requests.ok\", registry)\n\tnewMetrics.AdsCertRequestsFailure = metrics.GetOrRegisterMeter(\"ads_cert_requests.failed\", registry)\n\tnewMetrics.adsCertSignTimer = metrics.GetOrRegisterTimer(\"ads_cert_sign_time\", registry)\n\n\tfor module, stages := range moduleStageNames {\n\t\tregisterModuleMetrics(registry, module, stages, newMetrics.ModuleMetrics[module])\n\t}\n\n\treturn newMetrics\n}", "func (cg *ConsumerGroup) CG() *cluster.Consumer {\n\treturn cg.consumer\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tInputBytesTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_input_bytes_total\",\n\t\t\t\tHelp: \"Total number of bytes received\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tOutputBytesTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_output_bytes_total\",\n\t\t\t\tHelp: \"Total number of bytes sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tInputPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_input_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets received\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tOutputPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_output_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tDroppedPacketsTotal: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_dropped_pkts_total\",\n\t\t\t\tHelp: \"Total number of packets dropped by the router. This metric reports \" +\n\t\t\t\t\t\"the number of packets that were dropped because of errors.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tInterfaceUp: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_interface_up\",\n\t\t\t\tHelp: \"Either zero or one depending on whether the interface is up.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDInterfaceStateChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_state_changes_total\",\n\t\t\t\tHelp: \"Total number of BFD state changes.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDPacketsSent: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sent_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets sent.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tBFDPacketsReceived: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_received_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets received.\",\n\t\t\t},\n\t\t\t[]string{\"interface\", \"isd_as\", \"neighbor_isd_as\"},\n\t\t),\n\t\tServiceInstanceCount: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_service_instance_count\",\n\t\t\t\tHelp: \"Number of service instances known by the data plane.\",\n\t\t\t},\n\t\t\t[]string{\"service\", \"isd_as\"},\n\t\t),\n\t\tServiceInstanceChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_service_instance_changes_total\",\n\t\t\t\tHelp: \"Number of total service instance changes. Both addition and removal of a \" +\n\t\t\t\t\t\"service instance is accumulated.\",\n\t\t\t},\n\t\t\t[]string{\"service\", \"isd_as\"},\n\t\t),\n\t\tSiblingReachable: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tName: \"router_sibling_reachable\",\n\t\t\t\tHelp: \"Either zero or one depending on whether a sibling router \" +\n\t\t\t\t\t\"instance is reachable.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDPacketsSent: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sent_sibling_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets sent to sibling router instance.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDPacketsReceived: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_received_sibling_packets_total\",\n\t\t\t\tHelp: \"Number of BFD packets received from sibling router instance.\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t\tSiblingBFDStateChanges: promauto.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tName: \"router_bfd_sibling_state_changes_total\",\n\t\t\t\tHelp: \"Total number of BFD state changes for sibling router instances\",\n\t\t\t},\n\t\t\t[]string{\"sibling\", \"isd_as\"},\n\t\t),\n\t}\n}", "func (it *Mcmc7intmcMetricsIterator) Create(key uint64) (*Mcmc7intmcMetrics, error) {\n\ttmtr := &Mcmc7intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc7intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (it *Mcmc1intmcMetricsIterator) Create(key uint64) (*Mcmc1intmcMetrics, error) {\n\ttmtr := &Mcmc1intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc1intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func New(opts ...Option) *Metric {\n\tvar options Options\n\tfor _, opt := range opts {\n\t\topt(&options)\n\t}\n\tm := &Metric{\n\t\tOptions: options,\n\t\thistograms: make(map[string]metrics.Histogram),\n\t\tkeyLabels: make(map[string]map[string]string),\n\t}\n\tgo m.watch()\n\treturn m\n}", "func NewClientMetrics(scope tally.Scope) *ClientMetrics {\n\tclientAPIScope := scope.SubScope(\"api\")\n\tclientSuccessScope := scope.Tagged(map[string]string{\"result\": \"success\"})\n\tclientFailScope := scope.Tagged(map[string]string{\"result\": \"fail\"})\n\treturn &ClientMetrics{\n\t\tEventsConsumed: scope.Counter(\"eventsConsumed\"),\n\t\tStreamIDChange: scope.Counter(\"streamIdChange\"),\n\t\tPurgeOffset: scope.Gauge(\"purgeOffset\"),\n\t\tInitStreamAPI: clientAPIScope.Counter(\"initStream\"),\n\t\tInitStreamSuccess: clientSuccessScope.Counter(\"initStream\"),\n\t\tInitStreamFail: clientFailScope.Counter(\"initStream\"),\n\t\tWaitForEventsAPI: clientAPIScope.Counter(\"waitForEvents\"),\n\t\tWaitForEventsSuccess: clientSuccessScope.Counter(\"waitForEvents\"),\n\t\tWaitForEventsFailed: clientFailScope.Counter(\"waitForEvents\"),\n\t}\n}", "func newConsumerGroupCache(destPath string, cgDesc shared.ConsumerGroupDescription, cgLogger bark.Logger, h *OutputHost) *consumerGroupCache {\n\tcgCache := &consumerGroupCache{\n\t\tdestPath: destPath,\n\t\tcachedCGDesc: cgDesc,\n\t\tcachedTime: common.Now(),\n\t\textentCache: make(map[string]*extentCache),\n\t\tmsgsCh: make(chan *cherami.ConsumerMessage, defaultPrefetchBufferSize),\n\t\tmsgsRedeliveryCh: make(chan *cherami.ConsumerMessage, defaultPrefetchBufferSize),\n\t\tpriorityMsgsRedeliveryCh: make(chan *cherami.ConsumerMessage, 1),\n\t\tnotifyConsCloseCh: make(chan int, 5),\n\t\tnotifyReplicaCloseCh: make(chan string, 5),\n\t\tconnections: make(map[int]*consConnection),\n\t\tmsgCacheCh: make(chan cacheMsg, defaultPrefetchBufferSize),\n\t\tmsgCacheRedeliveredCh: make(chan cacheMsg, defaultPrefetchBufferSize),\n\t\tackMsgCh: make(chan timestampedAckID, ackChannelSize), // Have a buffer of at least 1, so that we aren't necessarily synchronous with the counterparty\n\t\tnackMsgCh: make(chan timestampedAckID, ackChannelSize),\n\t\tcloseChannel: make(chan struct{}),\n\t\toutputHostUUID: h.GetHostUUID(),\n\t\ttClients: h.GetClientFactory(),\n\t\twsConnector: h.GetWSConnector(),\n\t\tmetaClient: h.metaClient,\n\t\tshutdownWG: &h.shutdownWG,\n\t\tlogger: cgLogger.WithField(common.TagModule, `cgCache`),\n\t\tnotifyUnloadCh: h.unloadCacheCh,\n\t\tm3Client: h.m3Client,\n\t\tnotifier: newNotifier(),\n\t\tcreditNotifyCh: make(chan int32, 50),\n\t\tcreditRequestCh: make(chan string, 50),\n\t\tlastDisconnectTime: time.Now(),\n\t\tsessionID: h.sessionID,\n\t\tackIDGen: h.ackMgrIDGen,\n\t\tackMgrLoadCh: h.ackMgrLoadCh,\n\t\tackMgrUnloadCh: h.ackMgrUnloadCh,\n\t\tloadReporterFactory: h.GetLoadReporterDaemonFactory(),\n\t\thostMetrics: h.hostMetrics,\n\t\tcgMetrics: load.NewCGMetrics(),\n\t\tcfgMgr: h.cfgMgr,\n\t}\n\n\tcgCache.consumerM3Client = metrics.NewClientWithTags(h.m3Client, metrics.Outputhost, cgCache.getConsumerGroupTags())\n\tcgCache.loadReporter = cgCache.loadReporterFactory.CreateReporter(consGroupLoadReportingInterval, cgCache, cgLogger)\n\tcgCache.loadReporter.Start()\n\treturn cgCache\n}", "func createMetrics(\n\t_ context.Context,\n\tset receiver.CreateSettings,\n\tcfg component.Config,\n\tconsumer consumer.Metrics,\n) (receiver.Metrics, error) {\n\toCfg := cfg.(*Config)\n\tr, err := receivers.GetOrAdd(oCfg, func() (*otlpReceiver, error) {\n\t\treturn newOtlpReceiver(oCfg, set)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = r.Unwrap().registerMetricsConsumer(consumer); err != nil {\n\t\treturn nil, err\n\t}\n\treturn r, nil\n}", "func New(config *Config) (*Metrics, error) {\n\tm := &Metrics{\n\t\tconfig: config,\n\t\tcounters: make(map[string]prometheus.Counter),\n\t\tcounterVecs: make(map[string]*prometheus.CounterVec),\n\t\tgauges: make(map[string]prometheus.Gauge),\n\t\tgaugeVecs: make(map[string]*prometheus.GaugeVec),\n\t}\n\n\tif config.Enable {\n\t\tgo func() {\n\t\t\thttp.Handle(\"/metrics\", promhttp.Handler())\n\t\t\terr := http.ListenAndServe(config.Addr, http.DefaultServeMux)\n\t\t\tlog.Error().Err(err).Msg(\"could not start metrics HTTP server\")\n\t\t}()\n\t}\n\n\treturn m, nil\n}", "func newGauge(namespace, subsystem, name string, labelNames []string, client *statsd.Statter, isPrometheusEnabled bool) *Gauge {\n\topts := prometheus.GaugeOpts{\n\t\tNamespace: namespace,\n\t\tSubsystem: subsystem,\n\t\tName: name,\n\t}\n\tvec := prometheus.NewGaugeVec(opts, labelNames)\n\tif isPrometheusEnabled {\n\t\tprometheus.MustRegister(vec)\n\t}\n\n\treturn &Gauge{\n\t\twatcher: vec,\n\t\tlabels: labelNames,\n\t\tclient: client,\n\t\tprefix: strings.Join([]string{namespace, subsystem, name}, \".\"),\n\t}\n}", "func NewMetrics(factory promutil.Factory) *Metrics {\n\treturn &Metrics{\n\t\tImporterEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"importer_engine\",\n\t\t\t\tHelp: \"counting open and closed importer engines\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tIdleWorkersGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"idle_workers\",\n\t\t\t\tHelp: \"counting idle workers\",\n\t\t\t}, []string{\"name\"}),\n\n\t\tKvEncoderCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"kv_encoder\",\n\t\t\t\tHelp: \"counting kv open and closed kv encoder\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tTableCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"tables\",\n\t\t\t\tHelp: \"count number of tables processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tProcessedEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"engines\",\n\t\t\t\tHelp: \"count number of engines processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tChunkCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunks\",\n\t\t\t\tHelp: \"count number of chunks processed\",\n\t\t\t}, []string{\"state\"}),\n\n\t\tBytesCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"bytes\",\n\t\t\t\tHelp: \"count of total bytes\",\n\t\t\t}, []string{\"state\"}),\n\t\t// state can be one of:\n\t\t// - estimated (an estimation derived from the file size)\n\t\t// - pending\n\t\t// - running\n\t\t// - finished\n\t\t// - failed\n\n\t\tRowsCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"rows\",\n\t\t\t\tHelp: \"count of total rows\",\n\t\t\t}, []string{\"state\", \"table\"}),\n\n\t\tImportSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"import_seconds\",\n\t\t\t\tHelp: \"time needed to import a table\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.125, 2, 6),\n\t\t\t}),\n\n\t\tChunkParserReadBlockSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunk_parser_read_block_seconds\",\n\t\t\t\tHelp: \"time needed for chunk parser read a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tApplyWorkerSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"apply_worker_seconds\",\n\t\t\t\tHelp: \"time needed to apply a worker\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}, []string{\"name\"}),\n\n\t\tRowReadSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_seconds\",\n\t\t\t\tHelp: \"time needed to parse a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 7),\n\t\t\t}),\n\n\t\tRowReadBytesHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_bytes\",\n\t\t\t\tHelp: \"number of bytes being read out from data source\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1024, 2, 8),\n\t\t\t}),\n\n\t\tRowEncodeSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_encode_seconds\",\n\t\t\t\tHelp: \"time needed to encode a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tRowKVDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_kv_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver kvs of a single row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tBlockDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tBlockDeliverBytesHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_bytes\",\n\t\t\t\tHelp: \"number of bytes being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(512, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tBlockDeliverKVPairsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_kv_pairs\",\n\t\t\t\tHelp: \"number of KV pairs being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tChecksumSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"checksum_seconds\",\n\t\t\t\tHelp: \"time needed to complete the checksum stage\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}),\n\t\tSSTSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"sst_seconds\",\n\t\t\t\tHelp: \"time needed to complete the sst operations\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}, []string{\"kind\"}),\n\n\t\tLocalStorageUsageBytesGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"local_storage_usage_bytes\",\n\t\t\t\tHelp: \"disk/memory size currently occupied by intermediate files in local backend\",\n\t\t\t}, []string{\"medium\"}),\n\n\t\tProgressGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"progress\",\n\t\t\t\tHelp: \"progress of lightning phase\",\n\t\t\t}, []string{\"phase\"}),\n\t}\n}", "func newMonitoringMetrics() monitoringMetrics {\n\treturn monitoringMetrics{\n\t\tCSR: csrCounts.With(prometheus.Labels{}),\n\t\tAuthnError: authnErrorCounts.With(prometheus.Labels{}),\n\t\tSuccess: successCounts.With(prometheus.Labels{}),\n\t\tCSRError: csrParsingErrorCounts.With(prometheus.Labels{}),\n\t\tIDExtractionError: idExtractionErrorCounts.With(prometheus.Labels{}),\n\t\tcertSignErrors: certSignErrorCounts,\n\t}\n}", "func CreateMetrics(protocol string, host string, port int, tag string) (*Metrics, error) {\n\tvar m *Metrics\n\tswitch protocol {\n\tcase \"tcp\":\n\t\tm = &Metrics{Host: host, Port: port, Protocol: \"tcp\", Tag: tag}\n\tcase \"udp\":\n\t\tm = &Metrics{Host: host, Port: port, Protocol: \"udp\", Tag: tag}\n\t}\n\t// Initialize values\n\tm.Timeout = 0\n\tm.conn = nil\n\tm.Counters = make(map[string]Counter)\n\t// Connect\n\tif err := m.Connect(); err != nil {\n\t\treturn m, err\n\t}\n\tm.Ready = true\n\treturn m, nil\n}", "func NewMetrics(ns string) *Metrics {\n\tres := &Metrics{\n\t\tInfo: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"info\",\n\t\t\t\tHelp: \"Informations about given repository, value always 1\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"goversion\"},\n\t\t),\n\t\tDeprecated: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"deprecated\",\n\t\t\t\tHelp: \"Number of days since given dependency of repository is out-of-date\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"dependency\", \"type\", \"current\", \"latest\"},\n\t\t),\n\t\tReplaced: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"replaced\",\n\t\t\t\tHelp: \"Give information about module replacements\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"dependency\", \"type\", \"replacement\", \"version\"},\n\t\t),\n\t\tStatus: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"status\",\n\t\t\t\tHelp: \"Status of last analysis of given repository, 0 for error\",\n\t\t\t},\n\t\t\t[]string{\"repository\"},\n\t\t),\n\t\tDuration: promauto.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"duration\",\n\t\t\t\tHelp: \"Duration of last analysis in second\",\n\t\t\t},\n\t\t),\n\t\tRegistry: prometheus.NewRegistry(),\n\t}\n\n\tres.Registry.Register(res.Info)\n\tres.Registry.Register(res.Deprecated)\n\tres.Registry.Register(res.Replaced)\n\tres.Registry.Register(res.Status)\n\tres.Registry.Register(res.Duration)\n\treturn res\n}", "func NewGCM() *gcm {\n\treturn &gcm{\n\t\trandReader: cryptoRand.Reader,\n\t}\n}", "func NewMetrics(name string, r prometheus.Registerer) *Metrics {\n\treg := prometheus.WrapRegistererWith(prometheus.Labels{\"controller\": name}, r)\n\tm := Metrics{\n\t\treg: reg,\n\t\treconcileErrorsCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_reconcile_errors_total\",\n\t\t\tHelp: \"Number of errors that occurred while reconciling the statefulset\",\n\t\t}),\n\t\ttriggerByCounter: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_triggered_total\",\n\t\t\tHelp: \"Number of times a Kubernetes object add, delete or update event\" +\n\t\t\t\t\" triggered the Prometheus Operator to reconcile an object\",\n\t\t}, []string{\"triggered_by\", \"action\"}),\n\t\tstsDeleteCreateCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_reconcile_sts_delete_create_total\",\n\t\t\tHelp: \"Number of times that reconciling a statefulset required deleting and re-creating it\",\n\t\t}),\n\t\tlistCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_list_operations_total\",\n\t\t\tHelp: \"Total number of list operations\",\n\t\t}),\n\t\tlistFailedCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_list_operations_failed_total\",\n\t\t\tHelp: \"Total number of list operations that failed\",\n\t\t}),\n\t\twatchCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_watch_operations_total\",\n\t\t\tHelp: \"Total number of watch operations\",\n\t\t}),\n\t\twatchFailedCounter: prometheus.NewCounter(prometheus.CounterOpts{\n\t\t\tName: \"prometheus_operator_watch_operations_failed_total\",\n\t\t\tHelp: \"Total number of watch operations that failed\",\n\t\t}),\n\t}\n\tm.reg.MustRegister(\n\t\tm.reconcileErrorsCounter,\n\t\tm.triggerByCounter,\n\t\tm.stsDeleteCreateCounter,\n\t\tm.listCounter,\n\t\tm.listFailedCounter,\n\t\tm.watchCounter,\n\t\tm.watchFailedCounter,\n\t)\n\treturn &m\n}", "func (jbobject *MetricsMetricsSystem) CreateMetricsSystem(a string, b SparkConfInterface, c SecurityManagerInterface) *MetricsMetricsSystem {\n\tconv_a := javabind.NewGoToJavaString()\n\tconv_b := javabind.NewGoToJavaCallable()\n\tconv_c := javabind.NewGoToJavaCallable()\n\tif err := conv_a.Convert(a); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := conv_b.Convert(b); err != nil {\n\t\tpanic(err)\n\t}\n\tif err := conv_c.Convert(c); err != nil {\n\t\tpanic(err)\n\t}\n\tjret, err := javabind.GetEnv().CallStaticMethod(\"org/apache/spark/metrics/MetricsSystem\", \"createMetricsSystem\", \"org/apache/spark/metrics/MetricsSystem\", conv_a.Value().Cast(\"java/lang/String\"), conv_b.Value().Cast(\"org/apache/spark/SparkConf\"), conv_c.Value().Cast(\"org/apache/spark/SecurityManager\"))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tconv_a.CleanUp()\n\tconv_b.CleanUp()\n\tconv_c.CleanUp()\n\tretconv := javabind.NewJavaToGoCallable()\n\tdst := &javabind.Callable{}\n\tretconv.Dest(dst)\n\tif err := retconv.Convert(javabind.ObjectRef(jret)); err != nil {\n\t\tpanic(err)\n\t}\n\tretconv.CleanUp()\n\tunique_x := &MetricsMetricsSystem{}\n\tunique_x.Callable = dst\n\treturn unique_x\n}", "func NewMetrics(registry metrics.Registry) Metrics {\n\treturn &defaultMetrics{registry: registry}\n}", "func newReconciler(mgr manager.Manager) reconcile.Reconciler {\n\tlogger := logrus.WithFields(logrus.Fields{\"controller\": \"controller_cloudmetrics\"})\n\n\treturn &ReconcileCloudMetrics{\n\t\tclient: mgr.GetClient(),\n\t\tscheme: mgr.GetScheme(),\n\t\tlogger: logger,\n\t}\n}", "func createMetricsToMetrics(\n\t_ context.Context,\n\tset connector.CreateSettings,\n\tcfg component.Config,\n\tnextConsumer consumer.Metrics,\n) (connector.Metrics, error) {\n\treturn nil, nil\n}", "func New() *ClrInstGrp {\n\tvar m ClrInstGrp\n\treturn &m\n}", "func (it *Mxmx0inteccMetricsIterator) Create(key uint64) (*Mxmx0inteccMetrics, error) {\n\ttmtr := &Mxmx0inteccMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mxmx0inteccMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func newGCM(cipher goCipher.Block) (aeadIf, error) {\n\treturn newGCMWithNonceAndTagSize(cipher, gcmStandardNonceSize, gcmTagSize)\n}", "func New(name string, rate float64, tags ...string) Metric {\n\treturn Metric{name, rate, tags}\n}", "func New() handler.MetricHandler {\n\treturn &collectdMetricsHandler{}\n}", "func createMetricsReceiver(\n\tctx context.Context,\n\tparams component.ReceiverCreateSettings,\n\tconfig config.Receiver,\n\tconsumer consumer.Metrics,\n) (component.MetricsReceiver, error) {\n\n\tcfg := config.(*Config)\n\tcollector := NewVMAgeCollector(cfg.ExportInterval, cfg.BuildDate, cfg.VMImageName, cfg.VMStartTime, cfg.VMReadyTime, consumer, params.Logger)\n\n\treceiver := &Receiver{\n\t\tvmAgeCollector: collector,\n\t}\n\treturn receiver, nil\n}", "func New(client *redis.Client, group, consumer string, options ...Option) *Consumer {\n\tcfg := &config{\n\t\tgroup: group,\n\t\tconsumer: consumer,\n\t}\n\tfor _, opt := range options {\n\t\topt(cfg)\n\t}\n\tlastIDs := make(map[string]string)\n\tfor _, stream := range cfg.streams {\n\t\tlastIDs[stream] = \"0-0\"\n\t}\n\n\treturn &Consumer{\n\t\tclient: client,\n\t\tcfg: cfg,\n\t\tlastIDs: lastIDs,\n\t}\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\n\tconfig := struct{}{}\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tJmxClient: psoft.GetPsoftJMXClient(),\n\t}, nil\n}", "func NewMetric(rtype string) Metric {\n\treturn Metric{\n\t\tType: rtype,\n\t\tCurrent: map[string]int{},\n\t\tOwners: map[string]int{},\n\t}\n}", "func newHttpMetrics() *httpMetrics {\n\treturn &httpMetrics{\n\t\tRequestsTotal: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_requests_total\",\n\t\t\tHelp: \"Total number of HTTP requests.\",\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t\tRequestDurationHistogram: promauto.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_request_duration_seconds\",\n\t\t\tHelp: \"Seconds spent serving HTTP requests.\",\n\t\t\tBuckets: prometheus.DefBuckets,\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t}\n}", "func New(ctx context.Context, concurrency int) (*Group, context.Context) {\n\tif concurrency < 1 {\n\t\tconcurrency = 1\n\t}\n\n\tparent, ctx := errgroup.WithContext(ctx)\n\treturn &Group{\n\t\tlimiter: make(chan struct{}, concurrency),\n\t\tparent: parent,\n\t\tctx: ctx,\n\t}, ctx\n}", "func newCMLogger(name string, chainId string, logger *zap.SugaredLogger, logLevel log.LOG_LEVEL) *CMLogger {\n\treturn &CMLogger{name: name, chainId: chainId, SugaredLogger: logger, logLevel: logLevel}\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\treadyScope := scope.SubScope(\"ready\")\n\ttrackerScope := scope.SubScope(\"tracker\")\n\ttaskStateScope := scope.SubScope(\"tasks_state\")\n\n\treconcilerScope := scope.SubScope(\"reconciler\")\n\tleakScope := reconcilerScope.SubScope(\"leaks\")\n\tsuccessScope := reconcilerScope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := reconcilerScope.Tagged(map[string]string{\"result\": \"fail\"})\n\n\treturn &Metrics{\n\t\tReadyQueueLen: readyScope.Gauge(\"ready_queue_length\"),\n\t\tTasksCountInTracker: trackerScope.Gauge(\"task_len_tracker\"),\n\t\tTaskStatesGauge: map[task.TaskState]tally.Gauge{\n\t\t\ttask.TaskState_PENDING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_pending\"),\n\t\t\ttask.TaskState_READY: taskStateScope.Gauge(\n\t\t\t\t\"task_state_ready\"),\n\t\t\ttask.TaskState_PLACING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placing\"),\n\t\t\ttask.TaskState_PLACED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placed\"),\n\t\t\ttask.TaskState_LAUNCHING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launching\"),\n\t\t\ttask.TaskState_LAUNCHED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launched\"),\n\t\t\ttask.TaskState_RUNNING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_running\"),\n\t\t\ttask.TaskState_SUCCEEDED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_succeeded\"),\n\t\t\ttask.TaskState_FAILED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_failed\"),\n\t\t\ttask.TaskState_KILLED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_killed\"),\n\t\t\ttask.TaskState_LOST: taskStateScope.Gauge(\n\t\t\t\t\"task_state_lost\"),\n\t\t\ttask.TaskState_PREEMPTING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_preempting\"),\n\t\t},\n\t\tResourcesHeldByTaskState: map[task.TaskState]scalar.GaugeMaps{\n\t\t\ttask.TaskState_READY: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_ready\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placing\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placed\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launching\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launched\"),\n\t\t\t),\n\t\t\ttask.TaskState_RUNNING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_running\"),\n\t\t\t),\n\t\t\ttask.TaskState_STARTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_starting\"),\n\t\t\t),\n\t\t\ttask.TaskState_PREEMPTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_preempting\"),\n\t\t\t),\n\t\t\ttask.TaskState_KILLING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_killing\"),\n\t\t\t),\n\t\t},\n\t\tLeakedResources: scalar.NewGaugeMaps(leakScope),\n\t\tReconciliationSuccess: successScope.Counter(\"run\"),\n\t\tReconciliationFail: failScope.Counter(\"run\"),\n\t\tOrphanTasks: scope.Gauge(\"orphan_tasks\"),\n\t}\n}", "func (it *Msmsintprp5MetricsIterator) Create(key uint64) (*Msmsintprp5Metrics, error) {\n\ttmtr := &Msmsintprp5Metrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Msmsintprp5Metrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewGC(s *Storage, d time.Duration) *GarbageCollector {\n\treturn &GarbageCollector{s: s, period: d}\n}", "func New(dir string) *GC {\n\treturn &GC{\n\t\tdir: dir,\n\t\tmaxAge: DefaultMaxAge,\n\t\tinterval: DefaultInterval,\n\t}\n}", "func New() helper.MetricSeter {\n\treturn &MetricSeter{\n\t\tredisPools: map[string]*rd.Pool{},\n\t}\n}", "func New() *CPUCollector {\n\tcpuMetrics := newCPUMetrics()\n\tcpuStats := cpuclient.New()\n\n\treturn &CPUCollector{\n\t\tcpuMetrics: cpuMetrics,\n\t\tcpuStats: cpuStats,\n\t}\n}", "func (it *Mxmx1inteccMetricsIterator) Create(key uint64) (*Mxmx1inteccMetrics, error) {\n\ttmtr := &Mxmx1inteccMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mxmx1inteccMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}" ]
[ "0.63078237", "0.62615216", "0.61143553", "0.6085217", "0.6078891", "0.60658216", "0.605734", "0.6025629", "0.60183597", "0.5992243", "0.5982877", "0.59673464", "0.58899546", "0.5868422", "0.58567595", "0.5794954", "0.57837653", "0.5779299", "0.57310873", "0.5682949", "0.56576365", "0.56490177", "0.5646576", "0.56306577", "0.56246746", "0.5615454", "0.5611775", "0.56113887", "0.56039023", "0.5593502", "0.5582887", "0.55660623", "0.5548123", "0.55168486", "0.55164325", "0.55118006", "0.5511628", "0.55070686", "0.5501182", "0.54778296", "0.54709274", "0.5451282", "0.54506344", "0.54453576", "0.5432845", "0.5430783", "0.54301584", "0.5428831", "0.54159886", "0.54148054", "0.5412446", "0.53976154", "0.5395767", "0.53953916", "0.5391247", "0.5377524", "0.5373915", "0.53577316", "0.535557", "0.5352488", "0.5343767", "0.53435683", "0.53330225", "0.53281915", "0.52853334", "0.52813023", "0.5256713", "0.5246317", "0.5242852", "0.5238429", "0.5224102", "0.5216709", "0.5203575", "0.520282", "0.5175067", "0.51684296", "0.5167586", "0.51562226", "0.51535934", "0.5141289", "0.5124964", "0.5079903", "0.5076469", "0.50760186", "0.506106", "0.5059629", "0.50547665", "0.50382596", "0.5031044", "0.50235194", "0.5019826", "0.50169325", "0.50166076", "0.5014677", "0.5009686", "0.5004959", "0.5002147", "0.49984914", "0.49945983", "0.4992238" ]
0.7325726
0
NewExtentMetrics returns a new instance of ExtentMetrics. Should be one instance per extent uuid.
func NewExtentMetrics() *ExtentMetrics { return &ExtentMetrics{ CounterBank: common.NewCounterBank(numExtentMetrics), } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseImpl: newBaseImpl(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func NewMetrics(consume ConsumeMetricsFunc, options ...Option) (consumer.Metrics, error) {\n\tif consume == nil {\n\t\treturn nil, errNilFunc\n\t}\n\treturn &baseMetrics{\n\t\tbaseConsumer: newBaseConsumer(options...),\n\t\tConsumeMetricsFunc: consume,\n\t}, nil\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{items: make(map[string]*metric), rm: &sync.RWMutex{}}\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{}\n\tm.Reset()\n\treturn m\n}", "func newMetrics() *metrics {\n\treturn new(metrics)\n}", "func NewMetrics(period time.Duration, maxQueueSize int) (*Metrics, error) {\n\tmetrics := &Metrics{\n\t\tmaxQueueSize: maxQueueSize,\n\t\tperiod: period,\n\t\tinitialized: true,\n\t\tqueue: make([]Measurement, 0),\n\t\tlastSendingDate: -1,\n\t}\n\n\tif UseGlobalEngine {\n\t\tmetrics.Engine = Engine\n\t} else {\n\t\tmetrics.Engine = &req.Engine{}\n\t}\n\n\terr := validateMetrics(metrics)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif sources == nil {\n\t\tsources = make([]DataSource, 0)\n\t\tgo sendingLoop()\n\t}\n\n\tsources = append(sources, metrics)\n\n\treturn metrics, nil\n}", "func NewMetrics() *MetricsHolder {\n\tm := &MetricsHolder{\n\t\tlines: make(map[string]*Reading),\n\t\tchannel: make(chan interface{}),\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\tw, ok := <-m.channel\n\t\t\treading := w.(*Reading)\n\t\t\tif !ok {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif val, ok := m.lines[reading.Key]; ok {\n\t\t\t\tm.lines[reading.Key] = val.Accept(reading)\n\t\t\t} else {\n\t\t\t\tm.lines[reading.Key] = reading\n\t\t\t}\n\t\t}\n\t}()\n\treturn m\n}", "func NewMetrics() *Metrics {\n\tm := &Metrics{\n\t\tTimeMetrics: make(map[string]*TimeStats),\n\t\tNumberMetrics: make(map[string]*NumberStats),\n\t\tBoolMetrics: make(map[string]*BoolStats),\n\t}\n\treturn m\n}", "func (f *factory) NewMetricsAspect(env adapter.Env, cfg adapter.Config, metrics map[string]*adapter.MetricDefinition) (adapter.MetricsAspect, error) {\n\tvar serverErr error\n\tf.once.Do(func() {\n\t\tserverErr = f.srv.Start(env, promhttp.HandlerFor(f.registry, promhttp.HandlerOpts{}))\n\t})\n\tif serverErr != nil {\n\t\treturn nil, fmt.Errorf(\"could not start prometheus server: %v\", serverErr)\n\t}\n\n\tvar metricErr *multierror.Error\n\n\tmetricsMap := make(map[string]prometheus.Collector, len(metrics))\n\tfor _, m := range metrics {\n\t\tswitch m.Kind {\n\t\tcase adapter.Gauge:\n\t\t\tc, err := registerOrGet(f.registry, newGaugeVec(m.Name, m.Description, m.Labels))\n\t\t\tif err != nil {\n\t\t\t\tmetricErr = multierror.Append(metricErr, fmt.Errorf(\"could not register metric: %v\", err))\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tmetricsMap[m.Name] = c\n\t\tcase adapter.Counter:\n\t\t\tc, err := registerOrGet(f.registry, newCounterVec(m.Name, m.Description, m.Labels))\n\t\t\tif err != nil {\n\t\t\t\tmetricErr = multierror.Append(metricErr, fmt.Errorf(\"could not register metric: %v\", err))\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tmetricsMap[m.Name] = c\n\t\tdefault:\n\t\t\tmetricErr = multierror.Append(metricErr, fmt.Errorf(\"unknown metric kind (%d); could not register metric\", m.Kind))\n\t\t}\n\t}\n\n\treturn &prom{metricsMap}, metricErr.ErrorOrNil()\n}", "func newMetrics() *Metrics {\n\treturn newMetricsFrom(DefaultMetricsOpts)\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tPath: defaultPath,\n\t\tAddr: defaultAddr,\n\t\textraLabels: []extraLabel{},\n\t}\n}", "func NewMetrics(p fabricmetrics.Provider) *Metrics {\n\treturn &Metrics{\n\t\tRefreshTimer: p.NewHistogram(refreshTimer),\n\t}\n}", "func NewMetrics(component string, sampleRate float64, client metrics.Client) BaseMetrics {\n\treturn BaseMetrics{\n\t\tcomponent: component,\n\t\trate: sampleRate,\n\t\tmetrics: client,\n\t\tmetMap: map[string]string{\n\t\t\t\"latency\": \"comp.\" + component + \".requests.latency\",\n\t\t\t\"request\": \"comp.\" + component + \".requests.%d\",\n\t\t\t\"mLatency\": \"comp.\" + component + \".requests.%s.latency\",\n\t\t\t\"mRequest\": \"comp.\" + component + \".requests.%s.%d\",\n\t\t},\n\t}\n}", "func NewMetrics(factory metrics.Factory, globalTags map[string]string) *Metrics {\n\tm := &Metrics{}\n\tmetrics.Init(m, factory.Namespace(\"jaeger\", nil), globalTags)\n\treturn m\n}", "func NewMetricsWrapper(ctx context.Context, base api.Backend) api.Backend {\n\tmetricsOnce.Do(func() {\n\t\tprometheus.MustRegister(registeryCollectors...)\n\t})\n\n\t// XXX: When the registry backends support node deregistration,\n\t// handle this on the metrics side.\n\n\twrapper := &metricsWrapper{\n\t\tBackend: base,\n\t\tcloseCh: make(chan struct{}),\n\t\tclosedCh: make(chan struct{}),\n\t}\n\n\twrapper.updatePeriodicMetrics(ctx)\n\tgo wrapper.worker(ctx)\n\n\treturn wrapper\n}", "func NewMetrics(namespace string, logger Logger) Metrics {\n\tlog := logger.GetLogger()\n\n\treturn &metricsImpl{\n\t\tinternalMetrics: metrics.NewMetrics(\"\", log),\n\t\texternalMetrics: metrics.NewMetrics(strings.ToLower(namespace), log),\n\t}\n}", "func NewMetrics(scope tally.Scope) Metrics {\n\tscope = scope.SubScope(\"aggregation\")\n\treturn Metrics{\n\t\tCounter: newCounterMetrics(scope.SubScope(\"counters\")),\n\t\tGauge: newGaugeMetrics(scope.SubScope(\"gauges\")),\n\t}\n}", "func NewMetrics(ctx context.Context, output string, tenant string, refreshRate time.Duration) Metrics {\n\treturn Metrics{\n\t\tDaemonSupport: utils.NewDaemonSupport(ctx, \"metrics\"),\n\t\tstorage: localfs.NewPlaintextStorage(output),\n\t\ttenant: tenant,\n\t\trefreshRate: refreshRate,\n\t\tpromisesAccepted: metrics.NewCounter(),\n\t\tcommitsAccepted: metrics.NewCounter(),\n\t\trollbacksAccepted: metrics.NewCounter(),\n\t\tcreatedAccounts: metrics.NewCounter(),\n\t\tupdatedSnapshots: metrics.NewMeter(),\n\t\tsnapshotCronLatency: metrics.NewTimer(),\n\t}\n}", "func New(d *docker.Docker, containersInterval, metricsInterval time.Duration) *Metrics {\n\treturn &Metrics{\n\t\tdocker: d,\n\t\tmetricsMap: &metricsMap{\n\t\t\tmetrics: make(map[string]*docker.ContainerStats),\n\t\t},\n\t\tcInterval: containersInterval,\n\t\tmInterval: metricsInterval,\n\t}\n}", "func (p *MetadataServiceClient) CreateExtent(request *shared.CreateExtentRequest) (r *shared.CreateExtentResult_, err error) {\n if err = p.sendCreateExtent(request); err != nil { return }\n return p.recvCreateExtent()\n}", "func NewMetrics(reg prometheus.Registerer) *Metrics {\n\tvar m Metrics\n\tm.reg = reg\n\n\tm.dockerEntries = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_entries_total\",\n\t\tHelp: \"Total number of successful entries sent to the Docker target\",\n\t})\n\tm.dockerErrors = prometheus.NewCounter(prometheus.CounterOpts{\n\t\tNamespace: \"promtail\",\n\t\tName: \"docker_target_parsing_errors_total\",\n\t\tHelp: \"Total number of parsing errors while receiving Docker messages\",\n\t})\n\n\tif reg != nil {\n\t\treg.MustRegister(\n\t\t\tm.dockerEntries,\n\t\t\tm.dockerErrors,\n\t\t)\n\t}\n\n\treturn &m\n}", "func NewMetrics() *Metrics {\n\treturn &Metrics{\n\t\tRangeFeedCatchupScanNanos: metric.NewCounter(metaRangeFeedCatchupScanNanos),\n\t\tRangeFeedSlowClosedTimestampLogN: log.Every(5 * time.Second),\n\t\tRangeFeedSlowClosedTimestampNudgeSem: make(chan struct{}, 1024),\n\t}\n}", "func NewMetrics(subsystem string) *Metrics {\n\tbase := metrics.NewBase(subsystem, \"\")\n\treturn &Metrics{\n\t\tBase: base,\n\t\tFailedDatabaseMethods: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"failed_database_operations\",\n\t\t\tHelp: \"Tracks the number of database failures\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}, []string{methodLabel}),\n\t\tDepositIDMismatch: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"deposit_id_mismatch\",\n\t\t\tHelp: \"Set to 1 when the postgres and the disrburser contract \" +\n\t\t\t\t\"disagree on the next deposit id, and 0 otherwise\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tMissingDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"missing_disbursements\",\n\t\t\tHelp: \"Number of deposits that are missing disbursements in \" +\n\t\t\t\t\"postgres below our supposed next deposit id\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tSuccessfulDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"successful_disbursements\",\n\t\t\tHelp: \"Number of disbursements that emit a success event \" +\n\t\t\t\t\"from a given tx\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tFailedDisbursements: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"failed_disbursements\",\n\t\t\tHelp: \"Number of disbursements that emit a failed event \" +\n\t\t\t\t\"from a given tx\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tPostgresLastDisbursedID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"postgres_last_disbursed_id\",\n\t\t\tHelp: \"Latest recorded disbursement id in postgres\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tContractNextDisbursementID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"contract_next_disbursement_id\",\n\t\t\tHelp: \"Next disbursement id expected by the disburser contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tContractNextDepositID: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"contract_next_deposit_id\",\n\t\t\tHelp: \"next deposit id expected by the deposit contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tDisburserBalance: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"disburser_balance\",\n\t\t\tHelp: \"Balance in Wei of Teleportr's disburser wallet\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tDepositContractBalance: promauto.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"deposit_contract_balance\",\n\t\t\tHelp: \"Balance in Wei of Teleportr's deposit contract\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}),\n\t\tFailedTXSubmissions: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: \"failed_tx_submissions\",\n\t\t\tHelp: \"Number of failed transaction submissions\",\n\t\t\tSubsystem: base.SubsystemName(),\n\t\t}, []string{\n\t\t\t\"type\",\n\t\t}),\n\t}\n}", "func newCustomMetrics() ICustomMetrics {\n\n\tcounters := make(map[string]prometheus.Counter)\n\tgauges := make(map[string]prometheus.Gauge)\n\tsummaries := make(map[string]prometheus.Summary)\n\thistograms := make(map[string]prometheus.Histogram)\n\n\treturn &customMetrics{\n\t\tcounters: counters,\n\t\tgauges: gauges,\n\t\tsummaries: summaries,\n\t\thistograms: histograms,\n\t}\n}", "func newListMetrics() *listMetrics {\n\treturn new(listMetrics)\n}", "func NewMetrics() *Metrics {\n\tmtrcs := &Metrics{\n\t\tcounters: make(map[MetricName]int),\n\t\tSidecarSyncErrors: SidecarSyncErrors,\n\t\tSidecarVaultTokenErrors: SidecarVaultTokenErrors,\n\t\tSidecarSecretErrors: SidecarSecretErrors,\n\t}\n\n\treturn mtrcs\n}", "func New() *Metrics {\n\treturn &Metrics{\n\t\tSectionCounts: make(map[string]int),\n\t}\n}", "func NewMetrics(registry metrics.Registry) Metrics {\n\treturn &defaultMetrics{registry: registry}\n}", "func (r *Replicator) CreateExtent(ctx thrift.Context, createRequest *shared.CreateExtentRequest) (*shared.CreateExtentResult_, error) {\n\tlcllg := r.logger.WithFields(bark.Fields{\n\t\tcommon.TagDst: common.FmtDst(createRequest.GetExtent().GetDestinationUUID()),\n\t\tcommon.TagExt: common.FmtExt(createRequest.GetExtent().GetExtentUUID()),\n\t\tcommon.TagZoneName: common.FmtZoneName(createRequest.GetExtent().GetOriginZone()),\n\t})\n\tr.m3Client.IncCounter(metrics.ReplicatorCreateExtentScope, metrics.ReplicatorRequests)\n\n\tcontroller, err := r.GetClientFactory().GetControllerClient()\n\tif err != nil {\n\t\tlcllg.WithField(common.TagErr, err).Error(`Error getting controller client`)\n\t\tr.m3Client.IncCounter(metrics.ReplicatorCreateExtentScope, metrics.ReplicatorFailures)\n\t\treturn nil, err\n\t}\n\tres, err := controller.CreateRemoteZoneExtent(ctx, createRequest)\n\tif err != nil {\n\t\tlcllg.WithField(common.TagErr, err).Error(`Error calling controller to create extent`)\n\t\tr.m3Client.IncCounter(metrics.ReplicatorCreateExtentScope, metrics.ReplicatorFailures)\n\t\treturn nil, err\n\t}\n\n\tlcllg.Info(`Called controller to create extent`)\n\treturn res, nil\n}", "func newMetricsMetadata(cluster *string, containerInstance *string) *ecstcs.MetricsMetadata {\n\treturn &ecstcs.MetricsMetadata{\n\t\tCluster: cluster,\n\t\tContainerInstance: containerInstance,\n\t}\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tm := &Metrics{\n\t\tProcedures: map[string]*PerProcedureMetrics{},\n\t}\n\tfor _, procedure := range _procedures {\n\t\tresponseCodes := make(map[api.ResponseCode]*PerResponseCodeMetrics)\n\t\tfor _, responseCode := range api.ResponseCode_Values() {\n\t\t\tresponseCodeText, exists := _responseCodeToText[responseCode]\n\t\t\tif !exists {\n\t\t\t\tresponseCodeText = \"unknown-error\"\n\t\t\t}\n\t\t\ttag := map[string]string{\n\t\t\t\tTagProcedure: procedure,\n\t\t\t\tTagResponseCode: responseCodeText,\n\t\t\t\t// Fill empty string here so that prometheus won't panic\n\t\t\t\t// when the number of tags is changed inside subscope\n\t\t\t\tTagService: \"\",\n\t\t\t}\n\t\t\tsubscope := scope.Tagged(tag)\n\t\t\tresponseCodes[responseCode] = &PerResponseCodeMetrics{\n\t\t\t\tScope: subscope,\n\t\t\t\tCalls: subscope.Counter(MetricNameCalls),\n\t\t\t\tCallLatency: subscope.Timer(MetricNameCallLatency),\n\t\t\t}\n\t\t}\n\t\tm.Procedures[procedure] = &PerProcedureMetrics{\n\t\t\tResponseCodes: responseCodes,\n\t\t}\n\t}\n\treturn m\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\tsuccessScope := scope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := scope.Tagged(map[string]string{\"result\": \"fail\"})\n\ttimeoutScope := scope.Tagged(map[string]string{\"result\": \"timeout\"})\n\tapiScope := scope.SubScope(\"api\")\n\tserverScope := scope.SubScope(\"server\")\n\tplacement := scope.SubScope(\"placement\")\n\trecovery := scope.SubScope(\"recovery\")\n\n\treturn &Metrics{\n\t\tAPIEnqueueGangs: apiScope.Counter(\"enqueue_gangs\"),\n\t\tEnqueueGangSuccess: successScope.Counter(\"enqueue_gang\"),\n\t\tEnqueueGangFail: failScope.Counter(\"enqueue_gang\"),\n\n\t\tAPIDequeueGangs: apiScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangSuccess: successScope.Counter(\"dequeue_gangs\"),\n\t\tDequeueGangTimeout: timeoutScope.Counter(\"dequeue_gangs\"),\n\n\t\tAPIGetPreemptibleTasks: apiScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksSuccess: successScope.Counter(\"get_preemptible_tasks\"),\n\t\tGetPreemptibleTasksTimeout: timeoutScope.Counter(\"get_preemptible_tasks\"),\n\n\t\tAPISetPlacements: apiScope.Counter(\"set_placements\"),\n\t\tSetPlacementSuccess: successScope.Counter(\"set_placements\"),\n\t\tSetPlacementFail: failScope.Counter(\"set_placements\"),\n\n\t\tAPIGetPlacements: apiScope.Counter(\"get_placements\"),\n\t\tGetPlacementSuccess: successScope.Counter(\"get_placements\"),\n\t\tGetPlacementFail: failScope.Counter(\"get_placements\"),\n\n\t\tAPILaunchedTasks: apiScope.Counter(\"launched_tasks\"),\n\n\t\tRecoverySuccess: successScope.Counter(\"recovery\"),\n\t\tRecoveryFail: failScope.Counter(\"recovery\"),\n\t\tRecoveryRunningSuccessCount: successScope.Counter(\"task_count\"),\n\t\tRecoveryRunningFailCount: failScope.Counter(\"task_count\"),\n\t\tRecoveryEnqueueFailedCount: failScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryEnqueueSuccessCount: successScope.Counter(\"enqueue_task_count\"),\n\t\tRecoveryTimer: recovery.Timer(\"running_tasks\"),\n\n\t\tPlacementQueueLen: placement.Gauge(\"placement_queue_length\"),\n\t\tPlacementFailed: placement.Counter(\"fail\"),\n\n\t\tElected: serverScope.Gauge(\"elected\"),\n\t}\n}", "func NewMetricsMiddleware(metricsReporter MetricsReporter) endpoint.Middleware {\n\treturn func(next endpoint.Endpoint) endpoint.Endpoint {\n\t\treturn func(ctx context.Context, request interface{}) (resp interface{}, err error) {\n\t\t\tdefer func(startTime time.Time) {\n\t\t\t\tmetricsReporter.ReportRequest(ctx, startTime, ctxRequestPath(ctx), err)\n\t\t\t}(time.Now())\n\t\t\treturn next(ctx, request)\n\t\t}\n\t}\n}", "func NewMetrics(factory promutil.Factory) *Metrics {\n\treturn &Metrics{\n\t\tImporterEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"importer_engine\",\n\t\t\t\tHelp: \"counting open and closed importer engines\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tIdleWorkersGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"idle_workers\",\n\t\t\t\tHelp: \"counting idle workers\",\n\t\t\t}, []string{\"name\"}),\n\n\t\tKvEncoderCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"kv_encoder\",\n\t\t\t\tHelp: \"counting kv open and closed kv encoder\",\n\t\t\t}, []string{\"type\"}),\n\n\t\tTableCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"tables\",\n\t\t\t\tHelp: \"count number of tables processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tProcessedEngineCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"engines\",\n\t\t\t\tHelp: \"count number of engines processed\",\n\t\t\t}, []string{\"state\", \"result\"}),\n\n\t\tChunkCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunks\",\n\t\t\t\tHelp: \"count number of chunks processed\",\n\t\t\t}, []string{\"state\"}),\n\n\t\tBytesCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"bytes\",\n\t\t\t\tHelp: \"count of total bytes\",\n\t\t\t}, []string{\"state\"}),\n\t\t// state can be one of:\n\t\t// - estimated (an estimation derived from the file size)\n\t\t// - pending\n\t\t// - running\n\t\t// - finished\n\t\t// - failed\n\n\t\tRowsCounter: factory.NewCounterVec(\n\t\t\tprometheus.CounterOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"rows\",\n\t\t\t\tHelp: \"count of total rows\",\n\t\t\t}, []string{\"state\", \"table\"}),\n\n\t\tImportSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"import_seconds\",\n\t\t\t\tHelp: \"time needed to import a table\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.125, 2, 6),\n\t\t\t}),\n\n\t\tChunkParserReadBlockSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"chunk_parser_read_block_seconds\",\n\t\t\t\tHelp: \"time needed for chunk parser read a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tApplyWorkerSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"apply_worker_seconds\",\n\t\t\t\tHelp: \"time needed to apply a worker\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}, []string{\"name\"}),\n\n\t\tRowReadSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_seconds\",\n\t\t\t\tHelp: \"time needed to parse a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 7),\n\t\t\t}),\n\n\t\tRowReadBytesHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_read_bytes\",\n\t\t\t\tHelp: \"number of bytes being read out from data source\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1024, 2, 8),\n\t\t\t}),\n\n\t\tRowEncodeSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_encode_seconds\",\n\t\t\t\tHelp: \"time needed to encode a row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tRowKVDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"row_kv_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver kvs of a single row\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\n\t\tBlockDeliverSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_seconds\",\n\t\t\t\tHelp: \"time needed to deliver a block\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(0.001, 3.1622776601683795, 10),\n\t\t\t}),\n\t\tBlockDeliverBytesHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_bytes\",\n\t\t\t\tHelp: \"number of bytes being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(512, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tBlockDeliverKVPairsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"block_deliver_kv_pairs\",\n\t\t\t\tHelp: \"number of KV pairs being sent out to importer\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2, 10),\n\t\t\t}, []string{\"kind\"}),\n\t\tChecksumSecondsHistogram: factory.NewHistogram(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"checksum_seconds\",\n\t\t\t\tHelp: \"time needed to complete the checksum stage\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}),\n\t\tSSTSecondsHistogram: factory.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"sst_seconds\",\n\t\t\t\tHelp: \"time needed to complete the sst operations\",\n\t\t\t\tBuckets: prometheus.ExponentialBuckets(1, 2.2679331552660544, 10),\n\t\t\t}, []string{\"kind\"}),\n\n\t\tLocalStorageUsageBytesGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"local_storage_usage_bytes\",\n\t\t\t\tHelp: \"disk/memory size currently occupied by intermediate files in local backend\",\n\t\t\t}, []string{\"medium\"}),\n\n\t\tProgressGauge: factory.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: \"lightning\",\n\t\t\t\tName: \"progress\",\n\t\t\t\tHelp: \"progress of lightning phase\",\n\t\t\t}, []string{\"phase\"}),\n\t}\n}", "func NewMetrics(app, metricsPrefix, version, hash, date string) *Metrics {\n\tlabels := map[string]string{\n\t\t\"app\": app,\n\t\t\"version\": version,\n\t\t\"hash\": hash,\n\t\t\"buildTime\": date,\n\t}\n\n\tif metricsPrefix != \"\" {\n\t\tmetricsPrefix += \"_\"\n\t}\n\n\tpm := &Metrics{\n\t\tresponseTime: prometheus.NewHistogramVec(\n\t\t\tprometheus.HistogramOpts{\n\t\t\t\tName: metricsPrefix + \"response_time_seconds\",\n\t\t\t\tHelp: \"Description\",\n\t\t\t\tConstLabels: labels,\n\t\t\t},\n\t\t\t[]string{\"endpoint\"},\n\t\t),\n\t\ttotalRequests: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_total\",\n\t\t\tHelp: \"number of requests\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tduration: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_duration_seconds\",\n\t\t\tHelp: \"duration of a requests in seconds\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\", \"endpoint\"}),\n\t\tresponseSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"response_size_bytes\",\n\t\t\tHelp: \"size of the responses in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\trequestSize: prometheus.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tName: metricsPrefix + \"requests_size_bytes\",\n\t\t\tHelp: \"size of the requests in bytes\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"code\", \"method\"}),\n\t\thandlerStatuses: prometheus.NewCounterVec(prometheus.CounterOpts{\n\t\t\tName: metricsPrefix + \"requests_statuses_total\",\n\t\t\tHelp: \"count number of responses per status\",\n\t\t\tConstLabels: labels,\n\t\t}, []string{\"method\", \"status_bucket\"}),\n\t}\n\n\terr := prometheus.Register(pm)\n\tif e := new(prometheus.AlreadyRegisteredError); errors.As(err, e) {\n\t\treturn pm\n\t} else if err != nil {\n\t\tpanic(err)\n\t}\n\n\tgrpcPrometheus.EnableHandlingTimeHistogram()\n\n\treturn pm\n}", "func NewMeasurement(name string) Measurement {\n\tattrs := make(map[string]interface{})\n\treturn Measurement{\n\t\tName: name,\n\t\tAttributes: attrs,\n\t}\n}", "func NewExtentInCore(name string, extentID uint64) *Extent {\n\te := new(Extent)\n\te.extentID = extentID\n\te.filePath = name\n\n\treturn e\n}", "func NewHostMetrics() *HostMetrics {\n\tthis := HostMetrics{}\n\treturn &this\n}", "func NewBaseMetrics() (*BaseMetrics, error) {\n\tm, err := metrics.NewBaseMemoryStats(nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &BaseMetrics{innerMetrics: m}, nil\n}", "func NewMetricsExtracor(holder *MetricsHolder) *MetricsExtracor {\n\tchannel := make(chan interface{})\n\tm := &MetricsExtracor{\n\t\tchannel: channel,\n\t\tholder: holder,\n\t}\n\tgo func() {\n\t\tdone := false\n\t\tfor !done {\n\t\t\tselect {\n\t\t\tcase o := <-channel:\n\t\t\t\tswitch v := o.(type) {\n\t\t\t\tcase apiv1.Pod:\n\t\t\t\t\tm.propagatePodSpecs(v)\n\t\t\t\tcase metricsv1beta1.PodMetrics:\n\t\t\t\t\tm.propagatePodMetrics(v)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}()\n\treturn m\n}", "func NewServerMetrics(enableHistogram bool) *ServerMetrics {\n\treturn &ServerMetrics{\n\t\tenableHistogram,\n\t}\n}", "func NewMeasurement(name string, idx *DatabaseIndex) *Measurement {\n\treturn &Measurement{\n\t\tName: name,\n\t\tfieldNames: make(map[string]struct{}),\n\t\tindex: idx,\n\n\t\tseries: make(map[string]*Series),\n\t\tseriesByID: make(map[uint64]*Series),\n\t\tseriesByTagKeyValue: make(map[string]map[string]seriesIDs),\n\t\tseriesIDs: make(seriesIDs, 0),\n\t}\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t}, nil\n}", "func NewMetrics(reg *prometheus.Registry, namespace, subsystem string, methodsFrom interface{}) (metric Metrics) {\n\tmetric.callErrTotal = prometheus.NewCounterVec(\n\t\tprometheus.CounterOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"errors_total\",\n\t\t\tHelp: \"Amount of DAL errors.\",\n\t\t},\n\t\t[]string{methodLabel},\n\t)\n\treg.MustRegister(metric.callErrTotal)\n\tmetric.callDuration = prometheus.NewHistogramVec(\n\t\tprometheus.HistogramOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"call_duration_seconds\",\n\t\t\tHelp: \"DAL call latency.\",\n\t\t},\n\t\t[]string{methodLabel},\n\t)\n\treg.MustRegister(metric.callDuration)\n\n\tfor _, methodName := range reflectx.MethodsOf(methodsFrom) {\n\t\tl := prometheus.Labels{\n\t\t\tmethodLabel: methodName,\n\t\t}\n\t\tmetric.callErrTotal.With(l)\n\t\tmetric.callDuration.With(l)\n\t}\n\n\treturn metric\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\treturn &MetricSet{base}, nil\n}", "func New(evt *eventbus.EventBus) *TileStats {\n\tret := &TileStats{\n\t\tevt: evt,\n\t\tstats: map[string]*TraceStats{},\n\t}\n\tevt.SubscribeAsync(db.NEW_TILE_AVAILABLE_EVENT, func(it interface{}) {\n\t\ttile := it.(*tiling.Tile)\n\t\tglog.Info(\"TileStats: Beginning.\")\n\t\tret.calcStats(tile)\n\t\tglog.Info(\"TileStats: Finished.\")\n\t})\n\n\treturn ret\n}", "func newMonitoringMetrics() monitoringMetrics {\n\treturn monitoringMetrics{\n\t\tCSR: csrCounts.With(prometheus.Labels{}),\n\t\tAuthnError: authnErrorCounts.With(prometheus.Labels{}),\n\t\tSuccess: successCounts.With(prometheus.Labels{}),\n\t\tCSRError: csrParsingErrorCounts.With(prometheus.Labels{}),\n\t\tIDExtractionError: idExtractionErrorCounts.With(prometheus.Labels{}),\n\t\tcertSignErrors: certSignErrorCounts,\n\t}\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\n\tconfig := struct{}{}\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tJmxClient: psoft.GetPsoftJMXClient(),\n\t}, nil\n}", "func NewMeasurement(name string) Measurement {\n\treturn Measurement{\n\t\tname: name,\n\t\ttagSet: map[string]string{},\n\t\tfieldSet: map[string]interface{}{},\n\t\ttimestamp: time.Now(),\n\t}\n}", "func NewMetrics(registry metrics.Registry, exchanges []openrtb_ext.BidderName, disableAccountMetrics config.DisabledMetrics, syncerKeys []string, moduleStageNames map[string][]string) *Metrics {\n\tnewMetrics := NewBlankMetrics(registry, exchanges, disableAccountMetrics, moduleStageNames)\n\tnewMetrics.ConnectionCounter = metrics.GetOrRegisterCounter(\"active_connections\", registry)\n\tnewMetrics.TMaxTimeoutCounter = metrics.GetOrRegisterCounter(\"tmax_timeout\", registry)\n\tnewMetrics.ConnectionAcceptErrorMeter = metrics.GetOrRegisterMeter(\"connection_accept_errors\", registry)\n\tnewMetrics.ConnectionCloseErrorMeter = metrics.GetOrRegisterMeter(\"connection_close_errors\", registry)\n\tnewMetrics.ImpMeter = metrics.GetOrRegisterMeter(\"imps_requested\", registry)\n\n\tnewMetrics.ImpsTypeBanner = metrics.GetOrRegisterMeter(\"imp_banner\", registry)\n\tnewMetrics.ImpsTypeVideo = metrics.GetOrRegisterMeter(\"imp_video\", registry)\n\tnewMetrics.ImpsTypeAudio = metrics.GetOrRegisterMeter(\"imp_audio\", registry)\n\tnewMetrics.ImpsTypeNative = metrics.GetOrRegisterMeter(\"imp_native\", registry)\n\n\tnewMetrics.NoCookieMeter = metrics.GetOrRegisterMeter(\"no_cookie_requests\", registry)\n\tnewMetrics.AppRequestMeter = metrics.GetOrRegisterMeter(\"app_requests\", registry)\n\tnewMetrics.DebugRequestMeter = metrics.GetOrRegisterMeter(\"debug_requests\", registry)\n\tnewMetrics.RequestTimer = metrics.GetOrRegisterTimer(\"request_time\", registry)\n\tnewMetrics.DNSLookupTimer = metrics.GetOrRegisterTimer(\"dns_lookup_time\", registry)\n\tnewMetrics.TLSHandshakeTimer = metrics.GetOrRegisterTimer(\"tls_handshake_time\", registry)\n\tnewMetrics.PrebidCacheRequestTimerSuccess = metrics.GetOrRegisterTimer(\"prebid_cache_request_time.ok\", registry)\n\tnewMetrics.PrebidCacheRequestTimerError = metrics.GetOrRegisterTimer(\"prebid_cache_request_time.err\", registry)\n\tnewMetrics.StoredResponsesMeter = metrics.GetOrRegisterMeter(\"stored_responses\", registry)\n\tnewMetrics.OverheadTimer = makeOverheadTimerMetrics(registry)\n\tnewMetrics.BidderServerResponseTimer = metrics.GetOrRegisterTimer(\"bidder_server_response_time_seconds\", registry)\n\n\tfor _, dt := range StoredDataTypes() {\n\t\tfor _, ft := range StoredDataFetchTypes() {\n\t\t\ttimerName := fmt.Sprintf(\"stored_%s_fetch_time.%s\", string(dt), string(ft))\n\t\t\tnewMetrics.StoredDataFetchTimer[dt][ft] = metrics.GetOrRegisterTimer(timerName, registry)\n\t\t}\n\t\tfor _, e := range StoredDataErrors() {\n\t\t\tmeterName := fmt.Sprintf(\"stored_%s_error.%s\", string(dt), string(e))\n\t\t\tnewMetrics.StoredDataErrorMeter[dt][e] = metrics.GetOrRegisterMeter(meterName, registry)\n\t\t}\n\t}\n\n\tnewMetrics.AmpNoCookieMeter = metrics.GetOrRegisterMeter(\"amp_no_cookie_requests\", registry)\n\n\tnewMetrics.CookieSyncMeter = metrics.GetOrRegisterMeter(\"cookie_sync_requests\", registry)\n\tfor _, s := range CookieSyncStatuses() {\n\t\tnewMetrics.CookieSyncStatusMeter[s] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"cookie_sync_requests.%s\", s), registry)\n\t}\n\n\tnewMetrics.SetUidMeter = metrics.GetOrRegisterMeter(\"setuid_requests\", registry)\n\tfor _, s := range SetUidStatuses() {\n\t\tnewMetrics.SetUidStatusMeter[s] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"setuid_requests.%s\", s), registry)\n\t}\n\n\tfor _, syncerKey := range syncerKeys {\n\t\tnewMetrics.SyncerRequestsMeter[syncerKey] = make(map[SyncerCookieSyncStatus]metrics.Meter)\n\t\tfor _, status := range SyncerRequestStatuses() {\n\t\t\tnewMetrics.SyncerRequestsMeter[syncerKey][status] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"syncer.%s.request.%s\", syncerKey, status), registry)\n\t\t}\n\n\t\tnewMetrics.SyncerSetsMeter[syncerKey] = make(map[SyncerSetUidStatus]metrics.Meter)\n\t\tfor _, status := range SyncerSetUidStatuses() {\n\t\t\tnewMetrics.SyncerSetsMeter[syncerKey][status] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"syncer.%s.set.%s\", syncerKey, status), registry)\n\t\t}\n\t}\n\n\tfor _, a := range exchanges {\n\t\tregisterAdapterMetrics(registry, \"adapter\", string(a), newMetrics.AdapterMetrics[a])\n\t}\n\n\tfor typ, statusMap := range newMetrics.RequestStatuses {\n\t\tfor stat := range statusMap {\n\t\t\tstatusMap[stat] = metrics.GetOrRegisterMeter(\"requests.\"+string(stat)+\".\"+string(typ), registry)\n\t\t}\n\t}\n\n\tfor _, cacheRes := range CacheResults() {\n\t\tnewMetrics.StoredReqCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"stored_request_cache_%s\", string(cacheRes)), registry)\n\t\tnewMetrics.StoredImpCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"stored_imp_cache_%s\", string(cacheRes)), registry)\n\t\tnewMetrics.AccountCacheMeter[cacheRes] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"account_cache_%s\", string(cacheRes)), registry)\n\t}\n\n\tnewMetrics.RequestsQueueTimer[\"video\"][true] = metrics.GetOrRegisterTimer(\"queued_requests.video.accepted\", registry)\n\tnewMetrics.RequestsQueueTimer[\"video\"][false] = metrics.GetOrRegisterTimer(\"queued_requests.video.rejected\", registry)\n\n\tnewMetrics.TimeoutNotificationSuccess = metrics.GetOrRegisterMeter(\"timeout_notification.ok\", registry)\n\tnewMetrics.TimeoutNotificationFailure = metrics.GetOrRegisterMeter(\"timeout_notification.failed\", registry)\n\n\tnewMetrics.PrivacyCCPARequest = metrics.GetOrRegisterMeter(\"privacy.request.ccpa.specified\", registry)\n\tnewMetrics.PrivacyCCPARequestOptOut = metrics.GetOrRegisterMeter(\"privacy.request.ccpa.opt-out\", registry)\n\tnewMetrics.PrivacyCOPPARequest = metrics.GetOrRegisterMeter(\"privacy.request.coppa\", registry)\n\tnewMetrics.PrivacyLMTRequest = metrics.GetOrRegisterMeter(\"privacy.request.lmt\", registry)\n\tfor _, version := range TCFVersions() {\n\t\tnewMetrics.PrivacyTCFRequestVersion[version] = metrics.GetOrRegisterMeter(fmt.Sprintf(\"privacy.request.tcf.%s\", string(version)), registry)\n\t}\n\n\tnewMetrics.AdsCertRequestsSuccess = metrics.GetOrRegisterMeter(\"ads_cert_requests.ok\", registry)\n\tnewMetrics.AdsCertRequestsFailure = metrics.GetOrRegisterMeter(\"ads_cert_requests.failed\", registry)\n\tnewMetrics.adsCertSignTimer = metrics.GetOrRegisterTimer(\"ads_cert_sign_time\", registry)\n\n\tfor module, stages := range moduleStageNames {\n\t\tregisterModuleMetrics(registry, module, stages, newMetrics.ModuleMetrics[module])\n\t}\n\n\treturn newMetrics\n}", "func newMetrics(hostAndPort string) *metrics {\n\tm := metrics{\n\t\tmetricsCh: make(chan metricType),\n\t\thostAndPort: hostAndPort,\n\t}\n\n\treturn &m\n}", "func NewMeasurement(name string) *Measurement {\n\treturn &Measurement{\n\t\tname: name,\n\t\tfields: make(map[string]*Field),\n\t\tseries: make(map[uint32]*Series),\n\t}\n}", "func NewSetMetrics(name string) *SetMetrics {\n\tlabels := prometheus.Labels{\n\t\t\"query\": name,\n\t}\n\n\tm := &SetMetrics{\n\t\tName: name,\n\n\t\tSize: prometheus.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"set_size\",\n\t\t\tHelp: \"Size of the set.\",\n\t\t\tConstLabels: labels,\n\t\t}),\n\n\t\tAdditions: prometheus.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"set_additions\",\n\t\t\tHelp: \"Number of additions to the set.\",\n\t\t\tConstLabels: labels,\n\t\t}),\n\n\t\tRemovals: prometheus.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"set_removals\",\n\t\t\tHelp: \"Number of removals from the set.\",\n\t\t\tConstLabels: labels,\n\t\t}),\n\n\t\tChanges: prometheus.NewGauge(prometheus.GaugeOpts{\n\t\t\tName: \"set_changes\",\n\t\t\tHelp: \"Number of item that changed in the set.\",\n\t\t\tConstLabels: labels,\n\t\t}),\n\t}\n\n\tprometheus.MustRegister(m.Size)\n\tprometheus.MustRegister(m.Additions)\n\tprometheus.MustRegister(m.Removals)\n\tprometheus.MustRegister(m.Changes)\n\n\treturn m\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tvar config Config\n\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tcfg: config,\n\t}, nil\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\tcfgwarn.Experimental(\"The lxd info metricset is experimental.\")\n\n\tconfig := lxd.DefaultConfig()\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\tserverConnection, err := lxd.NewLXDClient(base.HostData().URI, config, base.Module().Config().Timeout)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t\tlxdClient: serverConnection,\n\t}, nil\n}", "func newMetricsWriter(w http.ResponseWriter, r *http.Request, collector collector) *metricWriter {\n\tinfo := &Info{TimeStart: time.Now(), Request: r, Header: w.Header()}\n\treturn &metricWriter{w: w, info: info, collector: collector}\n}", "func newProcessMetrics(id string) *processMetrics {\n\tcommonTags := tags{TAG_INGESTER_ID: id, TAG_INGESTER_SOURCE: \"poll\"}\n\treturn &processMetrics{\n\t\tignoredByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"ignored\"}),\n\t\tprocessedByPollingGauge: metrics2.GetInt64Metric(MEASUREMENT_INGESTION, commonTags, tags{TAG_INGESTION_METRIC: \"processed\"}),\n\t\tliveness: metrics2.NewLiveness(id, tags{TAG_INGESTER_SOURCE: \"poll\", TAG_INGESTION_METRIC: \"since-last-run\"}),\n\t}\n}", "func createMetricsToMetrics(\n\t_ context.Context,\n\tset connector.CreateSettings,\n\tcfg component.Config,\n\tnextConsumer consumer.Metrics,\n) (connector.Metrics, error) {\n\treturn nil, nil\n}", "func (m *Manager) WithAddedMetrics(m2 metrics.Type) interop.Manager { return m }", "func NewMetricsWrapper(registry metrics.Registry, service string) *MetricsWrapper {\n\tvar metricsWrapper = MetricsWrapper{\n\t\tregistry: registry,\n\t\tserviceName: service,\n\t}\n\n\treturn &metricsWrapper\n}", "func newTenantLimits(manager *runtimeconfig.Manager) validation.TenantLimits {\n\treturn &runtimeConfigTenantLimits{\n\t\tmanager: manager,\n\t}\n}", "func New(config *Config) (*Metrics, error) {\n\tm := &Metrics{\n\t\tconfig: config,\n\t\tcounters: make(map[string]prometheus.Counter),\n\t\tcounterVecs: make(map[string]*prometheus.CounterVec),\n\t\tgauges: make(map[string]prometheus.Gauge),\n\t\tgaugeVecs: make(map[string]*prometheus.GaugeVec),\n\t}\n\n\tif config.Enable {\n\t\tgo func() {\n\t\t\thttp.Handle(\"/metrics\", promhttp.Handler())\n\t\t\terr := http.ListenAndServe(config.Addr, http.DefaultServeMux)\n\t\t\tlog.Error().Err(err).Msg(\"could not start metrics HTTP server\")\n\t\t}()\n\t}\n\n\treturn m, nil\n}", "func New(base mb.BaseMetricSet) (mb.MetricSet, error) {\n\n\tconfig := struct{}{}\n\n\tif err := base.Module().UnpackConfig(&config); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &MetricSet{\n\t\tBaseMetricSet: base,\n\t}, nil\n}", "func NewDimensions(dimensions string) *Dimensions {\n\tparts := strings.Split(dimensions, \"x\")\n\tdw, _ := strconv.Atoi(parts[0])\n\tdh, _ := strconv.Atoi(parts[1])\n\n\treturn &Dimensions{\n\t\tString: dimensions,\n\t\tWidth: dw,\n\t\tHeight: dh,\n\t}\n}", "func New(name string, rate float64, tags ...string) Metric {\n\treturn Metric{name, rate, tags}\n}", "func NewMetrics(scope tally.Scope) *Metrics {\n\treadyScope := scope.SubScope(\"ready\")\n\ttrackerScope := scope.SubScope(\"tracker\")\n\ttaskStateScope := scope.SubScope(\"tasks_state\")\n\n\treconcilerScope := scope.SubScope(\"reconciler\")\n\tleakScope := reconcilerScope.SubScope(\"leaks\")\n\tsuccessScope := reconcilerScope.Tagged(map[string]string{\"result\": \"success\"})\n\tfailScope := reconcilerScope.Tagged(map[string]string{\"result\": \"fail\"})\n\n\treturn &Metrics{\n\t\tReadyQueueLen: readyScope.Gauge(\"ready_queue_length\"),\n\t\tTasksCountInTracker: trackerScope.Gauge(\"task_len_tracker\"),\n\t\tTaskStatesGauge: map[task.TaskState]tally.Gauge{\n\t\t\ttask.TaskState_PENDING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_pending\"),\n\t\t\ttask.TaskState_READY: taskStateScope.Gauge(\n\t\t\t\t\"task_state_ready\"),\n\t\t\ttask.TaskState_PLACING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placing\"),\n\t\t\ttask.TaskState_PLACED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_placed\"),\n\t\t\ttask.TaskState_LAUNCHING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launching\"),\n\t\t\ttask.TaskState_LAUNCHED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_launched\"),\n\t\t\ttask.TaskState_RUNNING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_running\"),\n\t\t\ttask.TaskState_SUCCEEDED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_succeeded\"),\n\t\t\ttask.TaskState_FAILED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_failed\"),\n\t\t\ttask.TaskState_KILLED: taskStateScope.Gauge(\n\t\t\t\t\"task_state_killed\"),\n\t\t\ttask.TaskState_LOST: taskStateScope.Gauge(\n\t\t\t\t\"task_state_lost\"),\n\t\t\ttask.TaskState_PREEMPTING: taskStateScope.Gauge(\n\t\t\t\t\"task_state_preempting\"),\n\t\t},\n\t\tResourcesHeldByTaskState: map[task.TaskState]scalar.GaugeMaps{\n\t\t\ttask.TaskState_READY: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_ready\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placing\"),\n\t\t\t),\n\t\t\ttask.TaskState_PLACED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_placed\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launching\"),\n\t\t\t),\n\t\t\ttask.TaskState_LAUNCHED: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_launched\"),\n\t\t\t),\n\t\t\ttask.TaskState_RUNNING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_running\"),\n\t\t\t),\n\t\t\ttask.TaskState_STARTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_starting\"),\n\t\t\t),\n\t\t\ttask.TaskState_PREEMPTING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_preempting\"),\n\t\t\t),\n\t\t\ttask.TaskState_KILLING: scalar.NewGaugeMaps(\n\t\t\t\tscope.SubScope(\"task_state_killing\"),\n\t\t\t),\n\t\t},\n\t\tLeakedResources: scalar.NewGaugeMaps(leakScope),\n\t\tReconciliationSuccess: successScope.Counter(\"run\"),\n\t\tReconciliationFail: failScope.Counter(\"run\"),\n\t\tOrphanTasks: scope.Gauge(\"orphan_tasks\"),\n\t}\n}", "func newMetricsFrom(opts *MetricsOpts) *Metrics {\n\tmetrics := &Metrics{\n\t\tcounters: make(map[string]prometheus.Counter, 512),\n\t\tgauges: make(map[string]prometheus.Gauge, 512),\n\t\thistorams: make(map[string]prometheus.Histogram, 512),\n\t\tsummaries: make(map[string]prometheus.Summary, 512),\n\t\tdefBuckets: opts.DefBuckets,\n\t\tdefQuantile: opts.DefQuantile,\n\t\tregistry: prometheus.NewRegistry(),\n\t}\n\treturn metrics\n}", "func NewMetricsHandler() http.Handler {\n\treturn promhttp.Handler()\n}", "func NewMetricsMiddleware(metrics *util.Metrics) Middleware {\n\treturn &metricsMiddleware{\n\t\thistogram: metrics.NewHistogram(\n\t\t\tfalse,\n\t\t\thistogramName,\n\t\t\t\"duration histogram of http requests\",\n\t\t\tdefaultBuckets,\n\t\t\tdefaultLabels,\n\t\t),\n\t\tsummary: metrics.NewSummary(\n\t\t\tfalse,\n\t\t\tsummaryName,\n\t\t\t\"duration summary of http requests\",\n\t\t\tdefaultQuantiles,\n\t\t\tdefaultLabels,\n\t\t),\n\t}\n}", "func NewProcMetrics() *ProcMetrics {\n\treturn NewProcMetricsWith(stats.DefaultEngine, os.Getpid())\n}", "func (it *Msmsintprp5MetricsIterator) Create(key uint64) (*Msmsintprp5Metrics, error) {\n\ttmtr := &Msmsintprp5Metrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Msmsintprp5Metrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func New(metrics ...interface{}) Master {\n\tvar sentries []Sentry\n\tvar entries []Metrics\n\n\tfor _, item := range metrics {\n\t\tswitch rItem := item.(type) {\n\t\tcase Metrics:\n\t\t\tentries = append(entries, rItem)\n\t\tcase Sentry:\n\t\t\tsentries = append(sentries, rItem)\n\t\t}\n\t}\n\n\treturn Master{\n\t\tmetrics: append(entries, Sentries(sentries...)),\n\t}\n}", "func NewMetrics(ns string) *Metrics {\n\tres := &Metrics{\n\t\tInfo: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"info\",\n\t\t\t\tHelp: \"Informations about given repository, value always 1\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"goversion\"},\n\t\t),\n\t\tDeprecated: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"deprecated\",\n\t\t\t\tHelp: \"Number of days since given dependency of repository is out-of-date\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"dependency\", \"type\", \"current\", \"latest\"},\n\t\t),\n\t\tReplaced: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"replaced\",\n\t\t\t\tHelp: \"Give information about module replacements\",\n\t\t\t},\n\t\t\t[]string{\"module\", \"dependency\", \"type\", \"replacement\", \"version\"},\n\t\t),\n\t\tStatus: promauto.NewGaugeVec(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"status\",\n\t\t\t\tHelp: \"Status of last analysis of given repository, 0 for error\",\n\t\t\t},\n\t\t\t[]string{\"repository\"},\n\t\t),\n\t\tDuration: promauto.NewGauge(\n\t\t\tprometheus.GaugeOpts{\n\t\t\t\tNamespace: ns,\n\t\t\t\tName: \"duration\",\n\t\t\t\tHelp: \"Duration of last analysis in second\",\n\t\t\t},\n\t\t),\n\t\tRegistry: prometheus.NewRegistry(),\n\t}\n\n\tres.Registry.Register(res.Info)\n\tres.Registry.Register(res.Deprecated)\n\tres.Registry.Register(res.Replaced)\n\tres.Registry.Register(res.Status)\n\tres.Registry.Register(res.Duration)\n\treturn res\n}", "func newMigratorMetrics(registerFunc func(k8smetrics.Registerable) error) *migratorMetrics {\n\t// objectMigrates is defined in kube-storave-version-migrator\n\tobjectsMigrated := k8smetrics.NewCounterVec(\n\t\t&k8smetrics.CounterOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"migrated_objects\",\n\t\t\tHelp: \"The total number of objects that have been migrated, labeled with the full resource name\",\n\t\t}, []string{\"resource\"})\n\tregisterFunc(objectsMigrated)\n\n\t// migration is defined in kube-storave-version-migrator\n\tmigration := k8smetrics.NewCounterVec(\n\t\t&k8smetrics.CounterOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"migrations\",\n\t\t\tHelp: \"The total number of completed migration, labeled with the full resource name, and the status of the migration (failed or succeeded)\",\n\t\t}, []string{\"resource\", \"status\"})\n\tregisterFunc(migration)\n\n\t// migrationDuration is not defined upstream but uses the same Namespace and Subsystem\n\t// as the other metrics that are defined in kube-storave-version-migrator\n\tmigrationDuration := k8smetrics.NewHistogramVec(\n\t\t&k8smetrics.HistogramOpts{\n\t\t\tNamespace: namespace,\n\t\t\tSubsystem: subsystem,\n\t\t\tName: \"migration_duration_seconds\",\n\t\t\tHelp: \"How long a successful migration takes in seconds, labeled with the full resource name\",\n\t\t\tBuckets: prometheus.ExponentialBuckets(120, 2, 7),\n\t\t}, []string{\"resource\"})\n\tregisterFunc(migrationDuration)\n\n\treturn &migratorMetrics{\n\t\tobjectsMigrated: objectsMigrated,\n\t\tmigration: migration,\n\t\tmigrationDuration: migrationDuration,\n\t}\n}", "func NewRequestMetrics(registerer prometheus.Registerer) alice.Constructor {\n\treturn func(next http.Handler) http.Handler {\n\t\t// Counter for all requests\n\t\t// This is bucketed based on the response code we set\n\t\tcounterHandler := func(next http.Handler) http.Handler {\n\t\t\treturn promhttp.InstrumentHandlerCounter(registerRequestsCounter(registerer), next)\n\t\t}\n\n\t\t// Gauge to all requests currently being handled\n\t\tinFlightHandler := func(next http.Handler) http.Handler {\n\t\t\treturn promhttp.InstrumentHandlerInFlight(registerInflightRequestsGauge(registerer), next)\n\t\t}\n\n\t\t// The latency of all requests bucketed by HTTP method\n\t\tdurationHandler := func(next http.Handler) http.Handler {\n\t\t\treturn promhttp.InstrumentHandlerDuration(registerRequestsLatencyHistogram(registerer), next)\n\t\t}\n\n\t\treturn alice.New(counterHandler, inFlightHandler, durationHandler).Then(next)\n\t}\n}", "func (it *Mcmc5intmcMetricsIterator) Create(key uint64) (*Mcmc5intmcMetrics, error) {\n\ttmtr := &Mcmc5intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc5intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func newHttpMetrics() *httpMetrics {\n\treturn &httpMetrics{\n\t\tRequestsTotal: promauto.NewCounterVec(prometheus.CounterOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_requests_total\",\n\t\t\tHelp: \"Total number of HTTP requests.\",\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t\tRequestDurationHistogram: promauto.NewHistogramVec(prometheus.HistogramOpts{\n\t\t\tSubsystem: \"provider\",\n\t\t\tName: \"http_request_duration_seconds\",\n\t\t\tHelp: \"Seconds spent serving HTTP requests.\",\n\t\t\tBuckets: prometheus.DefBuckets,\n\t\t}, []string{\"code\", \"method\", \"path\"}),\n\t}\n}", "func (it *MsmsintmsMetricsIterator) Create(key uint64) (*MsmsintmsMetrics, error) {\n\ttmtr := &MsmsintmsMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &MsmsintmsMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func (b *AdapterBase) WithExternalMetrics(p provider.ExternalMetricsProvider) {\n\tb.emProvider = p\n}", "func (m *podMetrics) New() runtime.Object {\n\treturn &metrics.PodMetrics{}\n}", "func NewMetric(rtype string) Metric {\n\treturn Metric{\n\t\tType: rtype,\n\t\tCurrent: map[string]int{},\n\t\tOwners: map[string]int{},\n\t}\n}", "func New(issueser getIssueser, metricser metricser, queries map[string]string) *Monitoring {\n\tlastActiveIssues := make(map[string]map[string]model.Issue)\n\tfor queryName := range queries {\n\t\tlastActiveIssues[queryName] = make(map[string]model.Issue)\n\t}\n\n\treturn &Monitoring{\n\t\tissueser: issueser,\n\t\tmetricser: metricser,\n\t\tlastActiveIssues: lastActiveIssues,\n\t\tqueries: queries,\n\t}\n}", "func (d *Dashboard) CreateMetricWithBufSize(target string, size int) (*Metric, error) {\n\treturn d.srv.metrics.Create(target, size)\n}", "func (in *Metrics) DeepCopy() *Metrics {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Metrics)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func NewExtentReader(inode uint64, key *proto.ExtentKey, dp *wrapper.DataPartition, followerRead bool) *ExtentReader {\n\treturn &ExtentReader{\n\t\tinode: inode,\n\t\tkey: key,\n\t\tdp: dp,\n\t\tfollowerRead: followerRead,\n\t}\n}", "func NewHandlerMetrics(scope tally.Scope) *HandlerMetrics {\n\thandlerAPIScope := scope.SubScope(\"api\")\n\thandlerSuccessScope := scope.Tagged(map[string]string{\"result\": \"success\"})\n\thandlerFailScope := scope.Tagged(map[string]string{\"result\": \"fail\"})\n\treturn &HandlerMetrics{\n\t\tHead: scope.Gauge(\"head\"),\n\t\tTail: scope.Gauge(\"tail\"),\n\t\tSize: scope.Gauge(\"size\"),\n\t\tCapacity: scope.Gauge(\"capacity\"),\n\t\tUnexpectedClientError: scope.Counter(\"unexpectedClientError\"),\n\t\tPurgeEventError: scope.Counter(\"purgeEventError\"),\n\t\tInvalidStreamIDError: scope.Counter(\"invalidStreamIdError\"),\n\t\tAddEventAPI: handlerAPIScope.Counter(\"addEvent\"),\n\t\tAddEventSuccess: handlerSuccessScope.Counter(\"addEvent\"),\n\t\tAddEventFail: handlerFailScope.Counter(\"addEvent\"),\n\t\tAddEventDeDupe: handlerAPIScope.Counter(\"addEventDeDupe\"),\n\t\tInitStreamAPI: handlerAPIScope.Counter(\"initStream\"),\n\t\tInitStreamSuccess: handlerSuccessScope.Counter(\"initStream\"),\n\t\tInitStreamFail: handlerFailScope.Counter(\"initStream\"),\n\t\tWaitForEventsAPI: handlerAPIScope.Counter(\"waitForEvents\"),\n\t\tWaitForEventsSuccess: handlerSuccessScope.Counter(\"waitForEvents\"),\n\t\tWaitForEventsFailed: handlerFailScope.Counter(\"waitForEvents\"),\n\t}\n}", "func NewMetricsRecorder(clock *hlc.Clock) *MetricsRecorder {\n\tmr := &MetricsRecorder{\n\t\tnodeRegistry: metric.NewRegistry(),\n\t}\n\tmr.mu.storeRegistries = make(map[roachpb.StoreID]*metric.Registry)\n\tmr.mu.stores = make(map[roachpb.StoreID]storeMetrics)\n\tmr.mu.clock = clock\n\treturn mr\n}", "func New() *SystemMetrics {\n\treturn &SystemMetrics{}\n}", "func New() *Metrics {\n\tm := &Metrics{\n\t\tBuildInfo: prometheus.NewGaugeVec(prometheus.GaugeOpts{\n\t\t\tNamespace: Namespace,\n\t\t\tSubsystem: Subsystem,\n\t\t\tName: \"build_info\",\n\t\t\tHelp: \"Build information\",\n\t\t}, []string{\"version\"}),\n\t}\n\n\t_ = prometheus.Register(m.BuildInfo)\n\t// TODO: implement metrics\n\treturn m\n}", "func NewMetricsHandler(mc *MetricsConfig) Middleware {\n\treturn func(h http.Handler) http.Handler {\n\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\tstartTime := time.Now()\n\t\t\tsw := newStatusResponseWriter(w)\n\t\t\tl := prometheus.Labels{\"path\": r.URL.Path}\n\t\t\tif mc.reqGauge != nil {\n\t\t\t\tmc.reqGauge.With(l).Inc()\n\t\t\t}\n\t\t\th.ServeHTTP(sw, r)\n\t\t\tif mc.reqGauge != nil {\n\t\t\t\tmc.reqGauge.With(l).Dec()\n\t\t\t}\n\t\t\tl[\"status_code\"] = fmt.Sprint(sw.status)\n\t\t\tif mc.respCnt != nil {\n\t\t\t\tmc.respCnt.With(l).Inc()\n\t\t\t}\n\t\t\tif mc.reqTime != nil {\n\t\t\t\tmc.reqTime.With(l).Observe(time.Since(startTime).Seconds())\n\t\t\t}\n\t\t})\n\t}\n}", "func insertDomainMetrics(eventType string, domain string, beanAttrVals []*beanAttrValue, request *beanRequest, i *integration.Integration, host, port string) error {\n\t// Create an entity for the domain\n\tvar e *integration.Entity\n\tvar err error\n\tswitch {\n\tcase args.RemoteMonitoring:\n\t\turl := net.JoinHostPort(host, port)\n\t\tif args.ConnectionURL != \"\" {\n\t\t\turl = getConnectionURLSAP(args.ConnectionURL)\n\t\t}\n\t\te, err = newRemoteEntity(domain, url, i)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\tcase args.LocalEntity:\n\t\te = i.LocalEntity()\n\tdefault:\n\t\t// create task for consistency with remote_monitoring\n\t\tif args.ConnectionURL != \"\" {\n\t\t\thost, port = getConnectionURLHostPort(args.ConnectionURL)\n\t\t}\n\n\t\thostIDAttr := integration.NewIDAttribute(\"host\", host)\n\t\tportIDAttr := integration.NewIDAttribute(\"port\", port)\n\t\te, err = i.Entity(domain, \"jmx-domain\", hostIDAttr, portIDAttr)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Create a map of bean names to metric sets\n\tentityMetricSets := make(map[string]*metric.Set)\n\n\t// For each bean/attribute returned from this domain\n\tfor _, beanAttrVal := range beanAttrVals {\n\t\tbeanName, err := getBeanName(beanAttrVal.beanAttr)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Query the metric set from the map or create it\n\t\tmetricSet, err := getOrCreateMetricSet(entityMetricSets, e, request, beanName, eventType, domain)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// If we want to collect the metric, populate the metric list\n\t\tif err := insertMetric(beanAttrVal.beanAttr, beanAttrVal.value, beanAttrVal.attrRequest, metricSet); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func NewMockMetrics(ctrl *gomock.Controller) *MockMetrics {\n\tmock := &MockMetrics{ctrl: ctrl}\n\tmock.recorder = &MockMetricsMockRecorder{mock}\n\treturn mock\n}", "func NewMockMetrics(ctrl *gomock.Controller) *MockMetrics {\n\tmock := &MockMetrics{ctrl: ctrl}\n\tmock.recorder = &MockMetricsMockRecorder{mock}\n\treturn mock\n}", "func NewMockMetrics(ctrl *gomock.Controller) *MockMetrics {\n\tmock := &MockMetrics{ctrl: ctrl}\n\tmock.recorder = &MockMetricsMockRecorder{mock}\n\treturn mock\n}", "func NewMockMetrics(ctrl *gomock.Controller) *MockMetrics {\n\tmock := &MockMetrics{ctrl: ctrl}\n\tmock.recorder = &MockMetricsMockRecorder{mock}\n\treturn mock\n}", "func NewMockMetrics(ctrl *gomock.Controller) *MockMetrics {\n\tmock := &MockMetrics{ctrl: ctrl}\n\tmock.recorder = &MockMetricsMockRecorder{mock}\n\treturn mock\n}", "func (it *Mcmc4intmcMetricsIterator) Create(key uint64) (*Mcmc4intmcMetrics, error) {\n\ttmtr := &Mcmc4intmcMetrics{}\n\n\tmtr := it.iter.Create(gometrics.EncodeScalarKey(key), tmtr.Size())\n\n\ttmtr = &Mcmc4intmcMetrics{metrics: mtr, key: key}\n\ttmtr.Unmarshal()\n\treturn tmtr, nil\n}", "func NewDimensions(minX, minY, maxX, maxY int64) Dimensions {\n\tdim := Dimensions{\n\t\tBottomLeft: Point{\n\t\t\tX: minX,\n\t\t\tY: minY,\n\t\t},\n\t\tTopRight: Point{\n\t\t\tX: maxX,\n\t\t\tY: maxY,\n\t\t},\n\t}\n\tdim.Init()\n\treturn dim\n}", "func NewMeasurement(database, name string) Measurement {\n\tm := Measurement{\n\t\tDatabase: database,\n\t\tName: name,\n\t\tTimestamp: time.Now(),\n\t\tValues: make(map[string]string, 0),\n\t\tTags: make(map[string]string, 0),\n\t}\n\treturn m\n}" ]
[ "0.6300966", "0.6068661", "0.59511846", "0.58217686", "0.5793168", "0.5735585", "0.57330275", "0.56729937", "0.5660602", "0.5657255", "0.56353873", "0.5573205", "0.5563562", "0.55550414", "0.55103785", "0.55049866", "0.54903716", "0.5483203", "0.5401924", "0.5340381", "0.53305286", "0.522412", "0.51613116", "0.51351684", "0.5126163", "0.51036984", "0.5100934", "0.50989246", "0.50557864", "0.5045842", "0.50355357", "0.50336105", "0.5029314", "0.49791902", "0.4965521", "0.49333122", "0.49004993", "0.4896822", "0.48925596", "0.4886322", "0.4859346", "0.48493987", "0.48422283", "0.48421144", "0.4836497", "0.48211524", "0.48157424", "0.47926903", "0.4792586", "0.47880158", "0.47829297", "0.47780007", "0.47766876", "0.4760438", "0.4755347", "0.47394112", "0.47337118", "0.47314817", "0.4725941", "0.47178203", "0.47056514", "0.4702402", "0.469864", "0.46962094", "0.46819523", "0.466892", "0.46676204", "0.4665745", "0.46397933", "0.46372667", "0.46348965", "0.46157026", "0.46117166", "0.46068287", "0.46044755", "0.46034423", "0.45937207", "0.45903125", "0.45886937", "0.4588312", "0.45882183", "0.45852983", "0.45841557", "0.45773917", "0.45636833", "0.45610094", "0.45538113", "0.45440298", "0.45417866", "0.45348698", "0.45274612", "0.4527082", "0.4527082", "0.4527082", "0.4527082", "0.4527082", "0.4524062", "0.45026085", "0.45019126" ]
0.7863173
0
Build returns the main package for given sourcefiles.
func Build(path string, sourcefiles []string) (*ssa.Package, error) { var conf loader.Config srcfs := strings.Join(sourcefiles, ", ") conf.CreateFromFilenames(path, srcfs) lprog, err := conf.Load() if err != nil { return nil, errors.Errorf("fail to load config of path: %s and sourcefiles: %s", path, srcfs) } prog := ssautil.CreateProgram(lprog, ssa.SanityCheckFunctions) mainPkg := prog.Package(lprog.Created[0].Pkg) prog.Build() return mainPkg, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g Golang) Build(gopath string, meta Metadata, skipTargets string, local bool) (err error) {\n\tlogrus.Debugf(\"Checking to see that gox is installed.\")\n\n\t// Install gox if it's not already there\n\tif _, err := os.Stat(filepath.Join(gopath, \"bin/gox\")); os.IsNotExist(err) {\n\t\terr = GoxInstall(gopath)\n\t\tif err != nil {\n\t\t\terr = errors.Wrap(err, \"Failed to install gox\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\tvar wd string\n\n\tif local {\n\t\twd, err = os.Getwd()\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"failed getting CWD\")\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\twd = fmt.Sprintf(\"%s/src/%s\", gopath, meta.Package)\n\n\t\tlogrus.Debugf(\"Changing working directory to: %s\", wd)\n\n\t\terr = os.Chdir(wd)\n\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"changing working dir to %q\", wd)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tgox := fmt.Sprintf(\"%s/bin/gox\", gopath)\n\n\tlogrus.Debugf(\"Gox is: %s\", gox)\n\n\tvar metadatapath string\n\tif local {\n\t\tmetadatapath = fmt.Sprintf(\"%s/%s\", wd, METADATA_FILENAME)\n\n\t} else {\n\t\tmetadatapath = fmt.Sprintf(\"%s/src/%s/%s\", gopath, meta.Package, METADATA_FILENAME)\n\t}\n\n\tmd, err := ReadMetadata(metadatapath)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"Failed to read metadata file from checked out code\")\n\t\treturn err\n\t}\n\n\tskipTargetsMap := make(map[string]int)\n\n\tif skipTargets != \"\" {\n\t\ttargetsList := strings.Split(skipTargets, \",\")\n\n\t\tfor _, t := range targetsList {\n\t\t\tskipTargetsMap[t] = 1\n\t\t}\n\t}\n\n\tfor _, target := range md.BuildInfo.Targets {\n\t\t// skip this target if we're told to do so\n\t\t_, skip := skipTargetsMap[target.Name]\n\t\tif skip {\n\t\t\tcontinue\n\t\t}\n\n\t\tlogrus.Debugf(\"Building target: %q in dir %s\", target.Name, wd)\n\n\t\t// This gets weird because go's exec shell doesn't like the arg format that gox expects\n\t\t// Building it thusly keeps the various quoting levels straight\n\n\t\trunenv := os.Environ()\n\n\t\tif !local {\n\t\t\tgopathenv := fmt.Sprintf(\"GOPATH=%s\", gopath)\n\t\t\trunenv = append(runenv, gopathenv)\n\t\t}\n\n\t\t// allow user to turn off go modules\n\t\tif !target.Legacy {\n\t\t\trunenv = append(runenv, \"GO111MODULE=on\")\n\t\t}\n\n\t\tcgo := \"\"\n\t\t// build with cgo if we're told to do so.\n\t\tif target.Cgo {\n\t\t\tcgo = \" -cgo\"\n\t\t}\n\n\t\tfor k, v := range target.Flags {\n\t\t\trunenv = append(runenv, fmt.Sprintf(\"%s=%s\", k, v))\n\t\t\tlogrus.Debugf(\"Build Flag: %s=%s\", k, v)\n\t\t}\n\n\t\tldflags := \"\"\n\t\tif target.Ldflags != \"\" {\n\t\t\tldflags = fmt.Sprintf(\" -ldflags %q \", target.Ldflags)\n\t\t\tlogrus.Debugf(\"LD Flag: %s\", ldflags)\n\t\t}\n\n\t\t// Interesting idea, but breaks multiple binary builds such as dbt. To properly implement, we'd have to find and handle each binary instead of relying on the './...'.\n\t\t//outputTemplate := fmt.Sprintf(\"%s_{{.OS}}_{{.Arch}}\", meta.Name)\n\t\t//args := gox + cgo + ldflags + ` -osarch=\"` + target.Name + `\"` + ` -output=\"` + outputTemplate + `\"` + \" ./...\"\n\n\t\targs := gox + cgo + ldflags + ` -osarch=\"` + target.Name + `\"` + \" ./...\"\n\n\t\tlogrus.Debugf(\"Running gox with: %s in dir %s\", args, wd)\n\n\t\t// Calling it through sh makes everything happy\n\t\tcmd := exec.Command(\"sh\", \"-c\", args)\n\n\t\tcmd.Env = runenv\n\n\t\tcmd.Stdout = os.Stdout\n\t\tcmd.Stderr = os.Stderr\n\t\tcmd.Stdin = os.Stdin\n\n\t\terr = cmd.Run()\n\t\tif err != nil {\n\t\t\terr = errors.Wrapf(err, \"failed building target %s\", target.Name)\n\t\t\treturn err\n\t\t}\n\n\t\tlogrus.Debugf(\"Gox build of target %s complete and successful.\", target.Name)\n\t}\n\n\terr = BuildExtras(md, wd)\n\tif err != nil {\n\t\terr = errors.Wrapf(err, \"Failed to build extras\")\n\t\treturn err\n\t}\n\n\treturn err\n}", "func Build() error {\n\tfiles, err := ioutil.ReadDir(\"./\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tfor _, f := range files {\n\t\tif f.IsDir() {\n\t\t\tif _, err = os.Stat(path.Join(f.Name(), \"Flekszible\")); err == nil {\n\t\t\t\tlog.Println(\"Testing \" + f.Name())\n\t\t\t\terr = sh.Run(\"flekszible\", \"generate\", f.Name())\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\t}\n\treturn nil\n}", "func buildOne(ctx context.Context, log logging.Logger, tgt *Target) error {\n\tctx, st := timing.Start(ctx, filepath.Base(tgt.Pkg))\n\tdefer st.End()\n\n\tfor _, ws := range tgt.Workspaces {\n\t\tsrc := filepath.Join(ws, \"src\")\n\t\tif _, err := os.Stat(src); os.IsNotExist(err) {\n\t\t\treturn fmt.Errorf(\"invalid workspace %q (no src subdir)\", ws)\n\t\t} else if err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tarchEnvs, ok := archToEnvs[tgt.Arch]\n\tif !ok {\n\t\treturn fmt.Errorf(\"unknown arch %q\", tgt.Arch)\n\t}\n\n\tconst ldFlags = \"-ldflags=-s -w\"\n\tcmd := exec.Command(\"go\", \"build\", ldFlags, \"-o\", tgt.Out, tgt.Pkg)\n\tcmd.Env = append(os.Environ(),\n\t\t\"GOPATH=\"+strings.Join(tgt.Workspaces, \":\"),\n\t\t// Disable cgo and PIE on building Tast binaries. See:\n\t\t// https://crbug.com/976196\n\t\t// https://github.com/golang/go/issues/30986#issuecomment-475626018\n\t\t\"CGO_ENABLED=0\",\n\t\t\"GOPIE=0\")\n\tcmd.Env = append(cmd.Env, archEnvs...)\n\n\tlog.Status(\"Compiling \" + tgt.Pkg)\n\n\tif out, err := cmd.CombinedOutput(); err != nil {\n\t\twriteMultiline(log, string(out))\n\t\treturn err\n\t}\n\treturn nil\n}", "func Build(srcdir, dstdir string) error {\n\n\t// Get a slice of markdown files.\n\tmatches, err := filepath.Glob(srcdir + \"/*.md\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Convert markdown to HTML.\n\tfor _, md := range matches {\n\t\tdstFile := strings.Replace(filepath.Base(md), \".md\", \".html\", 1)\n\t\tlog.Printf(\"[DEBUG] markdown file src: %s\\n\", md)\n\t\tlog.Printf(\"[DEBUG] markdown file dst: %s/%s\\n\", dstdir, dstFile)\n\n\t\tdat, err := ioutil.ReadFile(md)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"[INFO] Error reading .md file: %s\\n\", err)\n\t\t}\n\n\t\t// TODO: Write full HTML webpage from **template** here.\n\t\terr = ioutil.WriteFile(filepath.Join(dstdir, dstFile), toHTML(dat), 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func Build(target *prog.Target, lang, src string) (string, error) {\n\tbin, err := ioutil.TempFile(\"\", \"syzkaller\")\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to create temp file: %v\", err)\n\t}\n\tbin.Close()\n\tsysTarget := targets.List[target.OS][target.Arch]\n\tcompiler := sysTarget.CCompilerPrefix + \"gcc\"\n\tif _, err := exec.LookPath(compiler); err != nil {\n\t\treturn \"\", NoCompilerErr\n\t}\n\tflags := []string{\n\t\t\"-x\", lang, \"-Wall\", \"-Werror\", \"-O1\", \"-g\", \"-o\", bin.Name(),\n\t\tsrc, \"-pthread\",\n\t}\n\tflags = append(flags, sysTarget.CrossCFlags...)\n\tif sysTarget.PtrSize == 4 {\n\t\t// We do generate uint64's for syscall arguments that overflow longs on 32-bit archs.\n\t\tflags = append(flags, \"-Wno-overflow\")\n\t}\n\tout, err := exec.Command(compiler, append(flags, \"-static\")...).CombinedOutput()\n\tif err != nil {\n\t\t// Some distributions don't have static libraries.\n\t\tout, err = exec.Command(compiler, flags...).CombinedOutput()\n\t}\n\tif err != nil {\n\t\tos.Remove(bin.Name())\n\t\tdata, _ := ioutil.ReadFile(src)\n\t\treturn \"\", fmt.Errorf(\"failed to build program:\\n%s\\n%s\\ncompiler invocation: %v %v\\n\",\n\t\t\tdata, out, compiler, flags)\n\t}\n\treturn bin.Name(), nil\n}", "func Build(appEnv, appName, appPath string, logger *plugins.ServerLog) {\n\trevel.Init(appEnv, appPath, \"\")\n\n\t// First, clear the generated files (to avoid them messing with ProcessSource).\n\tcleanSource(appName)\n\n\tsourceInfo, compileError := ProcessSource(revel.CodePaths)\n\tif compileError != nil {\n panic(compileError)\n\t}\n\n checkIsHaveArgs := func(ctrl []*TypeInfo) bool {\n result := false\n\n\t L: for _, c := range ctrl {\n\t\t for _, a:= range c.MethodSpecs {\n if len(a.Args) > 0 {\n result = true\n break L\n }\n\t\t\t}\n\t }\n\n return result\n }\n\n ctrlSpecs := sourceInfo.ControllerSpecs()\n\n\t// Generate two source files.\n\ttemplateArgs := map[string]interface{}{\n \"AppName\": appName,\n \"AppPath\": appPath,\n \"AppEnv\": appEnv,\n\t\t\"Controllers\": ctrlSpecs,\n\t\t\"ValidationKeys\": sourceInfo.ValidationKeys,\n\t\t\"ImportPaths\": calcImportAliases(sourceInfo),\n\t\t\"TestSuites\": sourceInfo.TestSuites(),\n \"IsArgInCtrl\": checkIsHaveArgs(ctrlSpecs),\n\t}\n\n\tgenSource(runtimePath, appName + \".go\", MAIN, templateArgs)\n\t// genSource(\"routes\", \"routes.go\", ROUTES, templateArgs)\n\n}", "func buildExecutable() {\n\tvar executables []string\n\tvar execFilled int\n\n\t// check if there's a main package:\n\tif goPackages.GetMainCount() == 0 {\n\t\tlogger.Error(\"No main package found.\\n\")\n\t\tos.Exit(1)\n\t}\n\n\t// multiple main, no command file from command line and no -a -> error\n\tif (goPackages.GetMainCount() > 1) && (flag.NArg() == 0) && !*flagBuildAll {\n\t\tlogger.Error(\"Multiple files found with main function.\\n\")\n\t\tlogger.ErrorContinue(\"Please specify one or more as command line parameter or\\n\")\n\t\tlogger.ErrorContinue(\"run gobuild with -a. Available main files are:\\n\")\n\t\tfor _, fn := range goPackages.GetMainFilenames() {\n\t\t\tlogger.ErrorContinue(\"\\t %s\\n\", fn)\n\t\t}\n\t\tos.Exit(1)\n\t}\n\n\t// compile all needed packages\n\tif flag.NArg() > 0 {\n\t\tif *flagRunExec {\n\t\t\texecutables = make([]string, flag.NArg())\n\t\t}\n\t\tfor _, fn := range flag.Args() {\n\t\t\tmainPack, exists := goPackages.GetMain(fn, !*flagSingleMainFile)\n\t\t\tif !exists {\n\t\t\t\tlogger.Error(\"File %s not found.\\n\", fn)\n\t\t\t\treturn // or os.Exit?\n\t\t\t}\n\n\t\t\tif compile(mainPack) {\n\t\t\t\t// link everything together\n\t\t\t\tif link(mainPack) {\n\t\t\t\t\tif *flagRunExec {\n\t\t\t\t\t\texecutables[execFilled] = outputDirPrefix + mainPack.OutputFile\n\t\t\t\t\t\texecFilled++\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tlinkErrors = true\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlogger.Error(\"Can't link executable because of compile errors.\\n\")\n\t\t\t\tcompileErrors = true\n\t\t\t}\n\t\t}\n\t} else {\n\t\tif *flagRunExec {\n\t\t\texecutables = make([]string, goPackages.GetMainCount())\n\t\t}\n\t\tfor _, mainPack := range goPackages.GetMainPackages(!*flagSingleMainFile) {\n\n\t\t\tif compile(mainPack) {\n\t\t\t\tif link(mainPack) {\n\t\t\t\t\tif *flagRunExec {\n\t\t\t\t\t\texecutables[execFilled] = outputDirPrefix + mainPack.OutputFile\n\t\t\t\t\t\texecFilled++\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tlinkErrors = true\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tlogger.Error(\"Can't link executable because of compile errors.\\n\")\n\t\t\t\tcompileErrors = true\n\t\t\t}\n\t\t}\n\t}\n\n\tif *flagRunExec && !linkErrors && !compileErrors {\n\t\tfor i := 0; i < execFilled; i++ {\n\t\t\trunExec([]string{executables[i]})\n\t\t}\n\t}\n}", "func compileFiles(paths []paths) error {\n\ttype file struct {\n\t\tPath string\n\t\tName string\n\t\tData string\n\t\tVar string\n\t\tLen int\n\t}\n\tv := struct {\n\t\tPkg string\n\t\tArgs string\n\t\tFiles []file\n\t}{Pkg: *pkgName, Args: strings.Join(os.Args[1:], \" \")}\n\n\tfor i, p := range paths {\n\t\tdata, err := os.ReadFile(p.path)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"reading file contents: %v\", err)\n\t\t}\n\t\tpacked, err := bits.Encode(data)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"encoding file contents: %v\", err)\n\t\t}\n\t\ttrimmed := strings.TrimPrefix(p.target, *trimPrefix)\n\t\tadded := filepath.Join(*addPrefix, trimmed)\n\t\tif *baseOnly {\n\t\t\tadded = filepath.Join(*addPrefix, filepath.Base(trimmed))\n\t\t}\n\n\t\tvar src bytes.Buffer\n\t\tif err := bits.ToSource(&src, packed); err != nil {\n\t\t\treturn fmt.Errorf(\"packing file contents: %v\", err)\n\t\t}\n\t\tv.Files = append(v.Files, file{\n\t\t\tPath: trimmed,\n\t\t\tName: added,\n\t\t\tVar: fmt.Sprintf(\"_fileData%d\", i+1),\n\t\t\tData: src.String(),\n\t\t\tLen: len(data),\n\t\t})\n\t}\n\n\tbuf := new(bytes.Buffer)\n\tif err := fileTemplate.Execute(buf, v); err != nil {\n\t\treturn fmt.Errorf(\"generating source: %v\", err)\n\t}\n\tcode, err := format.Source(buf.Bytes())\n\tif err != nil {\n\t\treturn fmt.Errorf(\"formatting source: %v\", err)\n\t}\n\n\treturn os.WriteFile(*outputPath, code, 0644)\n}", "func build(args []string) {\n\tbuildCmd := flag.NewFlagSet(\"build\", flag.ExitOnError)\n\tbuildCmd.Usage = func() {\n\t\tfmt.Printf(\"USAGE: %s build [options] <file>[.yml]\\n\\n\", os.Args[0])\n\t\tfmt.Printf(\"Options:\\n\")\n\t\tbuildCmd.PrintDefaults()\n\t}\n\tbuildName := buildCmd.String(\"name\", \"\", \"Name to use for output files\")\n\tbuildPull := buildCmd.Bool(\"pull\", false, \"Always pull images\")\n\n\tif err := buildCmd.Parse(args); err != nil {\n\t\tlog.Fatal(\"Unable to parse args\")\n\t}\n\tremArgs := buildCmd.Args()\n\n\tif len(remArgs) == 0 {\n\t\tfmt.Println(\"Please specify a configuration file\")\n\t\tbuildCmd.Usage()\n\t\tos.Exit(1)\n\t}\n\tconf := remArgs[0]\n\tif !(filepath.Ext(conf) == \".yml\" || filepath.Ext(conf) == \".yaml\") {\n\t\tconf = conf + \".yml\"\n\t}\n\n\tbuildInternal(*buildName, *buildPull, conf)\n}", "func (b *Cowbuilder) BuildPackage(a BuildArguments, output io.Writer) (*BuildResult, error) {\n\tb.acquire()\n\tdefer b.release()\n\n\t//checks we supports everything\n\tsupported := b.getAllImages()\n\tfor _, targetArch := range a.Archs {\n\t\tfound := false\n\t\tfor _, aArch := range supported[a.Dist] {\n\t\t\tif targetArch == aArch {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif found == false {\n\t\t\treturn nil, fmt.Errorf(\"Distribution %s-%s is not supported\", a.Dist, targetArch)\n\t\t}\n\t}\n\n\t//checks that the input exists\n\tdscFile := path.Join(a.SourcePackage.BasePath, a.SourcePackage.Filename())\n\tif _, err := os.Stat(dscFile); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil, fmt.Errorf(\"Expected file %s, does not exists\", dscFile)\n\t\t}\n\t\treturn nil, fmt.Errorf(\"Could not check existence of %s: %s\", dscFile, err)\n\t}\n\n\t// ensure that destination directory exists\n\tif err := os.MkdirAll(a.Dest, 0755); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// creates output buffers and result structures\n\tvar buf bytes.Buffer\n\tchangesFiles := []string{}\n\tvar writer io.Writer = &buf\n\tif output != nil {\n\t\twriter = io.MultiWriter(&buf, output)\n\t}\n\n\tvar lastBuildArch deb.Architecture\n\tfor i, arch := range a.Archs {\n\t\tdebbuildopts := []string{}\n\t\t//only the last will build architecture-independent package\n\t\tvar cmd *exec.Cmd\n\t\tvar err error\n\t\tif i == len(a.Archs)-1 {\n\t\t\tdebbuildopts = append(debbuildopts, \"-b\")\n\t\t} else {\n\t\t\t//if it produce only arch indep package we skip the build\n\t\t\tskip := true\n\t\t\tfor _, targetArch := range a.SourcePackage.Archs {\n\t\t\t\tif targetArch == deb.Any {\n\t\t\t\t\tskip = false\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tif targetArch == arch {\n\t\t\t\t\tskip = false\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif skip == true {\n\t\t\t\tfmt.Fprintf(writer, \"Skiping build for %s, as it will produce no package\\n\", arch)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tdebbuildopts = append(debbuildopts, \"-B\")\n\t\t}\n\t\tcmd, err = b.cowbuilderCommand(a.Dist, arch, a.Deps, \"--build\",\n\t\t\t\"--debbuildopts\", `\"`+strings.Join(debbuildopts, \" \")+`\"`,\n\t\t\t\"--buildresult\", a.Dest,\n\t\t\tdscFile)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tcmd.Stdin = nil\n\t\tcmd.Stderr = writer\n\t\tcmd.Stdout = writer\n\t\tfmt.Fprintf(writer, \"--- Execute:%v\\n--- Env:%v\\n\", cmd.Args, cmd.Env)\n\t\terr = cmd.Run()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tchangesFileName := path.Join(a.Dest, fmt.Sprintf(\"%s_%s.changes\", a.SourcePackage.Identifier, arch))\n\t\tif _, err = os.Stat(changesFileName); err != nil {\n\t\t\tif os.IsNotExist(err) {\n\t\t\t\treturn nil, fmt.Errorf(\"Missing expected result file %s\", changesFileName)\n\t\t\t}\n\t\t\treturn nil, fmt.Errorf(\"Could not check existence of %s: %s\", changesFileName, err)\n\t\t}\n\t\tchangesFiles = append(changesFiles, changesFileName)\n\t\tlastBuildArch = arch\n\t}\n\n\tres := &BuildResult{\n\t\tBasePath: a.Dest,\n\t}\n\tif len(changesFiles) == 0 {\n\t\treturn nil, fmt.Errorf(\"No architecture where build!\")\n\t}\n\n\tres.ChangesPath = path.Base(changesFiles[0])\n\tvar suffix = string(lastBuildArch)\n\tif len(changesFiles) > 1 {\n\t\t// in that case we make a multi-arch upload file\n\t\tcmd := exec.Command(\"mergechanges\", changesFiles...)\n\t\tcmd.Stdin = nil\n\t\tvar mergedChanges bytes.Buffer\n\t\tcmd.Stdout = &mergedChanges\n\t\tcmd.Stderr = writer\n\t\tfmt.Fprintf(writer, \"--- Execute:%v\\n--- Env:%v\\n\", cmd.Args, cmd.Env)\n\t\terr := cmd.Run()\n\t\tif err == nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tres.ChangesPath = fmt.Sprintf(\"%s_multi.changes\", a.SourcePackage.Identifier)\n\t\tf, err := os.Create(path.Join(res.BasePath, res.ChangesPath))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\t_, err = io.Copy(f, &mergedChanges)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsuffix = \"multi\"\n\t}\n\tres.BuildLog = Log(buf.String())\n\n\tcf, err := os.Open(path.Join(res.BasePath, res.ChangesPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres.Changes, err = deb.ParseChangeFile(cf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres.Changes.Ref.Suffix = suffix\n\n\treturn res, nil\n}", "func buildLibrary() {\n\tvar packNames []string\n\tvar pack *godata.GoPackage\n\tvar exists bool\n\n\tif goPackages.GetPackageCount() == 0 {\n\t\tlogger.Warn(\"No packages found to build.\\n\")\n\t\treturn\n\t}\n\n\t// check for there is at least one package that can be compiled\n\tvar hasNoCompilablePacks bool = true\n\tfor _, packName := range goPackages.GetPackageNames() {\n\t\tpack, _ := goPackages.Get(packName)\n\t\tif pack.Name == \"main\" {\n\t\t\tcontinue\n\t\t}\n\t\tif pack.Files.Len() > 0 && !pack.HasCGOFiles() {\n\t\t\thasNoCompilablePacks = false\n\t\t\tbreak\n\t\t}\n\t}\n\tif hasNoCompilablePacks {\n\t\tlogger.Warn(\"No packages found that could be compiled by gobuild.\\n\")\n\t\tcompileErrors = true\n\t\treturn\n\t}\n\n\t// check for command line parameters\n\tif flag.NArg() > 0 {\n\t\tpackNames = flag.Args()\n\t} else {\n\t\tpackNames = goPackages.GetPackageNames()\n\t}\n\n\t// loop over all packages, compile them and build a .a file\n\tfor _, name := range packNames {\n\n\t\tif name == \"main\" {\n\t\t\tcontinue // don't make this into a library\n\t\t}\n\n\t\tpack, exists = goPackages.Get(name)\n\t\tif !exists {\n\t\t\tlogger.Error(\"Package %s doesn't exist.\\n\", name)\n\t\t\tcontinue // or exit?\n\t\t}\n\n\t\t// don't compile remote packages or packages without files\n\t\tif pack.Type == godata.REMOTE_PACKAGE || pack.Files.Len() == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// these packages come from invalid/unhandled imports\n\t\tif pack.Files.Len() == 0 {\n\t\t\tlogger.Debug(\"Skipping package %s, no files to compile.\\n\", pack.Name)\n\t\t\tcontinue\n\t\t}\n\n\t\tif !pack.Compiled && !pack.HasErrors {\n\t\t\tcompileErrors = !compile(pack) || compileErrors\n\t\t}\n\n\t\tif pack.HasErrors {\n\t\t\tlogger.Error(\"Can't create library because of compile errors.\\n\")\n\t\t\tcompileErrors = true\n\t\t} else {\n\t\t\tpackLib(pack)\n\t\t}\n\t}\n}", "func Build(name, packageName string) func() error {\n\treturn func() error {\n\t\tmg.Deps(git.CollectInfo)\n\n\t\tfmt.Printf(\" > Building %s [%s]\\n\", name, packageName)\n\n\t\tversion, err := git.TagMatch(fmt.Sprintf(\"cmd/%s*\", name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tvarsSetByLinker := map[string]string{\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.Version\": version,\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.Revision\": git.Revision,\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.Branch\": git.Branch,\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.BuildUser\": os.Getenv(\"USER\"),\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.BuildDate\": time.Now().Format(time.RFC3339),\n\t\t\t\"github.com/Zenithar/go-monorepo/build/version.GoVersion\": runtime.Version(),\n\t\t}\n\t\tvar linkerArgs []string\n\t\tfor name, value := range varsSetByLinker {\n\t\t\tlinkerArgs = append(linkerArgs, \"-X\", fmt.Sprintf(\"%s=%s\", name, value))\n\t\t}\n\t\tlinkerArgs = append(linkerArgs, \"-s\", \"-w\")\n\n\t\treturn sh.RunWith(map[string]string{\n\t\t\t\"CGO_ENABLED\": \"0\",\n\t\t}, \"go\", \"build\", \"-buildmode=pie\", \"-ldflags\", strings.Join(linkerArgs, \" \"), \"-mod=vendor\", \"-o\", fmt.Sprintf(\"../../bin/%s\", name), packageName)\n\t}\n}", "func (g *Generator) Build() ([]byte, error) {\n\tpkg := g.Package\n\tpkg.PrepareBuild()\n\n\t//write .PKGINFO\n\terr := writePKGINFO(pkg)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to write .PKGINFO: %s\", err.Error())\n\t}\n\n\t//write .INSTALL\n\twriteINSTALL(pkg)\n\n\t//write mtree\n\terr = writeMTREE(pkg)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to write .MTREE: %s\", err.Error())\n\t}\n\n\t//compress package\n\tvar buf bytes.Buffer\n\terr = pkg.FSRoot.ToTarXZArchive(&buf, false, true)\n\treturn buf.Bytes(), err\n}", "func buildPackageForNamedFileInMultiPackageDir(bpkg *build.Package, m *build.MultiplePackageError, filename string) (*build.Package, error) {\n\tcopy := *bpkg\n\tbpkg = &copy\n\n\t// First, find which package name each filename is in.\n\tfileToPkgName := make(map[string]string, len(m.Files))\n\tfor i, f := range m.Files {\n\t\tfileToPkgName[f] = m.Packages[i]\n\t}\n\n\tpkgName := fileToPkgName[filename]\n\tif pkgName == \"\" {\n\t\treturn nil, fmt.Errorf(\"package %q in %s has no file %q\", bpkg.ImportPath, bpkg.Dir, filename)\n\t}\n\n\tfilterToFilesInPackage := func(files []string, pkgName string) []string {\n\t\tvar keep []string\n\t\tfor _, f := range files {\n\t\t\tif fileToPkgName[f] == pkgName {\n\t\t\t\tkeep = append(keep, f)\n\t\t\t}\n\t\t}\n\t\treturn keep\n\t}\n\n\t// Trim the *GoFiles fields to only those files in the same\n\t// package.\n\tbpkg.Name = pkgName\n\tif pkgName == \"main\" {\n\t\t// TODO(sqs): If the package name is \"main\", and there are\n\t\t// multiple main packages that are separate programs (and,\n\t\t// e.g., expected to be run directly run `go run main1.go\n\t\t// main2.go`), then this will break because it will try to\n\t\t// compile them all together. There's no good way to handle\n\t\t// that case that I can think of, other than with heuristics.\n\t}\n\tvar nonXTestPkgName, xtestPkgName string\n\tif strings.HasSuffix(pkgName, \"_test\") {\n\t\tnonXTestPkgName = strings.TrimSuffix(pkgName, \"_test\")\n\t\txtestPkgName = pkgName\n\t} else {\n\t\tnonXTestPkgName = pkgName\n\t\txtestPkgName = pkgName + \"_test\"\n\t}\n\tbpkg.GoFiles = filterToFilesInPackage(bpkg.GoFiles, nonXTestPkgName)\n\tbpkg.TestGoFiles = filterToFilesInPackage(bpkg.TestGoFiles, nonXTestPkgName)\n\tbpkg.XTestGoFiles = filterToFilesInPackage(bpkg.XTestGoFiles, xtestPkgName)\n\n\treturn bpkg, nil\n}", "func main() {\n\t// define commands for execution\n\tinFile := flag.String(\"source\", \"\", \"source\")\n\toutFile := flag.String(\"target\", \"\", \"target\")\n\tpkg := flag.String(\"package\", \"main\", \"package name\")\n\tname := flag.String(\"name\", \"File\", \"identifier to use for the embedded data\")\n\tflag.Parse()\n\terr := genConf(*inFile, *outFile, *pkg, *name)\n\thandleError(err)\n}", "func BuildIn(gopath string, packagePath string, args ...string) (compiledPath string, err error) {\n\treturn doBuild(gopath, packagePath, nil, args...)\n}", "func Build(manifest model.Manifest) error {\n\tif _, f := manifest.BuildOutputs[model.Docker]; f {\n\t\tif err := Docker(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to build Docker: %v\", err)\n\t\t}\n\t}\n\n\tif err := SanitizeAllCharts(manifest); err != nil {\n\t\treturn fmt.Errorf(\"failed to sanitize charts: %v\", err)\n\t}\n\tif util.IsValidSemver(manifest.Version) {\n\t\tif _, f := manifest.BuildOutputs[model.Helm]; f {\n\t\t\tif err := HelmCharts(manifest); err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to build HelmCharts: %v\", err)\n\t\t\t}\n\t\t}\n\t} else {\n\t\tlog.Warnf(\"Invalid Semantic Version. Skipping Charts build\")\n\t}\n\n\tif _, f := manifest.BuildOutputs[model.Debian]; f {\n\t\tif err := Debian(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to build Debian: %v\", err)\n\t\t}\n\t}\n\n\tif _, f := manifest.BuildOutputs[model.Rpm]; f {\n\t\tif err := Rpm(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to build Rpm: %v\", err)\n\t\t}\n\t}\n\n\tif _, f := manifest.BuildOutputs[model.Archive]; f {\n\t\tif err := Archive(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to build Archive: %v\", err)\n\t\t}\n\t}\n\n\tif _, f := manifest.BuildOutputs[model.Grafana]; f {\n\t\tif err := Grafana(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to build Grafana: %v\", err)\n\t\t}\n\t}\n\n\t// Bundle all sources used in the build\n\tcmd := util.VerboseCommand(\"tar\", \"-czf\", \"out/sources.tar.gz\", \"sources\")\n\tcmd.Dir = path.Join(manifest.Directory)\n\tif err := cmd.Run(); err != nil {\n\t\treturn fmt.Errorf(\"failed to bundle sources: %v\", err)\n\t}\n\n\tif err := writeManifest(manifest, manifest.OutDir()); err != nil {\n\t\treturn fmt.Errorf(\"failed to write manifest: %v\", err)\n\t}\n\n\tif err := writeLicense(manifest); err != nil {\n\t\treturn fmt.Errorf(\"failed to package license file: %v\", err)\n\t}\n\n\tif manifest.DockerOutput == model.DockerOutputContext {\n\t\tlog.Warnf(\"Docker output in 'context' mode; will not produce SBOM.\")\n\t} else if manifest.SkipGenerateBillOfMaterials {\n\t\tlog.Warnf(\"Input manifest set SkipGenerateBillOfMaterials; will not produce SBOM.\")\n\t} else {\n\t\tif err := GenerateBillOfMaterials(manifest); err != nil {\n\t\t\treturn fmt.Errorf(\"failed to generate sbom: %v\", err)\n\t\t}\n\t}\n\n\treturn nil\n}", "func build(c *king.Config, td string, args []string) error {\n\tvar (\n\t\tfs bool\n\t\tfi bool\n\t\tfd bool\n\t\tfT bool\n\t\tfq bool\n\t\tfn bool\n\t)\n\n\tbo := new(king.BuildOptions)\n\tlo := new(king.InstallOptions)\n\tdo := new(king.DownloadOptions)\n\n\tpf := pflag.NewFlagSet(\"\", pflag.ExitOnError)\n\n\t// TODO we need structure-based flag parser at some point to avoid this mess\n\tpf.StringVarP(&lo.ExtractDir, \"extract-dir\", \"X\", filepath.Join(td, \"extract\"), \"\")\n\tpf.StringVarP(&bo.PackageDir, \"package-dir\", \"P\", filepath.Join(td, \"pkg\"), \"\")\n\t// pf.StringVarP(&fO, \"output-dir\", \"O\", filepath.Join(cd, \"logs\"), \"\")\n\tpf.StringVarP(&bo.BuildDir, \"build-dir\", \"B\", filepath.Join(td, \"build\"), \"\")\n\tpf.StringVarP(&bo.Compression, \"compression\", \"C\", os.Getenv(\"KISS_COMPRESS\"), \"\")\n\tpf.BoolVarP(&fs, \"no-verify\", \"s\", false, \"\")\n\tpf.BoolVarP(&fd, \"debug\", \"d\", false, \"\")\n\tpf.BoolVarP(&do.Overwrite, \"force\", \"f\", false, \"\")\n\tpf.BoolVarP(&fn, \"no-bar\", \"n\", false, \"\")\n\tpf.BoolVarP(&log.NoPrompt, \"no-prompt\", \"y\", os.Getenv(\"KISS_PROMPT\") == \"1\", \"\")\n\t// pf.BoolVarP(&bo.NoStripBinaries, \"no-strip\", \"S\", os.Getenv(\"KISS_STRIP\") == \"0\", \"\")\n\tpf.BoolVarP(&fT, \"no-prebuilt\", \"T\", false, \"\")\n\tpf.BoolVarP(&fi, \"install\", \"i\", false, \"\")\n\tpf.BoolVarP(&fq, \"quiet\", \"q\", false, \"\")\n\n\tpf.SetInterspersed(true)\n\n\tpf.Usage = func() {\n\t\tfmt.Fprintln(os.Stderr, buildUsage)\n\t}\n\n\tpf.Parse(args[1:])\n\n\tif pf.NArg() == 0 {\n\t\tpf.Usage()\n\t\tos.Exit(2)\n\t}\n\n\tif fd {\n\t\tlo.Debug = true\n\t\tbo.Debug = true\n\t} else {\n\t\t// XXX\n\t\tdefer cleanup.Run(func() error {\n\t\t\treturn os.RemoveAll(td)\n\t\t})()\n\t}\n\n\tif !fn {\n\t\tdo.Progress = os.Stderr\n\t}\n\n\tif !fq {\n\t\tbo.Output = os.Stdout\n\t}\n\n\tbpp := make([]*king.Package, 0, pf.NArg())\n\n\tfor _, n := range pf.Args() {\n\t\tp, err := king.NewPackage(c, &king.PackageOptions{\n\t\t\tName: n,\n\t\t\tFrom: king.All,\n\t\t})\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tbpp = append(bpp, p)\n\t}\n\n\tepp, dpp, tpp, err := resolveDependencies(c, bpp, fT)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif len(dpp) > 0 || len(tpp) > 0 {\n\t\t// TODO tree?\n\t\tw := tabwriter.NewWriter(os.Stderr, 0, 0, 3, ' ', 0) // TODO doc\n\n\t\tfmt.Fprint(w, \"<package>\\t<type>\\t<action>\\n\")\n\n\t\tfor _, t := range tpp {\n\t\t\tfmt.Fprint(w, t.Name+\"\\t\", \"pre-built dependency\\t\", \"install\\n\")\n\t\t}\n\n\t\tfor _, p := range dpp {\n\t\t\t// TODO print make dependency\n\t\t\tfmt.Fprint(w, p.Name+\"\\t\", \"dependency\\t\", \"build && install\\n\")\n\t\t}\n\n\t\tfor _, p := range epp {\n\t\t\tfmt.Fprint(w, p.Name+\"\\t\", \"candidate\\t\", \"build\")\n\n\t\t\tif fi {\n\t\t\t\tfmt.Fprint(w, \"&& install\")\n\t\t\t}\n\n\t\t\tfmt.Fprint(w, \"\\n\")\n\t\t}\n\n\t\tw.Flush()\n\t\tlog.Prompt(\"proceed to build?\")\n\t}\n\n\tfor _, p := range append(dpp, epp...) {\n\t\tif err := downloadSources(p, do, fs, fn); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, t := range tpp {\n\t\tlog.Runningf(\"installing pre-built dependency %s\", t.Name)\n\n\t\t// TODO forcefully install\n\t\t// https://github.com/kiss-community/kiss/blob/edfb25aa2da44076dcb35b19f8e6cfddd5a66dfa/kiss#L659\n\t\tif _, err := t.Install(lo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, p := range dpp {\n\t\tlog.Runningf(\"building dependency %s\", p.Name)\n\n\t\tt, err := p.Build(bo)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tlog.Runningf(\"installing dependency %s\", t.Name)\n\n\t\tif _, err := t.Install(lo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfor _, p := range epp {\n\t\tlog.Runningf(\"building %s\", p.Name)\n\n\t\tt, err := p.Build(bo)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif !fi {\n\t\t\tcontinue\n\t\t}\n\n\t\tlog.Runningf(\"installing %s\", p.Name)\n\n\t\tif _, err := t.Install(lo); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *Rust) Build(out io.Writer, verbose bool) error {\n\t// Get binary name from Cargo.toml.\n\tvar m CargoManifest\n\tif err := m.Read(\"Cargo.toml\"); err != nil {\n\t\treturn fmt.Errorf(\"error reading Cargo.toml manifest: %w\", err)\n\t}\n\tbinName := m.Package.Name\n\n\tif len(r.toolchain) == 0 {\n\t\trustConstraint, err := semver.NewConstraint(r.config.File.Language.Rust.ToolchainConstraint)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error parsing rust toolchain constraint: %w\", err)\n\t\t}\n\n\t\terr = r.checkRustcVersion(rustConstraint)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tr.toolchain, err = r.getToolchain()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttoolchain := fmt.Sprintf(\"+%s\", r.toolchain)\n\n\targs := []string{\n\t\ttoolchain,\n\t\t\"build\",\n\t\t\"--bin\",\n\t\tbinName,\n\t\t\"--release\",\n\t\t\"--target\",\n\t\tr.config.File.Language.Rust.WasmWasiTarget,\n\t\t\"--color\",\n\t\t\"always\",\n\t}\n\tif verbose {\n\t\targs = append(args, \"--verbose\")\n\t}\n\t// Append debuginfo RUSTFLAGS to command environment to ensure DWARF debug\n\t// information (such as, source mappings) are compiled into the binary.\n\trustflags := \"-C debuginfo=2\"\n\tif val, ok := os.LookupEnv(\"RUSTFLAGS\"); ok {\n\t\tos.Setenv(\"RUSTFLAGS\", fmt.Sprintf(\"%s %s\", val, rustflags))\n\t} else {\n\t\tos.Setenv(\"RUSTFLAGS\", rustflags)\n\t}\n\n\t// Execute the `cargo build` commands with the Wasm WASI target, release\n\t// flags and env vars.\n\tcmd := fstexec.Streaming{\n\t\tCommand: \"cargo\",\n\t\tArgs: args,\n\t\tEnv: os.Environ(),\n\t\tOutput: out,\n\t}\n\tif r.timeout > 0 {\n\t\tcmd.Timeout = time.Duration(r.timeout) * time.Second\n\t}\n\tif err := cmd.Exec(); err != nil {\n\t\treturn err\n\t}\n\n\t// Get working directory.\n\tdir, err := os.Getwd()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"getting current working directory: %w\", err)\n\t}\n\tvar metadata CargoMetadata\n\tif err := metadata.Read(); err != nil {\n\t\treturn fmt.Errorf(\"error reading cargo metadata: %w\", err)\n\t}\n\tsrc := filepath.Join(metadata.TargetDirectory, r.config.File.Language.Rust.WasmWasiTarget, \"release\", fmt.Sprintf(\"%s.wasm\", binName))\n\tdst := filepath.Join(dir, \"bin\", \"main.wasm\")\n\n\t// Check if bin directory exists and create if not.\n\tbinDir := filepath.Join(dir, \"bin\")\n\tif err := filesystem.MakeDirectoryIfNotExists(binDir); err != nil {\n\t\treturn fmt.Errorf(\"creating bin directory: %w\", err)\n\t}\n\n\terr = filesystem.CopyFile(src, dst)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"copying wasm binary: %w\", err)\n\t}\n\n\treturn nil\n}", "func Build(objs ...interface{}) (c *cli.App) {\n\tc, err := build(objs...)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn\n}", "func (t Template) Build(norepo bool) error {\n\t// Make dirs\n\tfor _, dir := range t.Dirs {\n\t\tdir = filepath.Join(t.Root, dir)\n\t\terr := os.MkdirAll(dir, 0755)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Printf(\"> Created dir: %s\\n\", dir)\n\t}\n\t// Make files\n\tfor _, file := range t.Files {\n\t\t_, filename := splitFilename(file)\n\t\tcontent := loadContent(t.Name, filename)\n\t\tcontent = replaceTokens(content, t.Root)\n\t\tfile = filepath.Join(t.Root, file)\n\t\terr := ioutil.WriteFile(file, content, 0644)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Printf(\"> Created file: %s\\n\", file)\n\t}\n\tif norepo {\n\t\treturn nil\n\t}\n\treturn t.CreateRepo()\n}", "func (j *Javadoc) genSources(ctx android.ModuleContext, srcFiles android.Paths,\n\tflags droiddocBuilderFlags) android.Paths {\n\n\toutSrcFiles := make(android.Paths, 0, len(srcFiles))\n\tvar aidlSrcs android.Paths\n\n\taidlIncludeFlags := genAidlIncludeFlags(srcFiles)\n\n\tfor _, srcFile := range srcFiles {\n\t\tswitch srcFile.Ext() {\n\t\tcase \".aidl\":\n\t\t\taidlSrcs = append(aidlSrcs, srcFile)\n\t\tcase \".logtags\":\n\t\t\tjavaFile := genLogtags(ctx, srcFile)\n\t\t\toutSrcFiles = append(outSrcFiles, javaFile)\n\t\tdefault:\n\t\t\toutSrcFiles = append(outSrcFiles, srcFile)\n\t\t}\n\t}\n\n\t// Process all aidl files together to support sharding them into one or more rules that produce srcjars.\n\tif len(aidlSrcs) > 0 {\n\t\tsrcJarFiles := genAidl(ctx, aidlSrcs, flags.aidlFlags+aidlIncludeFlags, flags.aidlDeps)\n\t\toutSrcFiles = append(outSrcFiles, srcJarFiles...)\n\t}\n\n\treturn outSrcFiles\n}", "func dockerPackBuild() ([]byte, error) {\n\n\twd := conf.PlatformEnvs().WorkDir\n\n\t// bp at wd/bp\n\t// copy code context to wd/code\n\terr := filehelper.CheckExist(filepath.Join(wd, \"code\"), true)\n\tif err != nil {\n\t\tif err = os.RemoveAll(filepath.Join(wd, \"code\")); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif err = copy.Copy(conf.Params().Context, filepath.Join(wd, \"code\")); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tvar exactDockerfilePath = filepath.Join(wd, \"bp\", \"pack\", \"Dockerfile\")\n\tvar startFilePath = filepath.Join(wd, \"bp\", \"pack\", \"start.sh\")\n\tstartFileContent, err := ioutil.ReadFile(startFilePath)\n\tif err == nil {\n\t\tnewStartFileContent := template.Render(string(startFileContent),\n\t\t\tmap[string]string{\"JAVA_OPTS\": conf.Params().JavaOpts},\n\t\t)\n\t\tlogrus.Infof(\"new startfile :%s\", newStartFileContent)\n\t\terr := ioutil.WriteFile(startFilePath, []byte(newStartFileContent), os.ModePerm)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tvar exactWd = wd\n\t// dockerfile use user's Dockerfile\n\tif conf.Params().Language == types.LanguageDockerfile {\n\t\texactDockerfilePath = filepath.Join(wd, \"code\", \"Dockerfile\")\n\t\texactWd = filepath.Join(wd, \"code\")\n\t}\n\n\tdockerfileContent, err := ioutil.ReadFile(exactDockerfilePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tnewDockerfileContent := dockerfile.ReplaceOrInsertBuildArgToDockerfile(dockerfileContent, conf.Params().BpArgs)\n\tif err = filehelper.CreateFile(exactDockerfilePath, string(newDockerfileContent), 0644); err != nil {\n\t\treturn nil, err\n\t}\n\n\tcpu := conf.PlatformEnvs().CPU\n\tmemory := conf.PlatformEnvs().Memory\n\n\tcpuShares := cpu * 1024\n\tcpuPeriod := 100000\n\t//cpuQuota := cpu * float64(cpuPeriod) * 2 // 0.5 -> 1, 1 -> 2\n\tcpuQuota := 150000\n\n\toneImage := conf.Params().Modules[0].Image.Name\n\n\tdockerBuildNetwork := os.Getenv(\"BP_DOCKER_BUILD_NETWORK\")\n\tif dockerBuildNetwork == \"\" {\n\t\tdockerBuildNetwork = \"host\"\n\t}\n\n\tdockerBuildCmdArgs := []string{\n\t\t\"build\",\n\t\t// float\n\t\t\"--memory\", strconv.FormatFloat(float64(memory*1000000), 'f', 0, 64),\n\t\t// int strconv.ParseInt\n\t\t\"--cpu-shares\", strconv.FormatFloat(float64(cpuShares), 'f', 0, 64),\n\t\t// int\n\t\t\"--cpu-period\", strconv.FormatFloat(float64(cpuPeriod), 'f', 0, 64),\n\t\t// int\n\t\t\"--cpu-quota\", strconv.FormatFloat(float64(cpuQuota), 'f', 0, 64),\n\t\t\"--network\", dockerBuildNetwork,\n\n\t\t\"--pull\",\n\n\t\t\"-t\", oneImage,\n\t\t\"-f\", exactDockerfilePath,\n\n\t\t\"--build-arg\", \"DICE_VERSION=\" + conf.PlatformEnvs().DiceVersion,\n\n\t\texactWd,\n\t}\n\n\t// HTTP_PROXY & HTTPS_PROXY\n\tif conf.Params().HttpProxy != \"\" {\n\t\tdockerBuildCmdArgs = append(dockerBuildCmdArgs, \"--build-arg\", \"HTTP_PROXY=\"+conf.Params().HttpProxy)\n\t}\n\tif conf.Params().HttpsProxy != \"\" {\n\t\tdockerBuildCmdArgs = append(dockerBuildCmdArgs, \"--build-arg\", \"HTTPS_PROXY=\"+conf.Params().HttpsProxy)\n\t}\n\n\t// build\n\tdockerBuild := exec.Command(\"docker\", dockerBuildCmdArgs...)\n\n\tbplog.Println(strutil.Join(dockerBuild.Args, \" \", false))\n\tbplog.Printf(\"docker build network: %s\\n\", dockerBuildNetwork)\n\n\tdockerBuild.Dir = wd\n\tdockerBuild.Stdout = os.Stdout\n\tdockerBuild.Stderr = os.Stderr\n\tif err := dockerBuild.Run(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// 0. 给 image 打上 APP_DIR env\n\t// 1. multi module docker tag and push\n\t// 2. 写 pack-result 文件\n\n\tpackResult := make([]ModuleImage, 0)\n\tvar tagPushScript = []string{\"#!/bin/sh\"}\n\tfor _, m := range conf.Params().Modules {\n\t\tdockerfileForARG := []string{\n\t\t\tfmt.Sprintf(\"FROM %s AS base\", oneImage),\n\t\t\tfmt.Sprintf(\"ENV APP_DIR=%s\", m.Path),\n\t\t\tfmt.Sprintf(\"FROM base\"),\n\t\t}\n\t\tdockerfileForARGPath := filepath.Join(wd, \"Dockerfile.build.\"+m.Name)\n\t\tif err := filehelper.CreateFile(dockerfileForARGPath, strings.Join(dockerfileForARG, \"\\n\"), 0755); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ttagPushScript = append(tagPushScript,\n\t\t\tfmt.Sprintf(\"docker build -t %s -f %s .\", m.Image.Name, dockerfileForARGPath),\n\t\t\tfmt.Sprintf(\"docker push %s\", m.Image.Name),\n\t\t)\n\n\t\tpackResult = append(packResult, ModuleImage{m.Name, m.Image.Name})\n\t}\n\n\ttagPushScriptPath := filepath.Join(wd, \"repack_push.sh\")\n\tif err := filehelper.CreateFile(tagPushScriptPath, strings.Join(tagPushScript, \"\\n\"), 075); err != nil {\n\t\treturn nil, err\n\t}\n\tcmd := exec.Command(\"/bin/sh\", tagPushScriptPath)\n\tcmd.Dir = wd\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\tif err := cmd.Run(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tb, err := json.MarshalIndent(packResult, \"\", \" \")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = os.RemoveAll(wd)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := filehelper.CreateFile(filepath.Join(wd, \"pack-result\"), string(b), 0644); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn b, err\n}", "func build(buildEnvVars *map[string]string) (string, error) {\n\t_ = os.Setenv(\"GOOS\", \"linux\")\n\t_ = os.Setenv(\"GOARCH\", \"amd64\")\n\t// Don't set Go modules on by default, set in aegis.yaml\n\t// _ = os.Setenv(\"GO111MODULE\", \"on\")\n\tif buildEnvVars != nil {\n\t\tfor k, v := range *buildEnvVars {\n\t\t\tif k != \"\" {\n\t\t\t\tfmt.Printf(\"Setting environment variable: %s=%s\\n\", strings.ToUpper(k), v)\n\t\t\t\t_ = os.Setenv(strings.ToUpper(k), v)\n\t\t\t}\n\t\t}\n\t\tfmt.Println(\"\")\n\t}\n\tpath := getExecPath(\"go\")\n\tpwd, _ := os.Getwd()\n\n\t// Try to build a smaller binary.\n\t// This won't work on Windows. Though Windows remains untested in general, let's try this and fall back.\n\tcmd := exec.Command(\"sh\", \"-c\", path+` build -ldflags=\"-w -s\" -o `+aegisAppName)\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\tif err := cmd.Run(); err != nil {\n\t\t// If it failed, just build without all the fancy flags. The binary size will be a little larger though.\n\t\t// This should work on Windows. Right? TODO: Test. Better yet, figure out how to build Cmd with flags.\n\t\t// Spent over an hour trying every method of escaping known to man. Why???\n\t\targs := []string{path, \"build\", \"-o\", aegisAppName}\n\t\tcmd := exec.Cmd{\n\t\t\tPath: path,\n\t\t\tArgs: args,\n\t\t}\n\t\tif err := cmd.Run(); err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\tbuiltApp := filepath.Join(pwd, aegisAppName)\n\n\treturn builtApp, nil\n}", "func buildAndCopyProgram(src io.Reader) error {\n\t// FIXME: BuildProgram should probably be in some other package,\n\t// so that it can be used by both the compiler tests and the\n\t// command line client.\n\td, err := ioutil.TempDir(\"\", \"langbuild\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tif debug {\n\t\tlog.Println(\"Using temporary directory\", d, \"(WARNING: will not automatically delete in debug mode)\")\n\t}\n\tif !debug {\n\t\tdefer os.RemoveAll(d)\n\t}\n\texe, err := codegen.BuildProgram(d, src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif exe == \"\" {\n\t\treturn fmt.Errorf(\"No executable built.\")\n\t}\n\tcwd, err := os.Getwd()\n\tif err != nil {\n\t\treturn err\n\t}\n\tname := path.Base(cwd)\n\tif name == \".\" || name == \"\" || name == \"/\" {\n\t\tlog.Fatal(\"Could not determine appropriate executable name.\")\n\t}\n\treturn copyFile(d+\"/\"+exe, \"./\"+name)\n}", "func Build(ctx context.Context, configSourcesSettings map[string]ConfigSettings, params CreateParams, factories Factories) (map[string]configsource.ConfigSource, error) {\n\tcfgSources := make(map[string]configsource.ConfigSource, len(configSourcesSettings))\n\tfor fullName, cfgSrcSettings := range configSourcesSettings {\n\t\t// If we have the setting we also have the factory.\n\t\tfactory, ok := factories[cfgSrcSettings.Type()]\n\t\tif !ok {\n\t\t\treturn nil, &errUnknownType{\n\t\t\t\tfmt.Errorf(\"unknown %s config source type for %s\", cfgSrcSettings.Type(), fullName),\n\t\t\t}\n\t\t}\n\n\t\tparams.Logger = params.Logger.With(zap.String(\"config_source\", fullName))\n\t\tcfgSrc, err := factory.CreateConfigSource(ctx, params, cfgSrcSettings)\n\t\tif err != nil {\n\t\t\treturn nil, &errConfigSourceCreation{\n\t\t\t\tfmt.Errorf(\"failed to create config source %s: %w\", fullName, err),\n\t\t\t}\n\t\t}\n\n\t\tif cfgSrc == nil {\n\t\t\treturn nil, &errFactoryCreatedNil{\n\t\t\t\tfmt.Errorf(\"factory for %q produced a nil extension\", fullName),\n\t\t\t}\n\t\t}\n\n\t\tcfgSources[fullName] = cfgSrc\n\t}\n\n\treturn cfgSources, nil\n}", "func (p *Package) SourceFiles() []string {\n\treturn concat(p.GoFiles, p.CgoFiles, p.IgnoredGoFiles)\n}", "func Build(path string, nameservers, dnsSearch, dnsOptions []string) (*File, error) {\n\tcontent := bytes.NewBuffer(nil)\n\tif len(dnsSearch) > 0 {\n\t\tif searchString := strings.Join(dnsSearch, \" \"); strings.Trim(searchString, \" \") != \".\" {\n\t\t\tif _, err := content.WriteString(\"search \" + searchString + \"\\n\"); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\tfor _, dns := range nameservers {\n\t\tif _, err := content.WriteString(\"nameserver \" + dns + \"\\n\"); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif len(dnsOptions) > 0 {\n\t\tif optsString := strings.Join(dnsOptions, \" \"); strings.Trim(optsString, \" \") != \"\" {\n\t\t\tif _, err := content.WriteString(\"options \" + optsString + \"\\n\"); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\tif err := os.WriteFile(path, content.Bytes(), 0o644); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &File{Content: content.Bytes(), Hash: hashData(content.Bytes())}, nil\n}", "func BuildFromFile(filePath string) (*CodeOwners, []error) {\n\tentries, errors := BuildEntriesFromFile(filePath, false)\n\tif errors != nil {\n\t\treturn nil, errors\n\t}\n\tindex, err := createIndexFromEntries(entries)\n\tif err != nil {\n\t\treturn nil, []error{err}\n\t}\n\treturn index, nil\n}", "func Build(recipes []string, config *config.BuildConfig) error {\n\tlog.WithField(\"recipes\", recipes).Info(\"Starting build\")\n\tcacheDir := filepath.Join(config.BuildDir, \".cache\")\n\n\t// Ensure build, output, and cache directories exist.\n\tfor _, dir := range []string{config.BuildDir, config.OutputDir, cacheDir} {\n\t\tlog.WithField(\"dir\", dir).Debug(\"Ensuring directory exists\")\n\t\tif err := os.Mkdir(dir, 0700); err != nil {\n\t\t\tif !os.IsExist(err) {\n\t\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\t\"dir\": dir,\n\t\t\t\t\t\"err\": err,\n\t\t\t\t}).Error(\"Could not create directory\")\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// Get dependency order for all input recipes.\n\tdeps, err := getRecipeDeps(recipes, config.Platform, config.Arch)\n\tif err != nil {\n\t\tlog.WithField(\"err\", err).Error(\"Could not get recipe dependencies\")\n\t\treturn err\n\t}\n\n\tcache, err := newSourceCache(cacheDir)\n\tif err != nil {\n\t\tlog.WithField(\"err\", err).Error(\"Could not create source cache\")\n\t\treturn err\n\t}\n\n\t// Make our context\n\tctx := context{\n\t\trootEnv: env.FromOS(),\n\t\tconfig: config,\n\t\tcache: cache,\n\t\tpackageEnv: make(map[string]map[string]string),\n\t}\n\n\t// For each dependency, we build it.\n\tfor _, dep := range deps {\n\t\tif err = buildOne(dep, &ctx); err != nil {\n\t\t\tlog.WithFields(logrus.Fields{\n\t\t\t\t\"dep\": dep,\n\t\t\t\t\"err\": err,\n\t\t\t}).Error(\"Error building dependency\")\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Build(f *function.Function) (function.FilesMap, function.Config, error) {\n\tfMap, err := getFiles(f)\n\tif err != nil {\n\t\treturn nil, function.Config{}, err\n\t}\n\treturn fMap, getDefaultConfig(), nil\n}", "func Build() flaw.Flaw {\n\t// make http-specs directory\n\tcurDir, httpDir, err := getDirectories()\n\tif err != nil {\n\t\treturn flaw.From(err)\n\t}\n\n\t// assemble the specs\n\terr = assembleSpecs(curDir, httpDir)\n\tif err != nil {\n\t\treturn flaw.From(err)\n\t}\n\n\t// image/build-with-tls images-to-deploy/http-specs\n\terr = image.BuildGoWithTLS(\"images-to-deploy/http-specs\")\n\tif err != nil {\n\t\t// fmt.Println(\"build go err:\", err.String())\n\t\treturn flaw.From(err)\n\t}\n\n\t// clean up/delete specs\n\terr = deleteSpecs(httpDir)\n\tif err != nil {\n\t\treturn flaw.From(err)\n\t}\n\n\treturn nil\n}", "func BuildGoSource(packagePath, output string) (func() error, error) {\n\tcmdBuild := exec.Command(\"go\", \"build\", \"-o\", output, packagePath)\n\tif out, err := cmdBuild.CombinedOutput(); err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to build package: %v (%s)\", err, string(out))\n\t}\n\treturn func() error {\n\t\treturn os.Remove(output)\n\t}, nil\n}", "func Compile(ctx context.Context, ui *ui.UI, discovered *discover.Discovered) (*Binaries, error) {\n\tegrp, ctx := errgroup.WithContext(ctx)\n\tbinaries := &Binaries{}\n\tif discovered.Local != nil {\n\t\tegrp.Go(func() error {\n\t\t\tpath, err := compile(ctx, ui, discovered.Local)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbinaries.Local = Binary{Path: path}\n\t\t\treturn nil\n\t\t})\n\t}\n\tif discovered.Plan != nil {\n\t\tegrp.Go(func() error {\n\t\t\tpath, err := compile(ctx, ui, discovered.Plan)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tbinaries.Plan = Binary{Path: path}\n\t\t\treturn nil\n\t\t})\n\t}\n\n\tif err := egrp.Wait(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn binaries, nil\n}", "func Build(config map[string]interface{}) {\n}", "func buildFromGoFile() {\n\n\tflag.Parse()\n\tbuildCmd := exec.Command(\"go\", \"build\", \"-o\", os.Args[1], *fileName)\n\tbuildCmd.Stderr = os.Stderr\n\tif err := buildCmd.Run(); err != nil {\n\t\tcolor.Red.Println(\"error: unable to build go file.\")\n\t} else {\n\t\tfmt.Println(\"Building file:\", *fileName)\n\t}\n}", "func (l *LambdaClient) buildPackage() ([]byte, error) {\n\tlog.Info(\"Building Lambda function\")\n\tbinFile, err := crossCompile(\"lambda_artifact\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer os.RemoveAll(filepath.Dir(binFile)) // Remove temporary binary file\n\n\tlog.Debug(\"Opening recompiled binary to be zipped\")\n\tbinReader, err := os.Open(binFile)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tzipBuf := new(bytes.Buffer)\n\tarchive := zip.NewWriter(zipBuf)\n\theader := &zip.FileHeader{\n\t\tName: \"main\",\n\t\tExternalAttrs: (0777 << 16), // File permissions\n\t\tCreatorVersion: (3 << 8), // Magic number indicating a Unix creator\n\t}\n\n\tlog.Debug(\"Adding binary to zip archive\")\n\twriter, err := archive.CreateHeader(header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t_, err = io.Copy(writer, binReader)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbinReader.Close()\n\tarchive.Close()\n\n\tlog.Debugf(\"Final zipped function binary size: %s\", humanize.Bytes(uint64(len(zipBuf.Bytes()))))\n\treturn zipBuf.Bytes(), nil\n}", "func buildTestExecutable() {\n\t// this will create a file called \"_testmain.go\"\n\ttestPack := createTestPackage()\n\n\tif compile(testPack) {\n\t\tlinkErrors = !link(testPack) || linkErrors\n\t} else {\n\t\tlogger.Error(\"Can't link executable because of compile errors.\\n\")\n\t\tcompileErrors = true\n\t}\n\n\t// delete temporary _testmain.go file\n\t// \tos.Remove(\"_testmain.go\")\n\n\tif compileErrors || linkErrors {\n\t\treturn\n\t}\n\n\tif *flagRunExec {\n\t\tvar argvFilled int\n\t\tvar argc int = 1\n\t\tif *flagMatch != \"\" {\n\t\t\targc += 2\n\t\t}\n\t\tif *flagBenchmarks != \"\" {\n\t\t\targc += 2\n\t\t}\n\t\tif *flagVerboseMode {\n\t\t\targc++\n\t\t}\n\t\targv := make([]string, argc)\n\t\targv[argvFilled] = outputDirPrefix + testPack.OutputFile\n\t\targvFilled++\n\t\tif *flagMatch != \"\" {\n\t\t\targv[argvFilled] = \"-match\"\n\t\t\targvFilled++\n\t\t\targv[argvFilled] = *flagMatch\n\t\t\targvFilled++\n\t\t}\n\t\tif *flagBenchmarks != \"\" {\n\t\t\targv[argvFilled] = \"-benchmarks\"\n\t\t\targvFilled++\n\t\t\targv[argvFilled] = *flagBenchmarks\n\t\t\targvFilled++\n\t\t}\n\t\tif *flagVerboseMode {\n\t\t\targv[argvFilled] = \"-v\"\n\t\t\targvFilled++\n\t\t}\n\n\t\trunExec(argv)\n\t}\n}", "func Build() (app *App, compileError *revel.Error) {\n\t// First, clear the generated files (to avoid them messing with ProcessSource).\n\tcleanSource(\"tmp\", \"routes\")\n\n\tsourceInfo, compileError := ProcessSource(revel.CodePaths)\n\tif compileError != nil {\n\t\treturn nil, compileError\n\t}\n\n\t// Add the db.import to the import paths.\n\tif dbImportPath, found := revel.Config.String(\"db.import\"); found {\n\t\tsourceInfo.InitImportPaths = append(sourceInfo.InitImportPaths, dbImportPath)\n\t}\n\n\t// Generate two source files.\n\ttemplateArgs := map[string]interface{}{\n\t\t\"Controllers\": sourceInfo.ControllerSpecs(),\n\t\t\"ValidationKeys\": sourceInfo.ValidationKeys,\n\t\t\"ImportPaths\": calcImportAliases(sourceInfo),\n\t\t\"TestSuites\": sourceInfo.TestSuites(),\n\t}\n\tgenSource(\"tmp\", \"main.go\", MAIN, templateArgs)\n\tgenSource(\"routes\", \"routes.go\", ROUTES, templateArgs)\n\n\t// Read build config.\n\tbuildTags := revel.Config.StringDefault(\"build.tags\", \"\")\n\n\t// Build the user program (all code under app).\n\t// It relies on the user having \"go\" installed.\n\tgoPath, err := exec.LookPath(\"go\")\n\tif err != nil {\n\t\trevel.ERROR.Fatalf(\"Go executable not found in PATH.\")\n\t}\n\n\tpkg, err := build.Default.Import(revel.ImportPath, \"\", build.FindOnly)\n\tif err != nil {\n\t\trevel.ERROR.Fatalln(\"Failure importing\", revel.ImportPath)\n\t}\n\tbinName := path.Join(pkg.BinDir, path.Base(revel.BasePath))\n\tif runtime.GOOS == \"windows\" {\n\t\tbinName += \".exe\"\n\t}\n\n\tgotten := make(map[string]struct{})\n\tfor {\n\t\tappVersion := getAppVersion()\n\t\tversionLinkerFlags := fmt.Sprintf(\"-X %s/app.APP_VERSION \\\"%s\\\"\", revel.ImportPath, appVersion)\n\n\t\tbuildCmd := exec.Command(goPath, \"build\",\n\t\t\t\"-ldflags\", versionLinkerFlags,\n\t\t\t\"-tags\", buildTags,\n\t\t\t\"-o\", binName, path.Join(revel.ImportPath, \"app\", \"tmp\"))\n\t\trevel.TRACE.Println(\"Exec:\", buildCmd.Args)\n\t\toutput, err := buildCmd.CombinedOutput()\n\n\t\t// If the build succeeded, we're done.\n\t\tif err == nil {\n\t\t\treturn NewApp(binName), nil\n\t\t}\n\t\trevel.ERROR.Println(string(output))\n\n\t\t// See if it was an import error that we can go get.\n\t\tmatches := importErrorPattern.FindStringSubmatch(string(output))\n\t\tif matches == nil {\n\t\t\treturn nil, newCompileError(output)\n\t\t}\n\n\t\t// Ensure we haven't already tried to go get it.\n\t\tpkgName := matches[1]\n\t\tif _, alreadyTried := gotten[pkgName]; alreadyTried {\n\t\t\treturn nil, newCompileError(output)\n\t\t}\n\t\tgotten[pkgName] = struct{}{}\n\n\t\t// Execute \"go get <pkg>\"\n\t\tgetCmd := exec.Command(goPath, \"get\", pkgName)\n\t\trevel.TRACE.Println(\"Exec:\", getCmd.Args)\n\t\tgetOutput, err := getCmd.CombinedOutput()\n\t\tif err != nil {\n\t\t\trevel.ERROR.Println(string(getOutput))\n\t\t\treturn nil, newCompileError(output)\n\t\t}\n\n\t\t// Success getting the import, attempt to build again.\n\t}\n\trevel.ERROR.Fatalf(\"Not reachable\")\n\treturn nil, nil\n}", "func BuildTemplate(dir string, files ...string) error {\n\tif _, err := os.Stat(dir); err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.New(\"dir open err\")\n\t}\n\tself := &templateFile{\n\t\troot: dir,\n\t\tfiles: make(map[string][]string),\n\t}\n\terr := filepath.Walk(dir, func(path string, f os.FileInfo, err error) error {\n\t\treturn self.visit(path, f, err)\n\t})\n\tif err != nil {\n\t\tfmt.Printf(\"filepath.Walk() returned %v\\n\", err)\n\t\treturn err\n\t}\n\tbuildAllFiles := len(files) == 0\n\tfor _, v := range self.files {\n\t\tfor _, file := range v {\n\t\t\tif buildAllFiles || yeestrings.IsInSlice(files, file) {\n\t\t\t\ttemplatesLock.Lock()\n\t\t\t\text := filepath.Ext(file)\n\t\t\t\tvar t *template.Template\n\t\t\t\tif len(ext) == 0 {\n\t\t\t\t\tt, err = getTemplate(self.root, file, v...)\n\t\t\t\t} else {\n\t\t\t\t\tt, err = getTemplate(self.root, file, v...)\n\t\t\t\t}\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t} else {\n\t\t\t\t\tcacheTemplates[file] = t\n\t\t\t\t}\n\t\t\t\ttemplatesLock.Unlock()\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func BuildRun(r *cmd.Root, s *cmd.Sub) {\n\t// gFlags := r.Flags.(*GlobalFlags)\n\tflags := s.Flags.(*BuildFlags)\n\tlog.Infoln(\"Loading configuration\")\n\tconf, err := config.Load(config.Path(flags.Src))\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to load configuration: %q\\n\", err)\n\t}\n\tfmt.Printf(\"%#v\\n\", conf)\n\tlog.Goodln(\"Config Loaded.\")\n\n\tlog.Infoln(\"Loading templates\")\n\ttmpls, err := templates.Load(flags.Src)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to load templates: %q\\n\", err)\n\t}\n\tfmt.Printf(\"%#v\\n\", tmpls)\n\tlog.Goodln(\"Templates Loaded.\")\n\tlog.Infoln(\"Setting up the rendering process\")\n\tsite, err := render.NewSite(&conf, tmpls)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to set up rendering: %q\\n\", err)\n\t}\n\tfmt.Printf(\"%#v\\n\", site)\n}", "func Compile(source string) error {\n\tpage := buildPage()\n\n\tcreateRSS(page)\n\tcreateHome(page)\n\tcreateIssues(page)\n\n\tCopyDir(\"themes/yeo/assets\", \"public/assets\")\n\tCopyDir(\"static/\", \"public/\")\n\n\treturn nil\n}", "func Main() error {\n\tenv := golang.Default()\n\tif env.CgoEnabled {\n\t\t// TODO(bluecmd): Might need CGO for pcap if that should be included\n\t\t// Given that we already depend on a gcc being available for u-boot and\n\t\t// the linux kernel, this might be fine. Especially if we need to do the\n\t\t// yocto route down the line.\n\t\tlog.Printf(\"Disabling CGO for u-bmc...\")\n\t\tenv.CgoEnabled = false\n\t}\n\tlog.Printf(\"Build environment: %s\", env)\n\tif env.GOOS != \"linux\" {\n\t\tlog.Printf(\"GOOS is not linux. Did you mean to set GOOS=linux?\")\n\t}\n\n\tbuilder := builder.BBBuilder{}\n\tarchiver, err := initramfs.GetArchiver(\"cpio\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttempDir, err := ioutil.TempDir(\"\", \"u-root\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer os.RemoveAll(tempDir)\n\n\tw, err := archiver.OpenWriter(*outputPath, env.GOOS, env.GOARCH)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tpkgs := append(packages, fmt.Sprintf(\"github.com/u-root/u-bmc/platform/%s/cmd/*\", *platform))\n\n\tbaseFile := uroot.DefaultRamfs.Reader()\n\n\topts := uroot.Opts{\n\t\tEnv: env,\n\t\tCommands: []uroot.Commands{\n\t\t\t{\n\t\t\t\tBuilder: builder,\n\t\t\t\tPackages: pkgs,\n\t\t\t},\n\t\t},\n\t\tBaseArchive: baseFile,\n\t\tTempDir: tempDir,\n\t\tOutputFile: w,\n\t\tInitCmd: \"init\",\n\t\tDefaultShell: \"elvish\",\n\t}\n\tlogger := log.New(os.Stderr, \"\", log.LstdFlags)\n\treturn uroot.CreateInitramfs(logger, opts)\n}", "func main() {\n\tgopath := os.Getenv(\"GOPATH\")\n\tif gopath == \"\" {\n\t\tgopath = build.Default.GOPATH\n\t}\n\tbase := \"cd \" + gopath + \"/src/github.com/cavapoo2/eventsBoard/\"\n\tfmt.Println(\"base is \", base)\n\tinstallPaths := []string{\n\n\t\tbase + \"lib/configuration && go install\",\n\t\tbase + \"lib/helper/amqp && go install\",\n\n\t\tbase + \"lib/helper/kafka && go install\",\n\t\tbase + \"lib/msgqueue/amqp && go install\",\n\t\tbase + \"lib/msgqueue/builder && go install\",\n\t\tbase + \"lib/msgqueue/kafka && go install\",\n\t\tbase + \"lib/msgqueue && go install\",\n\n\t\tbase + \"lib/persistence/mongolayer && go install\",\n\t\tbase + \"lib/persistence/dblayer && go install\",\n\t\tbase + \"lib/persistence && go install\",\n\n\t\tbase + \"bookingservice/listener && go install\",\n\t\tbase + \"bookingservice/rest && go install\",\n\t\tbase + \"contracts && go install\",\n\t\tbase + \"eventService/rest && go install\",\n\t\tbase + \"eventService/listner && go install\",\n\t}\n\n\tbuildPaths := []string{\n\t\tbase + \"bookingservice && go build main.go\",\n\t\tbase + \"eventService && go build main.go\",\n\t\tbase + \"eventService && go build -o eventservice main.go\",\n\t\tbase + \"bookingservice && go build -o bookingservice main.go\",\n\t}\n\tfor _, p := range installPaths {\n\t\texecCommand(p)\n\t}\n\tfor _, p := range buildPaths {\n\t\texecCommand(p)\n\t}\n\n}", "func (p *TestPackage) Build() error {\n\tif p.build != nil {\n\t\treturn nil\n\t}\n\tif err := p.write(); err != nil {\n\t\treturn err\n\t}\n\tbasePackage, err := build.Default.ImportDir(p.Path, build.ImportComment)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar packageFiles []string\n\tfor _, filename := range basePackage.GoFiles {\n\t\tpackageFiles = append(packageFiles, path.Join(p.Path, filename))\n\t}\n\n\tconf := &packages.Config{\n\t\tMode: gosec.LoadMode,\n\t\tTests: false,\n\t}\n\tpkgs, err := packages.Load(conf, packageFiles...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tp.build = &buildObj{\n\t\tpkg: basePackage,\n\t\tconfig: conf,\n\t\tpkgs: pkgs,\n\t}\n\treturn nil\n}", "func doBuild(env env.Project, options *BuildOptions) (err error) {\n\tif options == nil {\n\t\toptions = &BuildOptions{}\n\t}\n\n\tif len(options.BuildDocker) > 0 {\n\t\tenv.SetDockerBuild()\n\t}\n\n\tif options.GenerationOnly {\n\t\t// Only perform prepare\n\t\treturn PrepareApp(env, options.PrepareOptions)\n\t}\n\n\tif !options.SkipPrepare && !options.NoGeneration {\n\t\terr = PrepareApp(env, options.PrepareOptions)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\terr = env.Build()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !options.EmbedConfig {\n\t\tfgutil.CopyFile(filepath.Join(env.GetRootDir(), config.FileDescriptor), filepath.Join(env.GetBinDir(), config.FileDescriptor))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tos.Remove(filepath.Join(env.GetBinDir(), config.FileDescriptor))\n\t}\n\n\t// To create a dockerfile this component executes four steps\n\t// 1. Check if flogo.json exists in bin folder (built without -e)\n\t// 2. Read flogo.json from ./flogo.json\n\t// 3. Output the dockerfile in ./bin/dockerfile\n\t// 4. Execute docker build\n\tif len(options.BuildDocker) > 0 {\n\t\tfmt.Println(\"docker:\", options.BuildDocker)\n\t\tconfig, err := jsonconfig.LoadAbstract(\"./flogo.json\", \"\")\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdata := make(map[string]interface{})\n\t\tfound := false\n\n\t\tfor _, value := range config[\"triggers\"].Arr {\n\t\t\tif value.Obj[\"id\"].Str == options.BuildDocker {\n\t\t\t\tfound = true\n\t\t\t\tdata[\"name\"] = config[\"name\"].Str\n\t\t\t\tdata[\"version\"] = config[\"version\"].Str\n\t\t\t\tdata[\"port\"] = value.Obj[\"settings.port\"].Str\n\t\t\t}\n\t\t}\n\n\t\tif options.BuildDocker == \"no-trigger\" {\n\t\t\tfound = true\n\t\t\tdata[\"name\"] = config[\"name\"].Str\n\t\t\tdata[\"version\"] = config[\"version\"].Str\n\t\t\tdata[\"port\"] = \"\"\n\t\t}\n\n\t\tif found {\n\t\t\tt := template.Must(template.New(\"email\").Parse(dockerfile))\n\t\t\tbuf := &bytes.Buffer{}\n\t\t\tif err := t.Execute(buf, data); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\ts := buf.String()\n\n\t\t\tif data[\"port\"] == \"\" {\n\t\t\t\ts = strings.Replace(s, \"EXPOSE \\n\", \"\", -1)\n\t\t\t}\n\n\t\t\tfile, err := os.Create(\"./bin/dockerfile\")\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer file.Close()\n\n\t\t\tfile.WriteString(s)\n\t\t\tfile.Sync()\n\n\t\t\tcmd := exec.Command(\"docker\", \"build\", \".\", \"-t\", strings.ToLower(config[\"name\"].Str)+\":\"+config[\"version\"].Str)\n\t\t\tcmd.Dir = \"./bin\"\n\t\t\tcmd.Stdout = os.Stdout\n\t\t\tcmd.Stderr = os.Stderr\n\t\t\terr = cmd.Run()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tfmt.Println(\"Your app doesn't contain the trigger you specified so we can't create a dockerfile for it\")\n\t\t}\n\t}\n\treturn\n}", "func (r Rust) SourceDirectory() string { return \"src\" }", "func BuildTemplate(dir string, files ...string) error {\n\tvar err error\n\tfs := beeTemplateFS()\n\tf, err := fs.Open(dir)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.New(\"dir open err\")\n\t}\n\tdefer f.Close()\n\n\tbeeTemplates, ok := beeViewPathTemplates[dir]\n\tif !ok {\n\t\tpanic(\"Unknown view path: \" + dir)\n\t}\n\tself := &templateFile{\n\t\troot: dir,\n\t\tfiles: make(map[string][]string),\n\t}\n\terr = Walk(fs, dir, self.visit)\n\tif err != nil {\n\t\tfmt.Printf(\"Walk() returned %v\\n\", err)\n\t\treturn err\n\t}\n\tbuildAllFiles := len(files) == 0\n\tfor _, v := range self.files {\n\t\tfor _, file := range v {\n\t\t\tif buildAllFiles || utils.InSlice(file, files) {\n\t\t\t\ttemplatesLock.Lock()\n\t\t\t\text := filepath.Ext(file)\n\t\t\t\tvar t *template.Template\n\t\t\t\tif len(ext) == 0 {\n\t\t\t\t\tt, err = getTemplate(self.root, fs, file, v...)\n\t\t\t\t} else if fn, ok := beeTemplateEngines[ext[1:]]; ok {\n\t\t\t\t\tt, err = fn(self.root, file, beegoTplFuncMap)\n\t\t\t\t} else {\n\t\t\t\t\tt, err = getTemplate(self.root, fs, file, v...)\n\t\t\t\t}\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogs.Error(\"parse template err:\", file, err)\n\t\t\t\t\ttemplatesLock.Unlock()\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tbeeTemplates[file] = t\n\t\t\t\ttemplatesLock.Unlock()\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (c *TestCommand) build() (string, error) {\n\tpath := \"go\" // go must be in $PATH\n\t// -c compiles without running the test\n\targs := append([]string{\"test\", \"-c\"})\n\tcmd := exec.Command(path, args...)\n\tpkgDir := c.config.Wd()\n\tcmd.Dir = pkgDir\n\tcmd.Env = CrossEnv()\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\terr := cmd.Run()\n\t_, pkgName := filepath.Split(pkgDir)\n\tfileName := filepath.Join(pkgDir, pkgName+\".test\")\n\treturn fileName, err\n}", "func run(src string, tags string) error {\n\t// Create a temp folder.\n\ttempDir, err := ioutil.TempDir(\"\", \"vfsgendev_\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer func() {\n\t\terr := os.RemoveAll(tempDir)\n\t\tif err != nil {\n\t\t\tfmt.Fprintln(os.Stderr, \"warning: error removing temp dir:\", err)\n\t\t}\n\t}()\n\n\t// Write the source code file.\n\ttempFile := filepath.Join(tempDir, \"generate.go\")\n\terr = ioutil.WriteFile(tempFile, []byte(src), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Compile and run the program.\n\tcmd := exec.Command(\"go\", \"run\", \"-tags=\"+tags, tempFile)\n\tcmd.Stdout = os.Stdout\n\tcmd.Stderr = os.Stderr\n\treturn cmd.Run()\n}", "func main(){\n subcmd := \"build\"\n if(len(os.Args)>=2){ subcmd = os.Args[1] }\n fmt.Println(subcmd+\"ing \"+pName+\"...\");\n //Now run the appropriate type of operation\n var err error\n err = nil\n switch(subcmd){\n case \"build\":\n err = doBuild()\n\n case \"clean\":\n doClean()\n\n case \"package\":\n err = doPackage()\n\n case \"install\":\n err = doInstall()\n\n default:\n fmt.Println(\"Unknown action: \", subcmd)\n\tfmt.Println(\"Available actions are:\")\n\tfmt.Println(\" - make build:\", \"Compile the tools for the current system OS/ARCH\")\n\tfmt.Println(\" - make clean:\", \"Cleanup all the build files\")\n\tfmt.Println(\" - make package:\", \"Create a sterile \\\"dist\\\" directory ready to be copied/installed someplace\")\n\tfmt.Println(\" - make install:\", \"Install the package output to the designated directory\")\n os.Exit(1)\n }\n if(err != nil){ \n fmt.Println(\"[Error]\", err)\n os.Exit(1) \n } else {\n fmt.Println(\"[Success]\")\n os.Exit(0)\n }\n}", "func mainPackage(prog *ssa.Program, tests bool) (*ssa.Package, error) {\n\tpkgs := prog.AllPackages()\n\n\t// TODO(adonovan): allow independent control over tests, mains and libraries.\n\t// TODO(adonovan): put this logic in a library; we keep reinventing it.\n\n\tif tests {\n\t\t// If -test, use all packages' tests.\n\t\tif len(pkgs) > 0 {\n\t\t\tif main := prog.CreateTestMainPackage(pkgs...); main != nil {\n\t\t\t\treturn main, nil\n\t\t\t}\n\t\t}\n\t\treturn nil, fmt.Errorf(\"no tests\")\n\t}\n\n\t// Otherwise, use the first package named main.\n\tfor _, pkg := range pkgs {\n\t\tif pkg.Pkg.Name() == \"main\" {\n\t\t\tif pkg.Func(\"main\") == nil {\n\t\t\t\treturn nil, fmt.Errorf(\"no func main() in main package\")\n\t\t\t}\n\t\t\treturn pkg, nil\n\t\t}\n\t}\n\n\treturn nil, fmt.Errorf(\"no main package\")\n}", "func mkbuildcfg(file string) {\n\tvar buf bytes.Buffer\n\tfmt.Fprintf(&buf, \"// Code generated by go tool dist; DO NOT EDIT.\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"package buildcfg\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"import \\\"runtime\\\"\\n\")\n\tfmt.Fprintln(&buf)\n\tfmt.Fprintf(&buf, \"const defaultGO386 = `%s`\\n\", go386)\n\tfmt.Fprintf(&buf, \"const defaultGOAMD64 = `%s`\\n\", goamd64)\n\tfmt.Fprintf(&buf, \"const defaultGOARM = `%s`\\n\", goarm)\n\tfmt.Fprintf(&buf, \"const defaultGOMIPS = `%s`\\n\", gomips)\n\tfmt.Fprintf(&buf, \"const defaultGOMIPS64 = `%s`\\n\", gomips64)\n\tfmt.Fprintf(&buf, \"const defaultGOPPC64 = `%s`\\n\", goppc64)\n\tfmt.Fprintf(&buf, \"const defaultGOEXPERIMENT = `%s`\\n\", goexperiment)\n\tfmt.Fprintf(&buf, \"const defaultGO_EXTLINK_ENABLED = `%s`\\n\", goextlinkenabled)\n\tfmt.Fprintf(&buf, \"const defaultGO_LDSO = `%s`\\n\", defaultldso)\n\tfmt.Fprintf(&buf, \"const version = `%s`\\n\", findgoversion())\n\tfmt.Fprintf(&buf, \"const defaultGOOS = runtime.GOOS\\n\")\n\tfmt.Fprintf(&buf, \"const defaultGOARCH = runtime.GOARCH\\n\")\n\n\twritefile(buf.String(), file, writeSkipSame)\n}", "func buildTemplates(filenames ...string) *template.Template {\n\t// Build a slice of paths\n\tfs := []string{}\n\tfor _, fn := range filenames {\n\t\tf := buildFilePath(templatesPath, fn)\n\t\tfs = append(fs, f)\n\t}\n\n\treturn template.Must(template.ParseFiles(fs...))\n}", "func loadSources(files ...string) (string, []byte, error) {\n\tvar sources []string\n\tbuf := new(bytes.Buffer)\n\n\treadFile := func(path string) error {\n\t\tif common.IsStrictPerms() {\n\t\t\tif err := common.OwnerHasExclusiveWritePerms(path); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\tf, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"failed to open file %v\", path)\n\t\t}\n\t\tdefer f.Close()\n\n\t\tif _, err = io.Copy(buf, f); err != nil {\n\t\t\treturn errors.Wrapf(err, \"failed to read file %v\", path)\n\t\t}\n\t\treturn nil\n\t}\n\n\tfor _, filePath := range files {\n\t\tfilePath = paths.Resolve(paths.Config, filePath)\n\n\t\tif hasMeta(filePath) {\n\t\t\tmatches, err := filepath.Glob(filePath)\n\t\t\tif err != nil {\n\t\t\t\treturn \"\", nil, err\n\t\t\t}\n\t\t\tsources = append(sources, matches...)\n\t\t} else {\n\t\t\tsources = append(sources, filePath)\n\t\t}\n\t}\n\n\tif len(sources) == 0 {\n\t\treturn \"\", nil, errors.Errorf(\"no sources were found in %v\",\n\t\t\tstrings.Join(files, \", \"))\n\t}\n\n\tfor _, name := range sources {\n\t\tif err := readFile(name); err != nil {\n\t\t\treturn \"\", nil, err\n\t\t}\n\t}\n\n\treturn strings.Join(sources, \";\"), buf.Bytes(), nil\n}", "func setupCompileBuilder(lc *fs_tool.LifeCycle, sdk pb.Sdk, executorConfig *environment.ExecutorConfig) *executors.CompileBuilder {\n\tfilePath := lc.GetAbsoluteExecutableFilePath()\n\tval := setupValidators(sdk, filePath)\n\n\tcompileBuilder := executors.NewExecutorBuilder().\n\t\tWithValidator().\n\t\tWithSdkValidators(val).\n\t\tWithCompiler()\n\n\tswitch sdk {\n\tcase pb.Sdk_SDK_JAVA:\n\t\tworkingDir := lc.GetAbsoluteExecutableFilesFolderPath()\n\n\t\tcompileBuilder = compileBuilder.\n\t\t\tWithCommand(executorConfig.CompileCmd).\n\t\t\tWithArgs(executorConfig.CompileArgs).\n\t\t\tWithFileName(filePath).\n\t\t\tWithWorkingDir(workingDir)\n\t}\n\treturn compileBuilder\n}", "func getPreprocessSources(inputFiles, clangFlags []string, cppCode bool) (\n\tout bytes.Buffer, err error) {\n\t// get temp dir\n\tdir, err := ioutil.TempDir(\"\", \"c4go-union\")\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer func() { _ = os.RemoveAll(dir) }()\n\n\t// file name union file\n\tvar unionFileName = dir + \"/\" + \"unionFileName.c\"\n\n\t// create a body for union file\n\tvar unionBody string\n\tfor i := range inputFiles {\n\t\tvar absPath string\n\t\tabsPath, err = filepath.Abs(inputFiles[i])\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tunionBody += fmt.Sprintf(\"#include \\\"%s\\\"\\n\", absPath)\n\t}\n\n\t// write a union file\n\terr = ioutil.WriteFile(unionFileName, []byte(unionBody), 0644)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Add open source defines\n\tclangFlags = append(clangFlags, \"-D_GNU_SOURCE\")\n\n\t// preprocessor clang\n\tvar stderr bytes.Buffer\n\n\tvar args []string\n\targs = append(args, \"-E\", \"-C\")\n\targs = append(args, clangFlags...)\n\targs = append(args, unionFileName) // All inputFiles\n\n\tvar outFile bytes.Buffer\n\tvar cmd *exec.Cmd\n\n\tcompiler, compilerFlag := Compiler(cppCode)\n\targs = append(compilerFlag, args...)\n\tcmd = exec.Command(compiler, args...)\n\n\tcmd.Stdout = &outFile\n\tcmd.Stderr = &stderr\n\terr = cmd.Run()\n\tif err != nil {\n\t\terr = fmt.Errorf(\"preprocess for file: %v\\nfailed: %v\\nStdErr = %v\", inputFiles, err, stderr.String())\n\t\treturn\n\t}\n\t_, err = out.Write(outFile.Bytes())\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func buildFilegroup(state *core.BuildState, target *core.BuildTarget) (bool, error) {\n\tif err := prepareDirectory(state.ProcessExecutor, target.OutDir(), false); err != nil {\n\t\treturn true, err\n\t}\n\tchanged := false\n\toutDir := target.OutDir()\n\tlocalSources := target.AllSourceLocalPaths(state.Graph)\n\tfor i, source := range target.AllSourceFullPaths(state.Graph) {\n\t\tout := filepath.Join(outDir, localSources[i])\n\t\tfileChanged, err := theFilegroupBuilder.Build(state, target, source, out)\n\t\tif err != nil {\n\t\t\treturn true, err\n\t\t}\n\t\tchanged = changed || fileChanged\n\t}\n\n\t// When src targets are in the same package as us, the `source` and `out` paths are the same so the files are\n\t// considered unchanged. We should consider ourselves changed though, as the sources Might indeed have changed.\n\tfor _, bi := range target.AllSources() {\n\t\tif changed {\n\t\t\tbreak\n\t\t}\n\t\tl, ok := bi.Label()\n\t\tif !ok || !target.Label.InSamePackageAs(l) {\n\t\t\tcontinue\n\t\t}\n\t\tif ok && state.Graph.TargetOrDie(l).State() < core.Unchanged {\n\t\t\tchanged = true\n\t\t}\n\t}\n\n\tif target.HasLabel(\"py\") && !target.IsBinary {\n\t\t// Pre-emptively create __init__.py files so the outputs can be loaded dynamically.\n\t\t// It's a bit cheeky to do non-essential language-specific logic but this enables\n\t\t// a lot of relatively normal Python workflows.\n\t\t// Errors are deliberately ignored.\n\t\tif pkg := state.Graph.PackageByLabel(target.Label); pkg == nil || !pkg.HasOutput(\"__init__.py\") {\n\t\t\t// Don't create this if someone else is going to create this in the package.\n\t\t\tcreateInitPy(outDir)\n\t\t}\n\t}\n\tif target.HasLabel(\"go\") {\n\t\t// Create a dummy go.mod file so Go tooling ignores the contents of plz-out.\n\t\tgoModOnce.Do(writeGoMod)\n\t}\n\treturn changed, nil\n}", "func (builder *builder) build(context *cli.Context, p *Project) error {\n\treturn builder.buildFunc(builder, context, p)\n}", "func createMain() error {\n\tpath := \"./\" + Name + \"/main.go\"\n\n\t// Create the main.go file content\n\tmain := fmt.Sprintf(`package main\n\nimport (\n\t\"%s/cmd/server\"\n\t\"%s/pkg/logs\"\n)\n\t\nfunc main() {\n\tlogs.StartLogger()\n\t\t\n\tserver := server.NewServer()\n\n\tserver.StartServer()\n}\n\t`, Repo, Repo)\n\n\t// Create the main.go file\n\treturn util.CreateFile(path, main)\n}", "func Main() {\n\tif err := smake(); err != nil {\n\t\tfmt.Fprintln(os.Stderr, err)\n\t\tos.Exit(1)\n\t}\n}", "func compileAndRun(code []byte) (string, error) {\n\tconst name = \"tmp\"\n\tos.RemoveAll(name)\n\terr := os.Mkdir(name, 0700)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// write src\n\terr = ioutil.WriteFile(name+\"/main.go\", code, 0700)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// compile\n\tcmd := exec.Command(\"sh\", \"-c\", \"go build\")\n\tcmd.Dir = name\n\tout, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"%v: %s\", err, out)\n\t}\n\n\t// execute\n\tcmd = exec.Command(\"sh\", \"-c\", \"./\"+name)\n\tcmd.Dir = name\n\tout, err = cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"%v: %s\", err, out)\n\t}\n\n\tos.RemoveAll(name)\n\treturn string(out), nil\n}", "func Build(args ...string) {\n wd, _ := os.Getwd()\n logger.Log(fmt.Sprintf(\"In %s to build.\", wd))\n if cfg != nil {\n logger.Log(fmt.Sprintf(\"Building...%s\\n\", cfg.App))\n cmd := exec.Command(\"docker\", \"build\", \"-t\", cfg.Container, cfg.BuildFile)\n cmd.Stdout = os.Stdout\n cmd.Stderr = os.Stderr\n cmd.Stdin = os.Stdin\n cmd.Run()\n } else {\n config.LoadConfigs()\n for _, process := range config.List() {\n SetProcess(process)\n SetConfig(config.Process(process))\n Build(args...)\n }\n }\n}", "func Build(platform Platform, artifactToplevelDir string, args []string) error {\n\tartifactDir := path.Join(artifactToplevelDir, platform.Os, platform.Arch)\n\n\tlog.Printf(\"building %s\\n\", artifactDir)\n\n\tif err := os.MkdirAll(artifactDir, 0755); err != nil {\n\t\treturn err\n\t}\n\n\tallPackagesPath := fmt.Sprintf(\".%c...\", os.PathSeparator)\n\n\tcmd := exec.Command(\"go\")\n\tcmd.Args = []string{\"go\", \"build\", \"-o\", artifactDir}\n\tcmd.Args = append(cmd.Args, args...)\n\tcmd.Args = append(cmd.Args, allPackagesPath)\n\tcmd.Env = os.Environ()\n\tcmd.Env = append(cmd.Env, fmt.Sprintf(\"GOOS=%s\", platform.Os))\n\tcmd.Env = append(cmd.Env, fmt.Sprintf(\"GOARCH=%s\", platform.Arch))\n\tcmd.Stderr = os.Stderr\n\tcmd.Stdout = os.Stdout\n\treturn cmd.Run()\n}", "func Build() error {\n\tif strings.Contains(runtime.Version(), \"1.8\") {\n\t\t// Go 1.8 doesn't play along with go test ./... and /vendor.\n\t\t// We could fix that, but that would take time.\n\t\tfmt.Printf(\"Skip Build on %s\\n\", runtime.Version())\n\t\treturn nil\n\t}\n\n\t// TODO: Add lint after fixing errors\n\tmg.Deps(Fmt, Vet, TestRace)\n\treturn nil\n}", "func (app *ClientApplication) Build(context string, makisuArgs []string) error {\n\ttarget, err := prepContext(app.LocalSharedPath, app.WorkerSharedPath, context)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to prepare context: %v\", err)\n\t}\n\tlocalPath := filepath.Join(app.LocalSharedPath, target)\n\tdefer os.RemoveAll(localPath)\n\tworkerPath := filepath.Join(app.WorkerSharedPath, target)\n\n\tstart := time.Now()\n\tfor time.Since(start) < app.WaitDuration {\n\t\tif err = app.client().Build(makisuArgs, workerPath); err == client.ErrWorkerBusy {\n\t\t\ttime.Sleep(250 * time.Millisecond)\n\t\t\tcontinue\n\t\t} else if err != nil {\n\t\t\treturn fmt.Errorf(\"build failed: %v\", err)\n\t\t}\n\t\treturn nil\n\t}\n\treturn err\n}", "func buildSrcPath(fromSrc string) string {\n\treturn filepath.Join(buildSrcDir, filepath.FromSlash(fromSrc))\n}", "func outCodeFiles(config *MainConfig, files map[string]*CodeFiles, keys []string, fileFunc FileResultFunc) (packInfos []PackageInfo, fileLinks []FileLink) {\n\n\tscanPath := config.ScanPath\n\tappendPath := config.OutAppendPath\n\n\tif 0 != len(appendPath) && '/' == appendPath[0] {\n\t\tappendPath = appendPath[1:len(appendPath)]\n\t}\n\n\t// source code ouput path operation\n\t// projectroot/doc/src/[appendpath/main.go]\n\tisLinkRoot := config.CodeLinkRoot\n\toutCodeDir := \"\"\n\tif config.CopyCode {\n\t\toutCodeDir = dirpathSRC(config)\n\t}\n\n\t// markdown file save directory\n\tmdDir := dirpathMarkdownDefault(config)\n\n\t// return result make\n\tpackInfos = make([]PackageInfo, 0, len(files))\n\tfileLinks = make([]FileLink, 0, 0)\n\n\t// sort keys all document file first place\n\tnewKeys := make([]string, 0, len(keys))\n\ttempKeys := make([]string, 0, len(keys))\n\tfor _, key := range keys {\n\t\tcodefiles := files[key]\n\t\tif codefiles.IsAllDocFile() {\n\t\t\tnewKeys = append(newKeys, key)\n\t\t} else {\n\t\t\ttempKeys = append(tempKeys, key)\n\t\t}\n\t}\n\tnewKeys = append(newKeys, tempKeys...)\n\n\t// 1.FOR Directory\n\tfor _, key := range newKeys {\n\t\t// dirPath, codefiles\n\t\tdirPath := key\n\t\tcodefiles := files[key]\n\n\t\tfilesLen := codefiles.FilesLen()\n\t\trelativeDirPath := \"\"\n\n\t\tif 0 == strings.Index(dirPath, scanPath) {\n\t\t\trelativeDirPath = dirPath[len(scanPath):]\n\t\t} else {\n\t\t\tif nil != fileFunc {\n\t\t\t\tfileFunc(dirPath, ResultDebugErr, errors.New(\"map CodeFiles save path error.\"))\n\t\t\t}\n\t\t\tfmt.Println(\"map CodeFiles save path error.\")\n\t\t\tfmt.Println(\"ScanPath:\", scanPath)\n\t\t\tfmt.Println(\"CodeFiles Dirpath:\", dirPath)\n\t\t\tcontinue\n\t\t}\n\n\t\t// TODO 左侧导航中菜单列表的分类名称暂时不使用,使用\"\"空字符串代替,主要考虑到不如何进行展示,并且在提取分配名考虑到别的语言可能不通用\n\t\t/*\n\t\t 以下为GO语言的展示方案,但是其他语言有待商议(目前分类名称以\"\"空字符串代替,这里只是演示想展示的想法不是目前实际的操作)\n\t\t 当目录下没有源文件或文件时以当前目录喂分类名称显示\n\n\t\t index.html\n\t\t gosfdoc\n\t\t github.com/slowfei/gosfdoc\n\t\t github.com/slowfei/gosfdoc/assets\n\t\t lang\n\t\t github.com/slowfei/gosfdoc/lang/golang\n\t\t github.com/slowfei/gosfdoc/lang/java\n\t\t github.com/slowfei/gosfdoc/lang/javascript\n\t\t github.com/slowfei/gosfdoc/lang/objc\n\n\t\t src.html\n\t\t gosfdoc\n\t\t github.com/slowfei/gosfdoc.go\n\t\t github.com/slowfei/config.go\n\t\t github.com/slowfei/parse.go\n\t\t github.com/slowfei/gosfdoc/assets/assets.go\n\t\t github.com/slowfei/gosfdoc/assets/html.go\n\t\t lang\n\t\t github.com/slowfei/gosfdoc/lang/golang/golang.go\n\t\t github.com/slowfei/gosfdoc/lang/java/java.go\n\t\t github.com/slowfei/gosfdoc/lang/javascript/javascript.go\n\t\t github.com/slowfei/gosfdoc/lang/objc/objc.go\n\t\t*/\n\t\tmenuName := \"package\" // 暂时设定空字符串\n\n\t\t// 如果全部是文档文件则视为帮助文档\n\t\tif codefiles.IsAllDocFile() {\n\t\t\tmenuName = \"help document\"\n\t\t}\n\n\t\tpreviews := make([]Preview, 0, 0)\n\t\tblocks := make([]CodeBlock, 0, 0)\n\t\tdocuments := make([]Document, 0, 0)\n\t\tfilesName := make([]string, 0, filesLen)\n\t\tpackStrList := make([]string, 0, filesLen)\n\n\t\t// 2.FOR Files\n\t\tcodefiles.Each(func(code CodeFile) bool {\n\t\t\t// 3. source code check\n\t\t\tif !code.PrivateCode {\n\t\t\t\tswitch code.parser.(type) {\n\t\t\t\tcase *nilDocParser:\n\t\t\t\tdefault:\n\t\t\t\t\tvar outErr error = nil\n\t\t\t\t\tfileName := code.FileCont.FileInfo().Name()\n\t\t\t\t\tjoinName := filepath.Join(appendPath, relativeDirPath, fileName)\n\n\t\t\t\t\tif 0 != len(outCodeDir) {\n\t\t\t\t\t\toutPath := filepath.Join(outCodeDir, joinName)\n\t\t\t\t\t\toutErr = code.FileCont.WriteFilepath(outPath)\n\t\t\t\t\t\tif nil == outErr && nil != fileFunc {\n\t\t\t\t\t\t\tfileFunc(outPath, ResultFileSuccess, nil)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\n\t\t\t\t\tif nil == outErr {\n\t\t\t\t\t\tif 0 != len(outCodeDir) || isLinkRoot {\n\t\t\t\t\t\t\tfilesName = append(filesName, fileName)\n\n\t\t\t\t\t\t\tfileLink := FileLink{}\n\t\t\t\t\t\t\tfileLink.Link = joinName\n\t\t\t\t\t\t\tfileLink.Filename = joinName\n\t\t\t\t\t\t\tfileLink.menuName = menuName\n\n\t\t\t\t\t\t\tfileLinks = append(fileLinks, fileLink)\n\t\t\t\t\t\t}\n\t\t\t\t\t} else if nil != fileFunc {\n\t\t\t\t\t\tfileFunc(code.FileCont.path, ResultFileOutFail, outErr)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// 4. parse Preview and CodeBlock and Document\n\t\t\tif !code.PrivateDoc {\n\t\t\t\tps := code.parser.ParsePreview(code.FileCont)\n\t\t\t\tbs := code.parser.ParseCodeblock(code.FileCont)\n\n\t\t\t\tif 0 != len(ps) {\n\t\t\t\t\tpreviews = append(previews, ps...)\n\t\t\t\t}\n\n\t\t\t\tif 0 != len(bs) {\n\t\t\t\t\tblocks = append(blocks, bs...)\n\t\t\t\t}\n\n\t\t\t\tif 0 != len(code.docs) {\n\t\t\t\t\tdocuments = append(documents, code.docs...)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\t// 5. parse package info\n\t\t\tpackInfo := code.parser.ParsePackageInfo(code.FileCont)\n\t\t\tpackInfo = strings.Trim(packInfo, \"\\n \")\n\t\t\tif 0 != len(packInfo) {\n\t\t\t\tpackStrList = append(packStrList, packInfo)\n\t\t\t}\n\n\t\t\treturn true\n\t\t})\n\n\t\t//\n\t\tsort.Sort(SortSet{previews: previews})\n\t\tsort.Sort(SortSet{codeBlocks: blocks})\n\t\tsort.Sort(SortSet{documents: documents})\n\n\t\t// markdown file name is directory base name + suffix\n\t\tmdFileName := filepath.Base(dirPath) + FILE_SUFFIX_MARKDOWN\n\n\t\t// handle source code link path\n\t\tbrowseSrcJoinPath := config.GithubLink(path.Join(relativeDirPath, mdFileName), false)\n\t\t// fmt.Println(\"browseSrcJoinPath: \", browseSrcJoinPath)\n\n\t\t// 5.output markdown\n\t\tmdBytes := ParseMarkdown(documents, previews, blocks, filesName, config.currentVersion, browseSrcJoinPath)\n\t\tif 0 != len(mdBytes) {\n\t\t\tmdOutPath := filepath.Join(mdDir, appendPath, relativeDirPath, mdFileName)\n\n\t\t\terr := SFFileManager.WirteFilepath(mdOutPath, mdBytes)\n\t\t\tresult := ResultFileSuccess\n\n\t\t\tif nil != err {\n\t\t\t\tresult = ResultFileOutFail\n\t\t\t} else {\n\t\t\t\tinfo := PackageInfo{}\n\n\t\t\t\tinfo.Name = path.Join(appendPath, relativeDirPath)\n\t\t\t\tinfo.Link = path.Join(appendPath, relativeDirPath, mdFileName)\n\n\t\t\t\tjoinStr := strings.Join(packStrList, \";\")\n\t\t\t\tnewStr := strings.Replace(joinStr, \"\\n\", \", \", -1)\n\t\t\t\tinfo.Desc = newStr\n\t\t\t\tinfo.menuName = menuName\n\n\t\t\t\tpackInfos = append(packInfos, info)\n\t\t\t}\n\n\t\t\tif nil != fileFunc {\n\t\t\t\tfileFunc(mdOutPath, result, err)\n\t\t\t}\n\t\t}\n\n\t} // end for dirPath, codefiles := range files\n\n\treturn\n}", "func PkgSource(source string) {\n\tc := bindata.NewConfig()\n\tinput := parseInput(source)\n\tinputs := []bindata.InputConfig{input}\n\tc.Input = inputs\n\tc.Output = \"target/elsa-package/asset.go\"\n\terr := bindata.Translate(c)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"bindata: %v\\n\", err)\n\t\tos.Exit(1)\n\t}\n\tentry := GeneratePkgSource(source)\n\tf, _ := os.Create(\"target/elsa-package/main.go\")\n\n\tdefer f.Close()\n\tf.WriteString(entry)\n\tExecBuild(\"target/elsa-package\")\n}", "func Build() error {\n\treturn devtools.Build(devtools.DefaultBuildArgs())\n}", "func main() {\n\tcl := flag.Bool(\"clean\", false, \"don't build, just clean the generated pages\")\n\t//run := flag.Bool(\"run\", false, \"run the generated executable after build\")\n\tflag.Parse()\n\tif *cl {\n\t\terr := clean()\n\t\tif err != nil {\n\t\t\tprintln(err.Error())\n\t\t}\n\t\treturn\n\t}\n\tsettings, err := util.LoadSettings() //inits the settings and generates the .go source files\n\tif err != nil {\n\t\tprintln(err.Error())\n\t\tos.Exit(1)\n\t\treturn\n\t}\n\tutil.Config = settings.Data //stores settings to accessible variable\n\tprintln(\"generated\", len(settings.Data[\"pages\"]), \"gopages\")\n\tprintln()\n\terr = util.AddHandlers(settings.Data[\"pages\"]) //add all handlers\n\tif err != nil {\n\t\tprintln(err.Error())\n\t\treturn\n\t}\n\tif len(os.Args) > 1 && os.Args[1] == \"get\" {\n\t\tbuild(\"\")\n\t} else if len(os.Args) > 1 {\n\t\tprintln(\"unrecognized \", \"'\"+os.Args[1]+\"'\")\n\t\tprintln(\" gopages get - build project with go get after generating pages\")\n\t} else {\n\t\tbuild(\"pages\")\n\t\tprintln(\"run \\\"gopages get\\\" to build project with go get after generating pages\")\n\t}\n}", "func GobuildArgs(args []string) error {\n\tif len(args) <= 0 {\n\t\treturn nil\n\t}\n\n\tdefer func() {\n\t\tif err := recover(); err != nil {\n\t\t\tlog.Printf(\"gobuild.Error: %+s\", err)\n\t\t}\n\t}()\n\n\tcmdline := []string{\"go\", \"build\"}\n\n\t// target := filepath.Join(dir, name)\n\tcmdline = append(cmdline, args...)\n\n\t//setup the executor and use a shard buffer\n\tcmd := exec.Command(\"go\", cmdline[1:]...)\n\tbuf := bytes.NewBuffer([]byte{})\n\n\tmsg, err := cmd.CombinedOutput()\n\n\tif !cmd.ProcessState.Success() {\n\t\treturn fmt.Errorf(\"go.build failed: %s: %s -> Msg: %s\", buf.String(), err.Error(), msg)\n\t}\n\n\treturn nil\n}", "func crateSource(ctx android.SingletonContext, rModule *Module, comp *baseCompiler) (string, bool) {\n\t// Basic libraries, executables and tests.\n\tsrcs := comp.Properties.Srcs\n\tif len(srcs) != 0 {\n\t\treturn path.Join(ctx.ModuleDir(rModule), srcs[0]), true\n\t}\n\t// SourceProvider libraries.\n\tif rModule.sourceProvider != nil {\n\t\treturn sourceProviderSource(ctx, rModule)\n\t}\n\treturn \"\", false\n}", "func Build(packagePath string, args ...string) (compiledPath string, err error) {\n\treturn doBuild(build.Default.GOPATH, packagePath, nil, args...)\n}", "func (self *app) build() {\n\tvar done = make(chan bool)\n\tcmd.Loading(done)\n\n\t// * try build the application into rex-bin(.exe)\n\tcommand := exec.Command(\"go\", \"build\", \"-o\", self.binary)\n\tcommand.Dir = self.dir\n\tif e := command.Run(); e != nil {\n\t\tlog.Fatalf(\"Failed to compile the application: %v\", e)\n\t}\n\n\tdone <- true\n}", "func main() {\n\n\tvar version string\n\n\t// Derive the commit message from -X main.commit=$YOUR_VALUE_HERE\n\t// if the build does not have the commit variable set externally,\n\t// fall back to unsupported custom build\n\tif commit != \"\" {\n\t\tversion = commit\n\t} else {\n\t\tversion = \"unsupported custom build\"\n\t}\n\n\t// let the user know that we are running within a docker container\n\tcheckRunningWithinDocker()\n\n\t// build the Command Line interface\n\t// https://github.com/urfave/cli/blob/master/docs/v2/manual.md\n\n\t// basic information\n\tapp := &cli.App{\n\t\tName: \"appimagetool\",\n\t\tAuthors: \t\t\t\t[]*cli.Author{{Name: \"AppImage Project\"}},\n\t\tVersion: version,\n\t\tUsage: \t\t\"An automatic tool to create AppImages\",\n\t\tEnableBashCompletion: false,\n\t\tHideHelp: false,\n\t\tHideVersion: false,\n\t\tCompiled: time.Time{},\n\t\tCopyright: \"MIT License\",\n\t\tAction: \t\t\t\tbootstrapAppImageBuild,\n\n\t}\n\n\t// define subcommands, like 'deploy', 'validate', ...\n\tapp.Commands = []*cli.Command{\n\t\t{\n\t\t\tName: \"deploy\",\n\t\t\tUsage: \"Turns PREFIX directory into AppDir by deploying dependencies and AppRun file\",\n\t\t\tAction: bootstrapAppImageDeploy,\n\t\t},\n\t\t{\n\t\t\tName: \"validate\",\n\t\t\tUsage: \"Calculate the sha256 digest and check whether the signature is valid\",\n\t\t\tAction: bootstrapValidateAppImage,\n\t\t},\n\t\t{\n\t\t\tName: \"setupsigning\",\n\t\t\tUsage: \"Prepare a git repository that is used with Travis CI for signing AppImages\",\n\t\t\tAction: bootstrapSetupSigning,\n\t\t},\n\t\t{\n\t\t\tName: \t\"sections\",\n\t\t\tUsage: \t\"\",\n\t\t\tAction:\tbootstrapAppImageSections,\n\t\t},\n\t}\n\n\t// define flags, such as --libapprun_hooks, --standalone here ...\n\tapp.Flags = []cli.Flag{\n\t\t&cli.BoolFlag{\n\t\t\tName: \"libapprun_hooks\",\n\t\t\tAliases: []string{\"l\"},\n\t\t\tUsage: \"Use libapprun_hooks\",\n\t\t},\n\t\t&cli.BoolFlag{\n\t\t\tName: \"overwrite\",\n\t\t\tAliases: []string{\"o\"},\n\t\t\tUsage: \"Overwrite existing files\",\n\t\t},\n\t\t&cli.BoolFlag{\n\t\t\tName: \"standalone\",\n\t\t\tAliases: []string{\"s\"},\n\t\t\tUsage: \"Make standalone self-contained bundle\",\n\t\t},\n\t}\n\n\t// TODO: move travis based Sections to travis.go in future\n\tif os.Getenv(\"TRAVIS_TEST_RESULT\") == \"1\" {\n\t\tlog.Fatal(\"$TRAVIS_TEST_RESULT is 1, exiting...\")\n\t}\n\n\terrRuntime := app.Run(os.Args)\n\tif errRuntime != nil {\n\t\tlog.Fatal(errRuntime)\n\t}\n\n}", "func createProgram(packages ...string) (*loader.Program, error) {\n\tvar conf loader.Config\n\n\tfor _, name := range packages {\n\t\tconf.CreateFromFilenames(name, getFileNames(name)...)\n\t}\n\treturn conf.Load()\n}", "func main() {\n\tflag.Var(&filenames, \"filename\", \"specified filename those you want to generate, if no filename be set, will parse all files under ($dir)\")\n\tflag.Parse()\n\n\texportDir, _ := filepath.Abs(*generateDir)\n\t*dir, _ = filepath.Abs(*dir)\n\tif *debug {\n\t\tlog.Println(\"fromDir:\", *dir)\n\t\tlog.Println(\"generateFilename:\", *generateFilename)\n\t\tlog.Println(\"exportDir:\", exportDir)\n\t}\n\n\t// set custom funcs\n\t// tools.SetCustomGenTagFunc(CustomGenerateTagFunc)\n\t// tools.SetCustomParseTagFunc(CustomParseTagFunc)\n\n\tif len(filenames) == 0 {\n\t\tfiles, _ := ioutil.ReadDir(*dir)\n\t\tfor _, file := range files {\n\t\t\tif file.IsDir() || !strings.HasSuffix(file.Name(), \".go\") {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tfilenames = append(filenames, file.Name())\n\t\t}\n\t}\n\n\tcfg := &tools.UsageCfg{\n\t\tExportDir: exportDir,\n\t\tExportFilename: *generateFilename,\n\t\tExportPkgName: *generatePkgName,\n\t\tExportStructSuffix: *generateStructSuffix,\n\t\tModelImportPath: *modelImportPath,\n\t\tStructSuffix: *modelStructSuffix,\n\t\tDebug: *debug,\n\t\tFilenames: filenames,\n\t\tDir: *dir,\n\t}\n\n\tif *debug {\n\t\tlog.Println(\"following filenames will be parsed\", filenames)\n\t}\n\n\tif err := tools.ParseAndGenerate(cfg); err != nil {\n\t\tpanic(err)\n\t}\n\n\tprintln(\"done!\")\n}", "func (b *Builder) Build() {\n\tlog.Printf(\n\t\t\"Building: %s %s %s xcn:%t xcnF:%t\",\n\t\tb.Cmd, b.OS, b.Arch, b.XCN, xcnFlag)\n\tvar env []string // cmd env\n\tvar args []string // cmd args\n\n\tenv = append(env, fmt.Sprintf(\"GOOS=%s\", b.OS))\n\tenv = append(env, fmt.Sprintf(\"GOARCH=%s\", b.Arch))\n\tenv = append(env, \"GO15VENDOREXPERIMENT=1\")\n\n\tif xcnFlag && b.XCN {\n\t\tosarch := fmt.Sprintf(\"%s-%s\", b.OS, b.Arch)\n\t\tosArchCCs := map[string]string{\n\t\t\t\"darwin-386\": \"CC=o32-clang\",\n\t\t\t\"darwin-amd64\": \"CC=o64-clang\",\n\t\t\t\"linux-386\": \"\",\n\t\t\t\"linux-amd64\": \"\",\n\t\t\t\"windows-386\": \"CC=i686-w64-mingw32-gcc\",\n\t\t\t\"windows-amd64\": \"CC=x86_64-w64-mingw32-gcc\",\n\t\t}\n\t\tccenv, ok := osArchCCs[osarch]\n\t\tif !ok {\n\t\t\tlog.Printf(\"%s is not supported\", osarch)\n\t\t\tpanic(\"\")\n\t\t}\n\t\tif ccenv != \"\" {\n\t\t\tenv = append(env, ccenv)\n\t\t}\n\t\tenv = append(env, \"CGO_ENABLED=1\")\n\t}\n\n\targs = append(args, \"go\", \"build\")\n\tvar ldflags []string\n\tif optimizeFlag || (xcnFlag && b.OS == \"darwin\") {\n\t\t// workaround because non existent dsymutil in current osxcross\n\t\t// https://github.com/golang/go/issues/11994\n\t\tldflags = append(ldflags, \"-s\")\n\t}\n\n\tif b.GUI && b.OS == \"windows\" {\n\t\tldflags = append(ldflags, \"-H=windowsgui\")\n\t}\n\tfor _, v := range b.Vars {\n\t\tldflags = append(ldflags, fmt.Sprintf(\"-X %s\", v))\n\t}\n\n\tldflagsStr := strings.Join(ldflags, \" \")\n\tif ldflagsStr != \"\" {\n\t\targs = append(args, \"-ldflags\", ldflagsStr)\n\t}\n\n\tif verboseFlag {\n\t\t// args = append(args, \"-v\")\n\t\t// args = append(args, \"-a\")\n\t\t// args = append(args, \"-x\")\n\t}\n\n\targs = append(args, \"-o\", b.Filename())\n\targs = append(args, fmt.Sprintf(\"cmd/%s/%s.go\", b.Cmd, b.Cmd))\n\n\tvar c *exec.Cmd\n\tif !dockerFlag {\n\t\tmergedEnv := os.Environ()\n\t\tmergedEnv = append(mergedEnv, env...)\n\t\tc = newCmd(args[0], args[1:]...)\n\t\tc.Env = mergedEnv\n\t} else {\n\t\tvar dargs []string\n\t\tdargs = append(dargs, \"docker\", \"run\", \"--rm\")\n\t\tu, err := user.Current()\n\t\tfailErr(err, \"Could not get user id\")\n\t\tdargs = append(dargs, \"-u\", u.Uid)\n\n\t\twd, err := os.Getwd()\n\t\tfailErr(err, \"Could not read working directory\")\n\t\tdargs = append(dargs, fmt.Sprintf(\"-v=%s:/go/src/github.com/alkasir/alkasir/\", wd))\n\n\t\tfor _, v := range env {\n\t\t\tdargs = append(dargs, \"-e\", v)\n\t\t}\n\t\tdargs = append(dargs, \"alkasir-docker-xcn-builder\")\n\t\targs = append(dargs, args...)\n\t\tc = newCmd(args[0], args[1:]...)\n\t}\n\tif verboseFlag {\n\t\tlog.Printf(\"*** args:%+v env:%+v\", args, env)\n\t}\n\terr := c.Run()\n\tfailErr(err, \"error executing builder.Run\")\n\tb.Built = true\n}", "func Build(ctx context.Context, cli client.CommonAPIClient, options Options) (io.Reader, error) {\n\tlogger := ctxlog.FromContext(ctx)\n\n\tlogger.Debug(\"gobuild\",\n\t\tzap.Any(\"options\", options),\n\t)\n\n\tif err := image.PullIfNotFound(ctx, cli, DockerImageName); err != nil {\n\t\treturn nil, err\n\t}\n\n\tcontainerConfig := &container.Config{\n\t\tImage: DockerImageName,\n\t\tCmd: options.Packages,\n\t\tEnv: []string{\"GO_LDFLAGS=\" + options.LDFlags},\n\t\tOpenStdin: true,\n\t\tStdinOnce: true,\n\t\tAttachStdin: true,\n\t\tAttachStdout: true,\n\t\tAttachStderr: true,\n\t\tLabels: map[string]string{\n\t\t\t\"ua.owned\": \"true\",\n\t\t},\n\t}\n\thostConfig := &container.HostConfig{\n\t\tAutoRemove: true,\n\t}\n\n\tc, err := cli.ContainerCreate(ctx, containerConfig, hostConfig, nil, \"\")\n\tif err != nil {\n\t\tlogger.Error(\"error creating gobuild container\",\n\t\t\tzap.Error(err),\n\t\t)\n\t\treturn nil, err\n\t}\n\tcontainerID := c.ID\n\n\tctx, logger = ctxlog.FromContextWith(ctx,\n\t\tzap.String(\"container_id\", containerID),\n\t)\n\n\tattachOptions := types.ContainerAttachOptions{\n\t\tStream: true,\n\t\tStdin: true,\n\t\tStdout: true,\n\t\tStderr: true,\n\t}\n\n\thj, err := cli.ContainerAttach(ctx, containerID, attachOptions)\n\tif err != nil {\n\t\tlogger.Error(\"error attaching to gobuild container\",\n\t\t\tzap.Error(err),\n\t\t)\n\n\t\ttryContainerRemove(ctx, cli, containerID)\n\n\t\treturn nil, err\n\t}\n\tdefer hj.Close()\n\n\tresultC, errC := cli.ContainerWait(ctx, containerID, container.WaitConditionRemoved)\n\n\tif err := cli.ContainerStart(ctx, containerID, types.ContainerStartOptions{}); err != nil {\n\t\tlogger.Error(\"error starting gobuild container\",\n\t\t\tzap.Error(err),\n\t\t)\n\n\t\ttryContainerRemove(ctx, cli, containerID)\n\n\t\treturn nil, err\n\t}\n\n\tgo func() {\n\t\tdefer func() {\n\t\t\tif err := hj.CloseWrite(); err != nil {\n\t\t\t\tlogger.Warn(\"error closing gobuild stdin\",\n\t\t\t\t\tzap.Error(err),\n\t\t\t\t)\n\t\t\t}\n\t\t}()\n\n\t\tsource, err := archive.Tar(options.SrcPath, archive.Uncompressed)\n\t\tif err != nil {\n\t\t\tlogger.Error(\"error tarring go source\",\n\t\t\t\tzap.Error(err),\n\t\t\t)\n\t\t\treturn\n\t\t}\n\n\t\tdefer func() {\n\t\t\tif err := source.Close(); err != nil {\n\t\t\t\tlogger.Warn(\"error closing gobuild source\",\n\t\t\t\t\tzap.Error(err),\n\t\t\t\t)\n\t\t\t}\n\t\t}()\n\n\t\tif _, err := io.Copy(hj.Conn, source); err != nil {\n\t\t\tlogger.Warn(\"error copying gobuild stdin\",\n\t\t\t\tzap.Error(err),\n\t\t\t)\n\t\t}\n\t}() // Exits when hj.Conn closes, or the source code has been sent.\n\n\tstderr := &bytes.Buffer{}\n\tstdout := &bytes.Buffer{}\n\n\tgo func() {\n\t\t// This is probably the most evil thing I've ever seen inside Docker.\n\t\t// For some stupid reason, stdout and stderr are multiplexed, but only\n\t\t// for this specific type of attach. Hours of debugging later, I read\n\t\t// enough of the Docker CLI source code (hijack.go) and decided to try\n\t\t// it after seeing such a weird function signature.\n\t\t//\n\t\t// Nothing could be further than a \"standard copy\" than this. !@#$%!\n\t\tif _, err := stdcopy.StdCopy(stdout, stderr, hj.Reader); err != nil {\n\t\t\tlogger.Error(\"stdcopy error\",\n\t\t\t\tzap.Error(err),\n\t\t\t)\n\t\t}\n\t}() // Exits when hj.Reader returns EOF.\n\n\tselect {\n\tcase result := <-resultC:\n\t\tif result.Error != nil {\n\t\t\treturn nil, fmt.Errorf(\"%s\", result.Error.Message)\n\t\t}\n\n\t\tif result.StatusCode != 0 {\n\t\t\treturn nil, fmt.Errorf(\"gobuild: status code %d\\n%s\", result.StatusCode, stderr.String())\n\t\t}\n\n\tcase err := <-errC:\n\t\tlogger.Error(\"gobuild errC\",\n\t\t\tzap.Error(err),\n\t\t)\n\t\treturn nil, err\n\t}\n\n\treturn stdout, nil\n}", "func (*Basic) Build(_ Generator, _ *MethodContext, sourceID *xtype.JenID, source, target *xtype.Type) ([]jen.Code, *xtype.JenID, *Error) {\n\tif target.Named || (!target.Named && source.Named) {\n\t\treturn nil, xtype.OtherID(target.TypeAsJen().Call(sourceID.Code)), nil\n\t}\n\treturn nil, sourceID, nil\n}", "func build(\n\tcommit string,\n\tversion string,\n\tosys string) (dir string, err error) {\n\tlog.Printf(\"Building version %s from %s.\", version, commit)\n\n\t// Create a directory to become GOCACHE below.\n\tvar gocache string\n\tgocache, err = os.MkdirTemp(\"\", \"package_gcsfuse_gocache\")\n\tif err != nil {\n\t\terr = fmt.Errorf(\"TempDir: %w\", err)\n\t\treturn\n\t}\n\tdefer os.RemoveAll(gocache)\n\n\t// Create a directory to hold our outputs. Kill it if we later return in\n\t// error.\n\tdir, err = os.MkdirTemp(\"\", \"package_gcsfuse_build\")\n\tif err != nil {\n\t\terr = fmt.Errorf(\"TempDir: %w\", err)\n\t\treturn\n\t}\n\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tos.RemoveAll(dir)\n\t\t}\n\t}()\n\n\t// Create another directory into which we will clone the git repo bloe.\n\tgitDir, err := os.MkdirTemp(\"\", \"package_gcsfuse_git\")\n\tif err != nil {\n\t\terr = fmt.Errorf(\"TempDir: %w\", err)\n\t\treturn\n\t}\n\n\tdefer os.RemoveAll(gitDir)\n\n\t// Clone the git repo, checking out the correct tag.\n\t{\n\t\tlog.Printf(\"Cloning into %s\", gitDir)\n\n\t\tcmd := exec.Command(\n\t\t\t\"git\",\n\t\t\t\"clone\",\n\t\t\t\"-b\", commit,\n\t\t\t\"https://github.com/GoogleCloudPlatform/gcsfuse.git\",\n\t\t\tgitDir)\n\n\t\tvar output []byte\n\t\toutput, err = cmd.CombinedOutput()\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"Cloning: %w\\nOutput:\\n%s\", err, output)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Build build_gcsfuse.\n\tbuildTool := path.Join(gitDir, \"build_gcsfuse\")\n\t{\n\t\tlog.Printf(\"Building build_gcsfuse...\")\n\n\t\tcmd := exec.Command(\n\t\t\t\"go\",\n\t\t\t\"build\",\n\t\t\t\"-o\", buildTool,\n\t\t)\n\n\t\tcmd.Dir = path.Join(gitDir, \"tools/build_gcsfuse\")\n\t\tcmd.Env = []string{\n\t\t\t\"GO15VENDOREXPERIMENT=1\",\n\t\t\tfmt.Sprintf(\"GOROOT=%s\", runtime.GOROOT()),\n\t\t\tfmt.Sprintf(\"GOCACHE=%s\", gocache),\n\t\t\t\"GOPATH=/does/not/exist\",\n\t\t}\n\n\t\tvar output []byte\n\t\toutput, err = cmd.CombinedOutput()\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"Building build_gcsfuse: %w\\nOutput:\\n%s\", err, output)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Run build_gcsfuse.\n\t{\n\t\tlog.Printf(\"Running build_gcsfuse...\")\n\n\t\tcmd := exec.Command(\n\t\t\tbuildTool,\n\t\t\tgitDir,\n\t\t\tdir,\n\t\t\tversion)\n\n\t\tvar output []byte\n\t\toutput, err = cmd.CombinedOutput()\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"go run build_gcsfuse: %w\\nOutput:\\n%s\", err, output)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// build_gcsfuse writes files like:\n\t//\n\t// bin/gcsfuse\n\t// sbin/mount.gcsfuse\n\t//\n\t// Which is what we want for e.g. a homebrew cellar. But for a Linux package,\n\t// we want the first to live in /usr/bin.\n\terr = os.MkdirAll(path.Join(dir, \"usr\"), 0755)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"MkdirAll: %w\", err)\n\t\treturn\n\t}\n\n\terr = os.Rename(path.Join(dir, \"bin\"), path.Join(dir, \"usr/bin\"))\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Rename: %w\", err)\n\t\treturn\n\t}\n\n\treturn\n}", "func GenerateBinaries(c Calls, provider fsProviderFn) map[string]string {\n\t// Load all binaries\n\tbinaries := make(map[string]string)\n\tfor project, config := range c {\n\t\tbinaries[project] = loadBinary(provider, *config)\n\t}\n\treturn binaries\n}", "func Build() error {\n\n\t// ldf, err := flags()\n\t// if err != nil {\n\t// \treturn err\n\t// }\n\n\tlog.Print(\"running go build\")\n\t// use -tags make so we can have different behavior for when we know we've built with mage.\n\t// return sh.Run(\"go\", \"build\", \"-tags\", \"make\", \"--ldflags=\"+ldf, \"gnorm.org/gnorm\")\n\treturn sh.RunWith(flagEnv(), \"go\", \"build\", \"-o\", \"build/unichem2index\", \"-ldflags\", ldflags, packageName)\n}", "func Run(bp *build.Package, pkgdir, tmpdir string, useabs bool) (files, displayFiles []string, err error) {\n\tcgoCPPFLAGS, _, _, _ := cflags(bp, true)\n\t_, cgoexeCFLAGS, _, _ := cflags(bp, false)\n\n\tif len(bp.CgoPkgConfig) > 0 {\n\t\tpcCFLAGS, err := pkgConfigFlags(bp)\n\t\tif err != nil {\n\t\t\treturn nil, nil, err\n\t\t}\n\t\tcgoCPPFLAGS = append(cgoCPPFLAGS, pcCFLAGS...)\n\t}\n\n\t// Allows including _cgo_export.h from .[ch] files in the package.\n\tcgoCPPFLAGS = append(cgoCPPFLAGS, \"-I\", tmpdir)\n\n\t// _cgo_gotypes.go (displayed \"C\") contains the type definitions.\n\tfiles = append(files, filepath.Join(tmpdir, \"_cgo_gotypes.go\"))\n\tdisplayFiles = append(displayFiles, \"C\")\n\tfor _, fn := range bp.CgoFiles {\n\t\t// \"foo.cgo1.go\" (displayed \"foo.go\") is the processed Go source.\n\t\tf := cgoRe.ReplaceAllString(fn[:len(fn)-len(\"go\")], \"_\")\n\t\tfiles = append(files, filepath.Join(tmpdir, f+\"cgo1.go\"))\n\t\tdisplayFiles = append(displayFiles, fn)\n\t}\n\n\tvar cgoflags []string\n\tif bp.Goroot && bp.ImportPath == \"runtime/cgo\" {\n\t\tcgoflags = append(cgoflags, \"-import_runtime_cgo=false\")\n\t}\n\tif bp.Goroot && bp.ImportPath == \"runtime/race\" || bp.ImportPath == \"runtime/cgo\" {\n\t\tcgoflags = append(cgoflags, \"-import_syscall=false\")\n\t}\n\n\tvar cgoFiles []string = bp.CgoFiles\n\tif useabs {\n\t\tcgoFiles = make([]string, len(bp.CgoFiles))\n\t\tfor i := range cgoFiles {\n\t\t\tcgoFiles[i] = filepath.Join(pkgdir, bp.CgoFiles[i])\n\t\t}\n\t}\n\n\targs := stringList(\n\t\t\"go\", \"tool\", \"cgo\", \"-objdir\", tmpdir, cgoflags, \"--\",\n\t\tcgoCPPFLAGS, cgoexeCFLAGS, cgoFiles,\n\t)\n\tif false {\n\t\tlog.Printf(\"Running cgo for package %q: %s (dir=%s)\", bp.ImportPath, args, pkgdir)\n\t}\n\tcmd := exec.Command(args[0], args[1:]...)\n\tcmd.Dir = pkgdir\n\tcmd.Env = append(os.Environ(), \"PWD=\"+pkgdir)\n\tcmd.Stdout = os.Stderr\n\tcmd.Stderr = os.Stderr\n\tif err := cmd.Run(); err != nil {\n\t\treturn nil, nil, fmt.Errorf(\"cgo failed: %s: %s\", args, err)\n\t}\n\n\treturn files, displayFiles, nil\n}", "func genSource(dir, filename, templateSource string, args map[string]interface{}) {\n\tsourceCode := revel.ExecuteTemplate(\n\t\ttemplate.Must(template.New(\"\").Parse(templateSource)),\n\t\targs)\n\n\t// Create a fresh dir.\n\t// tmpPath := path.Join(revel.AppPath, dir)\n\n\t// Create the file\n\tfile, err := os.Create(path.Join(dir, filename))\n\tdefer file.Close()\n\tif err != nil {\n\t\trevel.ERROR.Fatalf(\"Failed to create file: %v\", err)\n\t}\n\t_, err = file.WriteString(sourceCode)\n\tif err != nil {\n\t\trevel.ERROR.Fatalf(\"Failed to write to file: %v\", err)\n\t}\n}", "func buildStart(c *cli.Context) (err error) {\n\trepo := c.Args().First()\n\towner, name, err := parseRepo(repo)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnumber, err := strconv.Atoi(c.Args().Get(1))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tclient, err := newClient(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar build *drone.Build\n\tif c.Bool(\"fork\") {\n\t\tbuild, err = client.BuildStart(owner, name, number)\n\t} else {\n\t\tbuild, err = client.BuildFork(owner, name, number)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"Starting build %s/%s#%d\\n\", owner, name, build.Number)\n\treturn nil\n}", "func configureJsFramework(sourceDir string, config *ScannerConfig) (*SourceInfo, error) {\n\t// first ensure that there is a package.json\n\tif !checksPass(sourceDir, fileExists(\"package.json\")) {\n\t\treturn nil, nil\n\t}\n\n\t// ensure package.json has a main, module, or start script\n\tdata, err := os.ReadFile(\"package.json\")\n\n\tif err != nil {\n\t\treturn nil, nil\n\t} else {\n\t\terr = json.Unmarshal(data, &packageJson)\n\t\tif err != nil {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\t// see if package.json has a \"main\"\n\t\tmain, _ := packageJson[\"main\"].(string)\n\n\t\t// check for tyep=\"module\" and a module being defined\n\t\tptype, ok := packageJson[\"type\"].(string)\n\t\tif ok && ptype == \"module\" {\n\t\t\tmodule, ok := packageJson[\"type\"].(string)\n\t\t\tif ok {\n\t\t\t\tmain = module\n\t\t\t}\n\t\t}\n\n\t\t// check for a start script\n\t\tscripts, ok := packageJson[\"scripts\"].(map[string]interface{})\n\n\t\tif ok && scripts[\"start\"] != nil {\n\t\t\tstart, ok := scripts[\"start\"].(string)\n\t\t\tif ok {\n\t\t\t\tmain = start\n\t\t\t}\n\t\t}\n\n\t\t// bail if no entrypoint can be found\n\t\tif main == \"\" {\n\t\t\treturn nil, nil\n\t\t}\n\t}\n\n\tsrcInfo := &SourceInfo{\n\t\tFamily: \"NodeJS\",\n\t\tSkipDeploy: true,\n\t\tCallback: JsFrameworkCallback,\n\t}\n\n\t_, err = os.Stat(\"bun.lockb\")\n\tif errors.Is(err, fs.ErrNotExist) {\n\t\t// ensure node is in $PATH\n\t\tnode, err := exec.LookPath(\"node\")\n\t\tif err != nil && !errors.Is(err, exec.ErrDot) {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\t// resolve to absolute path, see: https://tip.golang.org/doc/go1.19#os-exec-path\n\t\tnode, err = filepath.Abs(node)\n\t\tif err != nil {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\t// ensure node version is at least 16.0.0\n\t\tout, err := exec.Command(node, \"-v\").Output()\n\t\tif err != nil {\n\t\t\treturn nil, nil\n\t\t} else {\n\t\t\tminVersion, err := semver.Make(\"16.0.0\")\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\tnodeVersionString := strings.TrimSpace(string(out))\n\t\t\tif nodeVersionString[:1] == \"v\" {\n\t\t\t\tnodeVersionString = nodeVersionString[1:]\n\t\t\t}\n\n\t\t\tnodeVersion, err := semver.Make(nodeVersionString)\n\n\t\t\tif err != nil || nodeVersion.LT(minVersion) {\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t}\n\t} else {\n\t\t// ensure bun is in $PATH\n\t\tbun, err := exec.LookPath(\"bun\")\n\t\tif err != nil && !errors.Is(err, exec.ErrDot) {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\t// resolve to absolute path, see: https://tip.golang.org/doc/go1.19#os-exec-path\n\t\tbun, err = filepath.Abs(bun)\n\t\tif err != nil {\n\t\t\treturn nil, nil\n\t\t}\n\n\t\t// ensure bun version is at least 0.5.3, as that's when docker images started\n\t\t// getting published: https://hub.docker.com/r/oven/bun/tags\n\t\tout, err := exec.Command(bun, \"-v\").Output()\n\t\tif err != nil {\n\t\t\treturn nil, nil\n\t\t} else {\n\t\t\tminVersion, err := semver.Make(\"0.5.3\")\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\n\t\t\tbunVersionString := strings.TrimSpace(string(out))\n\t\t\tbunVersion, err := semver.Make(bunVersionString)\n\n\t\t\tif err != nil || bunVersion.LT(minVersion) {\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t}\n\n\t\t// set family\n\t\tsrcInfo.Family = \"Bun\"\n\t}\n\n\t// extract deps\n\tdeps, ok := packageJson[\"dependencies\"].(map[string]interface{})\n\tif !ok || deps == nil {\n\t\tdeps = make(map[string]interface{})\n\t}\n\n\t// don't prompt for redis or postgres unless they are used\n\tif deps[\"pg\"] == nil && deps[\"redis\"] == nil {\n\t\tsrcInfo.SkipDatabase = true\n\t}\n\n\t// While redundant and requires dual matenance, it has been a point of\n\t// confusion for many when the framework detected is listed as \"NodeJS\"\n\t// See flyapps/dockerfile-node for the actual framework detction.\n\t// Also change PlatformMap in core.go if this list ever changes.\n\tif deps[\"@adonisjs/core\"] != nil {\n\t\tsrcInfo.Family = \"AdonisJS\"\n\t} else if deps[\"gatsby\"] != nil {\n\t\tsrcInfo.Family = \"Gatsby\"\n\t} else if deps[\"@nestjs/core\"] != nil {\n\t\tsrcInfo.Family = \"NestJS\"\n\t} else if deps[\"next\"] != nil {\n\t\tsrcInfo.Family = \"Next.js\"\n\t} else if deps[\"nust\"] != nil {\n\t\tsrcInfo.Family = \"Nust\"\n\t} else if deps[\"remix\"] != nil || deps[\"@remix-run/node\"] != nil {\n\t\tsrcInfo.Family = \"Remix\"\n\t}\n\n\treturn srcInfo, nil\n}", "func Build(ctx context.Context, cfg *Config, tgts []*Target) error {\n\tctx, st := timing.Start(ctx, \"build\")\n\tdefer st.End()\n\n\tif cfg.TastWorkspace != \"\" {\n\t\tif err := checkSourceCompat(cfg.TastWorkspace); err != nil {\n\t\t\treturn fmt.Errorf(\"tast is too old: %v; please run: sudo emerge --update --deep --jobs=16 chromeos-base/tast-cmd\", err)\n\t\t}\n\t}\n\n\tif cfg.CheckBuildDeps {\n\t\tcfg.Logger.Status(\"Checking build dependencies\")\n\t\tif missing, cmds, err := checkDeps(ctx, cfg.CheckDepsCachePath); err != nil {\n\t\t\treturn fmt.Errorf(\"failed checking build deps: %v\", err)\n\t\t} else if len(missing) > 0 {\n\t\t\tif !cfg.InstallPortageDeps {\n\t\t\t\tlogMissingDeps(cfg.Logger, missing, cmds)\n\t\t\t\treturn errors.New(\"missing build dependencies\")\n\t\t\t}\n\t\t\tif err := installMissingDeps(ctx, cfg.Logger, missing, cmds); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// Compile targets in parallel.\n\tg, ctx := errgroup.WithContext(ctx)\n\tfor _, tgt := range tgts {\n\t\ttgt := tgt // bind to iteration-scoped variable\n\t\tg.Go(func() error {\n\t\t\tif err := buildOne(ctx, cfg.Logger, tgt); err != nil {\n\t\t\t\treturn fmt.Errorf(\"failed to build %s: %v\", tgt.Pkg, err)\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t}\n\treturn g.Wait()\n}", "func Build(input PairSlice) (vm FstVM, err error) {\n\tm := buildMast(input)\n\treturn m.compile()\n}", "func (info *Info) BuildFromFilePath(root string) (err error) {\n\tinfo.Name = filepath.Base(root)\n\tinfo.Files = nil\n\terr = filepath.Walk(root, func(path string, fi os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif fi.IsDir() {\n\t\t\t// Directories are implicit in torrent files.\n\t\t\treturn nil\n\t\t} else if path == root {\n\t\t\t// The root is a file.\n\t\t\tinfo.Length = fi.Size()\n\t\t\treturn nil\n\t\t}\n\t\trelPath, err := filepath.Rel(root, path)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error getting relative path: %s\", err)\n\t\t}\n\t\tinfo.Files = append(info.Files, FileInfo{\n\t\t\tPath: strings.Split(relPath, string(filepath.Separator)),\n\t\t\tLength: fi.Size(),\n\t\t})\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn\n\t}\n\tslices.Sort(info.Files, func(l, r FileInfo) bool {\n\t\treturn strings.Join(l.Path, \"/\") < strings.Join(r.Path, \"/\")\n\t})\n\terr = info.GeneratePieces(func(fi FileInfo) (io.ReadCloser, error) {\n\t\treturn os.Open(filepath.Join(root, strings.Join(fi.Path, string(filepath.Separator))))\n\t})\n\tif err != nil {\n\t\terr = fmt.Errorf(\"error generating pieces: %s\", err)\n\t}\n\treturn\n}", "func compile(dest string, src string, vars []string) error {\n\targs := []string{\n\t\t\"build\",\n\t\t\"-o\", dest,\n\t}\n\n\tif len(vars) > 0 {\n\t\targs = append(args, \"-ldflags\")\n\n\t\tfor idx, val := range vars {\n\t\t\tvars[idx] = \"-X \" + val\n\t\t}\n\n\t\tif Debug {\n\t\t\tvars = append(vars, \"-X main.debug=true\")\n\t\t}\n\n\t\targs = append(args, strings.Join(vars, \" \"))\n\t}\n\n\tt := time.Now()\n\n\toutput, err := exec.Command(\"go\", append(args, src)...).CombinedOutput()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Compile of %s failed: %s\", src, output)\n\t}\n\n\tdebugf(\"Compile %#v finished in %s\", args, time.Now().Sub(t))\n\treturn nil\n}", "func buildGeneric(builder *builder, c *cli.Context, p *Project, build procedure) error {\n\tlog.Infof(\"Building %s project at %s.\\n\", builder.name, p.Path)\n\tif err := withCleanup(p.Path, build)(); err != nil {\n\t\tlog.Errorf(\"Project couldn't be built: %s\\n\", err.Error())\n\t\treturn err\n\t}\n\n\tlog.Info(\"Project built.\")\n\treturn nil\n}", "func pack(cmd *cobra.Command, args []string) {\n\tif !coreutils.ExecutableExists(\"tar\") { // Tar not on system\n\t\ttrunk.LogFatal(\"tar does not exist on the system.\")\n\t}\n\n\tos.RemoveAll(tmpDir) // Wipe our tmpDir\n\n\tvar projectsToPack map[string]NoodlesProject\n\n\tif packProject == \"\" {\n\t\ttrunk.LogInfo(\"Started packing.\")\n\t\tprojectsToPack = noodles.Projects\n\t} else {\n\t\tprojectsToPack = map[string]NoodlesProject{\n\t\t\tpackProject: noodles.Projects[packProject],\n\t\t}\n\t}\n\n\tif creationErr := os.Mkdir(tmpDir, 0755); creationErr != nil {\n\t\ttrunk.LogErrRaw(fmt.Errorf(\"Failed to create our temporary directory:\\n%s\", creationErr.Error()))\n\t\treturn\n\t}\n\n\tfor projectName, project := range projectsToPack { // For each project\n\t\ttrunk.LogInfo(\"Packing \" + projectName)\n\n\t\tif project.Plugin != \"\" { // If a plugin is defined\n\t\t\tprojectDestFolder := filepath.Dir(project.Destination)\n\t\t\tfileName := filepath.Base(project.Destination)\n\t\t\tfileNameNoExt := strings.TrimSuffix(fileName, filepath.Ext(fileName))\n\n\t\t\tfiles := []string{fileName} // Have an array of files we should copy, at minimum the specified fileName\n\n\t\t\tif project.TarballLocation == \"\" { // If no tarball location\n\t\t\t\ttrunk.LogWarn(\"No tarball location has been set for this project. We'll attempt to place this in a smart place.\")\n\n\t\t\t\tswitch project.Plugin {\n\t\t\t\tcase \"less\":\n\t\t\t\t\tproject.TarballLocation = \"css/\"\n\t\t\t\tcase \"typescript\":\n\t\t\t\t\tproject.TarballLocation = \"js/\"\n\t\t\t\t\tfiles = append(files, fileNameNoExt+\".d.ts\") // Add the definition file\n\n\t\t\t\t\tif project.Compress {\n\t\t\t\t\t\tfiles = append(files, fileNameNoExt+\".min.js\") // Add the minified file\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tfor _, file := range files {\n\t\t\t\tCopyFile(filepath.Join(projectDestFolder, file), filepath.Join(tmpDir, project.TarballLocation, file)) // Copy this specific file\n\t\t\t}\n\t\t}\n\t}\n\n\tTarContents()\n}", "func parsePackageFiles(ctxt *build.Context, fset *token.FileSet, path string, which string) ([]*ast.File, error) {\r\n\t// Set the \"!cgo\" go/build tag, preferring (dummy) Go to\r\n\t// native C implementations of net.cgoLookupHost et al.\r\n\tctxt2 := *ctxt\r\n\tctxt2.CgoEnabled = false\r\n\r\n\t// Import(srcDir=\"\") disables local imports, e.g. import \"./foo\".\r\n\tbp, err := ctxt2.Import(path, \"\", 0)\r\n\tif _, ok := err.(*build.NoGoError); ok {\r\n\t\treturn nil, nil // empty directory\r\n\t}\r\n\tif err != nil {\r\n\t\treturn nil, err // import failed\r\n\t}\r\n\r\n\tvar filenames []string\r\n\tfor _, c := range which {\r\n\t\tvar s []string\r\n\t\tswitch c {\r\n\t\tcase 'g':\r\n\t\t\ts = bp.GoFiles\r\n\t\tcase 't':\r\n\t\t\ts = bp.TestGoFiles\r\n\t\tcase 'x':\r\n\t\t\ts = bp.XTestGoFiles\r\n\t\tdefault:\r\n\t\t\tpanic(c)\r\n\t\t}\r\n\t\tfilenames = append(filenames, s...)\r\n\t}\r\n\treturn ParseFiles(fset, bp.Dir, filenames...)\r\n}", "func createSouceCodeZipFile(context *Context) error {\n\tzipFile, _ := filepath.Abs(context.sourceZipFileName)\n\n\t// create zip file\n\tnewZipFile, err := os.Create(zipFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// create zip writer\n\tzipWriter := zip.NewWriter(newZipFile)\n\n\t// Add everything in Data.Sources to zip file\n\tfor _, item := range context.sechubConfig.Data.Sources {\n\t\terr = appendToSourceCodeZipFile(zipFile, zipWriter, item, context.config.quiet, context.config.debug)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Also support legacy definition:\n\tif len(context.sechubConfig.CodeScan.FileSystem.Folders) > 0 {\n\t\tnamedCodeScanConfig := NamedCodeScanConfig{\n\t\t\tName: \"\",\n\t\t\tFileSystem: context.sechubConfig.CodeScan.FileSystem,\n\t\t\tExcludes: context.sechubConfig.CodeScan.Excludes,\n\t\t\tSourceCodePatterns: context.sechubConfig.CodeScan.SourceCodePatterns,\n\t\t}\n\t\terr = appendToSourceCodeZipFile(zipFile, zipWriter, namedCodeScanConfig, context.config.quiet, context.config.debug)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tzipWriter.Close()\n\tnewZipFile.Close()\n\n\t// Check if context.sourceZipFileName is an empty zip\n\t// For performance reasons we only look deeper into very small files\n\tif sechubUtil.GetFileSize(zipFile) < 300 {\n\t\tzipFileContent, _ := sechubUtil.ReadContentOfZipFile(zipFile)\n\t\tif len(zipFileContent) == 0 {\n\t\t\treturn errors.New(sechubUtil.ZipFileHasNoContent)\n\t\t}\n\t}\n\n\treturn nil\n}", "func Build(basepath string, opts *BuildOpts) (ops chan OpData, prog chan string) {\n\treturn BuildFs(OsFilesystem{}, basepath, opts)\n}", "func Resolve(unit *apb.CompilationUnit, f Fetcher, opts *ResolveOptions) (*PackageInfo, error) {\n\tsourceFiles := stringset.New(unit.SourceFile...)\n\n\timap := make(map[string]*spb.VName) // import path → vname\n\tsrcs := make(map[*ast.File]string) // file → text\n\tfmap := make(map[string]*apb.FileInfo) // import path → file info\n\tsmap := make(map[string]*ast.File) // file path → file (sources)\n\tfilev := make(map[*ast.File]*spb.VName) // file → vname\n\tfloc := make(map[*token.File]*ast.File) // file → ast\n\tfset := token.NewFileSet() // location info for the parser\n\tdetails := goDetails(unit)\n\tvar files []*ast.File // parsed sources\n\tvar rules []*Ruleset // parsed linkage rules\n\n\t// Classify the required inputs as either sources, which are to be parsed,\n\t// or dependencies, which are to be \"imported\" via the type-checker's\n\t// import mechanism. If successful, this populates fset and files with the\n\t// lexical and syntactic content of the package's own sources.\n\t//\n\t// The build context is used to check build tags.\n\tbc := &build.Context{\n\t\tGOOS: details.GetGoos(),\n\t\tGOARCH: details.GetGoarch(),\n\t\tBuildTags: details.GetBuildTags(),\n\t\tReleaseTags: build.Default.ReleaseTags,\n\t\tToolTags: build.Default.ToolTags,\n\t}\n\tfor _, ri := range unit.RequiredInput {\n\t\tif ri.Info == nil {\n\t\t\treturn nil, errors.New(\"required input file info missing\")\n\t\t}\n\n\t\t// Source inputs need to be parsed, so we can give their ASTs to the\n\t\t// type checker later on.\n\t\tfpath := ri.Info.Path\n\t\tif sourceFiles.Contains(fpath) {\n\t\t\tdata, err := f.Fetch(fpath, ri.Info.Digest)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"fetching %q (%s): %v\", fpath, ri.Info.Digest, err)\n\t\t\t}\n\t\t\tif !matchesBuildTags(fpath, data, bc) {\n\t\t\t\tlog.Infof(\"Skipped source file %q because build tags do not match\", fpath)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tvpath := ri.VName.GetPath()\n\t\t\tif vpath == \"\" {\n\t\t\t\tvpath = fpath\n\t\t\t}\n\t\t\tparsed, err := parser.ParseFile(fset, vpath, data, parser.AllErrors|parser.ParseComments)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, fmt.Errorf(\"parsing %q: %v\", fpath, err)\n\t\t\t}\n\n\t\t\t// Cache file VNames based on the required input.\n\t\t\tfiles = append(files, parsed)\n\t\t\tvname := proto.Clone(ri.VName).(*spb.VName)\n\t\t\tif vname == nil {\n\t\t\t\tvname = proto.Clone(unit.VName).(*spb.VName)\n\t\t\t\tvname.Signature = \"\"\n\t\t\t\tvname.Language = \"\"\n\t\t\t}\n\t\t\tvname.Path = vpath\n\t\t\tfilev[parsed] = vname\n\t\t\tsrcs[parsed] = string(data)\n\t\t\tsmap[fpath] = parsed\n\n\t\t\t// If the file has inlined encoded metadata, add to rules.\n\t\t\tvar lastComment string\n\t\t\tif len(parsed.Comments) > 0 {\n\t\t\t\tlastComment = parsed.Comments[len(parsed.Comments)-1].Text()\n\t\t\t}\n\t\t\tconst delimiter = \"gokythe-inline-metadata:\"\n\n\t\t\tif strings.HasPrefix(lastComment, delimiter) {\n\t\t\t\tencodedMetadata := strings.TrimPrefix(lastComment, delimiter)\n\t\t\t\tnewRule, err := loadInlineMetadata(ri, encodedMetadata)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Errorf(\"loading metadata in %q: %v\", ri.Info.GetPath(), err)\n\t\t\t\t} else {\n\t\t\t\t\trules = append(rules, newRule)\n\t\t\t\t}\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// Check for mapping metadata.\n\t\tif rs, err := opts.checkRules(ri, f); err != nil {\n\t\t\tlog.Errorf(\"checking rules in %q: %v\", fpath, err)\n\t\t} else if rs != nil {\n\t\t\tlog.Infof(\"Found %d metadata rules for path %q\", len(rs.Rules), rs.Path)\n\t\t\trules = append(rules, rs)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Files may be source or compiled archives with type information for\n\t\t// other packages, or may be other ancillary files like C headers to\n\t\t// support cgo. Use the vname to determine which import path for each\n\t\t// and save that mapping for use by the importer.\n\t\tif ri.VName == nil {\n\t\t\treturn nil, fmt.Errorf(\"missing vname for %q\", fpath)\n\t\t}\n\n\t\tvar ipath string\n\t\tif info := goPackageInfo(ri.Details); info != nil {\n\t\t\tipath = info.ImportPath\n\t\t} else {\n\t\t\tipath = govname.ImportPath(ri.VName, details.GetGoroot())\n\t\t}\n\t\timap[ipath] = ri.VName\n\t\tfmap[ipath] = ri.Info\n\t}\n\tif len(files) == 0 {\n\t\treturn nil, ErrNoSourceFiles\n\t}\n\n\t// Populate the location mapping. This relies on the fact that Iterate\n\t// reports its files in the order they were added to the set, which in turn\n\t// is their order in the files list.\n\ti := 0\n\tfset.Iterate(func(f *token.File) bool {\n\t\tfloc[f] = files[i]\n\t\ti++\n\t\treturn true\n\t})\n\n\tpi := &PackageInfo{\n\t\tName: files[0].Name.Name,\n\t\tFileSet: fset,\n\t\tFiles: files,\n\t\tInfo: opts.info(),\n\t\tSourceText: srcs,\n\t\tPackageVName: make(map[*types.Package]*spb.VName),\n\t\tDependencies: make(map[string]*types.Package), // :: import path → package\n\t\tVendored: make(map[string]string),\n\n\t\tfunction: make(map[ast.Node]*funcInfo),\n\t\tsigs: make(map[types.Object]string),\n\t\tpackageInit: make(map[*ast.File]*funcInfo),\n\t\tfileVName: filev,\n\t\ttypeVName: make(map[types.Type]*spb.VName),\n\t\ttypeEmitted: stringset.New(),\n\t\tfileLoc: floc,\n\t\tdetails: details,\n\t}\n\tif info := goPackageInfo(unit.Details); info != nil {\n\t\tpi.ImportPath = info.ImportPath\n\t} else {\n\t\tpi.ImportPath = govname.ImportPath(unit.VName, details.GetGoroot())\n\t}\n\n\t// If mapping rules were found, populate the corresponding field.\n\tif len(rules) != 0 {\n\t\tpi.Rules = make(map[*ast.File]metadata.Rules)\n\t\tfor _, rs := range rules {\n\t\t\tf, ok := smap[rs.Path]\n\t\t\tif ok {\n\t\t\t\tpi.Rules[f] = rs.Rules\n\t\t\t}\n\t\t}\n\t}\n\n\t// Run the type-checker and collect any errors it generates. Errors in the\n\t// type checker are not returned directly; the caller can read them from\n\t// the Errors field.\n\tc := &types.Config{\n\t\tFakeImportC: true, // so we can handle cgo\n\t\tDisableUnusedImportCheck: true, // this is not fatal to type-checking\n\t\tImporter: &packageImporter{\n\t\t\tdeps: pi.Dependencies,\n\t\t\tfileSet: pi.FileSet,\n\t\t\tfileMap: fmap,\n\t\t\tfetcher: f,\n\n\t\t\tpkgPath: pi.ImportPath,\n\t\t\tvendored: pi.Vendored,\n\t\t},\n\t\tError: func(err error) { pi.Errors = append(pi.Errors, err) },\n\t}\n\tpi.Package, _ = c.Check(pi.Name, pi.FileSet, pi.Files, pi.Info)\n\n\t// Fill in the mapping from packages to vnames.\n\tfor ip, vname := range imap {\n\t\tif pkg := pi.Dependencies[ip]; pkg != nil {\n\t\t\tpi.PackageVName[pkg] = proto.Clone(vname).(*spb.VName)\n\t\t\tpi.PackageVName[pkg].Signature = \"package\"\n\t\t\tpi.PackageVName[pkg].Language = govname.Language\n\t\t}\n\t}\n\tif _, ok := pi.Dependencies[\"unsafe\"]; ok {\n\t\tpi.PackageVName[types.Unsafe] = govname.ForStandardLibrary(\"unsafe\")\n\t}\n\n\t// Set this package's own vname.\n\tpi.VName = proto.Clone(unit.VName).(*spb.VName)\n\tpi.VName.Language = govname.Language\n\tpi.VName.Signature = \"package\"\n\tpi.PackageVName[pi.Package] = pi.VName\n\n\treturn pi, nil\n}", "func buildTransformer(tmplPath string, fSet *token.FileSet, pkgs *[]*packages.Package) (*eg.Transformer, error) {\n\t// find the template package in the processed packages according to the absolute file path\n\tvar tmplPkg *packages.Package\n\tfor i := 0; tmplPkg == nil && i < len(*pkgs); i++ {\n\t\tpkg := (*pkgs)[i]\n\t\tfor _, f := range pkg.GoFiles {\n\t\t\tif f == tmplPath {\n\t\t\t\ttmplPkg = pkg\n\t\t\t\t*pkgs = append((*pkgs)[:i], (*pkgs)[i+1:]...)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\tif tmplPkg == nil {\n\t\treturn nil, errors.New(\"didn't find template in module path\")\n\t}\n\n\tvar tmplFile *ast.File\n\tfor _, f := range tmplPkg.Syntax {\n\t\tif tmplPath == fSet.File(f.Pos()).Name() {\n\t\t\ttmplFile = f\n\t\t\tbreak\n\t\t}\n\t}\n\tif tmplFile == nil {\n\t\tpanic(\"didn't find template in template package somehow\")\n\t}\n\n\treturn eg.NewTransformer(fSet, tmplPkg.Types, tmplFile, tmplPkg.TypesInfo, *verboseFlag)\n}", "func (ts *Tester) Build() error {\n\t// no-op\n\treturn nil\n}" ]
[ "0.58292294", "0.57030225", "0.56895274", "0.5610833", "0.5584406", "0.55761385", "0.5510384", "0.5491371", "0.54894584", "0.54794395", "0.5463892", "0.54571146", "0.5348241", "0.5306004", "0.5278089", "0.52519447", "0.52158", "0.51755077", "0.5139383", "0.511165", "0.51094705", "0.5106139", "0.5098783", "0.5090583", "0.50805634", "0.5071059", "0.503292", "0.50229347", "0.50134665", "0.5009439", "0.50043535", "0.50005525", "0.49893007", "0.49889553", "0.49788532", "0.4955814", "0.49522173", "0.49521032", "0.49394783", "0.49271557", "0.49125087", "0.4905767", "0.4903172", "0.48940653", "0.48891494", "0.48873582", "0.48767695", "0.4875624", "0.4875168", "0.48738027", "0.48720872", "0.4869163", "0.48641157", "0.4858159", "0.48553765", "0.4854387", "0.485228", "0.4834548", "0.48334634", "0.4829201", "0.48114118", "0.48072812", "0.4804452", "0.48005044", "0.47963792", "0.47946733", "0.47929308", "0.47890958", "0.478739", "0.47786683", "0.47771314", "0.47744346", "0.47734594", "0.47681478", "0.47518396", "0.4748084", "0.47466585", "0.47401455", "0.4737901", "0.4737779", "0.47372136", "0.4723509", "0.47190112", "0.47182387", "0.47055376", "0.47045785", "0.47028393", "0.469074", "0.46862963", "0.4685757", "0.46801504", "0.46757463", "0.46710187", "0.46633774", "0.46617097", "0.46549955", "0.46522394", "0.46508133", "0.46419936", "0.4637743" ]
0.70778966
0