// Code generated by ent, DO NOT EDIT. package ent import ( "context" "fmt" "math" "wechat-api/ent/compapijob" "wechat-api/ent/predicate" "entgo.io/ent/dialect/sql" "entgo.io/ent/dialect/sql/sqlgraph" "entgo.io/ent/schema/field" ) // CompapiJobQuery is the builder for querying CompapiJob entities. type CompapiJobQuery struct { config ctx *QueryContext order []compapijob.OrderOption inters []Interceptor predicates []predicate.CompapiJob // intermediate query (i.e. traversal path). sql *sql.Selector path func(context.Context) (*sql.Selector, error) } // Where adds a new predicate for the CompapiJobQuery builder. func (cjq *CompapiJobQuery) Where(ps ...predicate.CompapiJob) *CompapiJobQuery { cjq.predicates = append(cjq.predicates, ps...) return cjq } // Limit the number of records to be returned by this query. func (cjq *CompapiJobQuery) Limit(limit int) *CompapiJobQuery { cjq.ctx.Limit = &limit return cjq } // Offset to start from. func (cjq *CompapiJobQuery) Offset(offset int) *CompapiJobQuery { cjq.ctx.Offset = &offset return cjq } // Unique configures the query builder to filter duplicate records on query. // By default, unique is set to true, and can be disabled using this method. func (cjq *CompapiJobQuery) Unique(unique bool) *CompapiJobQuery { cjq.ctx.Unique = &unique return cjq } // Order specifies how the records should be ordered. func (cjq *CompapiJobQuery) Order(o ...compapijob.OrderOption) *CompapiJobQuery { cjq.order = append(cjq.order, o...) return cjq } // First returns the first CompapiJob entity from the query. // Returns a *NotFoundError when no CompapiJob was found. func (cjq *CompapiJobQuery) First(ctx context.Context) (*CompapiJob, error) { nodes, err := cjq.Limit(1).All(setContextOp(ctx, cjq.ctx, "First")) if err != nil { return nil, err } if len(nodes) == 0 { return nil, &NotFoundError{compapijob.Label} } return nodes[0], nil } // FirstX is like First, but panics if an error occurs. func (cjq *CompapiJobQuery) FirstX(ctx context.Context) *CompapiJob { node, err := cjq.First(ctx) if err != nil && !IsNotFound(err) { panic(err) } return node } // FirstID returns the first CompapiJob ID from the query. // Returns a *NotFoundError when no CompapiJob ID was found. func (cjq *CompapiJobQuery) FirstID(ctx context.Context) (id uint64, err error) { var ids []uint64 if ids, err = cjq.Limit(1).IDs(setContextOp(ctx, cjq.ctx, "FirstID")); err != nil { return } if len(ids) == 0 { err = &NotFoundError{compapijob.Label} return } return ids[0], nil } // FirstIDX is like FirstID, but panics if an error occurs. func (cjq *CompapiJobQuery) FirstIDX(ctx context.Context) uint64 { id, err := cjq.FirstID(ctx) if err != nil && !IsNotFound(err) { panic(err) } return id } // Only returns a single CompapiJob entity found by the query, ensuring it only returns one. // Returns a *NotSingularError when more than one CompapiJob entity is found. // Returns a *NotFoundError when no CompapiJob entities are found. func (cjq *CompapiJobQuery) Only(ctx context.Context) (*CompapiJob, error) { nodes, err := cjq.Limit(2).All(setContextOp(ctx, cjq.ctx, "Only")) if err != nil { return nil, err } switch len(nodes) { case 1: return nodes[0], nil case 0: return nil, &NotFoundError{compapijob.Label} default: return nil, &NotSingularError{compapijob.Label} } } // OnlyX is like Only, but panics if an error occurs. func (cjq *CompapiJobQuery) OnlyX(ctx context.Context) *CompapiJob { node, err := cjq.Only(ctx) if err != nil { panic(err) } return node } // OnlyID is like Only, but returns the only CompapiJob ID in the query. // Returns a *NotSingularError when more than one CompapiJob ID is found. // Returns a *NotFoundError when no entities are found. func (cjq *CompapiJobQuery) OnlyID(ctx context.Context) (id uint64, err error) { var ids []uint64 if ids, err = cjq.Limit(2).IDs(setContextOp(ctx, cjq.ctx, "OnlyID")); err != nil { return } switch len(ids) { case 1: id = ids[0] case 0: err = &NotFoundError{compapijob.Label} default: err = &NotSingularError{compapijob.Label} } return } // OnlyIDX is like OnlyID, but panics if an error occurs. func (cjq *CompapiJobQuery) OnlyIDX(ctx context.Context) uint64 { id, err := cjq.OnlyID(ctx) if err != nil { panic(err) } return id } // All executes the query and returns a list of CompapiJobs. func (cjq *CompapiJobQuery) All(ctx context.Context) ([]*CompapiJob, error) { ctx = setContextOp(ctx, cjq.ctx, "All") if err := cjq.prepareQuery(ctx); err != nil { return nil, err } qr := querierAll[[]*CompapiJob, *CompapiJobQuery]() return withInterceptors[[]*CompapiJob](ctx, cjq, qr, cjq.inters) } // AllX is like All, but panics if an error occurs. func (cjq *CompapiJobQuery) AllX(ctx context.Context) []*CompapiJob { nodes, err := cjq.All(ctx) if err != nil { panic(err) } return nodes } // IDs executes the query and returns a list of CompapiJob IDs. func (cjq *CompapiJobQuery) IDs(ctx context.Context) (ids []uint64, err error) { if cjq.ctx.Unique == nil && cjq.path != nil { cjq.Unique(true) } ctx = setContextOp(ctx, cjq.ctx, "IDs") if err = cjq.Select(compapijob.FieldID).Scan(ctx, &ids); err != nil { return nil, err } return ids, nil } // IDsX is like IDs, but panics if an error occurs. func (cjq *CompapiJobQuery) IDsX(ctx context.Context) []uint64 { ids, err := cjq.IDs(ctx) if err != nil { panic(err) } return ids } // Count returns the count of the given query. func (cjq *CompapiJobQuery) Count(ctx context.Context) (int, error) { ctx = setContextOp(ctx, cjq.ctx, "Count") if err := cjq.prepareQuery(ctx); err != nil { return 0, err } return withInterceptors[int](ctx, cjq, querierCount[*CompapiJobQuery](), cjq.inters) } // CountX is like Count, but panics if an error occurs. func (cjq *CompapiJobQuery) CountX(ctx context.Context) int { count, err := cjq.Count(ctx) if err != nil { panic(err) } return count } // Exist returns true if the query has elements in the graph. func (cjq *CompapiJobQuery) Exist(ctx context.Context) (bool, error) { ctx = setContextOp(ctx, cjq.ctx, "Exist") switch _, err := cjq.FirstID(ctx); { case IsNotFound(err): return false, nil case err != nil: return false, fmt.Errorf("ent: check existence: %w", err) default: return true, nil } } // ExistX is like Exist, but panics if an error occurs. func (cjq *CompapiJobQuery) ExistX(ctx context.Context) bool { exist, err := cjq.Exist(ctx) if err != nil { panic(err) } return exist } // Clone returns a duplicate of the CompapiJobQuery builder, including all associated steps. It can be // used to prepare common query builders and use them differently after the clone is made. func (cjq *CompapiJobQuery) Clone() *CompapiJobQuery { if cjq == nil { return nil } return &CompapiJobQuery{ config: cjq.config, ctx: cjq.ctx.Clone(), order: append([]compapijob.OrderOption{}, cjq.order...), inters: append([]Interceptor{}, cjq.inters...), predicates: append([]predicate.CompapiJob{}, cjq.predicates...), // clone intermediate query. sql: cjq.sql.Clone(), path: cjq.path, } } // GroupBy is used to group vertices by one or more fields/columns. // It is often used with aggregate functions, like: count, max, mean, min, sum. // // Example: // // var v []struct { // CreatedAt time.Time `json:"created_at,omitempty"` // Count int `json:"count,omitempty"` // } // // client.CompapiJob.Query(). // GroupBy(compapijob.FieldCreatedAt). // Aggregate(ent.Count()). // Scan(ctx, &v) func (cjq *CompapiJobQuery) GroupBy(field string, fields ...string) *CompapiJobGroupBy { cjq.ctx.Fields = append([]string{field}, fields...) grbuild := &CompapiJobGroupBy{build: cjq} grbuild.flds = &cjq.ctx.Fields grbuild.label = compapijob.Label grbuild.scan = grbuild.Scan return grbuild } // Select allows the selection one or more fields/columns for the given query, // instead of selecting all fields in the entity. // // Example: // // var v []struct { // CreatedAt time.Time `json:"created_at,omitempty"` // } // // client.CompapiJob.Query(). // Select(compapijob.FieldCreatedAt). // Scan(ctx, &v) func (cjq *CompapiJobQuery) Select(fields ...string) *CompapiJobSelect { cjq.ctx.Fields = append(cjq.ctx.Fields, fields...) sbuild := &CompapiJobSelect{CompapiJobQuery: cjq} sbuild.label = compapijob.Label sbuild.flds, sbuild.scan = &cjq.ctx.Fields, sbuild.Scan return sbuild } // Aggregate returns a CompapiJobSelect configured with the given aggregations. func (cjq *CompapiJobQuery) Aggregate(fns ...AggregateFunc) *CompapiJobSelect { return cjq.Select().Aggregate(fns...) } func (cjq *CompapiJobQuery) prepareQuery(ctx context.Context) error { for _, inter := range cjq.inters { if inter == nil { return fmt.Errorf("ent: uninitialized interceptor (forgotten import ent/runtime?)") } if trv, ok := inter.(Traverser); ok { if err := trv.Traverse(ctx, cjq); err != nil { return err } } } for _, f := range cjq.ctx.Fields { if !compapijob.ValidColumn(f) { return &ValidationError{Name: f, err: fmt.Errorf("ent: invalid field %q for query", f)} } } if cjq.path != nil { prev, err := cjq.path(ctx) if err != nil { return err } cjq.sql = prev } return nil } func (cjq *CompapiJobQuery) sqlAll(ctx context.Context, hooks ...queryHook) ([]*CompapiJob, error) { var ( nodes = []*CompapiJob{} _spec = cjq.querySpec() ) _spec.ScanValues = func(columns []string) ([]any, error) { return (*CompapiJob).scanValues(nil, columns) } _spec.Assign = func(columns []string, values []any) error { node := &CompapiJob{config: cjq.config} nodes = append(nodes, node) return node.assignValues(columns, values) } for i := range hooks { hooks[i](ctx, _spec) } if err := sqlgraph.QueryNodes(ctx, cjq.driver, _spec); err != nil { return nil, err } if len(nodes) == 0 { return nodes, nil } return nodes, nil } func (cjq *CompapiJobQuery) sqlCount(ctx context.Context) (int, error) { _spec := cjq.querySpec() _spec.Node.Columns = cjq.ctx.Fields if len(cjq.ctx.Fields) > 0 { _spec.Unique = cjq.ctx.Unique != nil && *cjq.ctx.Unique } return sqlgraph.CountNodes(ctx, cjq.driver, _spec) } func (cjq *CompapiJobQuery) querySpec() *sqlgraph.QuerySpec { _spec := sqlgraph.NewQuerySpec(compapijob.Table, compapijob.Columns, sqlgraph.NewFieldSpec(compapijob.FieldID, field.TypeUint64)) _spec.From = cjq.sql if unique := cjq.ctx.Unique; unique != nil { _spec.Unique = *unique } else if cjq.path != nil { _spec.Unique = true } if fields := cjq.ctx.Fields; len(fields) > 0 { _spec.Node.Columns = make([]string, 0, len(fields)) _spec.Node.Columns = append(_spec.Node.Columns, compapijob.FieldID) for i := range fields { if fields[i] != compapijob.FieldID { _spec.Node.Columns = append(_spec.Node.Columns, fields[i]) } } } if ps := cjq.predicates; len(ps) > 0 { _spec.Predicate = func(selector *sql.Selector) { for i := range ps { ps[i](selector) } } } if limit := cjq.ctx.Limit; limit != nil { _spec.Limit = *limit } if offset := cjq.ctx.Offset; offset != nil { _spec.Offset = *offset } if ps := cjq.order; len(ps) > 0 { _spec.Order = func(selector *sql.Selector) { for i := range ps { ps[i](selector) } } } return _spec } func (cjq *CompapiJobQuery) sqlQuery(ctx context.Context) *sql.Selector { builder := sql.Dialect(cjq.driver.Dialect()) t1 := builder.Table(compapijob.Table) columns := cjq.ctx.Fields if len(columns) == 0 { columns = compapijob.Columns } selector := builder.Select(t1.Columns(columns...)...).From(t1) if cjq.sql != nil { selector = cjq.sql selector.Select(selector.Columns(columns...)...) } if cjq.ctx.Unique != nil && *cjq.ctx.Unique { selector.Distinct() } for _, p := range cjq.predicates { p(selector) } for _, p := range cjq.order { p(selector) } if offset := cjq.ctx.Offset; offset != nil { // limit is mandatory for offset clause. We start // with default value, and override it below if needed. selector.Offset(*offset).Limit(math.MaxInt32) } if limit := cjq.ctx.Limit; limit != nil { selector.Limit(*limit) } return selector } // CompapiJobGroupBy is the group-by builder for CompapiJob entities. type CompapiJobGroupBy struct { selector build *CompapiJobQuery } // Aggregate adds the given aggregation functions to the group-by query. func (cjgb *CompapiJobGroupBy) Aggregate(fns ...AggregateFunc) *CompapiJobGroupBy { cjgb.fns = append(cjgb.fns, fns...) return cjgb } // Scan applies the selector query and scans the result into the given value. func (cjgb *CompapiJobGroupBy) Scan(ctx context.Context, v any) error { ctx = setContextOp(ctx, cjgb.build.ctx, "GroupBy") if err := cjgb.build.prepareQuery(ctx); err != nil { return err } return scanWithInterceptors[*CompapiJobQuery, *CompapiJobGroupBy](ctx, cjgb.build, cjgb, cjgb.build.inters, v) } func (cjgb *CompapiJobGroupBy) sqlScan(ctx context.Context, root *CompapiJobQuery, v any) error { selector := root.sqlQuery(ctx).Select() aggregation := make([]string, 0, len(cjgb.fns)) for _, fn := range cjgb.fns { aggregation = append(aggregation, fn(selector)) } if len(selector.SelectedColumns()) == 0 { columns := make([]string, 0, len(*cjgb.flds)+len(cjgb.fns)) for _, f := range *cjgb.flds { columns = append(columns, selector.C(f)) } columns = append(columns, aggregation...) selector.Select(columns...) } selector.GroupBy(selector.Columns(*cjgb.flds...)...) if err := selector.Err(); err != nil { return err } rows := &sql.Rows{} query, args := selector.Query() if err := cjgb.build.driver.Query(ctx, query, args, rows); err != nil { return err } defer rows.Close() return sql.ScanSlice(rows, v) } // CompapiJobSelect is the builder for selecting fields of CompapiJob entities. type CompapiJobSelect struct { *CompapiJobQuery selector } // Aggregate adds the given aggregation functions to the selector query. func (cjs *CompapiJobSelect) Aggregate(fns ...AggregateFunc) *CompapiJobSelect { cjs.fns = append(cjs.fns, fns...) return cjs } // Scan applies the selector query and scans the result into the given value. func (cjs *CompapiJobSelect) Scan(ctx context.Context, v any) error { ctx = setContextOp(ctx, cjs.ctx, "Select") if err := cjs.prepareQuery(ctx); err != nil { return err } return scanWithInterceptors[*CompapiJobQuery, *CompapiJobSelect](ctx, cjs.CompapiJobQuery, cjs, cjs.inters, v) } func (cjs *CompapiJobSelect) sqlScan(ctx context.Context, root *CompapiJobQuery, v any) error { selector := root.sqlQuery(ctx) aggregation := make([]string, 0, len(cjs.fns)) for _, fn := range cjs.fns { aggregation = append(aggregation, fn(selector)) } switch n := len(*cjs.selector.flds); { case n == 0 && len(aggregation) > 0: selector.Select(aggregation...) case n != 0 && len(aggregation) > 0: selector.AppendSelect(aggregation...) } rows := &sql.Rows{} query, args := selector.Query() if err := cjs.driver.Query(ctx, query, args, rows); err != nil { return err } defer rows.Close() return sql.ScanSlice(rows, v) }