mirror of
https://forgejo.merr.is/annika/isl-api.git
synced 2025-12-10 11:53:14 -05:00
I Guess I Need To Check In The Generated SQL For This To Work
This commit is contained in:
parent
89595b6f71
commit
8f3d38a066
2 changed files with 575 additions and 1 deletions
2
.gitignore
vendored
2
.gitignore
vendored
|
|
@ -1,2 +1,2 @@
|
|||
sql/**/*.go
|
||||
# sql/**/*.go
|
||||
docker/data
|
||||
|
|
|
|||
574
sql/powerItem/query.sql.go
Normal file
574
sql/powerItem/query.sql.go
Normal file
|
|
@ -0,0 +1,574 @@
|
|||
// Code generated by pggen. DO NOT EDIT.
|
||||
|
||||
package powerItem
|
||||
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"github.com/jackc/pgconn"
|
||||
"github.com/jackc/pgtype"
|
||||
"github.com/jackc/pgx/v4"
|
||||
)
|
||||
|
||||
// Querier is a typesafe Go interface backed by SQL queries.
|
||||
//
|
||||
// Methods ending with Batch enqueue a query to run later in a pgx.Batch. After
|
||||
// calling SendBatch on pgx.Conn, pgxpool.Pool, or pgx.Tx, use the Scan methods
|
||||
// to parse the results.
|
||||
type Querier interface {
|
||||
// GetAllItems finds all items.
|
||||
GetAllItems(ctx context.Context) ([]GetAllItemsRow, error)
|
||||
// GetAllItemsBatch enqueues a GetAllItems query into batch to be executed
|
||||
// later by the batch.
|
||||
GetAllItemsBatch(batch genericBatch)
|
||||
// GetAllItemsScan scans the result of an executed GetAllItemsBatch query.
|
||||
GetAllItemsScan(results pgx.BatchResults) ([]GetAllItemsRow, error)
|
||||
|
||||
// GetAllByType finds all items of a specific type.
|
||||
GetAllByType(ctx context.Context, itemType int32) ([]GetAllByTypeRow, error)
|
||||
// GetAllByTypeBatch enqueues a GetAllByType query into batch to be executed
|
||||
// later by the batch.
|
||||
GetAllByTypeBatch(batch genericBatch, itemType int32)
|
||||
// GetAllByTypeScan scans the result of an executed GetAllByTypeBatch query.
|
||||
GetAllByTypeScan(results pgx.BatchResults) ([]GetAllByTypeRow, error)
|
||||
|
||||
// FindByID finds items by the ID
|
||||
FindByID(ctx context.Context, id pgtype.UUID) (FindByIDRow, error)
|
||||
// FindByIDBatch enqueues a FindByID query into batch to be executed
|
||||
// later by the batch.
|
||||
FindByIDBatch(batch genericBatch, id pgtype.UUID)
|
||||
// FindByIDScan scans the result of an executed FindByIDBatch query.
|
||||
FindByIDScan(results pgx.BatchResults) (FindByIDRow, error)
|
||||
|
||||
// AddNewItem inserts a new power item
|
||||
AddNewItem(ctx context.Context, params AddNewItemParams) (AddNewItemRow, error)
|
||||
// AddNewItemBatch enqueues a AddNewItem query into batch to be executed
|
||||
// later by the batch.
|
||||
AddNewItemBatch(batch genericBatch, params AddNewItemParams)
|
||||
// AddNewItemScan scans the result of an executed AddNewItemBatch query.
|
||||
AddNewItemScan(results pgx.BatchResults) (AddNewItemRow, error)
|
||||
|
||||
// AddNewItemWithID inserts a new power item
|
||||
AddNewItemWithID(ctx context.Context, params AddNewItemWithIDParams) (AddNewItemWithIDRow, error)
|
||||
// AddNewItemWithIDBatch enqueues a AddNewItemWithID query into batch to be executed
|
||||
// later by the batch.
|
||||
AddNewItemWithIDBatch(batch genericBatch, params AddNewItemWithIDParams)
|
||||
// AddNewItemWithIDScan scans the result of an executed AddNewItemWithIDBatch query.
|
||||
AddNewItemWithIDScan(results pgx.BatchResults) (AddNewItemWithIDRow, error)
|
||||
}
|
||||
|
||||
type DBQuerier struct {
|
||||
conn genericConn // underlying Postgres transport to use
|
||||
types *typeResolver // resolve types by name
|
||||
}
|
||||
|
||||
var _ Querier = &DBQuerier{}
|
||||
|
||||
// genericConn is a connection to a Postgres database. This is usually backed by
|
||||
// *pgx.Conn, pgx.Tx, or *pgxpool.Pool.
|
||||
type genericConn interface {
|
||||
// Query executes sql with args. If there is an error the returned Rows will
|
||||
// be returned in an error state. So it is allowed to ignore the error
|
||||
// returned from Query and handle it in Rows.
|
||||
Query(ctx context.Context, sql string, args ...interface{}) (pgx.Rows, error)
|
||||
|
||||
// QueryRow is a convenience wrapper over Query. Any error that occurs while
|
||||
// querying is deferred until calling Scan on the returned Row. That Row will
|
||||
// error with pgx.ErrNoRows if no rows are returned.
|
||||
QueryRow(ctx context.Context, sql string, args ...interface{}) pgx.Row
|
||||
|
||||
// Exec executes sql. sql can be either a prepared statement name or an SQL
|
||||
// string. arguments should be referenced positionally from the sql string
|
||||
// as $1, $2, etc.
|
||||
Exec(ctx context.Context, sql string, arguments ...interface{}) (pgconn.CommandTag, error)
|
||||
}
|
||||
|
||||
// genericBatch batches queries to send in a single network request to a
|
||||
// Postgres server. This is usually backed by *pgx.Batch.
|
||||
type genericBatch interface {
|
||||
// Queue queues a query to batch b. query can be an SQL query or the name of a
|
||||
// prepared statement. See Queue on *pgx.Batch.
|
||||
Queue(query string, arguments ...interface{})
|
||||
}
|
||||
|
||||
// NewQuerier creates a DBQuerier that implements Querier. conn is typically
|
||||
// *pgx.Conn, pgx.Tx, or *pgxpool.Pool.
|
||||
func NewQuerier(conn genericConn) *DBQuerier {
|
||||
return NewQuerierConfig(conn, QuerierConfig{})
|
||||
}
|
||||
|
||||
type QuerierConfig struct {
|
||||
// DataTypes contains pgtype.Value to use for encoding and decoding instead
|
||||
// of pggen-generated pgtype.ValueTranscoder.
|
||||
//
|
||||
// If OIDs are available for an input parameter type and all of its
|
||||
// transitive dependencies, pggen will use the binary encoding format for
|
||||
// the input parameter.
|
||||
DataTypes []pgtype.DataType
|
||||
}
|
||||
|
||||
// NewQuerierConfig creates a DBQuerier that implements Querier with the given
|
||||
// config. conn is typically *pgx.Conn, pgx.Tx, or *pgxpool.Pool.
|
||||
func NewQuerierConfig(conn genericConn, cfg QuerierConfig) *DBQuerier {
|
||||
return &DBQuerier{conn: conn, types: newTypeResolver(cfg.DataTypes)}
|
||||
}
|
||||
|
||||
// WithTx creates a new DBQuerier that uses the transaction to run all queries.
|
||||
func (q *DBQuerier) WithTx(tx pgx.Tx) (*DBQuerier, error) {
|
||||
return &DBQuerier{conn: tx}, nil
|
||||
}
|
||||
|
||||
// preparer is any Postgres connection transport that provides a way to prepare
|
||||
// a statement, most commonly *pgx.Conn.
|
||||
type preparer interface {
|
||||
Prepare(ctx context.Context, name, sql string) (sd *pgconn.StatementDescription, err error)
|
||||
}
|
||||
|
||||
// PrepareAllQueries executes a PREPARE statement for all pggen generated SQL
|
||||
// queries in querier files. Typical usage is as the AfterConnect callback
|
||||
// for pgxpool.Config
|
||||
//
|
||||
// pgx will use the prepared statement if available. Calling PrepareAllQueries
|
||||
// is an optional optimization to avoid a network round-trip the first time pgx
|
||||
// runs a query if pgx statement caching is enabled.
|
||||
func PrepareAllQueries(ctx context.Context, p preparer) error {
|
||||
if _, err := p.Prepare(ctx, getAllItemsSQL, getAllItemsSQL); err != nil {
|
||||
return fmt.Errorf("prepare query 'GetAllItems': %w", err)
|
||||
}
|
||||
if _, err := p.Prepare(ctx, getAllByTypeSQL, getAllByTypeSQL); err != nil {
|
||||
return fmt.Errorf("prepare query 'GetAllByType': %w", err)
|
||||
}
|
||||
if _, err := p.Prepare(ctx, findByIDSQL, findByIDSQL); err != nil {
|
||||
return fmt.Errorf("prepare query 'FindByID': %w", err)
|
||||
}
|
||||
if _, err := p.Prepare(ctx, addNewItemSQL, addNewItemSQL); err != nil {
|
||||
return fmt.Errorf("prepare query 'AddNewItem': %w", err)
|
||||
}
|
||||
if _, err := p.Prepare(ctx, addNewItemWithIDSQL, addNewItemWithIDSQL); err != nil {
|
||||
return fmt.Errorf("prepare query 'AddNewItemWithID': %w", err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// typeResolver looks up the pgtype.ValueTranscoder by Postgres type name.
|
||||
type typeResolver struct {
|
||||
connInfo *pgtype.ConnInfo // types by Postgres type name
|
||||
}
|
||||
|
||||
func newTypeResolver(types []pgtype.DataType) *typeResolver {
|
||||
ci := pgtype.NewConnInfo()
|
||||
for _, typ := range types {
|
||||
if txt, ok := typ.Value.(textPreferrer); ok && typ.OID != unknownOID {
|
||||
typ.Value = txt.ValueTranscoder
|
||||
}
|
||||
ci.RegisterDataType(typ)
|
||||
}
|
||||
return &typeResolver{connInfo: ci}
|
||||
}
|
||||
|
||||
// findValue find the OID, and pgtype.ValueTranscoder for a Postgres type name.
|
||||
func (tr *typeResolver) findValue(name string) (uint32, pgtype.ValueTranscoder, bool) {
|
||||
typ, ok := tr.connInfo.DataTypeForName(name)
|
||||
if !ok {
|
||||
return 0, nil, false
|
||||
}
|
||||
v := pgtype.NewValue(typ.Value)
|
||||
return typ.OID, v.(pgtype.ValueTranscoder), true
|
||||
}
|
||||
|
||||
// setValue sets the value of a ValueTranscoder to a value that should always
|
||||
// work and panics if it fails.
|
||||
func (tr *typeResolver) setValue(vt pgtype.ValueTranscoder, val interface{}) pgtype.ValueTranscoder {
|
||||
if err := vt.Set(val); err != nil {
|
||||
panic(fmt.Sprintf("set ValueTranscoder %T to %+v: %s", vt, val, err))
|
||||
}
|
||||
return vt
|
||||
}
|
||||
|
||||
const getAllItemsSQL = `SELECT
|
||||
id,
|
||||
itemType,
|
||||
iconURL,
|
||||
itemName,
|
||||
minItemPower,
|
||||
maxItemPower,
|
||||
rarity,
|
||||
origin,
|
||||
tooltip,
|
||||
isEventItem
|
||||
FROM
|
||||
powerItem;`
|
||||
|
||||
type GetAllItemsRow struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
Itemtype *int32 `json:"itemtype"`
|
||||
Iconurl *string `json:"iconurl"`
|
||||
Itemname *string `json:"itemname"`
|
||||
Minitempower *int32 `json:"minitempower"`
|
||||
Maxitempower *int32 `json:"maxitempower"`
|
||||
Rarity *int32 `json:"rarity"`
|
||||
Origin *string `json:"origin"`
|
||||
Tooltip *string `json:"tooltip"`
|
||||
Iseventitem *bool `json:"iseventitem"`
|
||||
}
|
||||
|
||||
// GetAllItems implements Querier.GetAllItems.
|
||||
func (q *DBQuerier) GetAllItems(ctx context.Context) ([]GetAllItemsRow, error) {
|
||||
ctx = context.WithValue(ctx, "pggen_query_name", "GetAllItems")
|
||||
rows, err := q.conn.Query(ctx, getAllItemsSQL)
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("query GetAllItems: %w", err)
|
||||
}
|
||||
defer rows.Close()
|
||||
items := []GetAllItemsRow{}
|
||||
for rows.Next() {
|
||||
var item GetAllItemsRow
|
||||
if err := rows.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return nil, fmt.Errorf("scan GetAllItems row: %w", err)
|
||||
}
|
||||
items = append(items, item)
|
||||
}
|
||||
if err := rows.Err(); err != nil {
|
||||
return nil, fmt.Errorf("close GetAllItems rows: %w", err)
|
||||
}
|
||||
return items, err
|
||||
}
|
||||
|
||||
// GetAllItemsBatch implements Querier.GetAllItemsBatch.
|
||||
func (q *DBQuerier) GetAllItemsBatch(batch genericBatch) {
|
||||
batch.Queue(getAllItemsSQL)
|
||||
}
|
||||
|
||||
// GetAllItemsScan implements Querier.GetAllItemsScan.
|
||||
func (q *DBQuerier) GetAllItemsScan(results pgx.BatchResults) ([]GetAllItemsRow, error) {
|
||||
rows, err := results.Query()
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("query GetAllItemsBatch: %w", err)
|
||||
}
|
||||
defer rows.Close()
|
||||
items := []GetAllItemsRow{}
|
||||
for rows.Next() {
|
||||
var item GetAllItemsRow
|
||||
if err := rows.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return nil, fmt.Errorf("scan GetAllItemsBatch row: %w", err)
|
||||
}
|
||||
items = append(items, item)
|
||||
}
|
||||
if err := rows.Err(); err != nil {
|
||||
return nil, fmt.Errorf("close GetAllItemsBatch rows: %w", err)
|
||||
}
|
||||
return items, err
|
||||
}
|
||||
|
||||
const getAllByTypeSQL = `SELECT
|
||||
id,
|
||||
itemType,
|
||||
iconURL,
|
||||
itemName,
|
||||
minItemPower,
|
||||
maxItemPower,
|
||||
rarity,
|
||||
origin,
|
||||
tooltip,
|
||||
isEventItem
|
||||
FROM
|
||||
powerItem
|
||||
WHERE
|
||||
itemType = $1;`
|
||||
|
||||
type GetAllByTypeRow struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
Itemtype int32 `json:"itemtype"`
|
||||
Iconurl string `json:"iconurl"`
|
||||
Itemname string `json:"itemname"`
|
||||
Minitempower int32 `json:"minitempower"`
|
||||
Maxitempower int32 `json:"maxitempower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip *string `json:"tooltip"`
|
||||
Iseventitem *bool `json:"iseventitem"`
|
||||
}
|
||||
|
||||
// GetAllByType implements Querier.GetAllByType.
|
||||
func (q *DBQuerier) GetAllByType(ctx context.Context, itemType int32) ([]GetAllByTypeRow, error) {
|
||||
ctx = context.WithValue(ctx, "pggen_query_name", "GetAllByType")
|
||||
rows, err := q.conn.Query(ctx, getAllByTypeSQL, itemType)
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("query GetAllByType: %w", err)
|
||||
}
|
||||
defer rows.Close()
|
||||
items := []GetAllByTypeRow{}
|
||||
for rows.Next() {
|
||||
var item GetAllByTypeRow
|
||||
if err := rows.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return nil, fmt.Errorf("scan GetAllByType row: %w", err)
|
||||
}
|
||||
items = append(items, item)
|
||||
}
|
||||
if err := rows.Err(); err != nil {
|
||||
return nil, fmt.Errorf("close GetAllByType rows: %w", err)
|
||||
}
|
||||
return items, err
|
||||
}
|
||||
|
||||
// GetAllByTypeBatch implements Querier.GetAllByTypeBatch.
|
||||
func (q *DBQuerier) GetAllByTypeBatch(batch genericBatch, itemType int32) {
|
||||
batch.Queue(getAllByTypeSQL, itemType)
|
||||
}
|
||||
|
||||
// GetAllByTypeScan implements Querier.GetAllByTypeScan.
|
||||
func (q *DBQuerier) GetAllByTypeScan(results pgx.BatchResults) ([]GetAllByTypeRow, error) {
|
||||
rows, err := results.Query()
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("query GetAllByTypeBatch: %w", err)
|
||||
}
|
||||
defer rows.Close()
|
||||
items := []GetAllByTypeRow{}
|
||||
for rows.Next() {
|
||||
var item GetAllByTypeRow
|
||||
if err := rows.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return nil, fmt.Errorf("scan GetAllByTypeBatch row: %w", err)
|
||||
}
|
||||
items = append(items, item)
|
||||
}
|
||||
if err := rows.Err(); err != nil {
|
||||
return nil, fmt.Errorf("close GetAllByTypeBatch rows: %w", err)
|
||||
}
|
||||
return items, err
|
||||
}
|
||||
|
||||
const findByIDSQL = `SELECT
|
||||
id,
|
||||
itemType,
|
||||
iconURL,
|
||||
itemName,
|
||||
minItemPower,
|
||||
maxItemPower,
|
||||
rarity,
|
||||
origin,
|
||||
tooltip,
|
||||
isEventItem
|
||||
FROM
|
||||
powerItem
|
||||
WHERE
|
||||
id = $1;`
|
||||
|
||||
type FindByIDRow struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
Itemtype int32 `json:"itemtype"`
|
||||
Iconurl string `json:"iconurl"`
|
||||
Itemname string `json:"itemname"`
|
||||
Minitempower int32 `json:"minitempower"`
|
||||
Maxitempower int32 `json:"maxitempower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip *string `json:"tooltip"`
|
||||
Iseventitem *bool `json:"iseventitem"`
|
||||
}
|
||||
|
||||
// FindByID implements Querier.FindByID.
|
||||
func (q *DBQuerier) FindByID(ctx context.Context, id pgtype.UUID) (FindByIDRow, error) {
|
||||
ctx = context.WithValue(ctx, "pggen_query_name", "FindByID")
|
||||
row := q.conn.QueryRow(ctx, findByIDSQL, id)
|
||||
var item FindByIDRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("query FindByID: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
// FindByIDBatch implements Querier.FindByIDBatch.
|
||||
func (q *DBQuerier) FindByIDBatch(batch genericBatch, id pgtype.UUID) {
|
||||
batch.Queue(findByIDSQL, id)
|
||||
}
|
||||
|
||||
// FindByIDScan implements Querier.FindByIDScan.
|
||||
func (q *DBQuerier) FindByIDScan(results pgx.BatchResults) (FindByIDRow, error) {
|
||||
row := results.QueryRow()
|
||||
var item FindByIDRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("scan FindByIDBatch row: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
const addNewItemSQL = `INSERT INTO
|
||||
powerItem (
|
||||
itemType,
|
||||
iconURL,
|
||||
itemName,
|
||||
minItemPower,
|
||||
maxItemPower,
|
||||
rarity,
|
||||
origin,
|
||||
tooltip,
|
||||
isEventItem
|
||||
)
|
||||
VALUES
|
||||
(
|
||||
$1,
|
||||
$2,
|
||||
$3,
|
||||
$4,
|
||||
$5,
|
||||
$6,
|
||||
$7,
|
||||
$8,
|
||||
$9
|
||||
) RETURNING *;`
|
||||
|
||||
type AddNewItemParams struct {
|
||||
ItemType int32 `json:"itemType"`
|
||||
IconUrl string `json:"iconUrl"`
|
||||
ItemName string `json:"itemName"`
|
||||
MinItemPower int32 `json:"minItemPower"`
|
||||
MaxItemPower int32 `json:"maxItemPower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip string `json:"tooltip"`
|
||||
IsEventItem bool `json:"isEventItem"`
|
||||
}
|
||||
|
||||
type AddNewItemRow struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
Itemtype int32 `json:"itemtype"`
|
||||
Iconurl string `json:"iconurl"`
|
||||
Itemname string `json:"itemname"`
|
||||
Minitempower int32 `json:"minitempower"`
|
||||
Maxitempower int32 `json:"maxitempower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip *string `json:"tooltip"`
|
||||
Iseventitem *bool `json:"iseventitem"`
|
||||
}
|
||||
|
||||
// AddNewItem implements Querier.AddNewItem.
|
||||
func (q *DBQuerier) AddNewItem(ctx context.Context, params AddNewItemParams) (AddNewItemRow, error) {
|
||||
ctx = context.WithValue(ctx, "pggen_query_name", "AddNewItem")
|
||||
row := q.conn.QueryRow(ctx, addNewItemSQL, params.ItemType, params.IconUrl, params.ItemName, params.MinItemPower, params.MaxItemPower, params.Rarity, params.Origin, params.Tooltip, params.IsEventItem)
|
||||
var item AddNewItemRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("query AddNewItem: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
// AddNewItemBatch implements Querier.AddNewItemBatch.
|
||||
func (q *DBQuerier) AddNewItemBatch(batch genericBatch, params AddNewItemParams) {
|
||||
batch.Queue(addNewItemSQL, params.ItemType, params.IconUrl, params.ItemName, params.MinItemPower, params.MaxItemPower, params.Rarity, params.Origin, params.Tooltip, params.IsEventItem)
|
||||
}
|
||||
|
||||
// AddNewItemScan implements Querier.AddNewItemScan.
|
||||
func (q *DBQuerier) AddNewItemScan(results pgx.BatchResults) (AddNewItemRow, error) {
|
||||
row := results.QueryRow()
|
||||
var item AddNewItemRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("scan AddNewItemBatch row: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
const addNewItemWithIDSQL = `INSERT INTO
|
||||
powerItem (
|
||||
id,
|
||||
itemType,
|
||||
iconURL,
|
||||
itemName,
|
||||
minItemPower,
|
||||
maxItemPower,
|
||||
rarity,
|
||||
origin,
|
||||
tooltip,
|
||||
isEventItem
|
||||
)
|
||||
VALUES
|
||||
(
|
||||
$1,
|
||||
$2,
|
||||
$3,
|
||||
$4,
|
||||
$5,
|
||||
$6,
|
||||
$7,
|
||||
$8,
|
||||
$9,
|
||||
$10
|
||||
) RETURNING *;`
|
||||
|
||||
type AddNewItemWithIDParams struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
ItemType int32 `json:"itemType"`
|
||||
IconUrl string `json:"iconUrl"`
|
||||
ItemName string `json:"itemName"`
|
||||
MinItemPower int32 `json:"minItemPower"`
|
||||
MaxItemPower int32 `json:"maxItemPower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip string `json:"tooltip"`
|
||||
IsEventItem bool `json:"isEventItem"`
|
||||
}
|
||||
|
||||
type AddNewItemWithIDRow struct {
|
||||
ID pgtype.UUID `json:"id"`
|
||||
Itemtype int32 `json:"itemtype"`
|
||||
Iconurl string `json:"iconurl"`
|
||||
Itemname string `json:"itemname"`
|
||||
Minitempower int32 `json:"minitempower"`
|
||||
Maxitempower int32 `json:"maxitempower"`
|
||||
Rarity int32 `json:"rarity"`
|
||||
Origin string `json:"origin"`
|
||||
Tooltip *string `json:"tooltip"`
|
||||
Iseventitem *bool `json:"iseventitem"`
|
||||
}
|
||||
|
||||
// AddNewItemWithID implements Querier.AddNewItemWithID.
|
||||
func (q *DBQuerier) AddNewItemWithID(ctx context.Context, params AddNewItemWithIDParams) (AddNewItemWithIDRow, error) {
|
||||
ctx = context.WithValue(ctx, "pggen_query_name", "AddNewItemWithID")
|
||||
row := q.conn.QueryRow(ctx, addNewItemWithIDSQL, params.ID, params.ItemType, params.IconUrl, params.ItemName, params.MinItemPower, params.MaxItemPower, params.Rarity, params.Origin, params.Tooltip, params.IsEventItem)
|
||||
var item AddNewItemWithIDRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("query AddNewItemWithID: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
// AddNewItemWithIDBatch implements Querier.AddNewItemWithIDBatch.
|
||||
func (q *DBQuerier) AddNewItemWithIDBatch(batch genericBatch, params AddNewItemWithIDParams) {
|
||||
batch.Queue(addNewItemWithIDSQL, params.ID, params.ItemType, params.IconUrl, params.ItemName, params.MinItemPower, params.MaxItemPower, params.Rarity, params.Origin, params.Tooltip, params.IsEventItem)
|
||||
}
|
||||
|
||||
// AddNewItemWithIDScan implements Querier.AddNewItemWithIDScan.
|
||||
func (q *DBQuerier) AddNewItemWithIDScan(results pgx.BatchResults) (AddNewItemWithIDRow, error) {
|
||||
row := results.QueryRow()
|
||||
var item AddNewItemWithIDRow
|
||||
if err := row.Scan(&item.ID, &item.Itemtype, &item.Iconurl, &item.Itemname, &item.Minitempower, &item.Maxitempower, &item.Rarity, &item.Origin, &item.Tooltip, &item.Iseventitem); err != nil {
|
||||
return item, fmt.Errorf("scan AddNewItemWithIDBatch row: %w", err)
|
||||
}
|
||||
return item, nil
|
||||
}
|
||||
|
||||
// textPreferrer wraps a pgtype.ValueTranscoder and sets the preferred encoding
|
||||
// format to text instead binary (the default). pggen uses the text format
|
||||
// when the OID is unknownOID because the binary format requires the OID.
|
||||
// Typically occurs if the results from QueryAllDataTypes aren't passed to
|
||||
// NewQuerierConfig.
|
||||
type textPreferrer struct {
|
||||
pgtype.ValueTranscoder
|
||||
typeName string
|
||||
}
|
||||
|
||||
// PreferredParamFormat implements pgtype.ParamFormatPreferrer.
|
||||
func (t textPreferrer) PreferredParamFormat() int16 { return pgtype.TextFormatCode }
|
||||
|
||||
func (t textPreferrer) NewTypeValue() pgtype.Value {
|
||||
return textPreferrer{ValueTranscoder: pgtype.NewValue(t.ValueTranscoder).(pgtype.ValueTranscoder), typeName: t.typeName}
|
||||
}
|
||||
|
||||
func (t textPreferrer) TypeName() string {
|
||||
return t.typeName
|
||||
}
|
||||
|
||||
// unknownOID means we don't know the OID for a type. This is okay for decoding
|
||||
// because pgx call DecodeText or DecodeBinary without requiring the OID. For
|
||||
// encoding parameters, pggen uses textPreferrer if the OID is unknown.
|
||||
const unknownOID = 0
|
||||
Loading…
Add table
Add a link
Reference in a new issue