File size: 7,461 Bytes
1e92f2d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 |
// Package dataloader provides a generic batching and caching mechanism for loading data
// efficiently. It reduces the number of database queries by batching multiple individual
// requests together and caching results to avoid duplicate fetches.
//
// The primary use case is in GraphQL resolvers where you might have N+1 query problems.
// Instead of making individual database calls for each item, the dataloader batches
// requests together and executes them in a single operation.
//
// Basic usage example:
//
// // Create a loader for User entities
// userLoader := dataloader.NewStoreLoader(ctx, userStore.FindMany, func(u User) string { return u.ID })
//
// // Use in resolvers - these calls will be batched together
// user1, err := userLoader.FetchOne(ctx, "user-id-1")
// user2, err := userLoader.FetchOne(ctx, "user-id-2")
// user3, err := userLoader.FetchOne(ctx, "user-id-3")
//
// // For database operations with a DB connection:
// dbLoader := dataloader.NewStoreLoaderWithDB(ctx, db, dbStore.FindMany, func(u User) string { return u.ID })
//
// For parameterized queries, use FetcherParam:
//
// // Create a parameterized loader
// type UserParams struct { Active bool }
// paramLoader := dataloader.NewStoreLoaderParam(ctx,
// func(ctx context.Context, param UserParams, ids []string) ([]User, error) {
// return userStore.FindManyFiltered(ctx, ids, param.Active)
// },
// func(u User) string { return u.ID },
// )
// activeUser, err := paramLoader.FetchOneParam(ctx, "user-id", UserParams{Active: true})
package dataloader
import (
"context"
"sync"
"time"
"github.com/target/goalert/gadb"
)
type Loader[K comparable, V any] = Fetcher[K, V]
type (
FetchFunc[K comparable, V any] func(context.Context, []K) ([]V, error)
IDFunc[K comparable, V any] func(V) K
)
// NewStoreLoader creates a new Fetcher for loading data from a store without parameters.
// It's a convenience function for the common case where you only need to batch by ID.
//
// The fetchMany function should return values in any order, and the Fetcher will
// map them back to the correct requests using the ID field.
func NewStoreLoader[V any, K comparable](ctx context.Context, fetchMany FetchFunc[K, V], idFunc IDFunc[K, V]) *Fetcher[K, V] {
return &Fetcher[K, V]{
MaxBatch: 100,
Delay: 5 * time.Millisecond,
FetchFunc: func(ctx context.Context, ids []K) ([]V, error) { return fetchMany(ctx, ids) },
IDFunc: idFunc,
}
}
// NewStoreLoaderWithDB creates a new Fetcher that automatically passes a database
// connection to the fetch function. This is a convenience wrapper around NewStoreLoader
// for database-based loading operations.
func NewStoreLoaderWithDB[V any, K comparable](
ctx context.Context,
db gadb.DBTX,
fetchMany func(context.Context, gadb.DBTX, []K) ([]V, error),
idFunc IDFunc[K, V],
) *Fetcher[K, V] {
return NewStoreLoader(ctx, func(ctx context.Context, ids []K) ([]V, error) {
return fetchMany(ctx, db, ids)
}, idFunc)
}
// Fetcher provides batched loading of data with caching. It batches individual requests
// together to reduce the number of calls to the underlying data source, and caches
// results to avoid duplicate fetches within the same Fetcher instance.
//
// Type parameters:
// - K: The type of the unique identifier (key) for items being fetched
// - V: The type of values being fetched
//
// The Fetcher is safe for concurrent use. All methods can be called from multiple
// goroutines simultaneously.
type Fetcher[K comparable, V any] struct {
// FetchFunc is called to retrieve data for a batch of IDs.
// It should return values in any order - the Fetcher will map them back to requests
// using the ID extracted via IDFunc.
FetchFunc func(ctx context.Context, ids []K) ([]V, error)
// IDFunc extracts the unique identifier from a value. This function is required
// and is used to match returned values back to their corresponding requests.
IDFunc func(V) K
// MaxBatch sets the maximum number of IDs to include in a single batch.
// When this limit is reached, a new batch is started immediately.
MaxBatch int
// Delay specifies how long to wait before executing a batch. This allows
// multiple requests to accumulate into a single batch, improving efficiency.
Delay time.Duration
cache map[K]*result[K, V]
currentBatch *batch[K, V]
mx sync.Mutex
wg sync.WaitGroup
}
// result holds the outcome of fetching a single item, including the value, any error,
// and a channel to signal completion.
type result[K comparable, V any] struct {
id K
value *V
err error
done chan struct{}
}
// wait blocks until the result is available or the context is cancelled.
// It returns the fetched value and any error that occurred during fetching.
func (r *result[K, V]) wait(ctx context.Context) (*V, error) {
select {
case <-ctx.Done():
return nil, ctx.Err()
case <-r.done:
return r.value, r.err
}
}
// Close waits for all pending batches to complete. This should be called when
// the Fetcher is no longer needed to ensure proper cleanup and prevent
// goroutine leaks.
func (f *Fetcher[K, V]) Close() {
// Wait for all batches to complete
f.wg.Wait()
}
// fetchAll executes a batch request and distributes the results to waiting goroutines.
// It calls FetchFunc with all the IDs in the batch, then matches returned values
// back to their corresponding requests using IDFunc.
func (f *Fetcher[K, V]) fetchAll(ctx context.Context, batch *batch[K, V]) {
defer f.wg.Done()
f.mx.Lock()
// Before fetching, we check if this is the current batch, if so,
// we remove the pointer so that it isn't changed by another goroutine.
if f.currentBatch == batch {
f.currentBatch = nil // reset for next batch
}
f.mx.Unlock()
// Fetch ouside of the lock to avoid blocking other requests.
values, err := f.FetchFunc(ctx, batch.ids)
f.mx.Lock()
defer f.mx.Unlock()
if err != nil {
// In the error case, we close all results with the error since there is no value to return.
for _, res := range batch.results {
res.err = err
close(res.done)
}
return
}
// Match returned values to their corresponding requests
for _, val := range values {
id := f.IDFunc(val)
res, ok := f.cache[id]
if !ok {
// we didn't ask for this ID, ignore it
continue
}
res.value = &val
}
// Close all result channels - missing values will have nil value
for _, res := range batch.results {
close(res.done)
}
}
// FetchOne retrieves a single value by its ID. Multiple calls to FetchOne within
// the same batch window (defined by Delay) will be batched together into a single
// call to FetchFunc.
//
// The method returns a pointer to the value if found, or nil if not found.
// An error is returned if the fetch operation fails or the context is cancelled.
func (f *Fetcher[K, V]) FetchOne(ctx context.Context, id K) (*V, error) {
f.mx.Lock()
if f.cache == nil {
f.cache = make(map[K]*result[K, V])
}
if res, ok := f.cache[id]; ok {
// easy path, already cached
f.mx.Unlock()
return res.wait(ctx)
}
// create the entry for the new ID
if f.currentBatch == nil {
newBatch := new(batch[K, V])
f.currentBatch = newBatch
f.wg.Add(1)
time.AfterFunc(f.Delay, func() {
f.fetchAll(ctx, newBatch)
})
}
res := f.currentBatch.Add(ctx, id)
f.cache[id] = res
if len(f.currentBatch.ids) >= f.MaxBatch {
f.currentBatch = nil // reset for next batch
}
f.mx.Unlock()
return res.wait(ctx)
}
|