// Package dataloader provides a generic batching and caching mechanism for loading data // efficiently. It reduces the number of database queries by batching multiple individual // requests together and caching results to avoid duplicate fetches. // // The primary use case is in GraphQL resolvers where you might have N+1 query problems. // Instead of making individual database calls for each item, the dataloader batches // requests together and executes them in a single operation. // // Basic usage example: // // // Create a loader for User entities // userLoader := dataloader.NewStoreLoader(ctx, userStore.FindMany, func(u User) string { return u.ID }) // // // Use in resolvers - these calls will be batched together // user1, err := userLoader.FetchOne(ctx, "user-id-1") // user2, err := userLoader.FetchOne(ctx, "user-id-2") // user3, err := userLoader.FetchOne(ctx, "user-id-3") // // // For database operations with a DB connection: // dbLoader := dataloader.NewStoreLoaderWithDB(ctx, db, dbStore.FindMany, func(u User) string { return u.ID }) // // For parameterized queries, use FetcherParam: // // // Create a parameterized loader // type UserParams struct { Active bool } // paramLoader := dataloader.NewStoreLoaderParam(ctx, // func(ctx context.Context, param UserParams, ids []string) ([]User, error) { // return userStore.FindManyFiltered(ctx, ids, param.Active) // }, // func(u User) string { return u.ID }, // ) // activeUser, err := paramLoader.FetchOneParam(ctx, "user-id", UserParams{Active: true}) package dataloader import ( "context" "sync" "time" "github.com/target/goalert/gadb" ) type Loader[K comparable, V any] = Fetcher[K, V] type ( FetchFunc[K comparable, V any] func(context.Context, []K) ([]V, error) IDFunc[K comparable, V any] func(V) K ) // NewStoreLoader creates a new Fetcher for loading data from a store without parameters. // It's a convenience function for the common case where you only need to batch by ID. // // The fetchMany function should return values in any order, and the Fetcher will // map them back to the correct requests using the ID field. func NewStoreLoader[V any, K comparable](ctx context.Context, fetchMany FetchFunc[K, V], idFunc IDFunc[K, V]) *Fetcher[K, V] { return &Fetcher[K, V]{ MaxBatch: 100, Delay: 5 * time.Millisecond, FetchFunc: func(ctx context.Context, ids []K) ([]V, error) { return fetchMany(ctx, ids) }, IDFunc: idFunc, } } // NewStoreLoaderWithDB creates a new Fetcher that automatically passes a database // connection to the fetch function. This is a convenience wrapper around NewStoreLoader // for database-based loading operations. func NewStoreLoaderWithDB[V any, K comparable]( ctx context.Context, db gadb.DBTX, fetchMany func(context.Context, gadb.DBTX, []K) ([]V, error), idFunc IDFunc[K, V], ) *Fetcher[K, V] { return NewStoreLoader(ctx, func(ctx context.Context, ids []K) ([]V, error) { return fetchMany(ctx, db, ids) }, idFunc) } // Fetcher provides batched loading of data with caching. It batches individual requests // together to reduce the number of calls to the underlying data source, and caches // results to avoid duplicate fetches within the same Fetcher instance. // // Type parameters: // - K: The type of the unique identifier (key) for items being fetched // - V: The type of values being fetched // // The Fetcher is safe for concurrent use. All methods can be called from multiple // goroutines simultaneously. type Fetcher[K comparable, V any] struct { // FetchFunc is called to retrieve data for a batch of IDs. // It should return values in any order - the Fetcher will map them back to requests // using the ID extracted via IDFunc. FetchFunc func(ctx context.Context, ids []K) ([]V, error) // IDFunc extracts the unique identifier from a value. This function is required // and is used to match returned values back to their corresponding requests. IDFunc func(V) K // MaxBatch sets the maximum number of IDs to include in a single batch. // When this limit is reached, a new batch is started immediately. MaxBatch int // Delay specifies how long to wait before executing a batch. This allows // multiple requests to accumulate into a single batch, improving efficiency. Delay time.Duration cache map[K]*result[K, V] currentBatch *batch[K, V] mx sync.Mutex wg sync.WaitGroup } // result holds the outcome of fetching a single item, including the value, any error, // and a channel to signal completion. type result[K comparable, V any] struct { id K value *V err error done chan struct{} } // wait blocks until the result is available or the context is cancelled. // It returns the fetched value and any error that occurred during fetching. func (r *result[K, V]) wait(ctx context.Context) (*V, error) { select { case <-ctx.Done(): return nil, ctx.Err() case <-r.done: return r.value, r.err } } // Close waits for all pending batches to complete. This should be called when // the Fetcher is no longer needed to ensure proper cleanup and prevent // goroutine leaks. func (f *Fetcher[K, V]) Close() { // Wait for all batches to complete f.wg.Wait() } // fetchAll executes a batch request and distributes the results to waiting goroutines. // It calls FetchFunc with all the IDs in the batch, then matches returned values // back to their corresponding requests using IDFunc. func (f *Fetcher[K, V]) fetchAll(ctx context.Context, batch *batch[K, V]) { defer f.wg.Done() f.mx.Lock() // Before fetching, we check if this is the current batch, if so, // we remove the pointer so that it isn't changed by another goroutine. if f.currentBatch == batch { f.currentBatch = nil // reset for next batch } f.mx.Unlock() // Fetch ouside of the lock to avoid blocking other requests. values, err := f.FetchFunc(ctx, batch.ids) f.mx.Lock() defer f.mx.Unlock() if err != nil { // In the error case, we close all results with the error since there is no value to return. for _, res := range batch.results { res.err = err close(res.done) } return } // Match returned values to their corresponding requests for _, val := range values { id := f.IDFunc(val) res, ok := f.cache[id] if !ok { // we didn't ask for this ID, ignore it continue } res.value = &val } // Close all result channels - missing values will have nil value for _, res := range batch.results { close(res.done) } } // FetchOne retrieves a single value by its ID. Multiple calls to FetchOne within // the same batch window (defined by Delay) will be batched together into a single // call to FetchFunc. // // The method returns a pointer to the value if found, or nil if not found. // An error is returned if the fetch operation fails or the context is cancelled. func (f *Fetcher[K, V]) FetchOne(ctx context.Context, id K) (*V, error) { f.mx.Lock() if f.cache == nil { f.cache = make(map[K]*result[K, V]) } if res, ok := f.cache[id]; ok { // easy path, already cached f.mx.Unlock() return res.wait(ctx) } // create the entry for the new ID if f.currentBatch == nil { newBatch := new(batch[K, V]) f.currentBatch = newBatch f.wg.Add(1) time.AfterFunc(f.Delay, func() { f.fetchAll(ctx, newBatch) }) } res := f.currentBatch.Add(ctx, id) f.cache[id] = res if len(f.currentBatch.ids) >= f.MaxBatch { f.currentBatch = nil // reset for next batch } f.mx.Unlock() return res.wait(ctx) }