batched Rust macro utility for batching expensive async operations. Installation cargo add batched Or add this to your Cargo.toml : [ dependencies ] batched = " 0.2.7 " limit : Maximum amount of items that can be grouped and processed in a single batch. : Maximum amount of items that can be grouped and processed in a single batch. concurrent : Maximum amount of concurrent batched tasks running (default: Infinity ) : Maximum amount of concurrent batched tasks running (default: ) window : Minimum amount of time (in milliseconds) the background thread waits before processing a batch. : Minimum amount of time (in milliseconds) the background thread waits before processing a batch. window[x]: Minimum amount of time (in milliseconds) the background thread waits before processing a batch when latest buffer size is <= x The target function must have a single argument, a vector of items ( Vec ). The return value of the batched function is propagated (cloned) to all async calls of the batch, unless the batched function returns a Vec , in which case the return value for each call is pulled from the iterator in the same order of the input. If the return value is not an iterator, The target function return type must implement Clone to propagate the result. Use batched::error::SharedError to wrap your error types (if they don't implement Clone). Prerequisites Built for async environments (tokio), will not work without a tokio async runtime Target function must have async Not supported inside structs: struct A ; impl A { # [ batched ( window = 1000 , limit = 100 ) ] fn operation ( ) { .. . } } Tracing This feature automatically adds tracing spans to call functions for batched requests ( x , x_multiple ). This feature adds support for linking spans from callers to the inner batched call when using OpenTelemetry. Depending on whether your OpenTelemetry client supports it, you should be able to see the linked span to the batched call. Examples Simple add batch # [ batched ( window = 100 , limit = 1000 ) ] async fn add ( numbers : Vec < u32 > ) -> u32 { numbers . iter ( ) . sum ( ) } async fn main ( ) { for _ in 0 .. 99 { tokio :: task :: spawn ( async move { add ( 1 ) . await } ) ; } let result = add ( 1 ) . await ; assert_eq ! ( result , 100 ) ; } Batch insert rows use batched :: { batched , error :: SharedError } ; // Macros creates functions [`insert_message`] and [`insert_message_multiple`] # [ batched ( window = 100 , window1 = 10 , window5 = 20 , limit = 100_000 ) ] async fn insert_message ( messages : Vec < String > ) -> Result < ( ) , SharedError < anyhow :: Error > > { let pool = PgPool :: connect ( "postgres://user:password@localhost/dbname" ) . await ? ; let mut query = String :: from ( "INSERT INTO messages (content) VALUES " ) ; .. . } # [ post ( "/message" ) ] async fn service ( message : String ) -> Result < ( ) , anyhow :: Error > { insert_message ( message ) . await ? ; Ok ( ( ) ) } # [ post ( "/bulk_messages" ) ] async fn service ( messages : Vec < String > ) -> Result < ( ) , anyhow :: Error > { insert_message_multiple ( messages ) . await ? ; Ok ( ( ) ) } Batch insert rows and return them