platform: "pytorch_libtorch" max_batch_size: 128 input [ { name: "INPUT__0" data_type: TYPE_INT32 dims: [128] }, { name: "INPUT__1" data_type: TYPE_INT32 dims: [128] } ] output [ { name: "OUTPUT__0" data_type: TYPE_FP32 dims: [128, 768] }, { name: "1634__1" data_type: TYPE_FP32 dims: [768] } ] instance_group { count: 1 kind: KIND_GPU } dynamic_batching { preferred_batch_size: 16 }