1 #ifndef HeterogeneousCore_SonicTriton_TritonClient 2 #define HeterogeneousCore_SonicTriton_TritonClient 14 #include <unordered_map> 16 #include "grpc_client.h" 17 #include "grpc_service.pb.h" 44 void reset()
override;
60 const inference::ModelStatistics& end_status)
const;
78 std::unique_ptr<triton::client::InferenceServerGrpcClient>
client_;
uint64_t execution_count_
const std::string & debugName() const
bool setBatchSize(unsigned bsize)
std::vector< const triton::client::InferRequestedOutput * > outputsTriton_
std::unique_ptr< triton::client::InferenceServerGrpcClient > client_
TritonClient(const edm::ParameterSet ¶ms, const std::string &debugName)
uint64_t compute_infer_time_ns_
uint64_t inference_count_
ServerSideStats summarizeServerStats(const inference::ModelStatistics &start_status, const inference::ModelStatistics &end_status) const
TritonServerType serverType_
bool handle_exception(F &&call)
grpc_compression_algorithm compressionAlgo_
TritonServerType serverType() const
inference::ModelStatistics getServerSideStatus() const
triton::client::Headers headers_
unsigned long long uint64_t
std::vector< triton::client::InferInput * > inputsTriton_
void setUseSharedMemory(bool useShm)
triton::client::InferOptions options_
unsigned batchSize() const
void getResults(std::shared_ptr< triton::client::InferResult > results)
void reportServerSideStats(const ServerSideStats &stats) const
bool useSharedMemory() const
uint64_t compute_output_time_ns_
static void fillPSetDescription(edm::ParameterSetDescription &iDesc)
static uInt32 F(BLOWFISH_CTX *ctx, uInt32 x)
uint64_t compute_input_time_ns_