Go to the documentation of this file. 1 #ifndef HeterogeneousCore_SonicTriton_TritonClient
2 #define HeterogeneousCore_SonicTriton_TritonClient
14 #include <unordered_map>
16 #include "grpc_client.h"
17 #include "grpc_service.pb.h"
44 void reset()
override;
60 const inference::ModelStatistics& end_status)
const;
78 std::unique_ptr<triton::client::InferenceServerGrpcClient>
client_;
uint64_t inference_count_
std::vector< const triton::client::InferRequestedOutput * > outputsTriton_
bool useSharedMemory() const
ServerSideStats summarizeServerStats(const inference::ModelStatistics &start_status, const inference::ModelStatistics &end_status) const
bool setBatchSize(unsigned bsize)
const std::string & debugName() const
static uInt32 F(BLOWFISH_CTX *ctx, uInt32 x)
TritonServerType serverType() const
std::unique_ptr< triton::client::InferenceServerGrpcClient > client_
uint64_t compute_infer_time_ns_
uint64_t compute_input_time_ns_
grpc_compression_algorithm compressionAlgo_
TritonServerType serverType_
TritonClient(const edm::ParameterSet ¶ms, const std::string &debugName)
std::vector< triton::client::InferInput * > inputsTriton_
void reportServerSideStats(const ServerSideStats &stats) const
uint64_t execution_count_
uint64_t compute_output_time_ns_
bool handle_exception(F &&call)
static void fillPSetDescription(edm::ParameterSetDescription &iDesc)
triton::client::Headers headers_
unsigned long long uint64_t
inference::ModelStatistics getServerSideStatus() const
void getResults(std::shared_ptr< triton::client::InferResult > results)
unsigned batchSize() const
void setUseSharedMemory(bool useShm)
triton::client::InferOptions options_