1 #ifndef NuSonic_Triton_TritonClient 2 #define NuSonic_Triton_TritonClient 11 #include <unordered_map> 13 #include "grpc_client.h" 14 #include "grpc_service.pb.h" 52 bool getResults(std::shared_ptr<nvidia::inferenceserver::client::InferResult> results);
60 const inference::ModelStatistics& end_status)
const;
75 std::vector<nvidia::inferenceserver::client::InferInput*>
inputsTriton_;
76 std::vector<const nvidia::inferenceserver::client::InferRequestedOutput*>
outputsTriton_;
78 std::unique_ptr<nvidia::inferenceserver::client::InferenceServerGrpcClient>
client_;
80 nvidia::inferenceserver::client::InferOptions
options_;
std::vector< const nvidia::inferenceserver::client::InferRequestedOutput * > outputsTriton_
ServerSideStats summarizeServerStats(const inference::ModelStatistics &start_status, const inference::ModelStatistics &end_status) const
uint64_t compute_output_time_ns_
nvidia::inferenceserver::client::InferOptions options_
std::unordered_map< std::string, TritonOutputData > TritonOutputMap
void reportServerSideStats(const ServerSideStats &stats) const
const TritonOutputMap & output() const
uint64_t compute_infer_time_ns_
uint64_t compute_input_time_ns_
std::unique_ptr< nvidia::inferenceserver::client::InferenceServerGrpcClient > client_
bool getResults(std::shared_ptr< nvidia::inferenceserver::client::InferResult > results)
void finish(bool success)
inference::ModelStatistics getServerSideStatus() const
uint64_t inference_count_
unsigned batchSize() const
bool setBatchSize(unsigned bsize)
TritonClient(const fhicl::ParameterSet ¶ms)
uint64_t execution_count_
std::vector< nvidia::inferenceserver::client::InferInput * > inputsTriton_
std::unordered_map< std::string, TritonInputData > TritonInputMap