File inference_calculator.pb.h
FileList > calculators > tensor > inference_calculator.pb.h
Go to the source code of this file
#include <limits>
#include <string>
#include <google/protobuf/port_def.inc>
#include <google/protobuf/port_undef.inc>
#include <google/protobuf/io/coded_stream.h>
#include <google/protobuf/arena.h>
#include <google/protobuf/arenastring.h>
#include <google/protobuf/generated_message_bases.h>
#include <google/protobuf/generated_message_table_driven.h>
#include <google/protobuf/generated_message_util.h>
#include <google/protobuf/metadata_lite.h>
#include <google/protobuf/generated_message_reflection.h>
#include <google/protobuf/message.h>
#include <google/protobuf/repeated_field.h>
#include <google/protobuf/extension_set.h>
#include <google/protobuf/generated_enum_reflection.h>
#include <google/protobuf/unknown_field_set.h>
#include "mediapipe/framework/calculator.pb.h"
#include "mediapipe/framework/calculator_options.pb.h"
Namespaces
Type | Name |
---|---|
namespace | internal |
namespace | mediapipe |
Classes
Public Attributes
Type | Name |
---|---|
const ::PROTOBUF_NAMESPACE_ID::internal::DescriptorTable | descriptor_table_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto |
Public Functions
Macros
Type | Name |
---|---|
define | PROTOBUF_INTERNAL_EXPORT_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto |
Public Attributes Documentation
variable descriptor_table_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto
const ::PROTOBUF_NAMESPACE_ID::internal::DescriptorTable descriptor_table_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto;
Public Functions Documentation
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions >
template<>
PROTOBUF_NAMESPACE_OPEN ::mediapipe::InferenceCalculatorOptions * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate >
template<>
::mediapipe::InferenceCalculatorOptions_Delegate * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Gpu >
template<>
::mediapipe::InferenceCalculatorOptions_Delegate_Gpu * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Gpu > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Nnapi >
template<>
::mediapipe::InferenceCalculatorOptions_Delegate_Nnapi * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Nnapi > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_TfLite >
template<>
::mediapipe::InferenceCalculatorOptions_Delegate_TfLite * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_TfLite > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Xnnpack >
template<>
::mediapipe::InferenceCalculatorOptions_Delegate_Xnnpack * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_Delegate_Xnnpack > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig >
template<>
::mediapipe::InferenceCalculatorOptions_InputOutputConfig * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_FeedbackTensorLink >
template<>
::mediapipe::InferenceCalculatorOptions_InputOutputConfig_FeedbackTensorLink * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_FeedbackTensorLink > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorIndicesMap >
template<>
::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorIndicesMap * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorIndicesMap > (
Arena *
)
function CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorNamesMap >
template<>
::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorNamesMap * CreateMaybeMessage<::mediapipe::InferenceCalculatorOptions_InputOutputConfig_TensorNamesMap > (
Arena *
)
function GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_Api >
template<>
inline const EnumDescriptor * GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_Api > ()
function GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_CacheWritingBehavior >
template<>
inline const EnumDescriptor * GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_CacheWritingBehavior > ()
function GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_InferenceUsage >
template<>
inline const EnumDescriptor * GetEnumDescriptor< ::mediapipe::InferenceCalculatorOptions_Delegate_Gpu_InferenceUsage > ()
Macro Definition Documentation
define PROTOBUF_INTERNAL_EXPORT_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto
#define PROTOBUF_INTERNAL_EXPORT_mediapipe_2fcalculators_2ftensor_2finference_5fcalculator_2eproto
The documentation for this class was generated from the following file /home/friedel/devel/ILLIXR-plugins/hand_tracking/build/mediapipe/calculators/tensor/inference_calculator.pb.h