Lines Matching refs:input
361 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
365 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
366 success = floorPrepare(input.shape(), &outShape) && in executeOperation()
368 floorFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
377 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
381 if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
382 success = dequantizePrepare(input.shape(), &outShape) && in executeOperation()
385 reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
387 input.shape()); in executeOperation()
396 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
422 Shape inputShape = input.shape(); in executeOperation()
439 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
440 success = depthwiseConvPrepare(input.shape(), filter.shape(), bias.shape(), in executeOperation()
446 depthwiseConvFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
447 input.shape(), in executeOperation()
458 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
459 success = depthwiseConvPrepare(input.shape(), filter.shape(), bias.shape(), in executeOperation()
465 depthwiseConvQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
466 input.shape(), in executeOperation()
486 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
509 Shape inputShape = input.shape(); in executeOperation()
526 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
527 success = convPrepare(input.shape(), filter.shape(), bias.shape(), in executeOperation()
533 convFloat32(reinterpret_cast<const float*>(input.buffer), input.shape(), in executeOperation()
540 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
541 success = convPrepare(input.shape(), filter.shape(), bias.shape(), in executeOperation()
547 convQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
548 input.shape(), in executeOperation()
566 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
592 Shape inputShape = input.shape(); in executeOperation()
606 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
607 success = genericPoolingPrepare(input.shape(), in executeOperation()
614 averagePoolFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
615 input.shape(), in executeOperation()
622 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
623 success = genericPoolingPrepare(input.shape(), in executeOperation()
630 averagePoolQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
631 input.shape(), in executeOperation()
646 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
672 Shape inputShape = input.shape(); in executeOperation()
686 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
687 success = genericPoolingPrepare(input.shape(), in executeOperation()
694 l2PoolFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
695 input.shape(), in executeOperation()
710 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
736 Shape inputShape = input.shape(); in executeOperation()
750 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
751 success = genericPoolingPrepare(input.shape(), in executeOperation()
758 maxPoolFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
759 input.shape(), in executeOperation()
766 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
767 success = genericPoolingPrepare(input.shape(), in executeOperation()
774 maxPoolQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
775 input.shape(), in executeOperation()
789 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
793 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
794 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
796 reluFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
797 input.shape(), in executeOperation()
800 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
801 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
803 reluQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
804 input.shape(), in executeOperation()
813 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
817 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
818 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
820 relu1Float32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
821 input.shape(), in executeOperation()
824 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
825 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
827 relu1Quant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
828 input.shape(), in executeOperation()
837 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
841 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
842 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
844 relu6Float32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
845 input.shape(), in executeOperation()
848 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
849 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
851 relu6Quant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
852 input.shape(), in executeOperation()
861 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
865 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
866 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
868 tanhFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
869 input.shape(), in executeOperation()
878 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
882 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
883 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
885 logisticFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
886 input.shape(), in executeOperation()
889 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
890 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
892 logisticQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
893 input.shape(), in executeOperation()
902 RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
912 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
913 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
915 softmaxFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
916 input.shape(), in executeOperation()
920 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
921 success = genericActivationPrepare(input.shape(), &outShape) && in executeOperation()
923 softmaxQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
924 input.shape(), in executeOperation()
934 RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
943 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
944 success = fullyConnectedPrepare(input.shape(), weights.shape(), bias.shape(), in executeOperation()
947 fullyConnectedFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
948 input.shape(), in executeOperation()
956 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
957 success = fullyConnectedPrepare(input.shape(), weights.shape(), bias.shape(), in executeOperation()
960 fullyConnectedQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
961 input.shape(), in executeOperation()
987 RunTimeOperandInfo& input = mOperands[ins[i]]; in executeOperation() local
988 inputShapes[i] = input.shape(); in executeOperation()
989 inputDataPtrs[i] = reinterpret_cast<const float*>(input.buffer); in executeOperation()
1000 RunTimeOperandInfo& input = mOperands[ins[i]]; in executeOperation() local
1001 inputShapes[i] = input.shape(); in executeOperation()
1002 inputDataPtrs[i] = reinterpret_cast<const uint8_t*>(input.buffer); in executeOperation()
1015 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1019 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
1020 success = genericNormalizationPrepare(input.shape(), &outShape) && in executeOperation()
1022 l2normFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
1023 input.shape(), in executeOperation()
1026 } else if (input.type == OperandType::TENSOR_QUANT8_ASYMM) { in executeOperation()
1027 success = genericNormalizationPrepare(input.shape(), &outShape) && in executeOperation()
1029 l2normQuant8(reinterpret_cast<const uint8_t*>(input.buffer), in executeOperation()
1030 input.shape(), in executeOperation()
1039 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1048 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
1049 success = genericNormalizationPrepare(input.shape(), &outShape) && in executeOperation()
1051 localResponseNormFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
1052 input.shape(), in executeOperation()
1062 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1068 success = reshapePrepare(input.shape(), in executeOperation()
1073 reshapeGeneric(reinterpret_cast<const void*>(input.buffer), in executeOperation()
1074 input.shape(), in executeOperation()
1082 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1089 if (input.type == OperandType::TENSOR_FLOAT32) { in executeOperation()
1090 success = resizeBilinearPrepare(input.shape(), in executeOperation()
1094 resizeBilinearFloat32(reinterpret_cast<const float*>(input.buffer), in executeOperation()
1095 input.shape(), in executeOperation()
1104 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1110 success = depthToSpacePrepare(input.shape(), in executeOperation()
1114 depthToSpaceGeneric(input.buffer, in executeOperation()
1115 input.shape(), in executeOperation()
1124 const RunTimeOperandInfo& input = mOperands[ins[0]]; in executeOperation() local
1130 success = spaceToDepthPrepare(input.shape(), in executeOperation()
1134 spaceToDepthGeneric(input.buffer, in executeOperation()
1135 input.shape(), in executeOperation()