/aosp_15_r20/external/rust/android-crates-io/crates/tower-layer/src/ |
D | tuple.rs | 1 use crate::Layer; 3 impl<S> Layer<S> for () { implementation 6 fn layer(&self, service: S) -> Self::Service { in layer() method 11 impl<S, L1> Layer<S> for (L1,) impl 13 L1: Layer<S>, 17 fn layer(&self, service: S) -> Self::Service { in layer() function 19 l1.layer(service) in layer() 23 impl<S, L1, L2> Layer<S> for (L1, L2) impl 25 L1: Layer<L2::Service>, 26 L2: Layer<S>, [all …]
|
/aosp_15_r20/external/armnn/python/pyarmnn/src/pyarmnn/swig/modules/ |
H A D | armnn_network.i | 35 of the layer and any mismatch is reported. 70 An input connection slot for a layer. Slot lifecycle is managed by the layer. 72 The input slot can be connected to an output slot of the preceding layer in the graph. 83 Returns output slot of a preceding layer that is connected to the given input slot. 86 IOutputSlot: Borrowed reference to an output connection slot for a preceding layer. 95 An output connection slot for a layer. Slot lifecycle is managed by the layer. 200 Calculates the index of this slot for the layer. 210 Returns the index of the layer. Same value as `IConnectableLayer.GetGuid`. 213 int: Layer id. 234 Interface for a layer that is connectable to other layers via `IInputSlot` and `IOutputSlot`. [all …]
|
/aosp_15_r20/external/armnn/src/armnnTestUtils/ |
H A D | CreateWorkload.hpp | 32 // Calls CreateWorkload for a layer, and checks the returned pointer is of the correct type. 34 std::unique_ptr<Workload> MakeAndCheckWorkload(Layer& layer, in MakeAndCheckWorkload() argument 38 std::unique_ptr<IWorkload> workload = layer.CreateWorkload(factory); 42 layer.SetBackendId(factory.GetBackendId()); 43 CHECK(factory.IsLayerSupported(layer, layer.GetDataType(), reasonIfUnsupported, modelOptions)); 52 for (auto&& layer : graph.TopologicalSort()) in CreateTensorHandles() 54 layer->CreateTensorHandles(tmpRegistry, factory); in CreateTensorHandles() 69 // Creates the layer we're testing. in CreateActivationWorkloadTest() 75 ActivationLayer* const layer = graph.AddLayer<ActivationLayer>(layerDesc, "layer"); in CreateActivationWorkloadTest() local 78 Layer* const input = graph.AddLayer<InputLayer>(0, "input"); in CreateActivationWorkloadTest() [all …]
|
/aosp_15_r20/external/armnn/include/armnn/ |
H A D | INetwork.hpp | 22 /// @brief An input connection slot for a layer. 23 /// The input slot can be connected to an output slot of the preceding layer in the graph. 39 /// @brief An output connection slot for a layer. 67 /// @brief Interface for a layer that is connectable to other layers via InputSlots and OutputSlots. 71 /// Returns the name of the layer 95 /// Returns the unique id of the layer 98 /// Apply a visitor to this layer 101 /// Provide a hint for the optimizer as to which backend to prefer for this layer. 102 …// By providing a BackendSelectionHint there is no guarantee the input backend supports that layer. 105 /// layer (IsLayerSupported returns true for a specific backend). [all …]
|
/aosp_15_r20/frameworks/native/services/surfaceflinger/FrontEnd/ |
H A D | LayerLifecycleManager.cpp | 33 // Returns true if the layer is root of a display and can be mirrored by mirroringLayer 47 RequestedLayerState& layer = *newLayer.get(); in addLayers() local 48 auto [it, inserted] = mIdToLayer.try_emplace(layer.id, References{.owner = layer}); in addLayers() 50 "Duplicate layer id found. New layer: %s Existing layer: " in addLayers() 52 layer.getDebugString().c_str(), in addLayers() 56 layer.parentId = linkLayer(layer.parentId, layer.id); in addLayers() 57 layer.relativeParentId = linkLayer(layer.relativeParentId, layer.id); in addLayers() 58 if (layer.layerStackToMirror != ui::INVALID_LAYER_STACK) { in addLayers() 59 // Set mirror layer's default layer stack to -1 so it doesn't end up rendered on a in addLayers() 61 layer.layerStack = ui::INVALID_LAYER_STACK; in addLayers() [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/saving/saved_model/ |
H A D | save_impl.py | 71 def should_skip_serialization(layer): argument 72 """Skip serializing extra objects and functions if layer inputs aren't set.""" 73 saved_model_input_spec_set = (isinstance(layer, training_lib.Model) and 74 … layer._saved_model_inputs_spec is not None) # pylint: disable=protected-access 75 if not layer.built and not saved_model_input_spec_set: 76 logging.warning('Skipping full serialization of Keras layer {}, because ' 77 'it is not built.'.format(layer)) 82 def wrap_layer_objects(layer, serialization_cache): argument 83 """Returns extra trackable objects to attach to the serialized layer. 86 layer: Keras Layer object. [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/mixed_precision/ |
H A D | layer_test.py | 15 """Tests keras.layers.Layer works properly with mixed precision.""" 98 layer = mp_test_util.MultiplyLayer(assert_type=dtype) 99 self.assertEqual(layer.dtype, dtypes.float32) 100 self.assertEqual(get_layer_policy.get_layer_policy(layer).name, 102 y = layer(x) 103 self.assertEqual(layer.v.dtype, dtypes.float32) 105 self.assertEqual(layer.dtype_policy.name, policy_name) 106 self.assertIsInstance(layer.dtype_policy, policy.Policy) 107 self.assertEqual(layer.compute_dtype, dtype) 108 self.assertEqual(layer.dtype, dtypes.float32) [all …]
|
/aosp_15_r20/external/armnn/src/armnnSerializer/ |
H A D | Serializer.cpp | 114 // Build FlatBuffer for Input Layer 115 void SerializerStrategy::SerializeInputLayer(const armnn::IConnectableLayer* layer, LayerBindingId … in SerializeInputLayer() argument 120 auto flatBufferInputBaseLayer = CreateLayerBase(layer, serializer::LayerType::LayerType_Input); in SerializeInputLayer() 126 // Push layer binding id to outputIds. in SerializeInputLayer() 133 CreateAnyLayer(flatBufferInputLayer.o, serializer::Layer::Layer_InputLayer); in SerializeInputLayer() 136 // Build FlatBuffer for Output Layer 137 void SerializerStrategy::SerializeOutputLayer(const armnn::IConnectableLayer* layer, in SerializeOutputLayer() argument 143 … auto flatBufferOutputBaseLayer = CreateLayerBase(layer, serializer::LayerType::LayerType_Output); in SerializeOutputLayer() 149 // Push layer binding id to outputIds. in SerializeOutputLayer() 155 CreateAnyLayer(flatBufferOutputLayer.o, serializer::Layer::Layer_OutputLayer); in SerializeOutputLayer() [all …]
|
H A D | Serializer.hpp | 25 void ExecuteStrategy(const armnn::IConnectableLayer* layer, 57 /// Creates the Input Slots and Output Slots and LayerBase for the layer. 59 const armnn::IConnectableLayer* layer, 62 /// Creates the serializer AnyLayer for the layer and adds it to m_serializedLayers. 63 …void CreateAnyLayer(const flatbuffers::Offset<void>& layer, const armnnSerializer::Layer serialize… 78 /// Creates the serializer InputSlots for the layer. 80 const armnn::IConnectableLayer* layer); 82 /// Creates the serializer OutputSlots for the layer. 84 const armnn::IConnectableLayer* layer); 101 /// layer within our FlatBuffer index. [all …]
|
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/engine/ |
H A D | functional.py | 95 # The key of _layer_call_argspecs is a layer. tf.Module._flatten will fail to 96 # flatten the key since it is trying to convert Trackable/Layer to a string. 120 # layer is added or removed. 182 layer, node_index, tensor_index = x._keras_history # pylint: disable=protected-access 183 self._output_layers.append(layer) 184 self._output_coordinates.append((layer, node_index, tensor_index)) 188 layer, node_index, tensor_index = x._keras_history # pylint: disable=protected-access 189 # It's supposed to be an input layer, so only one node 193 self._input_layers.append(layer) 194 self._input_coordinates.append((layer, node_index, tensor_index)) [all …]
|
H A D | base_layer.py | 16 """Contains the base Layer class, from which all layers inherit.""" 88 # Prefix that is added to the TF op layer names. 97 @keras_export('keras.layers.Layer') 98 class Layer(module.Module, version_utils.LayerVersionSelector): class 101 A layer is a callable object that takes as input one or more tensors and 106 Users will just instantiate a layer and then treat it as a callable. 109 trainable: Boolean, whether the layer's variables should be trainable. 110 name: String name of the layer. 111 dtype: The dtype of the layer's computations and weights. Can also be a 116 dynamic: Set this to `True` if your layer should only be run eagerly, and [all …]
|
H A D | base_layer_v1.py | 16 """Contains the base Layer class, from which all layers inherit.""" 70 class Layer(base_layer.Layer): class 71 """Base layer class. 75 A layer is a class implementing common neural networks operations, such 77 losses, updates, and inter-layer connectivity. 79 Users will just instantiate a layer and then treat it as a callable. 81 We recommend that descendants of `Layer` implement the following methods: 90 once. Should actually perform the logic of applying the layer to the 94 trainable: Boolean, whether the layer's variables should be trainable. 95 name: String name of the layer. [all …]
|
/aosp_15_r20/frameworks/native/services/surfaceflinger/Tracing/ |
H A D | TransactionProtoParser.cpp | 97 auto& layer = resolvedComposerState.state; in toProto() local 99 proto.set_what(layer.what); in toProto() 101 if (layer.what & layer_state_t::ePositionChanged) { in toProto() 102 proto.set_x(layer.x); in toProto() 103 proto.set_y(layer.y); in toProto() 105 if (layer.what & layer_state_t::eLayerChanged) { in toProto() 106 proto.set_z(layer.z); in toProto() 109 if (layer.what & layer_state_t::eLayerStackChanged) { in toProto() 110 proto.set_layer_stack(layer.layerStack.id); in toProto() 112 if (layer.what & layer_state_t::eFlagsChanged) { in toProto() [all …]
|
/aosp_15_r20/external/rust/android-crates-io/crates/tracing-subscriber/src/layer/ |
D | mod.rs | 1 //! The [`Layer`] trait, a composable abstraction for building [`Subscriber`]s. 18 //! span IDs. The [`Layer`] trait represents this composable subset of the 24 //! Since a [`Layer`] does not implement a complete strategy for collecting 26 //! [`Layer`] trait is generic over a type parameter (called `S` in the trait 28 //! with. Thus, a [`Layer`] may be implemented that will only compose with a 30 //! added to constrain what types implementing `Subscriber` a `Layer` can wrap. 32 //! `Layer`s may be added to a `Subscriber` by using the [`SubscriberExt::with`] 35 //! `Layer` with the `Subscriber`. 39 //! use tracing_subscriber::Layer; 47 //! impl<S: Subscriber> Layer<S> for MyLayer { [all …]
|
D | layered.rs | 5 layer::{Context, Layer}, 17 /// [`Layer`]s. 19 /// [`Layer`]: crate::Layer 23 /// The layer. 24 layer: L, field 26 /// The inner value that `self.layer` was layered onto. 28 /// If this is also a `Layer`, then this `Layered` will implement `Layer`. 34 // level hints when per-layer filters are in use. 41 /// Does `self.layer` have per-layer filters? 46 /// `Layered`s have per-layer filters. [all …]
|
/aosp_15_r20/external/armnn/src/armnn/ |
H A D | Network.cpp | 8 #include "Layer.hpp" 757 const Layer* layer, in ReturnWithError() argument 762 failureMsg << "Layer of type " << GetLayerTypeAsCString(layer->GetType()) in ReturnWithError() 771 bool CheckScaleSetOnQuantizedType(Layer* layer, Optional<std::vector<std::string>&> errMessages) in CheckScaleSetOnQuantizedType() argument 774 unsigned int numOutputs = layer->GetNumOutputSlots(); in CheckScaleSetOnQuantizedType() 776 OutputSlot& outputSlot = layer->GetOutputSlot(i); in CheckScaleSetOnQuantizedType() 784 ss << "output " << i << " of layer " << GetLayerTypeAsCString(layer->GetType()) in CheckScaleSetOnQuantizedType() 785 << " (" << layer->GetNameStr() << ") is of type" in CheckScaleSetOnQuantizedType() 792 layer->GetType() == armnn::LayerType::Softmax) in CheckScaleSetOnQuantizedType() 795 ss << "Quantization parameters for Softmax layer (Scale: " << in CheckScaleSetOnQuantizedType() [all …]
|
H A D | Graph.cpp | 33 std::unordered_map<const Layer*, Layer*> otherToClonedMap; in Graph() 37 Layer* const layer = otherLayer->Clone(*this); in Graph() local 38 otherToClonedMap.emplace(otherLayer, layer); in Graph() 44 Layer* const thisLayer = otherToClonedMap[otherLayer]; in Graph() 51 const Layer& otherTgtLayer = otherInputSlot->GetOwningLayer(); in Graph() 52 Layer* const thisTgtLayer = otherToClonedMap[&otherTgtLayer]; in Graph() 100 const armnn::Layer *layer = it; in Print() local 102 auto outputTensorShape = layer->GetOutputSlots()[i].GetTensorInfo().GetShape(); in Print() 142 for (auto&& layer : m_Layers) in SerializeToDot() 144 DotNode node(stream, layer->GetGuid(), GetLayerTypeAsCString(layer->GetType())); in SerializeToDot() [all …]
|
/aosp_15_r20/external/intel-media-driver/media_softlet/agnostic/common/vp/hal/features/ |
H A D | vp_fc_filter.cpp | 87 MOS_STATUS VpFcFilter::InitLayer(VP_FC_LAYER &layer, bool isInputPipe, int index, SwFilterPipe &exe… in InitLayer() argument 94 layer.surf = surfHandle->second; in InitLayer() 96 VP_PUBLIC_CHK_NULL_RETURN(layer.surf); in InitLayer() 98 layer.layerID = index; in InitLayer() 99 layer.layerIDOrigin = index; in InitLayer() 102 …layer.scalingMode = scaling ? scaling->GetSwFilterParams().scalingMode : defaultScalingM… in InitLayer() 103 …layer.iscalingEnabled = scaling ? ISCALING_INTERLEAVED_TO_INTERLEAVED == scaling->GetSwFilte… in InitLayer() 104 …layer.fieldWeaving = scaling ? ISCALING_FIELD_TO_INTERLEAVED == scaling->GetSwFilterParam… in InitLayer() 107 …layer.rotation = rotation ? rotation->GetSwFilterParams().rotation : VPHAL_ROTATION_I… in InitLayer() 109 …layer.useSampleUnorm = false; // Force using sampler16 (dscaler) when compute walker in … in InitLayer() [all …]
|
/aosp_15_r20/external/armnn/src/armnnDeserializer/ |
H A D | Deserializer.cpp | 129 throw ParseException(fmt::format("{0} was called with an invalid layer index. " in CheckLayers() 130 "layers:{1} layer:{2} at {3}", in CheckLayers() 289 case Layer::Layer_AbsLayer: in GetBaseLayer() 291 case Layer::Layer_ActivationLayer: in GetBaseLayer() 293 case Layer::Layer_AdditionLayer: in GetBaseLayer() 295 case Layer::Layer_ArgMinMaxLayer: in GetBaseLayer() 297 case Layer::Layer_BatchMatMulLayer: in GetBaseLayer() 299 case Layer::Layer_BatchToSpaceNdLayer: in GetBaseLayer() 301 case Layer::Layer_BatchNormalizationLayer: in GetBaseLayer() 303 case Layer::Layer_CastLayer: in GetBaseLayer() [all …]
|
/aosp_15_r20/external/armnn/src/backends/backendsCommon/ |
H A D | WorkloadFactory.cpp | 6 #include <Layer.hpp> 26 using LayerList = std::list<Layer*>; 78 const Layer& layer = *(PolymorphicDowncast<const Layer*>(&connectableLayer)); in IsLayerConfigurationSupported() local 96 switch(layer.GetType()) in IsLayerConfigurationSupported() 100 auto cLayer = PolymorphicDowncast<const ActivationLayer*>(&layer); in IsLayerConfigurationSupported() 101 const TensorInfo& input = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported() 102 const TensorInfo& output = layer.GetOutputSlot(0).GetTensorInfo(); in IsLayerConfigurationSupported() 113 const TensorInfo& input0 = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported() 114 const TensorInfo& input1 = layer.GetInputSlot(1).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported() 115 const TensorInfo& output = layer.GetOutputSlot(0).GetTensorInfo(); in IsLayerConfigurationSupported() [all …]
|
/aosp_15_r20/development/tools/winscope/src/parsers/surface_flinger/computations/ |
H A D | visibility_properties_computation.ts | 59 for (const layer of topDownTraversal) { constant 60 let isVisible = this.getIsVisible(layer); 62 layer.addEagerProperty( 64 layer.id, 69 layer.addEagerProperty( 71 layer.id, 73 this.isHiddenByPolicy(layer), 77 const displaySize = this.getDisplaySize(layer); 83 this.getDefinedValue(layer, 'layerStack') 87 if (!this.layerContains(other, layer, displaySize)) { [all …]
|
/aosp_15_r20/external/armnn/src/backends/backendsCommon/test/ |
H A D | OptimizeSubgraphViewTests.cpp | 31 // Keep the layers organized by layer name 32 using LayerNameToLayerMap = std::unordered_map<std::string, Layer*>; 71 // Convenience function to add an input layer to a graph 72 Layer* AddInputLayer(Graph& graph, in AddInputLayer() 77 Layer* const inputLayer = graph.AddLayer<InputLayer>(inputId, layerName.c_str()); in AddInputLayer() 83 // Convenience function to add an output layer to a graph 84 Layer* AddOutputLayer(Graph& graph, in AddOutputLayer() 87 Layer* const outputLayer = graph.AddLayer<OutputLayer>(0, layerName.c_str()); in AddOutputLayer() 92 // Convenience function to add a convolution layer to a graph 106 // Convenience function to add a constant layer to a graph [all …]
|
/aosp_15_r20/frameworks/native/services/surfaceflinger/tests/unittests/ |
H A D | TransactionSurfaceFrameTest.cpp | 57 sp<Layer> createLayer() { in createLayer() 59 LayerCreationArgs args(mFlinger.flinger(), client, "layer", 0, LayerMetadata()); in createLayer() 60 return sp<Layer>::make(args); in createLayer() 63 void commitTransaction(Layer* layer) { layer->commitTransaction(); } in commitTransaction() argument 71 sp<Layer> layer = createLayer(); in PresentedSurfaceFrameForBufferlessTransaction() local 75 layer->setFrameTimelineVsyncForBufferlessTransaction(ftInfo, 10, in PresentedSurfaceFrameForBufferlessTransaction() 77 EXPECT_EQ(1u, layer->mDrawingState.bufferlessSurfaceFramesTX.size()); in PresentedSurfaceFrameForBufferlessTransaction() 78 ASSERT_TRUE(layer->mDrawingState.bufferSurfaceFrameTX == nullptr); in PresentedSurfaceFrameForBufferlessTransaction() 79 const auto surfaceFrame = layer->mDrawingState.bufferlessSurfaceFramesTX.at(/*token*/ 1); in PresentedSurfaceFrameForBufferlessTransaction() 80 commitTransaction(layer.get()); in PresentedSurfaceFrameForBufferlessTransaction() [all …]
|
/aosp_15_r20/device/generic/goldfish-opengl/system/hwc3/ |
D | Layer.cpp | 17 #include "Layer.h" 32 Layer::Layer() : mId(sNextId++) {} in Layer() function in aidl::android::hardware::graphics::composer3::impl::Layer 34 HWC3::Error Layer::setCursorPosition(const common::Point& position) { in setCursorPosition() 35 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in setCursorPosition() 46 common::Point Layer::getCursorPosition() const { in getCursorPosition() 47 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in getCursorPosition() 52 HWC3::Error Layer::setBuffer(buffer_handle_t buffer, const ndk::ScopedFileDescriptor& fence) { in setBuffer() 53 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in setBuffer() 64 FencedBuffer& Layer::getBuffer() { in getBuffer() 65 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in getBuffer() [all …]
|
/aosp_15_r20/external/armnn/src/armnn/test/ |
H A D | OptimizerTests.cpp | 42 LstmLayer* const layer = graph.AddLayer<LstmLayer>(layerDesc, "layer"); in CreateLSTMLayerHelper() local 48 layer->m_BasicParameters.m_InputToForgetWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 50 layer->m_BasicParameters.m_InputToCellWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 52 layer->m_BasicParameters.m_InputToOutputWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 54 layer->m_BasicParameters.m_RecurrentToForgetWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 56 layer->m_BasicParameters.m_RecurrentToCellWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 58 layer->m_BasicParameters.m_RecurrentToOutputWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 60 layer->m_BasicParameters.m_ForgetGateBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 62 layer->m_BasicParameters.m_CellBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() 64 layer->m_BasicParameters.m_OutputGateBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper() [all …]
|