Home
last modified time | relevance | path

Searched full:layer (Results 1 – 25 of 10740) sorted by relevance

12345678910>>...430

/aosp_15_r20/external/rust/android-crates-io/crates/tower-layer/src/
Dtuple.rs1 use crate::Layer;
3 impl<S> Layer<S> for () { implementation
6 fn layer(&self, service: S) -> Self::Service { in layer() method
11 impl<S, L1> Layer<S> for (L1,) impl
13 L1: Layer<S>,
17 fn layer(&self, service: S) -> Self::Service { in layer() function
19 l1.layer(service) in layer()
23 impl<S, L1, L2> Layer<S> for (L1, L2) impl
25 L1: Layer<L2::Service>,
26 L2: Layer<S>,
[all …]
/aosp_15_r20/external/armnn/python/pyarmnn/src/pyarmnn/swig/modules/
H A Darmnn_network.i35 of the layer and any mismatch is reported.
70 An input connection slot for a layer. Slot lifecycle is managed by the layer.
72 The input slot can be connected to an output slot of the preceding layer in the graph.
83 Returns output slot of a preceding layer that is connected to the given input slot.
86 IOutputSlot: Borrowed reference to an output connection slot for a preceding layer.
95 An output connection slot for a layer. Slot lifecycle is managed by the layer.
200 Calculates the index of this slot for the layer.
210 Returns the index of the layer. Same value as `IConnectableLayer.GetGuid`.
213 int: Layer id.
234 Interface for a layer that is connectable to other layers via `IInputSlot` and `IOutputSlot`.
[all …]
/aosp_15_r20/external/armnn/src/armnnTestUtils/
H A DCreateWorkload.hpp32 // Calls CreateWorkload for a layer, and checks the returned pointer is of the correct type.
34 std::unique_ptr<Workload> MakeAndCheckWorkload(Layer& layer, in MakeAndCheckWorkload() argument
38 std::unique_ptr<IWorkload> workload = layer.CreateWorkload(factory);
42 layer.SetBackendId(factory.GetBackendId());
43 CHECK(factory.IsLayerSupported(layer, layer.GetDataType(), reasonIfUnsupported, modelOptions));
52 for (auto&& layer : graph.TopologicalSort()) in CreateTensorHandles()
54 layer->CreateTensorHandles(tmpRegistry, factory); in CreateTensorHandles()
69 // Creates the layer we're testing. in CreateActivationWorkloadTest()
75 ActivationLayer* const layer = graph.AddLayer<ActivationLayer>(layerDesc, "layer"); in CreateActivationWorkloadTest() local
78 Layer* const input = graph.AddLayer<InputLayer>(0, "input"); in CreateActivationWorkloadTest()
[all …]
/aosp_15_r20/external/armnn/include/armnn/
H A DINetwork.hpp22 /// @brief An input connection slot for a layer.
23 /// The input slot can be connected to an output slot of the preceding layer in the graph.
39 /// @brief An output connection slot for a layer.
67 /// @brief Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
71 /// Returns the name of the layer
95 /// Returns the unique id of the layer
98 /// Apply a visitor to this layer
101 /// Provide a hint for the optimizer as to which backend to prefer for this layer.
102 …// By providing a BackendSelectionHint there is no guarantee the input backend supports that layer.
105 /// layer (IsLayerSupported returns true for a specific backend).
[all …]
/aosp_15_r20/frameworks/native/services/surfaceflinger/FrontEnd/
H A DLayerLifecycleManager.cpp33 // Returns true if the layer is root of a display and can be mirrored by mirroringLayer
47 RequestedLayerState& layer = *newLayer.get(); in addLayers() local
48 auto [it, inserted] = mIdToLayer.try_emplace(layer.id, References{.owner = layer}); in addLayers()
50 "Duplicate layer id found. New layer: %s Existing layer: " in addLayers()
52 layer.getDebugString().c_str(), in addLayers()
56 layer.parentId = linkLayer(layer.parentId, layer.id); in addLayers()
57 layer.relativeParentId = linkLayer(layer.relativeParentId, layer.id); in addLayers()
58 if (layer.layerStackToMirror != ui::INVALID_LAYER_STACK) { in addLayers()
59 // Set mirror layer's default layer stack to -1 so it doesn't end up rendered on a in addLayers()
61 layer.layerStack = ui::INVALID_LAYER_STACK; in addLayers()
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/saving/saved_model/
H A Dsave_impl.py71 def should_skip_serialization(layer): argument
72 """Skip serializing extra objects and functions if layer inputs aren't set."""
73 saved_model_input_spec_set = (isinstance(layer, training_lib.Model) and
74layer._saved_model_inputs_spec is not None) # pylint: disable=protected-access
75 if not layer.built and not saved_model_input_spec_set:
76 logging.warning('Skipping full serialization of Keras layer {}, because '
77 'it is not built.'.format(layer))
82 def wrap_layer_objects(layer, serialization_cache): argument
83 """Returns extra trackable objects to attach to the serialized layer.
86 layer: Keras Layer object.
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/mixed_precision/
H A Dlayer_test.py15 """Tests keras.layers.Layer works properly with mixed precision."""
98 layer = mp_test_util.MultiplyLayer(assert_type=dtype)
99 self.assertEqual(layer.dtype, dtypes.float32)
100 self.assertEqual(get_layer_policy.get_layer_policy(layer).name,
102 y = layer(x)
103 self.assertEqual(layer.v.dtype, dtypes.float32)
105 self.assertEqual(layer.dtype_policy.name, policy_name)
106 self.assertIsInstance(layer.dtype_policy, policy.Policy)
107 self.assertEqual(layer.compute_dtype, dtype)
108 self.assertEqual(layer.dtype, dtypes.float32)
[all …]
/aosp_15_r20/external/armnn/src/armnnSerializer/
H A DSerializer.cpp114 // Build FlatBuffer for Input Layer
115 void SerializerStrategy::SerializeInputLayer(const armnn::IConnectableLayer* layer, LayerBindingId … in SerializeInputLayer() argument
120 auto flatBufferInputBaseLayer = CreateLayerBase(layer, serializer::LayerType::LayerType_Input); in SerializeInputLayer()
126 // Push layer binding id to outputIds. in SerializeInputLayer()
133 CreateAnyLayer(flatBufferInputLayer.o, serializer::Layer::Layer_InputLayer); in SerializeInputLayer()
136 // Build FlatBuffer for Output Layer
137 void SerializerStrategy::SerializeOutputLayer(const armnn::IConnectableLayer* layer, in SerializeOutputLayer() argument
143 … auto flatBufferOutputBaseLayer = CreateLayerBase(layer, serializer::LayerType::LayerType_Output); in SerializeOutputLayer()
149 // Push layer binding id to outputIds. in SerializeOutputLayer()
155 CreateAnyLayer(flatBufferOutputLayer.o, serializer::Layer::Layer_OutputLayer); in SerializeOutputLayer()
[all …]
H A DSerializer.hpp25 void ExecuteStrategy(const armnn::IConnectableLayer* layer,
57 /// Creates the Input Slots and Output Slots and LayerBase for the layer.
59 const armnn::IConnectableLayer* layer,
62 /// Creates the serializer AnyLayer for the layer and adds it to m_serializedLayers.
63 …void CreateAnyLayer(const flatbuffers::Offset<void>& layer, const armnnSerializer::Layer serialize…
78 /// Creates the serializer InputSlots for the layer.
80 const armnn::IConnectableLayer* layer);
82 /// Creates the serializer OutputSlots for the layer.
84 const armnn::IConnectableLayer* layer);
101 /// layer within our FlatBuffer index.
[all …]
/aosp_15_r20/external/tensorflow/tensorflow/python/keras/engine/
H A Dfunctional.py95 # The key of _layer_call_argspecs is a layer. tf.Module._flatten will fail to
96 # flatten the key since it is trying to convert Trackable/Layer to a string.
120 # layer is added or removed.
182 layer, node_index, tensor_index = x._keras_history # pylint: disable=protected-access
183 self._output_layers.append(layer)
184 self._output_coordinates.append((layer, node_index, tensor_index))
188 layer, node_index, tensor_index = x._keras_history # pylint: disable=protected-access
189 # It's supposed to be an input layer, so only one node
193 self._input_layers.append(layer)
194 self._input_coordinates.append((layer, node_index, tensor_index))
[all …]
H A Dbase_layer.py16 """Contains the base Layer class, from which all layers inherit."""
88 # Prefix that is added to the TF op layer names.
97 @keras_export('keras.layers.Layer')
98 class Layer(module.Module, version_utils.LayerVersionSelector): class
101 A layer is a callable object that takes as input one or more tensors and
106 Users will just instantiate a layer and then treat it as a callable.
109 trainable: Boolean, whether the layer's variables should be trainable.
110 name: String name of the layer.
111 dtype: The dtype of the layer's computations and weights. Can also be a
116 dynamic: Set this to `True` if your layer should only be run eagerly, and
[all …]
H A Dbase_layer_v1.py16 """Contains the base Layer class, from which all layers inherit."""
70 class Layer(base_layer.Layer): class
71 """Base layer class.
75 A layer is a class implementing common neural networks operations, such
77 losses, updates, and inter-layer connectivity.
79 Users will just instantiate a layer and then treat it as a callable.
81 We recommend that descendants of `Layer` implement the following methods:
90 once. Should actually perform the logic of applying the layer to the
94 trainable: Boolean, whether the layer's variables should be trainable.
95 name: String name of the layer.
[all …]
/aosp_15_r20/frameworks/native/services/surfaceflinger/Tracing/
H A DTransactionProtoParser.cpp97 auto& layer = resolvedComposerState.state; in toProto() local
99 proto.set_what(layer.what); in toProto()
101 if (layer.what & layer_state_t::ePositionChanged) { in toProto()
102 proto.set_x(layer.x); in toProto()
103 proto.set_y(layer.y); in toProto()
105 if (layer.what & layer_state_t::eLayerChanged) { in toProto()
106 proto.set_z(layer.z); in toProto()
109 if (layer.what & layer_state_t::eLayerStackChanged) { in toProto()
110 proto.set_layer_stack(layer.layerStack.id); in toProto()
112 if (layer.what & layer_state_t::eFlagsChanged) { in toProto()
[all …]
/aosp_15_r20/external/rust/android-crates-io/crates/tracing-subscriber/src/layer/
Dmod.rs1 //! The [`Layer`] trait, a composable abstraction for building [`Subscriber`]s.
18 //! span IDs. The [`Layer`] trait represents this composable subset of the
24 //! Since a [`Layer`] does not implement a complete strategy for collecting
26 //! [`Layer`] trait is generic over a type parameter (called `S` in the trait
28 //! with. Thus, a [`Layer`] may be implemented that will only compose with a
30 //! added to constrain what types implementing `Subscriber` a `Layer` can wrap.
32 //! `Layer`s may be added to a `Subscriber` by using the [`SubscriberExt::with`]
35 //! `Layer` with the `Subscriber`.
39 //! use tracing_subscriber::Layer;
47 //! impl<S: Subscriber> Layer<S> for MyLayer {
[all …]
Dlayered.rs5 layer::{Context, Layer},
17 /// [`Layer`]s.
19 /// [`Layer`]: crate::Layer
23 /// The layer.
24 layer: L, field
26 /// The inner value that `self.layer` was layered onto.
28 /// If this is also a `Layer`, then this `Layered` will implement `Layer`.
34 // level hints when per-layer filters are in use.
41 /// Does `self.layer` have per-layer filters?
46 /// `Layered`s have per-layer filters.
[all …]
/aosp_15_r20/external/armnn/src/armnn/
H A DNetwork.cpp8 #include "Layer.hpp"
757 const Layer* layer, in ReturnWithError() argument
762 failureMsg << "Layer of type " << GetLayerTypeAsCString(layer->GetType()) in ReturnWithError()
771 bool CheckScaleSetOnQuantizedType(Layer* layer, Optional<std::vector<std::string>&> errMessages) in CheckScaleSetOnQuantizedType() argument
774 unsigned int numOutputs = layer->GetNumOutputSlots(); in CheckScaleSetOnQuantizedType()
776 OutputSlot& outputSlot = layer->GetOutputSlot(i); in CheckScaleSetOnQuantizedType()
784 ss << "output " << i << " of layer " << GetLayerTypeAsCString(layer->GetType()) in CheckScaleSetOnQuantizedType()
785 << " (" << layer->GetNameStr() << ") is of type" in CheckScaleSetOnQuantizedType()
792 layer->GetType() == armnn::LayerType::Softmax) in CheckScaleSetOnQuantizedType()
795 ss << "Quantization parameters for Softmax layer (Scale: " << in CheckScaleSetOnQuantizedType()
[all …]
H A DGraph.cpp33 std::unordered_map<const Layer*, Layer*> otherToClonedMap; in Graph()
37 Layer* const layer = otherLayer->Clone(*this); in Graph() local
38 otherToClonedMap.emplace(otherLayer, layer); in Graph()
44 Layer* const thisLayer = otherToClonedMap[otherLayer]; in Graph()
51 const Layer& otherTgtLayer = otherInputSlot->GetOwningLayer(); in Graph()
52 Layer* const thisTgtLayer = otherToClonedMap[&otherTgtLayer]; in Graph()
100 const armnn::Layer *layer = it; in Print() local
102 auto outputTensorShape = layer->GetOutputSlots()[i].GetTensorInfo().GetShape(); in Print()
142 for (auto&& layer : m_Layers) in SerializeToDot()
144 DotNode node(stream, layer->GetGuid(), GetLayerTypeAsCString(layer->GetType())); in SerializeToDot()
[all …]
/aosp_15_r20/external/intel-media-driver/media_softlet/agnostic/common/vp/hal/features/
H A Dvp_fc_filter.cpp87 MOS_STATUS VpFcFilter::InitLayer(VP_FC_LAYER &layer, bool isInputPipe, int index, SwFilterPipe &exe… in InitLayer() argument
94 layer.surf = surfHandle->second; in InitLayer()
96 VP_PUBLIC_CHK_NULL_RETURN(layer.surf); in InitLayer()
98 layer.layerID = index; in InitLayer()
99 layer.layerIDOrigin = index; in InitLayer()
102layer.scalingMode = scaling ? scaling->GetSwFilterParams().scalingMode : defaultScalingM… in InitLayer()
103layer.iscalingEnabled = scaling ? ISCALING_INTERLEAVED_TO_INTERLEAVED == scaling->GetSwFilte… in InitLayer()
104layer.fieldWeaving = scaling ? ISCALING_FIELD_TO_INTERLEAVED == scaling->GetSwFilterParam… in InitLayer()
107layer.rotation = rotation ? rotation->GetSwFilterParams().rotation : VPHAL_ROTATION_I… in InitLayer()
109layer.useSampleUnorm = false; // Force using sampler16 (dscaler) when compute walker in … in InitLayer()
[all …]
/aosp_15_r20/external/armnn/src/armnnDeserializer/
H A DDeserializer.cpp129 throw ParseException(fmt::format("{0} was called with an invalid layer index. " in CheckLayers()
130 "layers:{1} layer:{2} at {3}", in CheckLayers()
289 case Layer::Layer_AbsLayer: in GetBaseLayer()
291 case Layer::Layer_ActivationLayer: in GetBaseLayer()
293 case Layer::Layer_AdditionLayer: in GetBaseLayer()
295 case Layer::Layer_ArgMinMaxLayer: in GetBaseLayer()
297 case Layer::Layer_BatchMatMulLayer: in GetBaseLayer()
299 case Layer::Layer_BatchToSpaceNdLayer: in GetBaseLayer()
301 case Layer::Layer_BatchNormalizationLayer: in GetBaseLayer()
303 case Layer::Layer_CastLayer: in GetBaseLayer()
[all …]
/aosp_15_r20/external/armnn/src/backends/backendsCommon/
H A DWorkloadFactory.cpp6 #include <Layer.hpp>
26 using LayerList = std::list<Layer*>;
78 const Layer& layer = *(PolymorphicDowncast<const Layer*>(&connectableLayer)); in IsLayerConfigurationSupported() local
96 switch(layer.GetType()) in IsLayerConfigurationSupported()
100 auto cLayer = PolymorphicDowncast<const ActivationLayer*>(&layer); in IsLayerConfigurationSupported()
101 const TensorInfo& input = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported()
102 const TensorInfo& output = layer.GetOutputSlot(0).GetTensorInfo(); in IsLayerConfigurationSupported()
113 const TensorInfo& input0 = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported()
114 const TensorInfo& input1 = layer.GetInputSlot(1).GetConnection()->GetTensorInfo(); in IsLayerConfigurationSupported()
115 const TensorInfo& output = layer.GetOutputSlot(0).GetTensorInfo(); in IsLayerConfigurationSupported()
[all …]
/aosp_15_r20/development/tools/winscope/src/parsers/surface_flinger/computations/
H A Dvisibility_properties_computation.ts59 for (const layer of topDownTraversal) { constant
60 let isVisible = this.getIsVisible(layer);
62 layer.addEagerProperty(
64 layer.id,
69 layer.addEagerProperty(
71 layer.id,
73 this.isHiddenByPolicy(layer),
77 const displaySize = this.getDisplaySize(layer);
83 this.getDefinedValue(layer, 'layerStack')
87 if (!this.layerContains(other, layer, displaySize)) {
[all …]
/aosp_15_r20/external/armnn/src/backends/backendsCommon/test/
H A DOptimizeSubgraphViewTests.cpp31 // Keep the layers organized by layer name
32 using LayerNameToLayerMap = std::unordered_map<std::string, Layer*>;
71 // Convenience function to add an input layer to a graph
72 Layer* AddInputLayer(Graph& graph, in AddInputLayer()
77 Layer* const inputLayer = graph.AddLayer<InputLayer>(inputId, layerName.c_str()); in AddInputLayer()
83 // Convenience function to add an output layer to a graph
84 Layer* AddOutputLayer(Graph& graph, in AddOutputLayer()
87 Layer* const outputLayer = graph.AddLayer<OutputLayer>(0, layerName.c_str()); in AddOutputLayer()
92 // Convenience function to add a convolution layer to a graph
106 // Convenience function to add a constant layer to a graph
[all …]
/aosp_15_r20/frameworks/native/services/surfaceflinger/tests/unittests/
H A DTransactionSurfaceFrameTest.cpp57 sp<Layer> createLayer() { in createLayer()
59 LayerCreationArgs args(mFlinger.flinger(), client, "layer", 0, LayerMetadata()); in createLayer()
60 return sp<Layer>::make(args); in createLayer()
63 void commitTransaction(Layer* layer) { layer->commitTransaction(); } in commitTransaction() argument
71 sp<Layer> layer = createLayer(); in PresentedSurfaceFrameForBufferlessTransaction() local
75 layer->setFrameTimelineVsyncForBufferlessTransaction(ftInfo, 10, in PresentedSurfaceFrameForBufferlessTransaction()
77 EXPECT_EQ(1u, layer->mDrawingState.bufferlessSurfaceFramesTX.size()); in PresentedSurfaceFrameForBufferlessTransaction()
78 ASSERT_TRUE(layer->mDrawingState.bufferSurfaceFrameTX == nullptr); in PresentedSurfaceFrameForBufferlessTransaction()
79 const auto surfaceFrame = layer->mDrawingState.bufferlessSurfaceFramesTX.at(/*token*/ 1); in PresentedSurfaceFrameForBufferlessTransaction()
80 commitTransaction(layer.get()); in PresentedSurfaceFrameForBufferlessTransaction()
[all …]
/aosp_15_r20/device/generic/goldfish-opengl/system/hwc3/
DLayer.cpp17 #include "Layer.h"
32 Layer::Layer() : mId(sNextId++) {} in Layer() function in aidl::android::hardware::graphics::composer3::impl::Layer
34 HWC3::Error Layer::setCursorPosition(const common::Point& position) { in setCursorPosition()
35 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in setCursorPosition()
46 common::Point Layer::getCursorPosition() const { in getCursorPosition()
47 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in getCursorPosition()
52 HWC3::Error Layer::setBuffer(buffer_handle_t buffer, const ndk::ScopedFileDescriptor& fence) { in setBuffer()
53 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in setBuffer()
64 FencedBuffer& Layer::getBuffer() { in getBuffer()
65 DEBUG_LOG("%s: layer:%" PRId64, __FUNCTION__, mId); in getBuffer()
[all …]
/aosp_15_r20/external/armnn/src/armnn/test/
H A DOptimizerTests.cpp42 LstmLayer* const layer = graph.AddLayer<LstmLayer>(layerDesc, "layer"); in CreateLSTMLayerHelper() local
48 layer->m_BasicParameters.m_InputToForgetWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
50 layer->m_BasicParameters.m_InputToCellWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
52 layer->m_BasicParameters.m_InputToOutputWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
54 layer->m_BasicParameters.m_RecurrentToForgetWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
56 layer->m_BasicParameters.m_RecurrentToCellWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
58 layer->m_BasicParameters.m_RecurrentToOutputWeights = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
60 layer->m_BasicParameters.m_ForgetGateBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
62 layer->m_BasicParameters.m_CellBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
64 layer->m_BasicParameters.m_OutputGateBias = std::make_unique<ScopedTensorHandle> in CreateLSTMLayerHelper()
[all …]

12345678910>>...430