|
16 | 16 | #include <string> |
17 | 17 | #include <iostream> |
18 | 18 |
|
19 | | -namespace torch { |
20 | | -namespace executor { |
| 19 | +namespace executorch { |
| 20 | +namespace backends { |
| 21 | + |
| 22 | +using executorch::aten::Tensor; |
| 23 | +using executorch::runtime::ArrayRef; |
| 24 | +using executorch::runtime::Backend; |
| 25 | +using executorch::runtime::BackendExecutionContext; |
| 26 | +using executorch::runtime::BackendInitContext; |
| 27 | +using executorch::runtime::CompileSpec; |
| 28 | +using executorch::runtime::DelegateHandle; |
| 29 | +using executorch::runtime::EValue; |
| 30 | +using executorch::runtime::Error; |
| 31 | +using executorch::runtime::FreeableBuffer; |
| 32 | +using executorch::runtime::Result; |
21 | 33 |
|
22 | 34 | class MPSBackend final : public ::executorch::runtime::BackendInterface { |
23 | 35 | public: |
@@ -81,7 +93,7 @@ Error execute( |
81 | 93 | output_pointers.push_back(&args[i]->toTensor()); |
82 | 94 | } |
83 | 95 | } else if (args[i]->isTensorList()) { |
84 | | - const exec_aten::ArrayRef<exec_aten::Tensor>& tensorList = args[i]->toTensorList(); |
| 96 | + const executorch::aten::ArrayRef<executorch::aten::Tensor>& tensorList = args[i]->toTensorList(); |
85 | 97 | for (auto& tensor_ : tensorList) { |
86 | 98 | if (input_pointers.size() < executor->getNumInputs()) { |
87 | 99 | input_pointers.push_back(&tensor_); |
@@ -122,5 +134,5 @@ void destroy(DelegateHandle* handle) const override { |
122 | 134 | static auto success_with_compiler = register_backend(backend); |
123 | 135 | } // namespace |
124 | 136 |
|
125 | | -} // namespace executor |
126 | | -} // namespace torch |
| 137 | +} // namespace backends |
| 138 | +} // namespace executorch |
0 commit comments