Executorch
github.comKey Features
Tech Stack
Key Features
Tech Stack
This is still great, though. Previously, I thought a mobile model (eg speech/object recognition) would require me to learn both PyTorch and something like MLC in C++. Then, port them.
If this is as it appears, I could develop a small model that could run on mobile on my laptop, train it on cloud GPU's, test it locally, and use this tool to produce a mobile version (or save some steps?). That would keep us from having to learn C++ or MLC just to do mobile.
I mean, one still can learn other tools for their advantages. However, ML students and startups might benefit greatly from this by being able to rapidly develop or port mobile apps. Then, people learning other tools for their advantages build stuff that way. The overall ecosystem gets stronger with more competition.
in particular selective function scripting is essential!
Removing control flow isn't as easy as you'd think for some. It essentially means ripping large sections out of python and into separately implemented c++.
I bet a lot of trivial text capabilities (grammar checking, autocomplete, etc) will benefit from this rather than sending everything to a hosted model. It's possible right now with onnx / transformers.js / tensorflow.js - but isn't quite there yet in terms of efficiency. Given the target for microcontrollers, it'd be great to bring that efficiency to browsers as well.
While reading the README and related documentation, I noticed that Samsung Exynos NPU acceleration was listed, which immediately caught my attention. According to https://docs.pytorch.org/executorch/main/backends/samsung/sa..., Samsung has finally built and released an NPU SDK—so I followed the link to check it out.
Unfortunately, the experience was disappointing.
The so-called “version 1.0” SDK is available only for Ubuntu 22.04 / 20.04. There is no release date information per version, nor any visible roadmap. Even worse, downloading the SDK requires logging in. The product description page itself https://soc-developer.semiconductor.samsung.com/global/devel... does contain explanations, but they are provided almost entirely as images rather than text—presented in a style more reminiscent of corporate PR material than developer-facing technical documentation.
This is, regrettably, very typical of Samsung’s software support: opaque documentation, gated access, and little consideration for external developers. At this point, it is hard not to conclude that Exynos remains a poor choice, regardless of its theoretical hardware capabilities.
For comparison, Qualcomm and MediaTek actively collaborate with existing ecosystems, and their SDKs are generally available without artificial barriers. As a concrete example, see how LiteRT distributes its artifacts and references in this commit: https://github.com/google-ai-edge/LiteRT/commit/eaf7d635e1bc...
- Better microcontroller support is in our roadmap for 2026. There is a lot of development happening here from support for Arduino, STMicro and others. We will do this openly with the community as usual so if you are interested, feel free to join our discord and looped into the github repo.
- Better web support is also in the roadmap. There is some limited support already though not sure exactly what your usecase is. Feel free to open up a GH issue and we can see if there is a way to unblock you.
- Will take the feedback about Samsung to them. Seeing the user feedback first hand here will likely help them prioritize some of that. This is partially why we have not called this a production ready backend unlike the other backends like Qualcomm, Vulkan and a few others we ourselves are using in production.
I just want it on native desktop python.
Not affiliated with Hacker News or Y Combinator. We simply enrich the public API with analytics.