### Prerequisites - [x] I am running the latest code. Mention the version if possible as well. - [x] I carefully followed the [README.md](https://github.com/ggml-org/llama.cpp/blob/master/README.md). - [x] I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed). - [x] I reviewed the [Discussions](https://github.com/ggml-org/llama.cpp/discussions), and have a new and useful enhancement to share. ### Feature Description Support for Nvidia Nemotron-3-Nano-30B-A3B and architecture nemotron_h_moe which should be a moe extension of the existing nemotron_h. ### Motivation Benchmarks are very good for its size (i.e.: 38.8 in swe-bench) and speed is quite better than similar sized qwen3 models. ### Possible Implementation _No response_