-
Tuklus Labs
- Washington State
- https://linkedin.com/in/garyjduncan
Popular repositories Loading
-
kernel-anvil
kernel-anvil PublicProfile-guided GPU kernel optimizer for AMD/RDNA3. Auto-tunes llama.cpp MMVQ kernels per model shape. 2x decode speedup on 7900 XTX.
-
llama-cpp-turboquant
llama-cpp-turboquant PublicForked from TheTom/llama-cpp-turboquant
LLM inference in C/C++
C++ 1
-
hamm-r
hamm-r PublicHeadless Agent Mobile Management Relay -- Run Claude Code from your phone
Kotlin
-
moirai
moirai PublicThree-model local coding daemon: planner + coder + reviewer-orchestrator on a single GPU. Swap-on-demand llama-server, RO loop with tool-call discipline, turboquant KV compression on 7900 XTX.
Go
If the problem persists, check the GitHub status page or contact support.


