NextCoder-7B — RKLLM build for RK3588 boards
Author: @jamescallander
Source model: nextcoder/NextCoder-7b
Target: Rockchip RK3588 NPU via RKNN-LLM Runtime
This repository hosts a conversion of
NextCoder-7bfor use on Rockchip RK3588 single-board computers (Orange Pi 5 plus, Radxa Rock 5b+, Banana Pi M7, etc.). Conversion was performed using the RKNN-LLM toolkit
Conversion details
- RKLLM-Toolkit version: v1.2.1
 - NPU driver: v0.9.8
 - Python: 3.12
 - Quantization: 
w8a8_g128 - Output: single-file 
.rkllmartifact - Tokenizer: not required at runtime (UI handles prompt I/O)
 
⚠️ Code generation disclaimer
🛑 This model may produce incorrect, insecure, or non-optimal code.
- It is intended for research, educational, and prototyping purposes only.
 - Always review, test, and validate any generated code before using it in production.
 - The model does not guarantee compliance with security best practices or coding standards.
 - You are responsible for ensuring outputs meet your project’s requirements and legal obligations.
 
Intended use
- On-device coding assistant / code generation on RK3588 SBCs.
 - NextCoder-7B is tuned for software development and programming tasks, making it suitable for edge deployment where privacy and low power use are priorities.
 
Limitations
- Requires 8GB free memory.
 - Quantized build (
w8a8_g128) may show small quality differences vs. full-precision upstream. - Tested on Radxa Rock 5B+; other devices may require different drivers/toolkit versions.
 - Generated code should always be reviewed before use in production systems.
 
Quick start (RK3588)
1) Install runtime
The RKNN-LLM toolkit and instructions can be found on the specific development board's manufacturer website or from airockchip's github page.
Download and install the required packages as per the toolkit's instructions.
2) Simple Flask server deployment
The simplest way the deploy the .rkllm converted model is using an example script provided in the toolkit in this directory: rknn-llm/examples/rkllm_server_demo
python3 <TOOLKIT_PATH>/rknn-llm/examples/rkllm_server_demo/flask_server.py \
  --rkllm_model_path <MODEL_PATH>/NextCoder-7B_w8a8_g128_rk3588.rkllm \
  --target_platform rk3588
3) Sending a request
The basic format for a message request is:
{
    "model":"NextCoder-7b",
    "messages":[{
        "role":"user",
        "content":"<YOUR_PROMPT_HERE>"}],
    "stream":false
}
Example request using curl:
curl -s -X POST <SERVER_IP_ADDRESS>:8080/rkllm_chat \
    -H 'Content-Type: application/json' \
    -d '{"model":"NextCoder-7b","messages":[{"role":"user","content":"Explain in one sentence  what a static method is."}],"stream":false}'
The response is formated in the following way:
{
    "choices":[{
        "finish_reason":"stop",
        "index":0,
        "logprobs":null,
        "message":{
            "content":"<MODEL_REPLY_HERE">,
            "role":"assistant"}}],
        "created":null,
        "id":"rkllm_chat",
        "object":"rkllm_chat",
        "usage":{
            "completion_tokens":null,
            "prompt_tokens":null,
            "total_tokens":null}
}
Example response:
{"choices":[{"finish_reason":"stop","index":0,"logprobs":null,"message":{"content":"A static method is a method that belongs to a class rather than an instance of the class and can be called without creating an object of the class.","role":"assistant"}}],"created":null,"id":"rkllm_chat","object":"rkllm_chat","usage":{"completion_tokens":null,"prompt_tokens":null,"total_tokens":null}}
4) UI compatibility
This server exposes an OpenAI-compatible Chat Completions API.
You can connect it to any OpenAI-compatible client or UI (for example: Open WebUI)
- Configure your client with the API base: 
http://<SERVER_IP_ADDRESS>:8080and use the endpoint:/rkllm_chat - Make sure the 
modelfield matches the converted model’s name, for example: 
{
 "model": "NextCoder-7B",
 "messages": [{"role":"user","content":"Hello!"}],
 "stream": false
}
License
This conversion follows the license of the source model: mit
- Downloads last month
 - 369
 
Model tree for jamescallander/NextCoder-7B_w8a8_g128_rk3588.rkllm
Base model
Qwen/Qwen2.5-7B