from rkllm.api import RKLLM | |
modelpath = '.' | |
llm = RKLLM() | |
ret = llm.load_huggingface(model=modelpath, model_lora=None, device='cpu') | |
if ret != 0: | |
print('Load model failed!') | |
exit(ret) | |
qparams = None | |
ret = llm.build(do_quantization=True, optimization_level=1, quantized_dtype='w8a8_g128', | |
quantized_algorithm='normal', target_platform='rk3588', num_npu_core=3, extra_qparams=qparams) | |
if ret != 0: | |
print('Build model failed!') | |
exit(ret) | |
# Export rkllm model | |
ret = llm.export_rkllm("./qwen.rkllm") | |
if ret != 0: | |
print('Export model failed!') | |
exit(ret) | |