YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

This repository contains the full weights and LoRA weights for Zh-MT-LLM v1.0 which fine-tuned with ChatGLM3-6b-base.

Zh-MT-LLM

Zheng He Maritime Large Language Model (Zh-MT-LLM) is a vertical domain maritime Large Language Model developed by the Intelligent Technology Laboratory of Dalian Maritime University for practitioners, trainers and students in the maritime field, providing questions and answers on maritime laws and regulations, maritime education and training, and questions and answers on maritime expertise.

Corresponding to the above three segments, our model has the following three main characteristics:

  • Maritime Laws and Regulations Q&A:

    The model is trained on a wide range of maritime laws and regulations, providing consulting services for those in the maritime field.

  • Maritime Education and Training:

    The model learns from maritime professional test questions, vocational examination syllabi, and high-quality crew common Q&A to provide training knowledge.

  • Maritime Expertise Q&A:

    The model covers ship maintenance, safety management, port operations, maritime logistics, navigation technology, marine environmental protection, and scientific research to answer questions for maritime industry practitioners.

Zh-MT-SFT Dataset

The specific statistics of the dataset used for the above training are as follows:

Services Subtasks Data sets Data volume
Maritime Laws and Regulations Q&A Maritime Legal Advice CrimeKgAssitant 18,279
Zh-law-qa 59,244
The Court held Zh-law-court 2,684
Sentence projections Zh-law-predict 3,004
Maritime education and training Maritime Education Counseling Zh-edu-qa 41,052
Maritime Specialization Question Bank Zh-edu-qb 23,531
Maritime Expertise Q&A Ship Knowledge Zh-mt-qa 46,759
Navigational Knowledge
Port knowledge
Marine knowledge
Generic Dialogue moss-003-sft-data 300,000
Total 494,553

Code Usage

You can create a conversation using the Zh-MT-LLM model using the following codes:

>>>from transformers import AutoTokenizer, AutoModel
>>>tokenizer = AutoTokenizer.from_pretrained("ZhangFuXi/Zh-MT-LLM",trust_remote_code=True)
>>>model = AutoModel.from_pretrained("ZhangFuXi/Zh-MT-LLM", trust_remote_code=True).half().cuda()
>>>model = model.eval()
>>>response, history = model.chat(tokenizer, "你好", history=[])
>>>print(response)

Declaration

Due to factors such as the limitation of the number of model parameters and the degree of cleaning of the training data, the model open source in this project may have the following limitations:

  • Because it has not been harmlessly fine-tuned, it may result in discriminatory, harmful and unethical statements.

  • Lacking an accurate understanding of the real world, the model may produce hallucinatory responses that mislead the user.

  • The model's training data may contain biased data, and users should be cautious about potential bias in model responses.

  • Due to the limited number of model parameters, it may not be possible to cover all areas of knowledge, resulting in less accurate or complete responses on some topics.

  • When dealing with factual knowledge in a specific domain, models may provide incorrect answers due to insufficient information or misinterpretation, leading to misinformation or confusion.

Given the limitations of the above model, we request that the code, data, and model of this project not be used for socially harmful purposes and must follow the MODEL_LICENCE of the base model. We are not responsible for any problems, risks, or adverse consequences arising from the use of Zh-MT-LLM.

Licenses

The use of the source code in this repository complies with the Apache 2.0 License.

Downloads last month
9
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.