Spaces:
				
			
			
	
			
			
		Sleeping
		
	
	
	
			
			
	
	
	
	
		
		
		Sleeping
		
	
		binary-husky
		
	commited on
		
		
					加入LLAMA + 盘古 + RWKV本地模型
Browse files- docker-compose.yml +43 -0
    	
        docker-compose.yml
    CHANGED
    
    | @@ -72,3 +72,46 @@ services: | |
| 72 | 
             
                  bash -c " echo '[gpt-academic] 正在从github拉取最新代码...' &&
         | 
| 73 | 
             
                            git pull &&
         | 
| 74 | 
             
                            python3 -u main.py"
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 72 | 
             
                  bash -c " echo '[gpt-academic] 正在从github拉取最新代码...' &&
         | 
| 73 | 
             
                            git pull &&
         | 
| 74 | 
             
                            python3 -u main.py"
         | 
| 75 | 
            +
             | 
| 76 | 
            +
             | 
| 77 | 
            +
            ### ===================================================
         | 
| 78 | 
            +
            ### 【方案三】 如果需要运行ChatGPT + LLAMA + 盘古 + RWKV本地模型
         | 
| 79 | 
            +
            ### ===================================================
         | 
| 80 | 
            +
            version: '3'
         | 
| 81 | 
            +
            services:
         | 
| 82 | 
            +
              gpt_academic_with_rwkv:
         | 
| 83 | 
            +
                image: fuqingxu/gpt_academic:jittorllms    # [option 2] 如果需要运行ChatGPT + LLAMA + 盘古 + RWKV本地模型
         | 
| 84 | 
            +
                environment:
         | 
| 85 | 
            +
                  # 请查阅 `config.py` 以查看所有的配置信息
         | 
| 86 | 
            +
                  API_KEY:                  '    sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx,fkxxxxxx-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx  '
         | 
| 87 | 
            +
                  USE_PROXY:                '    True                                                                                           '
         | 
| 88 | 
            +
                  proxies:                  '    { "http": "socks5h://localhost:10880", "https": "socks5h://localhost:10880", }                 '
         | 
| 89 | 
            +
                  LLM_MODEL:                '    gpt-3.5-turbo                                                                                  '
         | 
| 90 | 
            +
                  AVAIL_LLM_MODELS:         '    ["gpt-3.5-turbo", "api2d-gpt-4", "newbing", "jittorllms_rwkv", "jittorllms_llama", "jittorllms_pangualpha"]                '
         | 
| 91 | 
            +
                  LOCAL_MODEL_DEVICE:       '    cuda                                                                                           '
         | 
| 92 | 
            +
                  DEFAULT_WORKER_NUM:       '    10                                                                                             '
         | 
| 93 | 
            +
                  WEB_PORT:                 '    32303                                                                                          '
         | 
| 94 | 
            +
                  ADD_WAIFU:                '    True                                                                                           '
         | 
| 95 | 
            +
                  AUTHENTICATION:           '    [("username", "passwd"), ("username2", "passwd2")]                                             '
         | 
| 96 | 
            +
             | 
| 97 | 
            +
                # 显卡的使用,nvidia0指第0个GPU
         | 
| 98 | 
            +
                runtime: nvidia
         | 
| 99 | 
            +
                devices:
         | 
| 100 | 
            +
                  - /dev/nvidia0:/dev/nvidia0
         | 
| 101 | 
            +
                  
         | 
| 102 | 
            +
                # 与宿主的网络融合
         | 
| 103 | 
            +
                network_mode: "host"
         | 
| 104 | 
            +
             | 
| 105 | 
            +
                # 使用代理网络拉取最新代码
         | 
| 106 | 
            +
                # command: >
         | 
| 107 | 
            +
                #   bash -c "       echo '[gpt-academic] 正在从github拉取最新代码...' &&
         | 
| 108 | 
            +
                #                   truncate -s -1 /etc/proxychains.conf &&
         | 
| 109 | 
            +
                #                   echo \"socks5 127.0.0.1 10880\" >> /etc/proxychains.conf &&
         | 
| 110 | 
            +
                #                   proxychains git pull &&
         | 
| 111 | 
            +
                #                   python3 -u main.py "
         | 
| 112 | 
            +
             | 
| 113 | 
            +
                # 不使用代理网络拉取最新代码
         | 
| 114 | 
            +
                command: >
         | 
| 115 | 
            +
                  bash -c " echo '[gpt-academic] 正在从github拉取最新代码...' &&
         | 
| 116 | 
            +
                            git pull &&
         | 
| 117 | 
            +
                            python3 -u main.py"
         |