Cline commited on
Commit
1112a05
·
1 Parent(s): 83b24e7

Added space configuration to README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -8
README.md CHANGED
@@ -1,15 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # Central AI Hub
2
 
3
  This is a Hugging Face Space for the Central AI Hub.
4
 
5
  ## How to Run
6
 
7
- 1. Make sure you have the required dependencies installed. You can install them using:
8
 
9
  ```bash
10
  pip install -r requirements.txt
11
  ```
12
- 2. To run the space, execute the following command:
13
 
14
  ```bash
15
  python app.py
@@ -17,7 +30,7 @@ This is a Hugging Face Space for the Central AI Hub.
17
 
18
  This will start the Gradio interface, which you can access in your browser.
19
 
20
- 3. You can optionally provide a local path to the Qwen2.5-14B GGUF model by setting the `LOCAL_MODEL_PATH` environment variable. For example:
21
 
22
  ```bash
23
  LOCAL_MODEL_PATH=/path/to/your/model.gguf python app.py
@@ -27,8 +40,8 @@ This is a Hugging Face Space for the Central AI Hub.
27
 
28
  ## Usage
29
 
30
- 1. Select a task type from the dropdown menu.
31
- 2. Enter the task content in the text box.
32
- 3. Enter any task requirements, separated by commas.
33
- 4. Click the "Submit Task" button.
34
- 5. The task status will be displayed in the output text box.
 
1
+ ---
2
+ title: AgenticAi
3
+ emoji: 🧠
4
+ colorFrom: blue
5
+ colorTo: purple
6
+ sdk: gradio
7
+ sdk_version: "4.18.0"
8
+ app_file: app.py
9
+ pinned: false
10
+ ---
11
+
12
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
13
+
14
  # Central AI Hub
15
 
16
  This is a Hugging Face Space for the Central AI Hub.
17
 
18
  ## How to Run
19
 
20
+ 1. Make sure you have the required dependencies installed. You can install them using:
21
 
22
  ```bash
23
  pip install -r requirements.txt
24
  ```
25
+ 2. To run the space, execute the following command:
26
 
27
  ```bash
28
  python app.py
 
30
 
31
  This will start the Gradio interface, which you can access in your browser.
32
 
33
+ 3. You can optionally provide a local path to the Qwen2.5-14B GGUF model by setting the `LOCAL_MODEL_PATH` environment variable. For example:
34
 
35
  ```bash
36
  LOCAL_MODEL_PATH=/path/to/your/model.gguf python app.py
 
40
 
41
  ## Usage
42
 
43
+ 1. Select a task type from the dropdown menu.
44
+ 2. Enter the task content in the text box.
45
+ 3. Enter any task requirements, separated by commas.
46
+ 4. Click the "Submit Task" button.
47
+ 5. The task status will be displayed in the output text box.