Upload wandb/output.log with huggingface_hub
Browse files- wandb/output.log +146 -0
wandb/output.log
ADDED
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Epoch 0: 100%|ββββββββββ| 1/1 [00:09<00:00, 9.95s/it, loss=2.3673]
|
2 |
+
Epoch 0 metrics: {'epoch': 0, 'loss': 2.3673205375671387, 'val_loss': 2.359609603881836}
|
3 |
+
Epoch 1: 100%|ββββββββββ| 1/1 [00:05<00:00, 5.72s/it, loss=2.3390]
|
4 |
+
Epoch 1 metrics: {'epoch': 1, 'loss': 2.33897066116333, 'val_loss': 2.3596091270446777}
|
5 |
+
Epoch 2: 100%|ββββββββββ| 1/1 [00:07<00:00, 7.63s/it, loss=2.3395]
|
6 |
+
Epoch 2 metrics: {'epoch': 2, 'loss': 2.3394534587860107, 'val_loss': 2.359609365463257}
|
7 |
+
Epoch 3: 100%|ββββββββββ| 1/1 [00:07<00:00, 7.26s/it, loss=2.3447]
|
8 |
+
Epoch 3 metrics: {'epoch': 3, 'loss': 2.3446695804595947, 'val_loss': 2.359609365463257}
|
9 |
+
Epoch 4: 100%|ββββββββββ| 1/1 [00:05<00:00, 5.76s/it, loss=2.3448]
|
10 |
+
Epoch 4 metrics: {'epoch': 4, 'loss': 2.3447532653808594, 'val_loss': 2.359609365463257}
|
11 |
+
Epoch 5: 100%|ββββββββββ| 1/1 [00:05<00:00, 5.67s/it, loss=2.3693]
|
12 |
+
Epoch 5 metrics: {'epoch': 5, 'loss': 2.3692800998687744, 'val_loss': 2.3596091270446777}
|
13 |
+
Epoch 6: 100%|ββββββββββ| 1/1 [00:06<00:00, 6.38s/it, loss=2.3505]
|
14 |
+
Epoch 6 metrics: {'epoch': 6, 'loss': 2.3504748344421387, 'val_loss': 2.359609603881836}
|
15 |
+
Epoch 7: 100%|ββββββββββ| 1/1 [00:06<00:00, 6.72s/it, loss=2.3486]
|
16 |
+
Epoch 7 metrics: {'epoch': 7, 'loss': 2.3485639095306396, 'val_loss': 2.359609603881836}
|
17 |
+
Epoch 8: 100%|ββββββββββ| 1/1 [00:05<00:00, 5.68s/it, loss=2.3788]
|
18 |
+
Epoch 8 metrics: {'epoch': 8, 'loss': 2.3788001537323, 'val_loss': 2.359609365463257}
|
19 |
+
Epoch 9: 100%|ββββββββββ| 1/1 [00:06<00:00, 6.99s/it, loss=2.3707]
|
20 |
+
Epoch 9 metrics: {'epoch': 9, 'loss': 2.3706822395324707, 'val_loss': 2.359609365463257}
|
21 |
+
Traceback (most recent call last):
|
22 |
+
File "<string>", line 1, in <module>
|
23 |
+
NameError: name 'prepare_model_for_serving' is not defined
|
24 |
+
<ipython-input-38-01a39a3e998e>:31: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
|
25 |
+
model = torch.load(model_path)
|
26 |
+
<ipython-input-39-9183a6f3edb8>:31: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
|
27 |
+
model = torch.load(model_path)
|
28 |
+
/bin/bash: line 1: torch-model-archiver: command not found
|
29 |
+
Collecting torchserve
|
30 |
+
Downloading torchserve-0.12.0-py3-none-any.whl.metadata (1.4 kB)
|
31 |
+
Collecting torch-model-archiver
|
32 |
+
Downloading torch_model_archiver-0.12.0-py3-none-any.whl.metadata (1.4 kB)
|
33 |
+
Requirement already satisfied: Pillow in /usr/local/lib/python3.11/dist-packages (from torchserve) (11.1.0)
|
34 |
+
Requirement already satisfied: psutil in /usr/local/lib/python3.11/dist-packages (from torchserve) (5.9.5)
|
35 |
+
Requirement already satisfied: packaging in /usr/local/lib/python3.11/dist-packages (from torchserve) (24.2)
|
36 |
+
Requirement already satisfied: wheel in /usr/local/lib/python3.11/dist-packages (from torchserve) (0.45.1)
|
37 |
+
Collecting enum-compat (from torch-model-archiver)
|
38 |
+
Downloading enum_compat-0.0.3-py3-none-any.whl.metadata (954 bytes)
|
39 |
+
Downloading torchserve-0.12.0-py3-none-any.whl (42.2 MB)
|
40 |
+
[2K [90mββββββββββββββββββββββββββββββββββββββββ[0m [32m42.2/42.2 MB[0m [31m8.0 MB/s[0m eta [36m0:00:00[0m:00:01[0m00:01[0m
|
41 |
+
[?25hDownloading torch_model_archiver-0.12.0-py3-none-any.whl (16 kB)
|
42 |
+
Downloading enum_compat-0.0.3-py3-none-any.whl (1.3 kB)
|
43 |
+
Installing collected packages: enum-compat, torchserve, torch-model-archiver
|
44 |
+
Successfully installed enum-compat-0.0.3 torch-model-archiver-0.12.0 torchserve-0.12.0
|
45 |
+
Traceback (most recent call last):
|
46 |
+
File "/usr/local/bin/torch-model-archiver", line 8, in <module>
|
47 |
+
sys.exit(generate_model_archive())
|
48 |
+
^^^^^^^^^^^^^^^^^^^^^^^^
|
49 |
+
File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging.py", line 72, in generate_model_archive
|
50 |
+
package_model(config, manifest=manifest)
|
51 |
+
File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging.py", line 45, in package_model
|
52 |
+
model_path = ModelExportUtils.copy_artifacts(
|
53 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
54 |
+
File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging_utils.py", line 215, in copy_artifacts
|
55 |
+
shutil.copy(path, model_path)
|
56 |
+
File "/usr/lib/python3.11/shutil.py", line 431, in copy
|
57 |
+
copyfile(src, dst, follow_symlinks=follow_symlinks)
|
58 |
+
File "/usr/lib/python3.11/shutil.py", line 256, in copyfile
|
59 |
+
with open(src, 'rb') as fsrc:
|
60 |
+
^^^^^^^^^^^^^^^
|
61 |
+
FileNotFoundError: [Errno 2] No such file or directory: 'titan_handler.py'
|
62 |
+
<ipython-input-43-06d5ff2de07d>:18: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
|
63 |
+
model = torch.load(model_path)
|
64 |
+
WARNING: sun.reflect.Reflection.getCallerClass is not supported. This will impact performance.
|
65 |
+
2025-01-18T18:01:30,224 [INFO ] main org.pytorch.serve.servingsdk.impl.PluginsManager - Loading snapshot serializer plugin...
|
66 |
+
nvidia-smi not available or failed: Cannot run program "nvidia-smi": error=2, No such file or directory
|
67 |
+
2025-01-18T18:01:30,334 [DEBUG] main org.pytorch.serve.util.ConfigManager - xpu-smi not available or failed: Cannot run program "xpu-smi": error=2, No such file or directory
|
68 |
+
2025-01-18T18:01:30,338 [WARN ] main org.pytorch.serve.util.ConfigManager - Your torchserve instance can access any URL to load models. When deploying to production, make sure to limit the set of allowed_urls in config.properties
|
69 |
+
2025-01-18T18:01:30,489 [INFO ] main org.pytorch.serve.util.TokenAuthorization -
|
70 |
+
######
|
71 |
+
TorchServe now enforces token authorization by default.
|
72 |
+
This requires the correct token to be provided when calling an API.
|
73 |
+
Key file located at /content/key_file.json
|
74 |
+
Check token authorization documenation for information: https://github.com/pytorch/serve/blob/master/docs/token_authorization_api.md
|
75 |
+
######
|
76 |
+
|
77 |
+
2025-01-18T18:01:30,489 [INFO ] main org.pytorch.serve.servingsdk.impl.PluginsManager - Initializing plugins manager...
|
78 |
+
2025-01-18T18:01:30,882 [INFO ] main org.pytorch.serve.metrics.configuration.MetricConfiguration - Successfully loaded metrics configuration from /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml
|
79 |
+
2025-01-18T18:01:31,174 [INFO ] main org.pytorch.serve.ModelServer -
|
80 |
+
Torchserve version: 0.12.0
|
81 |
+
TS Home: /usr/local/lib/python3.11/dist-packages
|
82 |
+
Current directory: /content
|
83 |
+
Temp directory: /tmp
|
84 |
+
Metrics config path: /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml
|
85 |
+
Number of GPUs: 0
|
86 |
+
Number of CPUs: 2
|
87 |
+
Max heap size: 3246 M
|
88 |
+
Python executable: /usr/bin/python3
|
89 |
+
Config file: N/A
|
90 |
+
Inference address: http://127.0.0.1:8080
|
91 |
+
Management address: http://127.0.0.1:8081
|
92 |
+
Metrics address: http://127.0.0.1:8082
|
93 |
+
Model Store: /content/model_store
|
94 |
+
Initial Models: titan=titan.mar
|
95 |
+
Log dir: /content/logs
|
96 |
+
Metrics dir: /content/logs
|
97 |
+
Netty threads: 0
|
98 |
+
Netty client threads: 0
|
99 |
+
Default workers per model: 2
|
100 |
+
Blacklist Regex: N/A
|
101 |
+
Maximum Response Size: 6553500
|
102 |
+
Maximum Request Size: 6553500
|
103 |
+
Limit Maximum Image Pixels: true
|
104 |
+
Prefer direct buffer: false
|
105 |
+
Allowed Urls: [file://.*|http(s)?://.*]
|
106 |
+
Custom python dependency for model allowed: false
|
107 |
+
Enable metrics API: true
|
108 |
+
Metrics mode: LOG
|
109 |
+
Disable system metrics: false
|
110 |
+
Workflow Store: /content/model_store
|
111 |
+
CPP log config: N/A
|
112 |
+
Model config: N/A
|
113 |
+
System metrics command: default
|
114 |
+
Model API enabled: false
|
115 |
+
2025-01-18T18:01:31,235 [INFO ] main org.pytorch.serve.ModelServer - Loading initial models: titan.mar
|
116 |
+
2025-01-18T18:02:24,534 [DEBUG] main org.pytorch.serve.wlm.ModelVersionedRefs - Adding new version 1.0 for model titan
|
117 |
+
2025-01-18T18:02:24,535 [DEBUG] main org.pytorch.serve.wlm.ModelVersionedRefs - Setting default version to 1.0 for model titan
|
118 |
+
2025-01-18T18:02:24,535 [INFO ] main org.pytorch.serve.wlm.ModelManager - Model titan loaded.
|
119 |
+
2025-01-18T18:02:24,536 [DEBUG] main org.pytorch.serve.wlm.ModelManager - updateModel: titan, count: 2
|
120 |
+
2025-01-18T18:02:24,575 [DEBUG] W-9000-titan_1.0 org.pytorch.serve.wlm.WorkerLifeCycle - Worker cmdline: [/usr/bin/python3, /usr/local/lib/python3.11/dist-packages/ts/model_service_worker.py, --sock-type, unix, --sock-name, /tmp/.ts.sock.9000, --metrics-config, /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml]
|
121 |
+
2025-01-18T18:02:24,581 [DEBUG] W-9001-titan_1.0 org.pytorch.serve.wlm.WorkerLifeCycle - Worker cmdline: [/usr/bin/python3, /usr/local/lib/python3.11/dist-packages/ts/model_service_worker.py, --sock-type, unix, --sock-name, /tmp/.ts.sock.9001, --metrics-config, /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml]
|
122 |
+
2025-01-18T18:02:24,602 [INFO ] main org.pytorch.serve.ModelServer - Initialize Inference server with: EpollServerSocketChannel.
|
123 |
+
2025-01-18T18:02:26,923 [INFO ] main org.pytorch.serve.ModelServer - Torchserve stopped.
|
124 |
+
java.io.IOException: Failed to bind to address: http://127.0.0.1:8080
|
125 |
+
at org.pytorch.serve.ModelServer.initializeServer(ModelServer.java:354)
|
126 |
+
at org.pytorch.serve.ModelServer.startRESTserver(ModelServer.java:415)
|
127 |
+
at org.pytorch.serve.ModelServer.startAndWait(ModelServer.java:124)
|
128 |
+
at org.pytorch.serve.ModelServer.main(ModelServer.java:105)
|
129 |
+
Caused by: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use
|
130 |
+
COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
|
131 |
+
node 6 root 21u IPv6 19939 0t0 TCP *:8080 (LISTEN)
|
132 |
+
node 6 root 26u IPv6 1016000 0t0 TCP 17c4989733c5:8080->172.28.0.1:39786 (ESTABLISHED)
|
133 |
+
node 6 root 28u IPv6 1017108 0t0 TCP 17c4989733c5:8080->172.28.0.1:45550 (ESTABLISHED)
|
134 |
+
node 6 root 29u IPv6 25417 0t0 TCP 17c4989733c5:8080->172.28.0.1:55124 (ESTABLISHED)
|
135 |
+
node 6 root 31u IPv6 867221 0t0 TCP 17c4989733c5:8080->172.28.0.1:50646 (ESTABLISHED)
|
136 |
+
usage: torchserve [-h] [-v | --start | --stop] [--ts-config TS_CONFIG] [--model-store MODEL_STORE]
|
137 |
+
[--workflow-store WORKFLOW_STORE]
|
138 |
+
[--models MODEL_PATH1 MODEL_NAME=MODEL_PATH2... [MODEL_PATH1 MODEL_NAME=MODEL_PATH2... ...]]
|
139 |
+
[--log-config LOG_CONFIG] [--cpp-log-config CPP_LOG_CONFIG] [--foreground]
|
140 |
+
[--no-config-snapshots] [--plugins-path PLUGINS_PATH] [--disable-token-auth]
|
141 |
+
[--enable-model-api]
|
142 |
+
torchserve: error: unrecognized arguments: --inference-address http://127.0.0.1:8083
|
143 |
+
/bin/bash: line 1: nano: command not found
|
144 |
+
/usr/local/lib/python3.11/dist-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'Repository' (from 'huggingface_hub.repository') is deprecated and will be removed from version '1.0'. Please prefer the http-based alternatives instead. Given its large adoption in legacy code, the complete removal is only planned on next major release.
|
145 |
+
For more details, please read https://huggingface.co/docs/huggingface_hub/concepts/git_vs_http.
|
146 |
+
warnings.warn(warning_message, FutureWarning)
|