rajveer43 commited on
Commit
dfb71eb
Β·
verified Β·
1 Parent(s): 8fe26d1

Upload wandb/output.log with huggingface_hub

Browse files
Files changed (1) hide show
  1. wandb/output.log +146 -0
wandb/output.log ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Epoch 0: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:09<00:00, 9.95s/it, loss=2.3673]
2
+ Epoch 0 metrics: {'epoch': 0, 'loss': 2.3673205375671387, 'val_loss': 2.359609603881836}
3
+ Epoch 1: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:05<00:00, 5.72s/it, loss=2.3390]
4
+ Epoch 1 metrics: {'epoch': 1, 'loss': 2.33897066116333, 'val_loss': 2.3596091270446777}
5
+ Epoch 2: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:07<00:00, 7.63s/it, loss=2.3395]
6
+ Epoch 2 metrics: {'epoch': 2, 'loss': 2.3394534587860107, 'val_loss': 2.359609365463257}
7
+ Epoch 3: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:07<00:00, 7.26s/it, loss=2.3447]
8
+ Epoch 3 metrics: {'epoch': 3, 'loss': 2.3446695804595947, 'val_loss': 2.359609365463257}
9
+ Epoch 4: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:05<00:00, 5.76s/it, loss=2.3448]
10
+ Epoch 4 metrics: {'epoch': 4, 'loss': 2.3447532653808594, 'val_loss': 2.359609365463257}
11
+ Epoch 5: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:05<00:00, 5.67s/it, loss=2.3693]
12
+ Epoch 5 metrics: {'epoch': 5, 'loss': 2.3692800998687744, 'val_loss': 2.3596091270446777}
13
+ Epoch 6: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:06<00:00, 6.38s/it, loss=2.3505]
14
+ Epoch 6 metrics: {'epoch': 6, 'loss': 2.3504748344421387, 'val_loss': 2.359609603881836}
15
+ Epoch 7: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:06<00:00, 6.72s/it, loss=2.3486]
16
+ Epoch 7 metrics: {'epoch': 7, 'loss': 2.3485639095306396, 'val_loss': 2.359609603881836}
17
+ Epoch 8: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:05<00:00, 5.68s/it, loss=2.3788]
18
+ Epoch 8 metrics: {'epoch': 8, 'loss': 2.3788001537323, 'val_loss': 2.359609365463257}
19
+ Epoch 9: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1/1 [00:06<00:00, 6.99s/it, loss=2.3707]
20
+ Epoch 9 metrics: {'epoch': 9, 'loss': 2.3706822395324707, 'val_loss': 2.359609365463257}
21
+ Traceback (most recent call last):
22
+ File "<string>", line 1, in <module>
23
+ NameError: name 'prepare_model_for_serving' is not defined
24
+ <ipython-input-38-01a39a3e998e>:31: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
25
+ model = torch.load(model_path)
26
+ <ipython-input-39-9183a6f3edb8>:31: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
27
+ model = torch.load(model_path)
28
+ /bin/bash: line 1: torch-model-archiver: command not found
29
+ Collecting torchserve
30
+ Downloading torchserve-0.12.0-py3-none-any.whl.metadata (1.4 kB)
31
+ Collecting torch-model-archiver
32
+ Downloading torch_model_archiver-0.12.0-py3-none-any.whl.metadata (1.4 kB)
33
+ Requirement already satisfied: Pillow in /usr/local/lib/python3.11/dist-packages (from torchserve) (11.1.0)
34
+ Requirement already satisfied: psutil in /usr/local/lib/python3.11/dist-packages (from torchserve) (5.9.5)
35
+ Requirement already satisfied: packaging in /usr/local/lib/python3.11/dist-packages (from torchserve) (24.2)
36
+ Requirement already satisfied: wheel in /usr/local/lib/python3.11/dist-packages (from torchserve) (0.45.1)
37
+ Collecting enum-compat (from torch-model-archiver)
38
+ Downloading enum_compat-0.0.3-py3-none-any.whl.metadata (954 bytes)
39
+ Downloading torchserve-0.12.0-py3-none-any.whl (42.2 MB)
40
+  ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.2/42.2 MB 8.0 MB/s eta 0:00:00:00:0100:01
41
+ [?25hDownloading torch_model_archiver-0.12.0-py3-none-any.whl (16 kB)
42
+ Downloading enum_compat-0.0.3-py3-none-any.whl (1.3 kB)
43
+ Installing collected packages: enum-compat, torchserve, torch-model-archiver
44
+ Successfully installed enum-compat-0.0.3 torch-model-archiver-0.12.0 torchserve-0.12.0
45
+ Traceback (most recent call last):
46
+ File "/usr/local/bin/torch-model-archiver", line 8, in <module>
47
+ sys.exit(generate_model_archive())
48
+ ^^^^^^^^^^^^^^^^^^^^^^^^
49
+ File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging.py", line 72, in generate_model_archive
50
+ package_model(config, manifest=manifest)
51
+ File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging.py", line 45, in package_model
52
+ model_path = ModelExportUtils.copy_artifacts(
53
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
54
+ File "/usr/local/lib/python3.11/dist-packages/model_archiver/model_packaging_utils.py", line 215, in copy_artifacts
55
+ shutil.copy(path, model_path)
56
+ File "/usr/lib/python3.11/shutil.py", line 431, in copy
57
+ copyfile(src, dst, follow_symlinks=follow_symlinks)
58
+ File "/usr/lib/python3.11/shutil.py", line 256, in copyfile
59
+ with open(src, 'rb') as fsrc:
60
+ ^^^^^^^^^^^^^^^
61
+ FileNotFoundError: [Errno 2] No such file or directory: 'titan_handler.py'
62
+ <ipython-input-43-06d5ff2de07d>:18: FutureWarning: You are using `torch.load` with `weights_only=False` (the current default value), which uses the default pickle module implicitly. It is possible to construct malicious pickle data which will execute arbitrary code during unpickling (See https://github.com/pytorch/pytorch/blob/main/SECURITY.md#untrusted-models for more details). In a future release, the default value for `weights_only` will be flipped to `True`. This limits the functions that could be executed during unpickling. Arbitrary objects will no longer be allowed to be loaded via this mode unless they are explicitly allowlisted by the user via `torch.serialization.add_safe_globals`. We recommend you start setting `weights_only=True` for any use case where you don't have full control of the loaded file. Please open an issue on GitHub for any issues related to this experimental feature.
63
+ model = torch.load(model_path)
64
+ WARNING: sun.reflect.Reflection.getCallerClass is not supported. This will impact performance.
65
+ 2025-01-18T18:01:30,224 [INFO ] main org.pytorch.serve.servingsdk.impl.PluginsManager - Loading snapshot serializer plugin...
66
+ nvidia-smi not available or failed: Cannot run program "nvidia-smi": error=2, No such file or directory
67
+ 2025-01-18T18:01:30,334 [DEBUG] main org.pytorch.serve.util.ConfigManager - xpu-smi not available or failed: Cannot run program "xpu-smi": error=2, No such file or directory
68
+ 2025-01-18T18:01:30,338 [WARN ] main org.pytorch.serve.util.ConfigManager - Your torchserve instance can access any URL to load models. When deploying to production, make sure to limit the set of allowed_urls in config.properties
69
+ 2025-01-18T18:01:30,489 [INFO ] main org.pytorch.serve.util.TokenAuthorization -
70
+ ######
71
+ TorchServe now enforces token authorization by default.
72
+ This requires the correct token to be provided when calling an API.
73
+ Key file located at /content/key_file.json
74
+ Check token authorization documenation for information: https://github.com/pytorch/serve/blob/master/docs/token_authorization_api.md
75
+ ######
76
+
77
+ 2025-01-18T18:01:30,489 [INFO ] main org.pytorch.serve.servingsdk.impl.PluginsManager - Initializing plugins manager...
78
+ 2025-01-18T18:01:30,882 [INFO ] main org.pytorch.serve.metrics.configuration.MetricConfiguration - Successfully loaded metrics configuration from /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml
79
+ 2025-01-18T18:01:31,174 [INFO ] main org.pytorch.serve.ModelServer -
80
+ Torchserve version: 0.12.0
81
+ TS Home: /usr/local/lib/python3.11/dist-packages
82
+ Current directory: /content
83
+ Temp directory: /tmp
84
+ Metrics config path: /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml
85
+ Number of GPUs: 0
86
+ Number of CPUs: 2
87
+ Max heap size: 3246 M
88
+ Python executable: /usr/bin/python3
89
+ Config file: N/A
90
+ Inference address: http://127.0.0.1:8080
91
+ Management address: http://127.0.0.1:8081
92
+ Metrics address: http://127.0.0.1:8082
93
+ Model Store: /content/model_store
94
+ Initial Models: titan=titan.mar
95
+ Log dir: /content/logs
96
+ Metrics dir: /content/logs
97
+ Netty threads: 0
98
+ Netty client threads: 0
99
+ Default workers per model: 2
100
+ Blacklist Regex: N/A
101
+ Maximum Response Size: 6553500
102
+ Maximum Request Size: 6553500
103
+ Limit Maximum Image Pixels: true
104
+ Prefer direct buffer: false
105
+ Allowed Urls: [file://.*|http(s)?://.*]
106
+ Custom python dependency for model allowed: false
107
+ Enable metrics API: true
108
+ Metrics mode: LOG
109
+ Disable system metrics: false
110
+ Workflow Store: /content/model_store
111
+ CPP log config: N/A
112
+ Model config: N/A
113
+ System metrics command: default
114
+ Model API enabled: false
115
+ 2025-01-18T18:01:31,235 [INFO ] main org.pytorch.serve.ModelServer - Loading initial models: titan.mar
116
+ 2025-01-18T18:02:24,534 [DEBUG] main org.pytorch.serve.wlm.ModelVersionedRefs - Adding new version 1.0 for model titan
117
+ 2025-01-18T18:02:24,535 [DEBUG] main org.pytorch.serve.wlm.ModelVersionedRefs - Setting default version to 1.0 for model titan
118
+ 2025-01-18T18:02:24,535 [INFO ] main org.pytorch.serve.wlm.ModelManager - Model titan loaded.
119
+ 2025-01-18T18:02:24,536 [DEBUG] main org.pytorch.serve.wlm.ModelManager - updateModel: titan, count: 2
120
+ 2025-01-18T18:02:24,575 [DEBUG] W-9000-titan_1.0 org.pytorch.serve.wlm.WorkerLifeCycle - Worker cmdline: [/usr/bin/python3, /usr/local/lib/python3.11/dist-packages/ts/model_service_worker.py, --sock-type, unix, --sock-name, /tmp/.ts.sock.9000, --metrics-config, /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml]
121
+ 2025-01-18T18:02:24,581 [DEBUG] W-9001-titan_1.0 org.pytorch.serve.wlm.WorkerLifeCycle - Worker cmdline: [/usr/bin/python3, /usr/local/lib/python3.11/dist-packages/ts/model_service_worker.py, --sock-type, unix, --sock-name, /tmp/.ts.sock.9001, --metrics-config, /usr/local/lib/python3.11/dist-packages/ts/configs/metrics.yaml]
122
+ 2025-01-18T18:02:24,602 [INFO ] main org.pytorch.serve.ModelServer - Initialize Inference server with: EpollServerSocketChannel.
123
+ 2025-01-18T18:02:26,923 [INFO ] main org.pytorch.serve.ModelServer - Torchserve stopped.
124
+ java.io.IOException: Failed to bind to address: http://127.0.0.1:8080
125
+ at org.pytorch.serve.ModelServer.initializeServer(ModelServer.java:354)
126
+ at org.pytorch.serve.ModelServer.startRESTserver(ModelServer.java:415)
127
+ at org.pytorch.serve.ModelServer.startAndWait(ModelServer.java:124)
128
+ at org.pytorch.serve.ModelServer.main(ModelServer.java:105)
129
+ Caused by: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use
130
+ COMMAND PID USER FD TYPE DEVICE SIZE/OFF NODE NAME
131
+ node 6 root 21u IPv6 19939 0t0 TCP *:8080 (LISTEN)
132
+ node 6 root 26u IPv6 1016000 0t0 TCP 17c4989733c5:8080->172.28.0.1:39786 (ESTABLISHED)
133
+ node 6 root 28u IPv6 1017108 0t0 TCP 17c4989733c5:8080->172.28.0.1:45550 (ESTABLISHED)
134
+ node 6 root 29u IPv6 25417 0t0 TCP 17c4989733c5:8080->172.28.0.1:55124 (ESTABLISHED)
135
+ node 6 root 31u IPv6 867221 0t0 TCP 17c4989733c5:8080->172.28.0.1:50646 (ESTABLISHED)
136
+ usage: torchserve [-h] [-v | --start | --stop] [--ts-config TS_CONFIG] [--model-store MODEL_STORE]
137
+ [--workflow-store WORKFLOW_STORE]
138
+ [--models MODEL_PATH1 MODEL_NAME=MODEL_PATH2... [MODEL_PATH1 MODEL_NAME=MODEL_PATH2... ...]]
139
+ [--log-config LOG_CONFIG] [--cpp-log-config CPP_LOG_CONFIG] [--foreground]
140
+ [--no-config-snapshots] [--plugins-path PLUGINS_PATH] [--disable-token-auth]
141
+ [--enable-model-api]
142
+ torchserve: error: unrecognized arguments: --inference-address http://127.0.0.1:8083
143
+ /bin/bash: line 1: nano: command not found
144
+ /usr/local/lib/python3.11/dist-packages/huggingface_hub/utils/_deprecation.py:131: FutureWarning: 'Repository' (from 'huggingface_hub.repository') is deprecated and will be removed from version '1.0'. Please prefer the http-based alternatives instead. Given its large adoption in legacy code, the complete removal is only planned on next major release.
145
+ For more details, please read https://huggingface.co/docs/huggingface_hub/concepts/git_vs_http.
146
+ warnings.warn(warning_message, FutureWarning)