Taha Aksu commited on
Commit
c1b8789
·
1 Parent(s): f2ed354

Update readme

Browse files
Files changed (1) hide show
  1. README.md +81 -6
README.md CHANGED
@@ -1,10 +1,85 @@
1
  ---
 
 
2
  tags:
3
- - model_hub_mixin
4
- - pytorch_model_hub_mixin
 
 
 
 
5
  ---
6
 
7
- This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
- - Code: [More Information Needed]
9
- - Paper: [More Information Needed]
10
- - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
+ pipeline_tag: time-series-forecasting
4
  tags:
5
+ - time series
6
+ - forecasting
7
+ - pretrained models
8
+ - foundation models
9
+ - time series foundation models
10
+ - time-series
11
  ---
12
 
13
+ # Moirai-2.0-R-Small
14
+
15
+ Moirai 2.0 is a decoder-only universal time series forecasting transformer Model pre-trained on:
16
+ - Subset of [GIFT-Eval Pretrain](https://huggingface.co/datasets/Salesforce/GiftEvalPretrain), and [Train](https://huggingface.co/datasets/Salesforce/GiftEval) datasets (Non-leaking historical context).
17
+ - Mixup data generated from non-leaking subsets of [Chronos Dataset](https://arxiv.org/abs/2403.07815).
18
+ - Synthetic time series produced via KernelSynth introduced in [Chronos paper](https://arxiv.org/abs/2403.07815).
19
+ - Internal Salesforce operational data.
20
+
21
+ We make significant improvements over the first version of Moirai (please refer to the [paper](https://arxiv.org/abs/2402.02592) for previous version):
22
+ - Switched from a distributional loss to a quantile loss formulation.
23
+ - Moved from single-token to multi-token prediction, improving efficiency and stability.
24
+ - Added a data filtering mechanism to filter out non-forecastable, low quality, time series during pretraining.
25
+ - Added a new patch token embedding which includes missing value information.
26
+ - Added patch-level random mask to improve robustness of the model during inference.
27
+
28
+ ## Usage
29
+ To perform inference with Moirai 2.0, install the uni2ts library from our [GitHub repo](https://github.com/SalesforceAIResearch/uni2ts).
30
+
31
+ 1. Clone repository:
32
+ ```shell
33
+ git clone https://github.com/SalesforceAIResearch/uni2ts.git
34
+ cd uni2ts
35
+ ```
36
+
37
+ 2) Create virtual environment:
38
+ ```shell
39
+ virtualenv venv
40
+ . venv/bin/activate
41
+ ```
42
+
43
+ 3) Build from source:
44
+ ```shell
45
+ pip install -e '.[notebook]'
46
+ ```
47
+
48
+ 4) Create a `.env` file:
49
+ ```shell
50
+ touch .env
51
+ ```
52
+
53
+ A simple notebook to get started: [github_notebook_link](https://github.com/SalesforceAIResearch/uni2ts/blob/main/example/moirai_forecast.ipynb)
54
+
55
+
56
+ ## The Moirai Family
57
+
58
+ | # Model | # Parameters |
59
+ | :---: | :---: |
60
+ | [Moirai-2.0-R-Small](https://huggingface.co/Salesforce/moirai-1.0-R-small) | 11m |
61
+ | [Moirai-1.0-R-Small](https://huggingface.co/Salesforce/moirai-1.0-R-small) | 14m |
62
+ | [Moirai-1.0-R-Base](https://huggingface.co/Salesforce/moirai-1.0-R-base) | 91m |
63
+ | [Moirai-1.0-R-Large](https://huggingface.co/Salesforce/moirai-1.0-R-large) | 311m |
64
+
65
+ ## Citation
66
+
67
+ If you're using Uni2TS in your research or applications, please cite it using this BibTeX:
68
+
69
+ ```markdown
70
+ @article{woo2024unified,
71
+ title={Unified Training of Universal Time Series Forecasting Transformers},
72
+ author={Woo, Gerald and Liu, Chenghao and Kumar, Akshat and Xiong, Caiming and Savarese, Silvio and Sahoo, Doyen},
73
+ journal={arXiv preprint arXiv:2402.02592},
74
+ year={2024}
75
+ }
76
+ ```
77
+
78
+ ## Ethical Considerations
79
+
80
+ This release is for research purposes only in support of an academic paper.
81
+ Our models, datasets, and code are not specifically designed or evaluated for all downstream purposes.
82
+ We strongly recommend users evaluate and address potential concerns related to accuracy, safety, and fairness before deploying this model.
83
+ We encourage users to consider the common limitations of AI, comply with applicable laws,
84
+ and leverage best practices when selecting use cases, particularly for high-risk scenarios where errors or misuse could significantly
85
+ impact people’s lives, rights, or safety. For further guidance on use cases, refer to our AUP and AI AUP.