vasilii-feofanov commited on
Commit
572b9c9
·
verified ·
1 Parent(s): 3647e4f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -27,14 +27,13 @@ Paris Noah's Ark Lab consists of 3 research teams that cover the following topic
27
  ### Preprints
28
 
29
  - [TAG: A Decentralized Framework for Multi-Agent Hierarchical Reinforcement Learning](https://huggingface.co/papers/2502.15425): distributed multi-agent hierarchical reinforcement learning framework.
30
- - [AdaPTS: Adapting Univariate Foundation Models to Probabilistic Multivariate Time Series Forecasting](https://arxiv.org/abs/2502.10235): simple yet powerful tricks to extend foundation models.
31
  - [SKADA-Bench: Benchmarking Unsupervised Domain Adaptation Methods with Realistic Validation](https://arxiv.org/abs/2407.11676): benchmark of shallow and deep domain adaptation method with realistic validation
32
  - [Clustering Head: A Visual Case Study of the Training Dynamics in Transformers](https://arxiv.org/abs/2410.24050): visual and theoretical understanding of training dynamics in transformers.
33
  - [Large Language Models as Markov Chains](https://huggingface.co/papers/2410.02724): theoretical insights on their generalization and convergence properties.
34
  - [A Systematic Study Comparing Hyperparameter Optimization Engines on Tabular Data](https://balazskegl.medium.com/navigating-the-maze-of-hyperparameter-optimization-insights-from-a-systematic-study-6019675ea96c): insights to navigate the maze of hyperopt techniques.
35
 
36
  ### 2025
37
-
38
  - *(ICLR'25)* - [Zero-shot Model-based Reinforcement Learning using Large Language Models](https://huggingface.co/papers/2410.11711): disentangled in-context learning for multivariate time series forecasting and model-based RL.
39
  - *(ICASSP'25)* - [Easing Optimization Paths: A Circuit Perspective](https://arxiv.org/abs/2501.02362): mechanistic study of training dynamics in transformers.
40
  - *(Neurocomputing)* - [Self-training: A survey](https://www.sciencedirect.com/science/article/pii/S0925231224016758): know more about pseudo-labeling strategies.
 
27
  ### Preprints
28
 
29
  - [TAG: A Decentralized Framework for Multi-Agent Hierarchical Reinforcement Learning](https://huggingface.co/papers/2502.15425): distributed multi-agent hierarchical reinforcement learning framework.
 
30
  - [SKADA-Bench: Benchmarking Unsupervised Domain Adaptation Methods with Realistic Validation](https://arxiv.org/abs/2407.11676): benchmark of shallow and deep domain adaptation method with realistic validation
31
  - [Clustering Head: A Visual Case Study of the Training Dynamics in Transformers](https://arxiv.org/abs/2410.24050): visual and theoretical understanding of training dynamics in transformers.
32
  - [Large Language Models as Markov Chains](https://huggingface.co/papers/2410.02724): theoretical insights on their generalization and convergence properties.
33
  - [A Systematic Study Comparing Hyperparameter Optimization Engines on Tabular Data](https://balazskegl.medium.com/navigating-the-maze-of-hyperparameter-optimization-insights-from-a-systematic-study-6019675ea96c): insights to navigate the maze of hyperopt techniques.
34
 
35
  ### 2025
36
+ - *(ICML'25)* - [AdaPTS: Adapting Univariate Foundation Models to Probabilistic Multivariate Time Series Forecasting](https://arxiv.org/abs/2502.10235): simple yet powerful tricks to extend foundation models.
37
  - *(ICLR'25)* - [Zero-shot Model-based Reinforcement Learning using Large Language Models](https://huggingface.co/papers/2410.11711): disentangled in-context learning for multivariate time series forecasting and model-based RL.
38
  - *(ICASSP'25)* - [Easing Optimization Paths: A Circuit Perspective](https://arxiv.org/abs/2501.02362): mechanistic study of training dynamics in transformers.
39
  - *(Neurocomputing)* - [Self-training: A survey](https://www.sciencedirect.com/science/article/pii/S0925231224016758): know more about pseudo-labeling strategies.