Skip to content

DATA MACHINA AI NEWSLETTER - A DEEP DIVE INTO AI/ ML, EVERY WEEK

Keep up with the latest in Artificial Intelligence / Machine Learning research, projects & repos

  • Home
  • What is this?
  • Why subscribe?

Data Machina #240

Foundation Models, Transformers and Time-Series. TimesFM. Lag-Llama for TSF.. MOMENT. Mixture of Linear Experts. BUD-E Open Voice Assistants. MoneyPrinter GenAI Videos. Multimodal llama Cookbook. 

Foundation Models, Transformers and Time-Series. Statisticians and econometricians have been searching for the Holy Grail of time-series forecasting (TSF) for more than 40 years. “Classical” models like ARIMA still work remarkably well in some TSF scenarios. But today, the cool stuff is all about transformer-based, DL & foundation models for TSF. How “good” are these new DL models for TSF? How do we evaluate these models? Do these new models really achieve SOTA performance as some papers claim? Are DL researchers cherrypicking ts datasets to easily fit a SOTA TSF DL model?…

Real world time-series data is complex, messy, and it has a lot of subtleties like: randomness of errors, trend, seasonality, linearity, and stability… Compiling a solid, large ts dataset that suits the pre-training of a large DL model is hard and expensive. Most time series data come with huge noise and poor context. Since transfer learning is one of the pillars of foundation models: How do we figure out what ”knowledge” can be actually transferred across different time series so that the foundation model learns? 

But first let’s see what’s the very latest in transformer-based, DL and foundations models for TSF.

A new Foundation Model for TSF v.3. A group of Google researchers recently introduced their latest time-series foundation model for forecasting. The researchers claim that the model has out-of-the-box zero-shot capability and near SOTA performance on a variety of public datasets. Paper: A Decoder-only Foundation Model for Time-series Forecasting


A new open source Foundation Model for TSF. A few days ago, a group of researchers working at ServiceNow, open sourced Lag-Llama: the first open-source foundation model for time series forecasting. Checkout the repo, paper and notebooks here: Lag-Llama: Foundation Models for Probabilistic Time Series Forecasting. 


A new family of open Foundation Models for TSF. This week, researchers at CMU introduced MOMENT. The paper describes how the model addresses the challenges of pre-training large models on time-series. Checkout the paper, and official Pycode implementation: MOMENT: A Family of Open Time-series Foundation Models.

Mixture-of-Experts for Long-term TSF. Near SOTA, or SOTA Linear-centric models for TSF are not able to adapt their prediction rules to periodic changes in time series patterns. To address this challenge, a group of Microsoft researchers proposed Mixture-of-Linear-Experts (MoLE): a Mixture-of-Experts-style augmentation for linear-centric models. The researchers claim that MoLE achieves SOTA in almost 70% of evaluations and datasets. Paper: Mixture-of-Linear-Experts for Long-term Time Series Forecasting


And now a few food for thought snippets below: 

  • Transformers is not what you need for TSF. Curated by @valeman, a very vocal, opinionated researcher specialised in conformal prediction (checkout a gentle intro to CP) and probabilistic TSF. This is a quite interesting repository showing papers on why possibly transformers don’t work in time series forecasting. Repo: Transformers_Are_What_You_Dont_Need
  • Deep Learning vs Statistics  for TSF. This is a great blogpost describing the pros and cons of statistical vs. DL methods for TSF, and tips on when to use DL or stats methods. Although not updated with the very latest in transformer-based, deep learning methods for TSF, the blogpost concludes that it’s early days in DL for TSF. Blogpost: Time-Series Forecasting: Deep Learning vs Statistics — Who Wins?
  • A tutorial on TSF evaluation for data scientists and ML people. This is a tutorial-like paper on well-proven times-series forecast evaluation techniques that are usually neglected by data scientists and ML researchers. The paper’s goal is to bridge the knowledge gap between traditional methods of forecasting and current state-of-the-art DL/ ML techniques. Forecast evaluation for data scientists: common pitfalls and best practices
  • A great free book on forecasting. Considered my many researchers and practitioners, “the bible” of forecasting. Written by probably one the world’s top expert in forecasting. Forecasting: Principles and Practice 3rd ed
  • A mega study on what works in time-series forecasting. Last year Nixtla, conducted a study on TSF with a huge ts dataset containing 100 billion time series points. Nixtla is a startup well known for its research towards achieving SOTA in TSF, and the developer behind TimeGPT, a Generative AI mode for time series. The study concluded that LightGBM significantly outperformed TimeGPT and all other DL models. What Truly Works in Time Service Forecasting: The Nixtla Mega Study

Have a nice week.

10 Link-o-Troned

  1. Emergent Deception & Emergent Optimisation in AI
  2. Visualising Representations: Deep Learning & Human Beings
  3. Thinking about HQ Human Data in DL Model Training
  4. A Guide to The Landscape of LLM Programming Abstractions
  5. BUD-E: A New OSS Project for Human-Like AI Voice Assistants
  6. Beyond Self-Attention: How a Small Model Predicts the Next Token
  7. Winners of the Vesuvius AI Challenge & The TimesSformer
  8. [brilliant 🙂 ]The World’s Most Responsible AI Model (blog, demo)
  9. Odyssey 2024 – Emotions Recognition Challenge with SSL Models
  10. [watch] Fully Autonomous Androids Run by a Vision NNet in Real-time

Share Data Machina with your friends


the ML Pythonista

  1. MoneyPrinter – Automated YouTube Video Generation
  2. How to: Efficient Linear Model Merging for LLMs
  3. The Multimodal Ollama-LlaVA Cookbook

Deep & Other Learning Bits

  1. Google Research – Graph Neural Networks in TensorFlow
  2. Physics-Informed NNets: An Application-Centric Guide
  3. Matryoshka Representation Learning (MRL) from the Ground Up

AI/ DL ResearchDocs

  1. Deepmind: Grandmaster-Level Chess Without Search
  2. Why Do Random Forests Work?
  3. Future Directions in Foundations of Graph Machine Learning

MLOps Untangled

  1. [free webinar] AI in Production
  2. Why You Need LLMOps
  3. Automate Insurance Claim Lifecycle with Agents, KBs & Amazon Bedrock

data v-i-s-i-o-n-s

  1. [interactive dataviz] London Traffic: The Worst in the World
  2. [new] The Quad-Tile Chart & Squaremap: Squarify Your Data
  3. Du Bois Visualization Challenge: 2024

AI startups -> radar

  1. Daedalus – AI for Advanced Precision Parts Manufacturing
  2. Unlearn.ai – AI + Digital Twins for Clinical Trials
  3. Beam AI – An AI Assistant for Construction Costs Estimation

ML Datasets & Stuff

  1. InstaGen: Improving Object Detection with AI Generated Synth Dataset
  2. 300+ Open Datasets for Beautiful News
  3. Ego-Exo4D- Large-scale Multi-modal, Multi-view, Video Dataset

Postscript, etc 

Keep up with the very latest in AI / Machine Learning research, projects & repos. A weekly digest packed with AI / ML insights & updates that you won’t find elsewhere

Submit your suggestions, feedback, posts and links to:

datamachina@datamachina.com

Published February 12, 2024By alg01
Categorized as AI Newsletter, Artificial Intelligence, Deep Learning, DL Newsletter, Machine Learning, ML Newsletter Tagged AI, AI Newsletter, Artificial Intelligence, Deep Learning, DL, DL NEwsletter, Machine Learning, ML, ML Newsletter

Post navigation

Previous post

Data Machina #239

Next post

Data Machina #241

  • Follow on X
DATA MACHINA AI NEWSLETTER – A DEEP DIVE INTO AI/ ML, EVERY WEEK
Proudly powered by WordPress.