Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

This template is brought to you by Optimizely, a leading experimentation platform.

...

Experiment plan and results

Replicating Results of Transformer Fusion Model using Favourita Dataset
https://arxiv.org/pdf/1912.09363.pdf

  • A. Batchwise Data Processing
    The favorita dataset is too large to fit into memory for in one go. So, it is processed in chunks of 300000.

  • B. Data Preprocessing
    In this following dataset each product number-store number pair is treated as a separate entity and is denoted by an embedding of the following variables:

    Code Block
    languagepy
    ['holiday_type',
    'locale',
    'locale_name',
    'description',
    'transferred',
    'city',
    'state',
    'store_type',
    'cluster',
    'family',
    'class',
    'perishable']
  • We treat each product number-store number pair as a separate entity

  • We include an additional ’open’ flag to denote whether data is present on a given day

  • Data is resampled at regular daily intervals,imputing any missing days using the last available observation

  • We apply a log-transform on the sales data, and adopt z-score normalization across all entities

  • Dropping where any record missing

  • The training set is made up of samples taken between 2015-01-01 to 2015-12-01. The validation set of samples from the 30 days after the training set. The test set of all entities over the 30-day horizon following the validation set.

  • We consider log sales, transactions, oil to be real-valued and the rest to be categorical.

    Experiment owner

    kriti mahajan

    Reviewers

    Approver

    •  

    Optimizely link

    Jira ticket(s)

    Status

    Status
    colourBlue
    titleIn review
    /
    Status
    colourYellow
    titleIn progress
    /
    Status
    titleComplete

    On this page

    Table of Contents
    maxLevel2

    ...

    Overview: Replicating Results of Transformer Fusion Model using Favourita Dataset
    https://arxiv.org/pdf/1912.09363.pdf

    Current Roadblock:
    - Fitting data into memory for modeling

    Tasks Completed Till Now

    • Step 1: Batchwise Data Processing for Preprocessing
      The favorita dataset is too large to fit into memory for in one go. So, it is processed in chunks of 300000.

    • Step 2: Data Preprocessing
      In this following dataset each product number-store number pair is treated as a separate entity and is denoted by an embedding of the following variables:

      Code Block
      languagepy
      ['holiday_type',
      'locale',
      'locale_name',
      'description',
      'transferred',
      'city',
      'state',
      'store_type',
      'cluster',
      'family',
      'class',
      'perishable']
    1. We treat each product number-store number pair as a separate entity

    2. We include an additional ’open’ flag to denote whether data is present on a given day

    3. Data is resampled at regular daily intervals,imputing any missing days using the last available observation

    4. We apply a log-transform on the sales data, and adopt z-score normalization across all entities

    5. Dropping where any record missing

    6. The training set is made up of samples taken between 2015-01-01 to 2015-12-01. The validation set of samples from the 30 days after the training set. The test set of all entities over the 30-day horizon following the validation set.

    7. We consider log sales, transactions, oil to be real-valued and the rest to be categorical.

    ...