Flieber, Syrup Tech, and SupChains Launch an AI-Driven Supply Chain Forecasting Competition
Welcome to the VN1 Forecasting - Accuracy Challenge! Here’s everything you need to know to participate.
🚀 Take on the VN1 Forecasting Accuracy Challenge and compete for a share of $20,000 in prizes! With a grand prize of $10,000m you’ll not only enhance your skills in predictive modeling and data analysis but also gain insights from top industry professionals. Plus, standout participants could unlock internship and placement opportunities. Showcase your talent, gather valuable feedback, and elevate your career today! 💡
Participants in this datathon are tasked with accurately forecasting future sales using
historical sales, inventory, and pricing data provided. The goal is to develop robust
predictive models that can anticipate sales trends for various products across different
clients and warehouses. Submissions will be evaluated based on their accuracy and bias
against actual sales figures. The competition is structured into two phases.
Phase 1 (12th of September onwards]
In this phase you will use the provided Phase 0 sales data to predict sales for Phase 1. This phase will last three weeks, during which there will be live leaderboard updates to track your progress and provide feedback on your predictions. At the end of Phase 1, you will receive the actual sales data for this phase.
Phase 2 [ 3rd of October2024 - 17th of October 2024]
Using both Phase 0 and Phase 1 data, you will now predict sales for Phase 2. This second phase will last two weeks, but unlike Phase 1, there will be no leaderboard updates until the competition ends . During Phase 2, participants can still submit predictions for Phase 1 to test their models on the public leaderboard.
Welcome to the VN1 Forecasting - Accuracy Challenge! Here’s everything you need to know to participate.
🚀 Take on the VN1 Forecasting Accuracy Challenge and compete for a share of $20,000 in prizes! With a grand prize of $10,000m you’ll not only enhance your skills in predictive modeling and data analysis but also gain insights from top industry professionals. Plus, standout participants could unlock internship and placement opportunities. Showcase your talent, gather valuable feedback, and elevate your career today! 💡
Participants in this datathon are tasked with accurately forecasting future sales using
historical sales, inventory, and pricing data provided. The goal is to develop robust
predictive models that can anticipate sales trends for various products across different
clients and warehouses. Submissions will be evaluated based on their accuracy and bias
against actual sales figures. The competition is structured into two phases.
Phase 1 (12th of September onwards]
In this phase you will use the provided Phase 0 sales data to predict sales for Phase 1. This phase will last three weeks, during which there will be live leaderboard updates to track your progress and provide feedback on your predictions. At the end of Phase 1, you will receive the actual sales data for this phase.
Phase 2 [ 3rd of October2024 - 17th of October 2024]
Using both Phase 0 and Phase 1 data, you will now predict sales for Phase 2. This second phase will last two weeks, but unlike Phase 1, there will be no leaderboard updates until the competition ends . During Phase 2, participants can still submit predictions for Phase 1 to test their models on the public leaderboard.
August at 19:20 UTC
October at 08:00 UTC
October at 08:00 UTC
In Phase 1, you will use the provided Phase 0 sales data to predict sales for Phase 1. This
phase will last three weeks, during which there will be live leaderboard updates to track
your progress and provide feedback on your predictions. At the end of Phase 1, you will
receive the actual sales data for this phase.
Following this, Phase 2 begins (THIS IS THE LINK TO ACCESS THE PHASE 2 WHEN THE TIME COMES). Using both Phase 0 and Phase 1 data, you will now predict
sales for Phase 2. This second phase will last two weeks, but unlike Phase 1, there will be
no leaderboard updates until the competition ends – only the last submission for Phase 2
will be taken into account.
During Phase 2, participants can still submit predictions for Phase 1 to test their models on the public leaderboard.
Phase 1 – Starts on Thursday, the 12th of September (08:00 AM UTC)
Phase 2 – Starts on Thursday, the 3rd of October (08:00 AM UTC) and ends on Thursday, the 17th of October (18:00 PM UTC)
Winner announcement – Last week of October (Week of 28th of October)
Online conference for winners to present their solutions & insights – November 2024
Your submissions will be evaluated based on a score combining accuracy and bias (both normalized as percentages). To ensure a fair comparison, we censor shortages: the score is only computed during weeks with at least four days of available inventory.
In practice, the following code will be used to evaluate submissions,
def data_competition_evaluation(phase="Phase 2", name=""):
# Submission should be loaded from a .csv file.
submission = pd.read_csv(name)
assert all(col in submission.columns for col in ["Client", "Warehouse", "Product"])
submission = submission.set_index(["Client", "Warehouse", "Product"])
submission.columns = pd.to_datetime(submission.columns)
assert (~submission.isnull().any().any())
# Load Objective
objective = pd.read_csv(f"{phase} - Sales.csv").set_index(["Client", "Warehouse", "Product"])
objective.columns = pd.to_datetime(objective.columns)
assert (submission.index == objective.index).all()
assert (submission.columns == objective.columns).all()
# Remove cases where we have 3 days of stock or less
# This is an important rule that we communicate to competitors.
in_stock = pd.read_csv(f"{phase} - Inventory.csv").set_index(["Client", "Warehouse", "Product"])
in_stock.columns = pd.to_datetime(in_stock.columns)
in_stock = in_stock.fillna(7) #If na we assume it's in stock.
in_stock = in_stock > 3
objective[~in_stock] = np.nan
# This is an important rule that we communicate to competitors.
abs_err = np.nansum(abs(submission - objective))
err = np.nansum((submission - objective))
score = abs_err + abs(err)
score /= objective.sum().sum()
print(f"{name}:,", score) #It's a percentage
Participants can use provided baselines for comparison. They serve as a starting point for participants to build and refine their models, aiming to surpass these initial benchmarks in terms of accuracy and efficiency.
By adhering to these evaluation metrics and leveraging the provided baselines, participants can demonstrate the robustness and effectiveness of their predictive models in forecasting sales trends effectively.
To submit a prediction, follow these rules,
Keep up to date by participating in our global community of data scientists and AI enthusiasts. We discuss the latest developments in data science competitions, new techniques for solving complex challenges, AI and machine learning models, and much more!