Full Simple Model


This page contains details of how you can build a simple model using NeuralProphet with minimal features.


After downloading the code repository (via git clone), change to the repository directory (cd neural_prophet) and install neuralprophet as python package with pip install .


If you plan to use the package in a Jupyter notebook, it is recommended to install the ‘live’ package version with pip install .[live]. This will allow you to enable plot_live_loss in the train function to get a live plot of train (and validation) loss.


The input data format expected by the neural_prophet package is the same as in original prophet. It should have two columns, ds which has the timestamps and y column which contains the observed values of the time series.













Throughout this documentation, we will be using the time series data of the log daily page views for the Peyton Manning Wikipedia page. The data can be imported as follows.

import pandas as pd

data_location = "https://raw.githubusercontent.com/ourownstory/neuralprophet-data/main/datasets/"

df = pd.read_csv(data_location + 'wp_log_peyton_manning.csv')

Simple Model

A simple model with neural_prophet for this dataset can be fitted by creating an object of the NeuralProphet class as follows and calling the fit function. This fits a model with the default settings in the model. Note that the frequency of data is set globally here. Valid timeseries frequency settings are pandas timeseries offset aliases.

m = NeuralProphet()
metrics = m.fit(df, freq="D")

Once the model is fitted, we can make predictions using the fitted model. Here we are predicting in-sample over our data to evaluate the model fit. We could do the same for a holdout set.

future = m.make_future_dataframe(df=df, periods=365)
forecast = m.predict(df=future)


Let’s visualize the obtained forecast:

fig_forecast = m.plot(forecast)

This is a simple model with a trend, a weekly seasonality and a yearly seasonality estimated by default. You can also look at the individual components separately as below.

fig_comp = m.plot_components(forecast)

The individual coefficient values can also be plotted as below to gain further insights.

fig_param = m.plot_parameters()


There are two ways to perform model validation in NeuralProphet:

1. Manual Split

Users can split the dataset manually to validate after the model fitting like below by specifying the fraction of validation data. Thereby, the validation set is reserved from the end of the series.

m = NeuralProphet()
df_train, df_test = m.split_df(df, valid_p=0.2)

You can now look at the training and validation metrics separately as below.

train_metrics = m.fit(df_train)
test_metrics = m.test(df_test)

2. Builtin Function

Alternatively, you can perform validation per every epoch during model fitting as below.

m = NeuralProphet()
metrics = m.fit(df_train, validation_df=df_test)


The variability of results comes from SGD finding different optima on different runs. The majority of the randomness comes from the random initialization of weights, different learning rates and different shuffling of the dataloader.

Although, NeuralProphet allows you to control the random number generator by setting it’s seed:

from neuralprophet import set_random_seed

This should lead to identical results every time you run the model. Note that you have to explicitly set the random seed to the same random number each time before fitting the model.


Congrats on completing the full simple model tutorial! 🥳 Let’s solve some real world applications and head over to the advanced Tutorials to your left! 🏄🏼‍♂️