Welcome to this next lesson on Saving and Loading a TensorFlow Model. By the end of this lesson, you’ll be able to understand the importance of saving and loading models, how to save a trained TensorFlow model, load it from the saved file format, and validate the loaded model. This will give you a full cycle understanding and hands-on knowledge on how to handle models when training is done. With the provided code examples that train, save, load, and test a model, let's start our lesson!
When building machine learning models, it's important to save your models for various reasons. The most obvious one is efficiency - once you trained an intricate model that could take hours or even days to train, you want to keep the learned weights to avoid re-training. So, you’d save the model for reuse later without the need to retrain.
Not only that, but the saved model can be shared with others - if you're collaborating with other professionals or even publishing your results, it aids in reproducibility of your results by others. Finally, when deploying a model to production you'll need to load the trained model to make predictions on new data.
In the previous lessons, we trained a TensorFlow model. Now, let's save it!
Before we focus on saving our model, let's briefly revisit the key steps we took to load our data and train the model. Here's the code snippet we used to preprocess the Iris dataset:
Python1import numpy as np 2from sklearn.datasets import load_iris 3from sklearn.model_selection import train_test_split 4from sklearn.preprocessing import StandardScaler, OneHotEncoder 5 6def load_preprocessed_data(): 7 # Load the Iris dataset 8 iris = load_iris() 9 X, y = iris.data, iris.target 10 11 # Split the dataset into training and testing sets 12 X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, stratify=y, random_state=42) 13 14 # Scale the features 15 scaler = StandardScaler().fit(X_train) 16 X_train_scaled = scaler.transform(X_train) 17 X_test_scaled = scaler.transform(X_test) 18 19 # One-hot encode the targets 20 encoder = OneHotEncoder(sparse_output=False).fit(y_train.reshape(-1, 1)) 21 y_train_encoded = encoder.transform(y_train.reshape(-1, 1)) 22 y_test_encoded = encoder.transform(y_test.reshape(-1, 1)) 23 24 return X_train_scaled, X_test_scaled, y_train_encoded, y_test_encoded
And to train a model with our preprocessed data:
Python1import tensorflow as tf 2from data_preprocessing import load_preprocessed_data 3 4# Load preprocessed data 5X_train, X_test, y_train, y_test = load_preprocessed_data() 6 7# Define the model 8model = tf.keras.Sequential([ 9 tf.keras.layers.Input(shape=(4,)), 10 tf.keras.layers.Dense(10, activation='relu'), 11 tf.keras.layers.Dense(10, activation='relu'), 12 tf.keras.layers.Dense(3, activation='softmax') 13]) 14 15# Compile the model 16model.compile(optimizer='adam', 17 loss='categorical_crossentropy', 18 metrics=['accuracy']) 19 20# Train the model 21history = model.fit(X_train, y_train, epochs=150, batch_size=5, validation_data=(X_test, y_test))
In summary:
- We began by loading the preprocessed data using the
load_preprocessed_data()
function from ourdata_preprocessing.py
file to get our datasets:X_train
,X_test
,y_train
, andy_test
. - We constructed a sequential model with TensorFlow's Keras API, with the input shape tailored to our dataset, including several dense layers with ReLU and Softmax activations.
- The model was then compiled using the Adam optimizer and the categorical crossentropy loss function, with accuracy as a metric.
- Finally, we trained the model for 150 epochs with a batch size of 5, validating its performance on the test data throughout the training process.
With our training steps revisited, let's move on to saving our well-trained model.
After training your model, you can save the model's architecture, learned weights, and the configuration of the model's optimizer, so you can resume training exactly where you left off.
To save a model in TensorFlow, the save()
method of the Model
class is used. This method accepts one argument: filepath
, a string that specifies the path and the filename where the model should be saved.
Using the provided solution code as guidance, let's save our model:
Python1# Saving the model 2model.save('iris_model.keras')
By running this code, TensorFlow will write a file named iris_model.keras
in the current working directory. This file is saved with the .keras
extension, which is a standard format used by TensorFlow for saving complete models. The file contains everything we need to use the model: its architecture, its learned parameters, and the configuration of its optimizer.
Now that the model is saved, we can load it at any time without needing to train it again or write its architecture manually. To load a model in TensorFlow, we use the load_model()
function from tensorflow.keras.models
module.
This function accepts one argument: filepath
, a string specifying the path and the filename of the model file. The function returns the loaded model. Let's load the model using the code from the script:
Python1from tensorflow.keras.models import load_model 2 3# Loading the model 4loaded_model = load_model('iris_model.keras')
And voila! You've just loaded a pretrained model. If you're thinking about using this model to make predictions, you're on the right track. But wait a second! Before using the loaded model for predictions, let's verify it first.
To verify whether the loaded model works as expected, we evaluate it using the same test data that we used when training the original model. In TensorFlow, we can use the evaluate()
method of the Model
class to evaluate a model. This method accepts test data and labels as arguments and returns the loss value and metrics values for the model in test mode.
Let's evaluate the loaded model using the test data:
Python1# Verify the model by evaluating it on test data 2loss, accuracy = loaded_model.evaluate(X_test, y_test, verbose=False) 3print(f'Loaded Model - Test Accuracy: {accuracy}, Test Loss: {loss}')
The output of the above code will be:
Plain text1Loaded Model - Test Accuracy: 0.9111, Test Loss: 0.2468
This indicates that the loaded model performs just as well as the initial model before it was saved, demonstrating successful model saving and loading operations.
Congratulations! We traversed the unique topic of saving and loading TensorFlow models. We started by understanding why saving and loading models is important, moved onto the process of saving a trained TensorFlow model, and loaded it back only to validate the loaded model against meaningful criteria.
This is pivotal in a real-world context, since saving and loading models allows us to reuse trained models without having to retrain them, enhancing efficiency, reproducibility, and sharing of models. A reminder that practice makes perfect, so get ready to dive into some hands-on activities to solidify the concepts covered in this lesson. Happy learning!