This page gives an overview of how to use Databricks Feature Store in a machine learning workflow for both online and batch use cases.
The typical machine learning workflow using Feature Store follows this path:
Write code to convert raw data into features and create a Spark DataFrame containing the desired features.
Write the DataFrame as a feature table in Feature Store.
Train a model using features from the feature store. When you do this, the model stores the specifications of features used for training. When the model is used for inference, it automatically joins features from the appropriate feature tables.
Register model in Model Registry.
You can now use the model to make predictions on new data.
For batch use cases, the model automatically retrieves the features it needs from Feature Store.
For real-time serving use cases, publish the features to an online store.
At inference time, the model reads pre-computed features from the online feature store and joins them with the data provided in the client request to the model serving endpoint.