Azure Machine Learning Notebooks Integration#
Connecting to the Feature Store from Azure Machine Learning Notebooks requires setting up a Feature Store API key for Azure Machine Learning Notebooks and installing the HSFS on the notebook. This guide explains step by step how to connect to the Feature Store from Azure Machine Learning Notebooks.
Network Connectivity
To be able to connect to the Feature Store, please ensure that the Network Security Group of your Hopsworks instance on Azure is configured to allow incoming traffic from your compute target on ports 443, 9083 and 9085 (443,9083,9085). See Network security groups for more information. If your compute target is not in the same VNet as your Hopsworks instance and the Hopsworks instance is not accessible from the internet then you will need to configure Virtual Network Peering.
Generate an API key#
In Hopsworks, click on your username in the top-right corner and select Settings to open the user settings. Select API keys. Give the key a name and select the job, featurestore and project scopes before creating the key. Copy the key into your clipboard for the next step.
Scopes
The API key should contain at least the following scopes:
- featurestore
- project
- job
Info
You are only ably to retrieve the API key once. If you did not manage to copy it to your clipboard, delete it again and create a new one.
Connect from an Azure Machine Learning Notebook#
To access the Feature Store from Azure Machine Learning, open a Python notebook and proceed with the following steps to install HSFS and connect to the Feature Store:
Install HSFS#
To be able to access the Hopsworks Feature Store, the HSFS
Python library needs to be installed. One way of achieving this is by opening a Python notebook in Azure Machine Learning and installing the HSFS
with a magic command and pip:
!pip install hsfs[hive]~=[HOPSWORKS_VERSION]
Hive Dependencies
By default, HSFS
assumes Spark is used as execution engine and therefore Hive dependencies are not installed. Hence, if you are using a regular Python Kernel without Spark, make sure to install the "hive" extra dependencies (hsfs[hive]
).
Matching Hopsworks version
The major version of HSFS
needs to match the major version of Hopsworks. Check PyPI for available releases.
Connect to the Feature Store#
You are now ready to connect to the Hopsworks Feature Store from the notebook:
import hsfs
# Put the API key into Key Vault for any production setup:
# See, https://docs.microsoft.com/en-us/azure/machine-learning/how-to-use-secrets-in-runs
#from azureml.core import Experiment, Run
#run = Run.get_context()
#secret_value = run.get_secret(name="fs-api-key")
secret_value = 'MY_API_KEY'
# Create a connection
conn = hsfs.connection(
host='MY_INSTANCE.cloud.hopsworks.ai', # DNS of your Feature Store instance
port=443, # Port to reach your Hopsworks instance, defaults to 443
project='MY_PROJECT', # Name of your Hopsworks Feature Store project
api_key_value=secret_value, # The API key to authenticate with Hopsworks
hostname_verification=True, # Disable for self-signed certificates
engine='hive' # Choose Hive as engine
)
# Get the feature store handle for the project's feature store
fs = conn.get_feature_store()
Next Steps#
For more information about how to use the Feature Store, see the Quickstart Guide.