Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Feast on Kubernetes is only supported using Feast 0.9 (and below). We are working to add support for Feast on Kubernetes with the latest release of Feast (0.10+). Please see our roadmap for more details.
If you would like to deploy a new installation of Feast, click on Install Feast
If you would like to connect to an existing Feast deployment, click on Connect to Feast
If you would like to learn more about Feast, click on Learn Feast
This guide installs Feast on GKE using our reference Terraform configuration.
The Terraform configuration used here is a greenfield installation that neither assumes anything about, nor integrates with, existing resources in your GCP account. The Terraform configuration presents an easy way to get started, but you may want to customize this set up before using Feast in production.
This Terraform configuration creates the following resources:
GKE cluster
Feast services running on GKE
Google Memorystore (Redis) as online store
Dataproc cluster
Kafka running on GKE, exposed to the dataproc cluster via internal load balancer
Install Terraform > = 0.12 (tested with 0.13.3)
Install Helm (tested with v3.3.4)
GCP authentication and sufficient privilege to create the resources listed above.
Create a .tfvars
file underfeast/infra/terraform/gcp
. Name the file. In our example, we use my_feast.tfvars
. You can see the full list of configuration variables in variables.tf
. Sample configurations are provided below:
After completing the configuration, initialize Terraform and apply:
This guide installs Feast on an existing Kubernetes cluster, and ensures the following services are running:
Feast Core
Feast Online Serving
Postgres
Redis
Feast Jupyter (Optional)
Prometheus (Optional)
Add the Feast Helm repository and download the latest charts:
Feast includes a Helm chart that installs all necessary components to run Feast Core, Feast Online Serving, and an example Jupyter notebook.
Feast Core requires Postgres to run, which requires a secret to be set on Kubernetes:
Install Feast using Helm. The pods may take a few minutes to initialize.
After all the pods are in a RUNNING
state, port-forward to the Jupyter Notebook Server in the cluster:
You can now connect to the bundled Jupyter Notebook Server at localhost:8888
and follow the example Jupyter notebook.
This guide installs Feast on an existing IBM Cloud Kubernetes cluster or Red Hat OpenShift on IBM Cloud , and ensures the following services are running:
Feast Core
Feast Online Serving
Postgres
Redis
Kafka (Optional)
Feast Jupyter (Optional)
Prometheus (Optional)
Install Kubectl that matches the major.minor versions of your IKS or Install the OpenShift CLI that matches your local operating system and OpenShift cluster version.
Install Helm 3
Install Kustomize
:warning: If you have Red Hat OpenShift Cluster on IBM Cloud skip to this section.
By default, IBM Cloud Kubernetes cluster uses IBM Cloud File Storage based on NFS as the default storage class, and non-root users do not have write permission on the volume mount path for NFS-backed storage. Some common container images in Feast, such as Redis, Postgres, and Kafka specify a non-root user to access the mount path in the images. When containers are deployed using these images, the containers fail to start due to insufficient permissions of the non-root user creating folders on the mount path.
IBM Cloud Block Storage allows for the creation of raw storage volumes and provides faster performance without the permission restriction of NFS-backed storage
Therefore, to deploy Feast we need to set up IBM Cloud Block Storage as the default storage class so that you can have all the functionalities working and get the best experience from Feast.
Follow the instructions to install the Helm version 3 client on your local machine.
Add the IBM Cloud Helm chart repository to the cluster where you want to use the IBM Cloud Block Storage plug-in.
Install the IBM Cloud Block Storage plug-in. When you install the plug-in, pre-defined block storage classes are added to your cluster.
Example output:
Verify that all block storage plugin pods are in a "Running" state.
Verify that the storage classes for Block Storage were added to your cluster.
Set the Block Storage as the default storageclass.
Example output:
Security Context Constraint Setup (OpenShift only)
By default, in OpenShift, all pods or containers will use the Restricted SCC which limits the UIDs pods can run with, causing the Feast installation to fail. To overcome this, you can allow Feast pods to run with any UID by executing the following:
Install Feast using kustomize. The pods may take a few minutes to initialize.
You may optionally enable the Feast Jupyter component which contains code examples to demonstrate Feast. Some examples require Kafka to stream real time features to the Feast online serving. To enable, edit the following properties in the values.yaml
under the manifests/contrib/feast
folder:
Then regenerate the resource manifests and deploy:
After all the pods are in a RUNNING
state, port-forward to the Jupyter Notebook Server in the cluster:
You can now connect to the bundled Jupyter Notebook Server at localhost:8888
and follow the example Jupyter notebook.
When running the minimal_ride_hailing_example Jupyter Notebook example the following errors may occur:
When running job = client.get_historical_features(...)
:
or
Add the following environment variable:
When running job.get_status()
Add the following environment variable:
When running job = client.start_stream_to_online_ingestion(...)
Add the following environment variable:
This guide installs Feast on AWS using our reference Terraform configuration.
The Terraform configuration used here is a greenfield installation that neither assumes anything about, nor integrates with, existing resources in your AWS account. The Terraform configuration presents an easy way to get started, but you may want to customize this set up before using Feast in production.
This Terraform configuration creates the following resources:
Kubernetes cluster on Amazon EKS (3x r3.large nodes)
Kafka managed by Amazon MSK (2x kafka.t3.small nodes)
Postgres database for Feast metadata, using serverless Aurora (min capacity: 2)
Redis cluster, using Amazon Elasticache (1x cache.t2.micro)
Amazon EMR cluster to run Spark (3x spot m4.xlarge)
Staging S3 bucket to store temporary data
Create an AWS account and configure credentials locally
Install Terraform > = 0.12 (tested with 0.13.3)
Install Helm (tested with v3.3.4)
Create a .tfvars
file underfeast/infra/terraform/aws
. Name the file. In our example, we use my_feast.tfvars
. You can see the full list of configuration variables in variables.tf
. At a minimum, you need to set name_prefix
and an AWS region:
After completing the configuration, initialize Terraform and apply:
Starting may take a minute. A kubectl configuration file is also created in this directory, and the file's name will start with kubeconfig_
and end with a random suffix.
After all pods are running, connect to the Jupyter Notebook Server running in the cluster.
To connect to the remote Feast server you just created, forward a port from the remote k8s cluster to your local machine. Replace kubeconfig_XXXXXXX
below with the kubeconfig file name Terraform generates for you.
You can now connect to the bundled Jupyter Notebook Server at localhost:8888
and follow the example Jupyter notebook.
A production deployment of Feast is deployed using Kubernetes.
This guide installs Feast into an existing Kubernetes cluster using Helm. The installation is not specific to any cloud platform or environment, but requires Kubernetes and Helm.
This guide installs Feast into an AWS environment using Terraform. The Terraform script is opinionated and intended to allow you to start quickly.
This guide installs Feast into an Azure AKS environment with Helm.
This guide installs Feast into an Azure environment using Terraform. The Terraform script is opinionated and intended to allow you to start quickly.
This guide installs Feast into a Google Cloud environment using Terraform. The Terraform script is opinionated and intended to allow you to start quickly.
This guide installs Feast into an existing IBM Cloud Kubernetes Service or Red Hat OpenShift on IBM Cloud using Kustomize.
The Feast Python SDK is used as a library to interact with a Feast deployment.
Define, register, and manage entities and features
Ingest data into Feast
Build and retrieve training datasets
Retrieve online features
The Feast CLI is a command line implementation of the Feast Python SDK.
Define, register, and manage entities and features from the terminal
Ingest data into Feast
Manage ingestion jobs
The following clients can be used to retrieve online feature values:
This guide is meant for exploratory purposes only. It allows users to run Feast locally using Docker Compose instead of Kubernetes. The goal of this guide is for users to be able to quickly try out the full Feast stack without needing to deploy to Kubernetes. It is not meant for production use.
This guide shows you how to deploy Feast using . Docker Compose allows you to explore the functionality provided by Feast while requiring only minimal infrastructure.
This guide includes the following containerized components:
Feast Core with Postgres
Feast Online Serving with Redis.
Feast Job Service
A Jupyter Notebook Server with built in Feast example(s). For demo purposes only.
A Kafka cluster for testing streaming ingestion. For demo purposes only.
Clone the latest stable version of Feast from the :
Create a new configuration file:
Start Feast with Docker Compose:
Wait until all all containers are in a running state:
You can now connect to the bundled Jupyter Notebook Server running at localhost:8888
and follow the example Jupyter notebook.
Please ensure that the following ports are available on your host machine:
6565
6566
8888
9094
5432
If some of the containers continue to restart, or you are unable to access a service, inspect the logs using the following command:
The Feast Docker Compose setup can be configured by modifying properties in your .env
file.
To access Google Cloud Storage as a data source, the Docker Compose installation requires access to a GCP service account.
Grant the service account access to your bucket(s).
Copy the service account to the path you have configured in .env
under GCP_SERVICE_ACCOUNT
.
Restart your Docker Compose setup of Feast.
Explore the following resources to learn more about Feast:
describes all important Feast API concepts.
provides guidance on completing Feast workflows.
contains Jupyter notebooks that you can run on your Feast deployment.
contains information about both advanced and operational aspects of Feast.
contains detailed API and design documents for advanced users.
contains resources for anyone who wants to contribute to Feast.
The best way to learn Feast is to use it. Jump over to our guide to have one of our examples running in no time at all!
This guide installs Feast on Azure Kubernetes cluster (known as AKS), and ensures the following services are running:
Feast Core
Feast Online Serving
Postgres
Redis
Spark
Kafka
Feast Jupyter (Optional)
Prometheus (Optional)
Install and configure
Install and configure
Install
Add the Feast Helm repository and download the latest charts:
Feast includes a Helm chart that installs all necessary components to run Feast Core, Feast Online Serving, and an example Jupyter notebook.
Feast Core requires Postgres to run, which requires a secret to be set on Kubernetes:
Install Feast using Helm. The pods may take a few minutes to initialize.
and ensure the service account used by Feast has permissions to manage Spark Application resources. This depends on your k8s setup, but typically you'd need to configure a Role and a RoleBinding like the one below:
After all the pods are in a RUNNING
state, port-forward to the Jupyter Notebook Server in the cluster:
You can now connect to the bundled Jupyter Notebook Server at localhost:8888
and follow the example Jupyter notebook.
This guide installs Feast on Azure using our .
The Terraform configuration used here is a greenfield installation that neither assumes anything about, nor integrates with, existing resources in your Azure account. The Terraform configuration presents an easy way to get started, but you may want to customize this set up before using Feast in production.
This Terraform configuration creates the following resources:
Kubernetes cluster on Azure AKS
Kafka managed by HDInsight
Postgres database for Feast metadata, running as a pod on AKS
Redis cluster, using Azure Cache for Redis
to run Spark
Staging Azure blob storage container to store temporary data
Create an Azure account and
Install (tested with 0.13.5)
Install (tested with v3.4.2)
Create a .tfvars
file underfeast/infra/terraform/azure
. Name the file. In our example, we use my_feast.tfvars
. You can see the full list of configuration variables in variables.tf
. At a minimum, you need to set name_prefix
and resource_group
:
After completing the configuration, initialize Terraform and apply:
After all pods are running, connect to the Jupyter Notebook Server running in the cluster.
To connect to the remote Feast server you just created, forward a port from the remote k8s cluster to your local machine.
You can now connect to the bundled Jupyter Notebook Server at localhost:8888
and follow the example Jupyter notebook.
If a port conflict cannot be resolved, you can modify the port mappings in the provided file to use different ports on the host.
If you are unable to resolve the problem, visit to create an issue.
Create a new and save a JSON key.
Create an AKS cluster with Azure CLI. The detailed steps can be found , and a high-level walk through includes:
Follow the documentation , and Feast documentation to
If you are running the , you may want to make sure the following environment variables are correctly set:
Install the Feast CLI using pip:
Configure the CLI to connect to your Feast Core deployment:
By default, all configuration is stored in ~/.feast/config
The CLI is a wrapper around the Feast Python SDK: