Skip to content

Commit d7d33c9

Browse files
authored
docs: Add references to the extras users need to install and quickstarts (#3249)
docs: Add references to the extras users need to install and quickstart templates Signed-off-by: Danny Chiao <[email protected]> Signed-off-by: Danny Chiao <[email protected]>
1 parent 025f00c commit d7d33c9

File tree

17 files changed

+311
-251
lines changed

17 files changed

+311
-251
lines changed

docs/reference/offline-stores/bigquery.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,9 @@ The BigQuery offline store provides support for reading [BigQuerySources](../dat
77
* All joins happen within BigQuery.
88
* Entity dataframes can be provided as a SQL query or can be provided as a Pandas dataframe. A Pandas dataframes will be uploaded to BigQuery as a table (marked for expiration) in order to complete join operations.
99

10+
## Getting started
11+
In order to use this offline store, you'll need to run `pip install 'feast[gcp]'`. You can get started by then running `feast init -t gcp`.
12+
1013
## Example
1114

1215
{% code title="feature_store.yaml" %}

docs/reference/offline-stores/mssql.md

Lines changed: 21 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,9 @@ The MsSQL offline store provides support for reading [MsSQL Sources](../data-sou
66

77
* Entity dataframes can be provided as a SQL query or can be provided as a Pandas dataframe.
88

9+
## Getting started
10+
In order to use this offline store, you'll need to run `pip install 'feast[azure]'`. You can get started by then following this [tutorial](https://github.com/feast-dev/feast/blob/master/docs/tutorials/azure/README.md).
11+
912
## Disclaimer
1013

1114
The MsSQL offline store does not achieve full test coverage.
@@ -34,26 +37,26 @@ offline_store:
3437
The set of functionality supported by offline stores is described in detail [here](overview.md#functionality).
3538
Below is a matrix indicating which functionality is supported by the Spark offline store.
3639
37-
| | MsSql |
38-
| :-------------------------------- | :-- |
39-
| `get_historical_features` (point-in-time correct join) | yes |
40-
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
41-
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
42-
| `offline_write_batch` (persist dataframes to offline store) | no |
43-
| `write_logged_features` (persist logged features to offline store) | no |
40+
| | MsSql |
41+
| :----------------------------------------------------------------- | :---- |
42+
| `get_historical_features` (point-in-time correct join) | yes |
43+
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
44+
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
45+
| `offline_write_batch` (persist dataframes to offline store) | no |
46+
| `write_logged_features` (persist logged features to offline store) | no |
4447

4548
Below is a matrix indicating which functionality is supported by `MsSqlServerRetrievalJob`.
4649

47-
| | MsSql |
48-
| --------------------------------- | --- |
49-
| export to dataframe | yes |
50-
| export to arrow table | yes |
51-
| export to arrow batches | no |
52-
| export to SQL | no |
53-
| export to data lake (S3, GCS, etc.) | no |
54-
| export to data warehouse | no |
55-
| local execution of Python-based on-demand transforms | no |
56-
| remote execution of Python-based on-demand transforms | no |
57-
| persist results in the offline store | yes |
50+
| | MsSql |
51+
| ----------------------------------------------------- | ----- |
52+
| export to dataframe | yes |
53+
| export to arrow table | yes |
54+
| export to arrow batches | no |
55+
| export to SQL | no |
56+
| export to data lake (S3, GCS, etc.) | no |
57+
| export to data warehouse | no |
58+
| local execution of Python-based on-demand transforms | no |
59+
| remote execution of Python-based on-demand transforms | no |
60+
| persist results in the offline store | yes |
5861

5962
To compare this set of functionality against other offline stores, please see the full [functionality matrix](overview.md#functionality-matrix).

docs/reference/offline-stores/postgres.md

Lines changed: 24 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,9 @@ The PostgreSQL offline store provides support for reading [PostgreSQLSources](..
1010
The PostgreSQL offline store does not achieve full test coverage.
1111
Please do not assume complete stability.
1212

13+
## Getting started
14+
In order to use this offline store, you'll need to run `pip install 'feast[postgres]'`. You can get started by then running `feast init -t postgres`.
15+
1316
## Example
1417

1518
{% code title="feature_store.yaml" %}
@@ -42,29 +45,29 @@ The full set of configuration options is available in [PostgreSQLOfflineStoreCon
4245
The set of functionality supported by offline stores is described in detail [here](overview.md#functionality).
4346
Below is a matrix indicating which functionality is supported by the PostgreSQL offline store.
4447

45-
| | Postgres |
46-
| :-------------------------------- | :-- |
47-
| `get_historical_features` (point-in-time correct join) | yes |
48-
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
49-
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
50-
| `offline_write_batch` (persist dataframes to offline store) | no |
51-
| `write_logged_features` (persist logged features to offline store) | no |
48+
| | Postgres |
49+
| :----------------------------------------------------------------- | :------- |
50+
| `get_historical_features` (point-in-time correct join) | yes |
51+
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
52+
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
53+
| `offline_write_batch` (persist dataframes to offline store) | no |
54+
| `write_logged_features` (persist logged features to offline store) | no |
5255

5356
Below is a matrix indicating which functionality is supported by `PostgreSQLRetrievalJob`.
5457

55-
| | Postgres |
56-
| --------------------------------- | --- |
57-
| export to dataframe | yes |
58-
| export to arrow table | yes |
59-
| export to arrow batches | no |
60-
| export to SQL | yes |
61-
| export to data lake (S3, GCS, etc.) | yes |
62-
| export to data warehouse | yes |
63-
| export as Spark dataframe | no |
64-
| local execution of Python-based on-demand transforms | yes |
65-
| remote execution of Python-based on-demand transforms | no |
66-
| persist results in the offline store | yes |
67-
| preview the query plan before execution | yes |
68-
| read partitioned data | yes |
58+
| | Postgres |
59+
| ----------------------------------------------------- | -------- |
60+
| export to dataframe | yes |
61+
| export to arrow table | yes |
62+
| export to arrow batches | no |
63+
| export to SQL | yes |
64+
| export to data lake (S3, GCS, etc.) | yes |
65+
| export to data warehouse | yes |
66+
| export as Spark dataframe | no |
67+
| local execution of Python-based on-demand transforms | yes |
68+
| remote execution of Python-based on-demand transforms | no |
69+
| persist results in the offline store | yes |
70+
| preview the query plan before execution | yes |
71+
| read partitioned data | yes |
6972

7073
To compare this set of functionality against other offline stores, please see the full [functionality matrix](overview.md#functionality-matrix).

docs/reference/offline-stores/redshift.md

Lines changed: 24 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,9 @@ The Redshift offline store provides support for reading [RedshiftSources](../dat
77
* All joins happen within Redshift.
88
* Entity dataframes can be provided as a SQL query or can be provided as a Pandas dataframe. A Pandas dataframes will be uploaded to Redshift temporarily in order to complete join operations.
99

10+
## Getting started
11+
In order to use this offline store, you'll need to run `pip install 'feast[aws]'`. You can get started by then running `feast init -t aws`.
12+
1013
## Example
1114

1215
{% code title="feature_store.yaml" %}
@@ -32,30 +35,30 @@ The full set of configuration options is available in [RedshiftOfflineStoreConfi
3235
The set of functionality supported by offline stores is described in detail [here](overview.md#functionality).
3336
Below is a matrix indicating which functionality is supported by the Redshift offline store.
3437
35-
| | Redshift |
36-
| :-------------------------------- | :-- |
37-
| `get_historical_features` (point-in-time correct join) | yes |
38-
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
39-
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
40-
| `offline_write_batch` (persist dataframes to offline store) | yes |
41-
| `write_logged_features` (persist logged features to offline store) | yes |
38+
| | Redshift |
39+
| :----------------------------------------------------------------- | :------- |
40+
| `get_historical_features` (point-in-time correct join) | yes |
41+
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
42+
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
43+
| `offline_write_batch` (persist dataframes to offline store) | yes |
44+
| `write_logged_features` (persist logged features to offline store) | yes |
4245

4346
Below is a matrix indicating which functionality is supported by `RedshiftRetrievalJob`.
4447

45-
| | Redshift |
46-
| --------------------------------- | --- |
47-
| export to dataframe | yes |
48-
| export to arrow table | yes |
49-
| export to arrow batches | yes |
50-
| export to SQL | yes |
51-
| export to data lake (S3, GCS, etc.) | no |
52-
| export to data warehouse | yes |
53-
| export as Spark dataframe | no |
54-
| local execution of Python-based on-demand transforms | yes |
55-
| remote execution of Python-based on-demand transforms | no |
56-
| persist results in the offline store | yes |
57-
| preview the query plan before execution | yes |
58-
| read partitioned data | yes |
48+
| | Redshift |
49+
| ----------------------------------------------------- | -------- |
50+
| export to dataframe | yes |
51+
| export to arrow table | yes |
52+
| export to arrow batches | yes |
53+
| export to SQL | yes |
54+
| export to data lake (S3, GCS, etc.) | no |
55+
| export to data warehouse | yes |
56+
| export as Spark dataframe | no |
57+
| local execution of Python-based on-demand transforms | yes |
58+
| remote execution of Python-based on-demand transforms | no |
59+
| persist results in the offline store | yes |
60+
| preview the query plan before execution | yes |
61+
| read partitioned data | yes |
5962

6063
To compare this set of functionality against other offline stores, please see the full [functionality matrix](overview.md#functionality-matrix).
6164

docs/reference/offline-stores/snowflake.md

Lines changed: 28 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,13 @@ The [Snowflake](https://trial.snowflake.com) offline store provides support for
66
* All joins happen within Snowflake.
77
* Entity dataframes can be provided as a SQL query or can be provided as a Pandas dataframe. A Pandas dataframes will be uploaded to Snowflake as a temporary table in order to complete join operations.
88

9+
## Getting started
10+
In order to use this offline store, you'll need to run `pip install 'feast[snowflake]'`.
11+
12+
If you're using a file based registry, then you'll also need to install the relevant cloud extra (`pip install 'feast[snowflake, CLOUD]'` where `CLOUD` is one of `aws`, `gcp`, `azure`)
13+
14+
You can get started by then running `feast init -t snowflake`.
15+
916
## Example
1017

1118
{% code title="feature_store.yaml" %}
@@ -31,29 +38,29 @@ The full set of configuration options is available in [SnowflakeOfflineStoreConf
3138
The set of functionality supported by offline stores is described in detail [here](overview.md#functionality).
3239
Below is a matrix indicating which functionality is supported by the Snowflake offline store.
3340
34-
| | Snowflake |
35-
| :-------------------------------- | :-- |
36-
| `get_historical_features` (point-in-time correct join) | yes |
37-
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
38-
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
39-
| `offline_write_batch` (persist dataframes to offline store) | yes |
40-
| `write_logged_features` (persist logged features to offline store) | yes |
41+
| | Snowflake |
42+
| :----------------------------------------------------------------- | :-------- |
43+
| `get_historical_features` (point-in-time correct join) | yes |
44+
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
45+
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
46+
| `offline_write_batch` (persist dataframes to offline store) | yes |
47+
| `write_logged_features` (persist logged features to offline store) | yes |
4148

4249
Below is a matrix indicating which functionality is supported by `SnowflakeRetrievalJob`.
4350

44-
| | Snowflake |
45-
| --------------------------------- | --- |
46-
| export to dataframe | yes |
47-
| export to arrow table | yes |
48-
| export to arrow batches | no |
49-
| export to SQL | yes |
50-
| export to data lake (S3, GCS, etc.) | yes |
51-
| export to data warehouse | yes |
52-
| export as Spark dataframe | no |
53-
| local execution of Python-based on-demand transforms | yes |
54-
| remote execution of Python-based on-demand transforms | no |
55-
| persist results in the offline store | yes |
56-
| preview the query plan before execution | yes |
57-
| read partitioned data | yes |
51+
| | Snowflake |
52+
| ----------------------------------------------------- | --------- |
53+
| export to dataframe | yes |
54+
| export to arrow table | yes |
55+
| export to arrow batches | no |
56+
| export to SQL | yes |
57+
| export to data lake (S3, GCS, etc.) | yes |
58+
| export to data warehouse | yes |
59+
| export as Spark dataframe | no |
60+
| local execution of Python-based on-demand transforms | yes |
61+
| remote execution of Python-based on-demand transforms | no |
62+
| persist results in the offline store | yes |
63+
| preview the query plan before execution | yes |
64+
| read partitioned data | yes |
5865

5966
To compare this set of functionality against other offline stores, please see the full [functionality matrix](overview.md#functionality-matrix).

docs/reference/offline-stores/spark.md

Lines changed: 24 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,9 @@ The Spark offline store provides support for reading [SparkSources](../data-sour
1111
The Spark offline store does not achieve full test coverage.
1212
Please do not assume complete stability.
1313

14+
## Getting started
15+
In order to use this offline store, you'll need to run `pip install 'feast[spark]'`. You can get started by then running `feast init -t spark`.
16+
1417
## Example
1518

1619
{% code title="feature_store.yaml" %}
@@ -39,29 +42,29 @@ The full set of configuration options is available in [SparkOfflineStoreConfig](
3942
The set of functionality supported by offline stores is described in detail [here](overview.md#functionality).
4043
Below is a matrix indicating which functionality is supported by the Spark offline store.
4144
42-
| | Spark |
43-
| :-------------------------------- | :-- |
44-
| `get_historical_features` (point-in-time correct join) | yes |
45-
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
46-
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
47-
| `offline_write_batch` (persist dataframes to offline store) | no |
48-
| `write_logged_features` (persist logged features to offline store) | no |
45+
| | Spark |
46+
| :----------------------------------------------------------------- | :---- |
47+
| `get_historical_features` (point-in-time correct join) | yes |
48+
| `pull_latest_from_table_or_query` (retrieve latest feature values) | yes |
49+
| `pull_all_from_table_or_query` (retrieve a saved dataset) | yes |
50+
| `offline_write_batch` (persist dataframes to offline store) | no |
51+
| `write_logged_features` (persist logged features to offline store) | no |
4952

5053
Below is a matrix indicating which functionality is supported by `SparkRetrievalJob`.
5154

52-
| | Spark |
53-
| --------------------------------- | --- |
54-
| export to dataframe | yes |
55-
| export to arrow table | yes |
56-
| export to arrow batches | no |
57-
| export to SQL | no |
58-
| export to data lake (S3, GCS, etc.) | no |
59-
| export to data warehouse | no |
60-
| export as Spark dataframe | yes |
61-
| local execution of Python-based on-demand transforms | no |
62-
| remote execution of Python-based on-demand transforms | no |
63-
| persist results in the offline store | yes |
64-
| preview the query plan before execution | yes |
65-
| read partitioned data | yes |
55+
| | Spark |
56+
| ----------------------------------------------------- | ----- |
57+
| export to dataframe | yes |
58+
| export to arrow table | yes |
59+
| export to arrow batches | no |
60+
| export to SQL | no |
61+
| export to data lake (S3, GCS, etc.) | no |
62+
| export to data warehouse | no |
63+
| export as Spark dataframe | yes |
64+
| local execution of Python-based on-demand transforms | no |
65+
| remote execution of Python-based on-demand transforms | no |
66+
| persist results in the offline store | yes |
67+
| preview the query plan before execution | yes |
68+
| read partitioned data | yes |
6669

6770
To compare this set of functionality against other offline stores, please see the full [functionality matrix](overview.md#functionality-matrix).

0 commit comments

Comments
 (0)