Sorry guys, I am super busy recently for other projects, I will come back to continue to improve maybe a month later (since Apr 15th), please create an issue if you have any problem.
Hive is not included in current Feast roadmap, this project intends to add Hive support for Offline Store.
For more details, can check this Feast issue.
The public releases have passed all integration tests, please create an issue if you got any problem.
- DONE [v0.1.1]
I am working on the first workable version, think it will be released in a couple of days. - DONE [v0.1.2]
Allow custom hive conf when connect to a HiveServer2 - DONE [v0.14.0]
Support Feast 0.14.x - DONE [v0.17.0]
Support Feast 0.17.0 - TODO It currently supports
insert into
for uploading entity_df, which is a little inefficient, gonna add extra parameters for people who are able to provide HDFS address in next version (for uploading to HDFS).
pip install feast
- Install stable version
pip install feast-hive
- Install develop version (not stable):
pip install git+https://github.com/baineng/feast-hive.git
feast init feature_repo
cd feature_repo
set offline_store
type to be feast_hive.HiveOfflineStore
project: ...
registry: ...
provider: local
offline_store:
type: feast_hive.HiveOfflineStore
host: localhost
port: 10000 # optional, default is `10000`
database: default # optional, default is `default`
hive_conf: # optional, hive conf overlay
hive.join.cache.size: 14797
hive.exec.max.dynamic.partitions: 779
... # other parameters
online_store:
...
- Upload
data/driver_stats.parquet
to HDFS
hdfs dfs -copyFromLocal ./data/driver_stats.parquet /tmp/
- Create Hive Table
CREATE TABLE driver_stats (
event_timestamp bigint,
driver_id bigint,
conv_rate float,
acc_rate float,
avg_daily_trips int,
created bigint
)
STORED AS PARQUET;
- Load data into the table
LOAD DATA INPATH '/tmp/driver_stats.parquet' INTO TABLE driver_stats;
# This is an example feature definition file
from google.protobuf.duration_pb2 import Duration
from feast import Entity, Feature, FeatureView, ValueType
from feast_hive import HiveSource
# Read data from Hive table
# Here we use a Query to reuse the original parquet data,
# but you can replace to your own Table or Query.
driver_hourly_stats = HiveSource(
# table='driver_stats',
query = """
SELECT Timestamp(cast(event_timestamp / 1000000 as bigint)) AS event_timestamp,
driver_id, conv_rate, acc_rate, avg_daily_trips,
Timestamp(cast(created / 1000000 as bigint)) AS created
FROM driver_stats
""",
event_timestamp_column="event_timestamp",
created_timestamp_column="created",
)
# Define an entity for the driver.
driver = Entity(name="driver_id", value_type=ValueType.INT64, description="driver id", )
# Define FeatureView
driver_hourly_stats_view = FeatureView(
name="driver_hourly_stats",
entities=["driver_id"],
ttl=Duration(seconds=86400 * 1),
features=[
Feature(name="conv_rate", dtype=ValueType.FLOAT),
Feature(name="acc_rate", dtype=ValueType.FLOAT),
Feature(name="avg_daily_trips", dtype=ValueType.INT64),
],
online=True,
batch_source=driver_hourly_stats,
tags={},
)
feast apply
The rest are as same as Feast Quickstart
git clone https://github.com/baineng/feast-hive.git
cd feast-hive
# creating virtual env ...
pip install -e ".[dev]"
# before commit
make format
make lint
pip install -e ".[test]"
pytest -n 6 --host=localhost --port=10000 --database=default