diff --git a/java/pom.xml b/java/pom.xml
index 245d2d7cc6..3a94e5f19c 100644
--- a/java/pom.xml
+++ b/java/pom.xml
@@ -61,7 +61,7 @@
1.5.24
3.14.7
3.10
- 2.12.6
+ 2.14.0
2.3.1
1.3.2
2.0.1.Final
diff --git a/sdk/python/feast/infra/materialization/contrib/bytewax/bytewax_materialization_dataflow.py b/sdk/python/feast/infra/materialization/contrib/bytewax/bytewax_materialization_dataflow.py
index bf5229303a..fe2a7f35c1 100644
--- a/sdk/python/feast/infra/materialization/contrib/bytewax/bytewax_materialization_dataflow.py
+++ b/sdk/python/feast/infra/materialization/contrib/bytewax/bytewax_materialization_dataflow.py
@@ -2,7 +2,6 @@
import pyarrow as pa
import pyarrow.parquet as pq
-import s3fs
from bytewax.dataflow import Dataflow # type: ignore
from bytewax.execution import cluster_main
from bytewax.inputs import ManualInputConfig, distribute
@@ -29,8 +28,7 @@ def __init__(
self._run_dataflow()
def process_path(self, path):
- fs = s3fs.S3FileSystem()
- dataset = pq.ParquetDataset(path, filesystem=fs, use_legacy_dataset=False)
+ dataset = pq.ParquetDataset(path, use_legacy_dataset=False)
batches = []
for fragment in dataset.fragments:
for batch in fragment.to_table().to_batches():
diff --git a/setup.py b/setup.py
index 047100f03e..9fbc2bc2cd 100644
--- a/setup.py
+++ b/setup.py
@@ -91,6 +91,7 @@
"google-cloud-datastore>=2.1.0,<3",
"google-cloud-storage>=1.34.0,<3",
"google-cloud-bigtable>=2.11.0,<3",
+ "gcsfs",
]
REDIS_REQUIRED = [
@@ -158,7 +159,7 @@
"moto",
"mypy>=0.981,<0.990",
"avro==1.10.0",
- "gcsfs>=0.4.0,<=2022.01.0",
+ "gcsfs",
"urllib3>=1.25.4,<2",
"psutil==5.9.0",
"py>=1.11.0", # https://github.com/pytest-dev/pytest/issues/10420