Skip to content

Commit

Permalink
Reduce the size of synthetic data used in Criteo test (#579)
Browse files Browse the repository at this point in the history
On an 11GB GPU, 1m rows causes an out-of-memory error, while 100k rows makes the test pass.
  • Loading branch information
karlhigley authored Aug 31, 2022
1 parent 3ef07f6 commit a677c05
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion tests/unit/examples/test_scaling_criteo_merlin_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ def test_func():
from merlin.datasets.synthetic import generate_data
train, valid = generate_data("criteo", int(1000000), set_sizes=(0.7, 0.3))
train, valid = generate_data("criteo", int(100000), set_sizes=(0.7, 0.3))
train.to_ddf().compute().to_parquet('/tmp/input/criteo/day_0.parquet')
valid.to_ddf().compute().to_parquet('/tmp/input/criteo/day_1.parquet')
Expand Down

0 comments on commit a677c05

Please sign in to comment.