Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
89 commits
Select commit Hold shift + click to select a range
a590076
cleanup
rjzamora Nov 6, 2024
7f1bec7
rename to parallel
rjzamora Nov 7, 2024
023e085
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 7, 2024
e7a2fce
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 7, 2024
69a3374
simplify solution
rjzamora Nov 7, 2024
6aa3694
Merge branch 'cudf-polars-dask-simple' of github.com:rjzamora/cudf in…
rjzamora Nov 7, 2024
ea22a9a
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 7, 2024
915a779
deeper dive
rjzamora Nov 8, 2024
bd9d783
improve simple agg reduction
rjzamora Nov 8, 2024
7363d91
cleanup fundamental bugs
rjzamora Nov 10, 2024
58ee5f4
move PartitionInfo
rjzamora Nov 10, 2024
ecc51ef
add Literal
rjzamora Nov 10, 2024
75eae0c
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 12, 2024
fb2d6bf
add lower_ir_graph
rjzamora Nov 12, 2024
c17564c
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
rjzamora Nov 12, 2024
6e66998
strip out most exploratory logic
rjzamora Nov 12, 2024
c41723d
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 12, 2024
d774f38
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 13, 2024
6886f8d
Add basic Dask evaluate test
pentschev Nov 13, 2024
29b2d7b
Replace environment variable with new `"executor"` config
pentschev Nov 13, 2024
3a68a6d
Add kwarg to specify executor in `assert_gpu_result_equal`
pentschev Nov 13, 2024
8079ac0
Add couple of Dask executor tests
pentschev Nov 13, 2024
6f7ccee
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
pentschev Nov 13, 2024
af4c5f5
Merge remote-tracking branch 'rjzamora/cudf-polars-dask-simple' into …
pentschev Nov 13, 2024
8aed94f
Improve `count` code
pentschev Nov 13, 2024
aadaf10
Pass `executor` to `GPUEngine` in `assert_gpu_result_equal`
pentschev Nov 13, 2024
c3a6907
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
pentschev Nov 14, 2024
4f67819
Merge branch 'branch-24.12' into cudf-polars-dask-simple
rjzamora Nov 14, 2024
c8ca09e
Clarify intent renaming executor to "dask-experimental"
pentschev Nov 14, 2024
3fd51bb
move PartitionInfo out of ir module
rjzamora Nov 14, 2024
bf182e4
Merge remote-tracking branch 'rjzamora/cudf-polars-dask-simple' into …
pentschev Nov 14, 2024
453e274
skip coverage on sanity-check errors
rjzamora Nov 14, 2024
2b74f28
Add `--executor` to pytest
pentschev Nov 14, 2024
6d3cd55
Merge remote-tracking branch 'rjzamora/cudf-polars-dask-simple' into …
pentschev Nov 14, 2024
2398a2e
Enable dask-experimental tests in CI, remove duplicates
pentschev Nov 14, 2024
9aa479a
Fix wrong protocol name in deserialization test
pentschev Nov 14, 2024
64ea98e
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
pentschev Nov 14, 2024
22678a5
Remove `executor` kwarg from `assert_gpu_result_equal`
pentschev Nov 14, 2024
0c196f5
experimental groupby logic
rjzamora Nov 14, 2024
41441ca
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
rjzamora Nov 15, 2024
efadb78
Reintroduce `executor` kwarg in `assert_gpu_result_equal`
pentschev Nov 15, 2024
9b78d8f
Add basic tests for all executors to ensure 100% coverage
pentschev Nov 15, 2024
c54c217
Merge remote-tracking branch 'rjzamora/cudf-polars-dask-simple' into …
pentschev Nov 15, 2024
70da7a9
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
pentschev Nov 15, 2024
3aeb1e4
Fix `executor` in `assert_gpu_result_equal`
pentschev Nov 18, 2024
485a161
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
pentschev Nov 18, 2024
eb41100
Merge remote-tracking branch 'upstream/branch-24.12' into cudf-polars…
rjzamora Nov 19, 2024
1536de8
Merge branch 'cudf-polars-dask-simple' into cudf-polars-dask-groupby-v2
rjzamora Nov 19, 2024
6c5c50e
add parallel scan
rjzamora Nov 19, 2024
9ad9d73
add basic broadcast join support
rjzamora Nov 19, 2024
15eb393
add Select and HConcat support
rjzamora Nov 19, 2024
2bd4937
support for simple tpch queries
rjzamora Nov 20, 2024
e675c9d
add pq statistcs sampling
rjzamora Nov 20, 2024
dcc7503
split large parquet files
rjzamora Nov 20, 2024
e91fdb9
Merge branch 'branch-25.02' into cudf-polars-dask-simple
pentschev Nov 20, 2024
cac9e4f
Merge remote-tracking branch 'upstream/branch-25.02' into cudf-polars…
rjzamora Nov 20, 2024
22486fd
Merge remote-tracking branch 'origin/cudf-polars-dask-simple' into cu…
rjzamora Nov 20, 2024
5865410
re-align with branch-25.02
rjzamora Nov 20, 2024
2129d08
make blocksize configurable
rjzamora Nov 20, 2024
476e1e1
add benchmarks
rjzamora Nov 20, 2024
4527907
Merge branch 'branch-25.02' into cudf-polars-dask-groupby-v2
rjzamora Nov 20, 2024
9dc4b4c
update bench
rjzamora Nov 20, 2024
78cc9af
Merge branch 'cudf-polars-dask-groupby-v2' of github.com:rjzamora/cud…
rjzamora Nov 20, 2024
c2543f5
Merge remote-tracking branch 'upstream/branch-25.02' into cudf-polars…
rjzamora Nov 20, 2024
3180065
proper tree reduction
rjzamora Nov 20, 2024
3fdd292
disable chunked reading in benchmark
rjzamora Nov 20, 2024
9aef9cc
add more queries
rjzamora Nov 21, 2024
cc9eb18
pause debugging
rjzamora Nov 21, 2024
addae40
Merge remote-tracking branch 'upstream/branch-25.02' into cudf-polars…
rjzamora Nov 21, 2024
dbf37d3
address code review - round 1
rjzamora Nov 21, 2024
4d21f7c
move sort tupling
rjzamora Nov 21, 2024
e241af3
remove need for stringify
rjzamora Nov 21, 2024
1064fcb
address code review - round 2
rjzamora Nov 21, 2024
bbddfb6
Merge remote-tracking branch 'upstream/branch-25.02' into cudf-polars…
rjzamora Nov 21, 2024
aeecd4d
remove global caching
rjzamora Nov 21, 2024
09c5217
use general StateInfo
rjzamora Nov 21, 2024
62f10bc
revert (for now)
rjzamora Nov 21, 2024
242e79d
Merge branch 'cudf-polars-dask-simple' into cudf-polars-dask-groupby-v2
rjzamora Nov 21, 2024
709d0d6
fix groupby
rjzamora Nov 21, 2024
186e2da
fix split parquet read
rjzamora Nov 21, 2024
9c46dad
simplify fix
rjzamora Nov 21, 2024
8ce3666
add basic scan and groupby tests
rjzamora Nov 22, 2024
ae3d07f
Merge remote-tracking branch 'upstream/branch-25.02' into cudf-polars…
rjzamora Nov 22, 2024
6b3d895
basic join support
rjzamora Nov 22, 2024
07b260e
add parallel dataframe scan support
rjzamora Nov 22, 2024
5a3373e
work around missing shuffle
rjzamora Nov 23, 2024
59c8956
add simple all-to-all shuffle
rjzamora Nov 23, 2024
1cbc411
fix hash partitioning
rjzamora Nov 24, 2024
10b7b99
save
rjzamora Nov 25, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions ci/run_cudf_polars_pytests.sh
Original file line number Diff line number Diff line change
Expand Up @@ -8,4 +8,8 @@ set -euo pipefail
# Support invoking run_cudf_polars_pytests.sh outside the script directory
cd "$(dirname "$(realpath "${BASH_SOURCE[0]}")")"/../python/cudf_polars/

# Test the default "cudf" executor
python -m pytest --cache-clear "$@" tests

# Test the "dask-experimental" executor
python -m pytest --cache-clear "$@" tests --executor dask-experimental
22 changes: 19 additions & 3 deletions python/cudf_polars/cudf_polars/callback.py
Original file line number Diff line number Diff line change
Expand Up @@ -135,6 +135,7 @@ def _callback(
*,
device: int | None,
memory_resource: int | None,
executor: str | None,
) -> pl.DataFrame:
assert with_columns is None
assert pyarrow_predicate is None
Expand All @@ -145,7 +146,14 @@ def _callback(
set_device(device),
set_memory_resource(memory_resource),
):
return ir.evaluate(cache={}).to_polars()
if executor is None or executor == "cudf":
return ir.evaluate(cache={}).to_polars()
elif executor == "dask-experimental":
from cudf_polars.experimental.parallel import evaluate_dask

return evaluate_dask(ir).to_polars()
else:
raise ValueError(f"Unknown executor '{executor}'")


def validate_config_options(config: dict) -> None:
Expand All @@ -162,13 +170,19 @@ def validate_config_options(config: dict) -> None:
ValueError
If the configuration contains unsupported options.
"""
if unsupported := (config.keys() - {"raise_on_fail", "parquet_options"}):
if unsupported := (
config.keys()
- {"raise_on_fail", "parquet_options", "parallel_options", "executor"}
):
raise ValueError(
f"Engine configuration contains unsupported settings: {unsupported}"
)
assert {"chunked", "chunk_read_limit", "pass_read_limit"}.issuperset(
assert {"blocksize", "chunked", "chunk_read_limit", "pass_read_limit"}.issuperset(
config.get("parquet_options", {})
)
assert {"parquet_blocksize", "num_rows_threshold"}.issuperset(
config.get("parallel_options", {})
)


def execute_with_cudf(nt: NodeTraverser, *, config: GPUEngine) -> None:
Expand Down Expand Up @@ -197,6 +211,7 @@ def execute_with_cudf(nt: NodeTraverser, *, config: GPUEngine) -> None:
device = config.device
memory_resource = config.memory_resource
raise_on_fail = config.config.get("raise_on_fail", False)
executor = config.config.get("executor", None)
validate_config_options(config.config)

with nvtx.annotate(message="ConvertIR", domain="cudf_polars"):
Expand Down Expand Up @@ -226,5 +241,6 @@ def execute_with_cudf(nt: NodeTraverser, *, config: GPUEngine) -> None:
ir,
device=device,
memory_resource=memory_resource,
executor=executor,
)
)
38 changes: 27 additions & 11 deletions python/cudf_polars/cudf_polars/dsl/ir.py
Original file line number Diff line number Diff line change
Expand Up @@ -683,26 +683,30 @@ class DataFrameScan(IR):
This typically arises from ``q.collect().lazy()``
"""

__slots__ = ("df", "projection", "predicate")
_non_child = ("schema", "df", "projection", "predicate")
__slots__ = ("df", "projection", "predicate", "config_options")
_non_child = ("schema", "df", "projection", "predicate", "config_options")
df: Any
"""Polars LazyFrame object."""
projection: tuple[str, ...] | None
"""List of columns to project out."""
predicate: expr.NamedExpr | None
"""Mask to apply."""
config_options: dict[str, Any]
"""GPU-specific configuration options"""

def __init__(
self,
schema: Schema,
df: Any,
projection: Sequence[str] | None,
predicate: expr.NamedExpr | None,
config_options: dict[str, Any],
):
self.schema = schema
self.df = df
self.projection = tuple(projection) if projection is not None else None
self.predicate = predicate
self.config_options = config_options
self._non_child_args = (schema, df, self.projection, predicate)
self.children = ()

Expand Down Expand Up @@ -1599,13 +1603,15 @@ def __init__(self, schema: Schema, name: str, options: Any, df: IR):
# polars requires that all to-explode columns have the
# same sub-shapes
raise NotImplementedError("Explode with more than one column")
self.options = (tuple(to_explode),)
elif self.name == "rename":
old, new, _ = self.options
old, new, strict = self.options
# TODO: perhaps polars should validate renaming in the IR?
if len(new) != len(set(new)) or (
set(new) & (set(df.schema.keys()) - set(old))
):
raise NotImplementedError("Duplicate new names in rename.")
self.options = (tuple(old), tuple(new), strict)
elif self.name == "unpivot":
indices, pivotees, variable_name, value_name = self.options
value_name = "value" if value_name is None else value_name
Expand All @@ -1623,13 +1629,20 @@ def __init__(self, schema: Schema, name: str, options: Any, df: IR):
self.options = (
tuple(indices),
tuple(pivotees),
(variable_name, schema[variable_name]),
(value_name, schema[value_name]),
variable_name,
value_name,
)
self._non_child_args = (name, self.options)
self._non_child_args = (schema, name, self.options)

def get_hashable(self) -> Hashable: # pragma: no cover; Needed by experimental
"""Hashable representation of the node."""
schema_hash = tuple(self.schema.items())
return (type(self), schema_hash, self.name, self.options, *self.children)

@classmethod
def do_evaluate(cls, name: str, options: Any, df: DataFrame) -> DataFrame:
def do_evaluate(
cls, schema: Schema, name: str, options: Any, df: DataFrame
) -> DataFrame:
"""Evaluate and return a dataframe."""
if name == "rechunk":
# No-op in our data model
Expand All @@ -1651,8 +1664,8 @@ def do_evaluate(cls, name: str, options: Any, df: DataFrame) -> DataFrame:
(
indices,
pivotees,
(variable_name, variable_dtype),
(value_name, value_dtype),
variable_name,
value_name,
) = options
npiv = len(pivotees)
index_columns = [
Expand All @@ -1669,15 +1682,18 @@ def do_evaluate(cls, name: str, options: Any, df: DataFrame) -> DataFrame:
plc.interop.from_arrow(
pa.array(
pivotees,
type=plc.interop.to_arrow(variable_dtype),
type=plc.interop.to_arrow(schema[variable_name]),
),
)
]
),
df.num_rows,
).columns()
value_column = plc.concatenate.concatenate(
[df.column_map[pivotee].astype(value_dtype).obj for pivotee in pivotees]
[
df.column_map[pivotee].astype(schema[value_name]).obj
for pivotee in pivotees
]
)
return DataFrame(
[
Expand Down
4 changes: 3 additions & 1 deletion python/cudf_polars/cudf_polars/dsl/translate.py
Original file line number Diff line number Diff line change
Expand Up @@ -263,6 +263,7 @@ def _(
translate_named_expr(translator, n=node.selection)
if node.selection is not None
else None,
translator.config.config.copy(),
)


Expand Down Expand Up @@ -633,9 +634,10 @@ def _(node: pl_expr.Sort, translator: Translator, dtype: plc.DataType) -> expr.E

@_translate_expr.register
def _(node: pl_expr.SortBy, translator: Translator, dtype: plc.DataType) -> expr.Expr:
options = node.sort_options
return expr.SortBy(
dtype,
node.sort_options,
(options[0], tuple(options[1]), tuple(options[2])),
translator.translate_expr(n=node.expr),
*(translator.translate_expr(n=n) for n in node.by),
)
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
# SPDX-FileCopyrightText: Copyright (c) 2024 NVIDIA CORPORATION & AFFILIATES.
# SPDX-License-Identifier: Apache-2.0

"""Experimental benchmarks."""

from __future__ import annotations

__all__: list[str] = []
Loading