Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion tests/frame/filter_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@ def test_filter_missing_column(
df = nw.from_native(constructor(data))

if "polars" in str(constructor):
msg = r"^unable to find column \"c\"; valid columns: \[\"a\", \"b\"\]"
msg = r"unable to find column \"c\"; valid columns: \[\"a\", \"b\"\]"
elif any(id_ == constructor_id for id_ in ("duckdb", "pyspark")):
msg = r"\n\nHint: Did you mean one of these columns: \['a', 'b'\]?"
else:
Expand Down
4 changes: 2 additions & 2 deletions tests/frame/select_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ def test_missing_columns(
# and we have no way to recover exactly which columns the user
# tried selecting. So, we just emit their message (which varies
# across versions...)
msg = r"^e"
msg = r"e"
elif constructor_id == "pyspark[connect]": # pragma: no cover
msg = r"^\[UNRESOLVED_COLUMN.WITH_SUGGESTION\]"
elif any(id_ == constructor_id for id_ in ("duckdb", "pyspark")):
Expand All @@ -116,7 +116,7 @@ def test_missing_columns(

# for the next two cases the error message is different in Polars
if constructor_id == "polars[lazy]":
msg = r"^fdfa"
msg = r"fdfa"
elif constructor_id == "polars[eager]":
msg = r"\n\nHint: Did you mean one of these columns: \['a', 'b', 'z'\]?"
with pytest.raises(ColumnNotFoundError, match=msg):
Expand Down
2 changes: 1 addition & 1 deletion tests/frame/with_columns_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -97,7 +97,7 @@ def test_with_columns_missing_column(
df = nw.from_native(constructor(data))

if "polars" in str(constructor):
msg = r"^c"
msg = r"c"
elif any(id_ == constructor_id for id_ in ("duckdb", "pyspark")):
msg = r"\n\nHint: Did you mean one of these columns: \['a', 'b'\]?"
elif constructor_id == "pyspark[connect]": # pragma: no cover
Expand Down
Loading