You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This commit was created on GitHub.com and signed with GitHub’s verified signature.
1.19.0 (2024-06-25)
Snowpark Python API Updates
Improvements
New Features
Added support for to_boolean function.
Added documentation pages for Index and its APIs.
Bug Fixes
Fixed a bug where python stored procedure with table return type fails when run in a task.
Fixed a bug where df.dropna fails due to RecursionError: maximum recursion depth exceeded when the DataFrame has more than 500 columns.
Fixed a bug where AsyncJob.result("no_result") doesn't wait for the query to finish execution.
Snowpark Local Testing Updates
New Features
Added support for the strict parameter when registering UDFs and Stored Procedures.
Bug Fixes
Fixed a bug in convert_timezone that made the setting the source_timezone parameter return an error.
Fixed a bug where creating DataFrame with empty data of type DateType raises AttributeError.
Fixed a bug that table merge fails when update clause exists but no update takes place.
Fixed a bug in mock implementation of to_char that raises IndexError when incoming column has nonconsecutive row index.
Fixed a bug in handling of CaseExpr expressions that raises IndexError when incoming column has nonconsecutive row index.
Fixed a bug in implementation of Column.like that raises IndexError when incoming column has nonconsecutive row index.
Improvements
Added support for type coercion in the implementation of DataFrame.replace, DataFrame.dropna and the mock function iff.
Snowpark pandas API Updates
New Features
Added partial support for DataFrame.pct_change and Series.pct_change without the freq and limit parameters.
Added support for Series.str.get.
Added support for Series.dt.dayofweek, Series.dt.day_of_week, Series.dt.dayofyear, and Series.dt.day_of_year.
Added support for Series.str.__getitem__ (Series.str[...]).
Added support for Series.str.lstrip and Series.str.rstrip.
Added support for DataFrameGroupby.size and SeriesGroupby.size.
Added support for DataFrame.expanding and Series.expanding for aggregations count, sum, min, max, mean, std, and var with axis=0.
Added support for DataFrame.rolling and Series.rolling for aggregation count with axis=0.
Added support for Series.str.match.
Added support for DataFrame.resample and Series.resample for aggregation size.
Bug Fixes
Fixed a bug that causes output of GroupBy.aggregate's columns to be ordered incorrectly.
Fixed a bug where DataFrame.describe on a frame with duplicate columns of differing dtypes could cause an error or incorrect results.
Fixed a bug in DataFrame.rolling and Series.rolling so window=0 now throws NotImplementedError instead of ValueError
Improvements
Added support for named aggregations in DataFrame.aggregate and Series.aggregate with axis=0.
pd.read_csv reads using the native pandas CSV parser, then uploads data to snowflake using parquet. This enables most of the parameters supported by read_csv including date parsing and numeric conversions. Uploading via parquet is roughly twice as fast as uploading via CSV.
Initial work to support an pd.Index directly in Snowpark pandas. Support for pd.Index as a first-class component of Snowpark pandas is coming soon.
Added a lazy index constructor and support for len, shape, size, empty, to_pandas() and names. For df.index, Snowpark pandas creates a lazy index object.
For df.columns, Snowpark pandas supports a non-lazy version of an Index since the data is already stored locally.