Skip to content

Releases: alteryx/woodwork

v0.11.1

04 Jan 19:46
4d32985
Compare
Choose a tag to compare

v0.11.1 Jan 4, 2022

  • Changes
    • Update inference process to only check for NaturalLanguage if no other type matches are found first (#1234)
  • Documentation Changes
    • Updating contributing doc with Spark installation instructions (#1232)
  • Testing Changes
    • Enable auto-merge for minimum and latest dependency merge requests (#1228, #1230, #1233)

Thanks to the following people for contributing to this release:
@gsheni, @thehomebrewnerd, @willsmithorg

v0.11.0

22 Dec 17:14
5f9ac56
Compare
Choose a tag to compare

v0.11.0 Dec 22, 2021

  • Enhancements
    • Add type inference for natural language (#1210)
  • Changes
    • Make public method get_subset_schema (#1218)

Thanks to the following people for contributing to this release:
@jeff-hernandez, @thehomebrewnerd, @tuethan1999

v0.10.0

30 Nov 21:30
b38dcfb
Compare
Choose a tag to compare

v0.10.0 Nov 30, 2020

  • Enhancements
    • Allow frequency inference on temporal (Datetime, Timedelta) columns of Woodwork DataFrame (#1202)
    • Update describe_dict to compute top_values for double columns that contain only integer values (#1206)
  • Changes
    • Return histogram bins as a list of floats instead of a pandas.Interval object (#1207)

Thanks to the following people for contributing to this release:
@tamargrey, @thehomebrewnerd

v0.9.1

19 Nov 19:35
1a5ded4
Compare
Choose a tag to compare

v0.9.1 Nov 19, 2021

  • Fixes
    • Fix bug that causes mutual_information to fail with certain index types (#1199)
  • Changes
    • Update pip to 21.3.1 for test requirements (#1196)
  • Documentation Changes
    • Update install page with updated minimum optional dependencies (#1193)

Thanks to the following people for contributing to this release:
@gsheni, @thehomebrewnerd

v0.9.0

11 Nov 20:16
e980524
Compare
Choose a tag to compare

v0.9.0 Nov 11, 2021

  • Enhancements
    • Added read_file parameter for replacing empty string values with NaN values (#1161)
  • Fixes
    • Set a maximum version for pyspark until we understand why :pr:1169 failed (#1179)
    • Require newer dask version (#1180)
  • Changes
    • Make box plot low/high indices/values optional to return in box_plot_dict (#1184)
  • Documentation Changes
    • Update docs dependencies (#1176)
  • Testing Changes
    • Add black linting package and remove autopep8 (#1164, #1183)
    • Updated notebook standardizer to standardize python versions (#1166)

Thanks to the following people for contributing to this release:
@bchen1116, @davesque, @gsheni, @rwedge, @tamargrey, @thehomebrewnerd

v0.8.2

12 Oct 19:04
d1836fd
Compare
Choose a tag to compare

v0.8.2 Oct 12, 2021

  • Fixes
    • Fixed an issue when inferring the format of datetime strings with day of week or meridiem placeholders (#1158)
    • Implements change in Datetime.transform to prevent initialization failure in some cases (#1162)
  • Testing Changes
    • Update reviewers for minimum and latest dependency checkers (#1150)
    • Added notebook standardizer to remove executed outputs (#1153)

Thanks to the following people for contributing to this release:
@bchen1116, @davesque, @jeff-hernandez, @thehomebrewnerd

v0.8.1

16 Sep 19:19
fea3ff7
Compare
Choose a tag to compare

v0.8.1 Sep 16, 2021

  • Changes
    • Update Datetime.transform to use default nrows value when calling _infer_datetime_format (#1137)
  • Documentation Changes
    • Hide spark config in Using Dask and Koalas Guide (#1139)

Thanks to the following people for contributing to this release:
@jeff-hernandez, @simha104, @thehomebrewnerd

v0.8.0

10 Sep 00:29
d3295e9
Compare
Choose a tag to compare
  • Enhancements
    • Add support for automatically inferring the URL and IPAddress logical types (#1122, #1124)
    • Add get_valid_mi_columns method to list columns that have valid logical types for mutual information calculation (#1129)
    • Add attribute to check if column has a nullable logical type (#1127)
  • Changes
    • Update get_invalid_schema_message to improve performance (#1132)
  • Documentation Changes
    • Fix typo in the "Get Started" documentation (#1126)
    • Clean up the logical types guide (#1134)

Thanks to the following people for contributing to this release:
@ajaypallekonda, @davesque, @jeff-hernandez, @thehomebrewnerd

v0.7.1

25 Aug 20:25
54b1ef8
Compare
Choose a tag to compare

v0.7.1 Aug 25, 2021

  • Fixes
    • Validate schema's index if being used in partial schema init (#1115)
    • Allow falsy index, time index, and name values to be set along with partial schema at init (#1115)

Thanks to the following people for contributing to this release:
@tamargrey

v0.7.0

25 Aug 14:53
debd491
Compare
Choose a tag to compare

v0.7.0 Aug 25, 2021

  • Enhancements
    • Add 'passthrough' and 'ignore' to tags in list_semantic_tags (#1094)
    • Add initialize with partial table schema (#1100)
    • Apply ordering specified by the Ordinal logical type to underlying series (#1097)
    • Add AgeFractional logical type (#1112)

Thanks to the following people for contributing to this release:
@davesque, @jeff-hernandez, @tamargrey, @tuethan1999

Breaking Changes

  • #1100:
    The behavior for init has changed. A full schema is a
    schema that contains all of the columns of the dataframe it describes
    whereas a partial schema only contains a subset. A full schema will also
    require that the schema is valid without having to make any changes to
    the DataFrame. Before, only a full schema was permitted by the init
    method so passing a partial schema would error. Additionally, any
    parameters like logical_types would be ignored if passing in a schema.
    Now, passing a partial schema to the init method calls the
    init_with_partial_schema method instead of throwing an error.
    Information from keyword arguments will override information from the
    partial schema. For example, if column a has the Integer Logical Type
    in the partial schema, it's possible to use the logical_type argument
    to reinfer it's logical type by passing {'a': None} or force a type by
    passing in {'a': Double}. These changes mean that Woodwork init is less
    restrictive. If no type inference takes place and no changes are required
    of the DataFrame at initialization, init_with_full_schema should be
    used instead of init. init_with_full_schema maintains the same
    functionality as when a schema was passed to the old init.