Skip to content

Commit dda22fe

Browse files
authored
RC 1.0.4 (#884)
* doc and version update * stop roxygen from making $[0, 1]$ to a link * stop roxygen from making $[0, 1]$ to a link
1 parent a482442 commit dda22fe

15 files changed

+61
-57
lines changed

DESCRIPTION

+6-6
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
Package: parsnip
22
Title: A Common API to Modeling and Analysis Functions
3-
Version: 1.0.3.9003
3+
Version: 1.0.4
44
Authors@R: c(
5-
person("Max", "Kuhn", , "max@rstudio.com", role = c("aut", "cre")),
6-
person("Davis", "Vaughan", , "davis@rstudio.com", role = "aut"),
5+
person("Max", "Kuhn", , "max@posit.co", role = c("aut", "cre")),
6+
person("Davis", "Vaughan", , "davis@posit.co", role = "aut"),
77
person("Emil", "Hvitfeldt", , "emilhhvitfeldt@gmail.com", role = "ctb"),
8-
person("RStudio", role = c("cph", "fnd"))
8+
person("Posit Software PBC", role = c("cph", "fnd"))
99
)
10-
Maintainer: Max Kuhn <max@rstudio.com>
10+
Maintainer: Max Kuhn <max@posit.co>
1111
Description: A common interface is provided to allow users to specify a
1212
model without having to remember the different argument names across
1313
different functions or computational engines (e.g. 'R', 'Spark',
@@ -76,4 +76,4 @@ Config/testthat/edition: 3
7676
Encoding: UTF-8
7777
LazyData: true
7878
Roxygen: list(markdown = TRUE)
79-
RoxygenNote: 7.2.3
79+
RoxygenNote: 7.2.3.9000

NEWS.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# parsnip (development version)
1+
# parsnip 1.0.4
22

33
* For censored regression models, a "reverse Kaplan-Meier" curve is computed for the censoring distribution. This can be used when evaluating this type of model (#855).
44

README.md

+25-25
Original file line numberDiff line numberDiff line change
@@ -77,22 +77,22 @@ between implementations.
7777

7878
In this example:
7979

80-
- the **type** of model is “random forest”,
81-
- the **mode** of the model is “regression” (as opposed to
82-
classification, etc), and
83-
- the computational **engine** is the name of the R package.
80+
- the **type** of model is “random forest”,
81+
- the **mode** of the model is “regression” (as opposed to
82+
classification, etc), and
83+
- the computational **engine** is the name of the R package.
8484

8585
The goals of parsnip are to:
8686

87-
- Separate the definition of a model from its evaluation.
88-
- Decouple the model specification from the implementation (whether
89-
the implementation is in R, spark, or something else). For example,
90-
the user would call `rand_forest` instead of `ranger::ranger` or
91-
other specific packages.
92-
- Harmonize argument names (e.g. `n.trees`, `ntrees`, `trees`) so that
93-
users only need to remember a single name. This will help *across*
94-
model types too so that `trees` will be the same argument across
95-
random forest as well as boosting or bagging.
87+
- Separate the definition of a model from its evaluation.
88+
- Decouple the model specification from the implementation (whether the
89+
implementation is in R, spark, or something else). For example, the
90+
user would call `rand_forest` instead of `ranger::ranger` or other
91+
specific packages.
92+
- Harmonize argument names (e.g. `n.trees`, `ntrees`, `trees`) so that
93+
users only need to remember a single name. This will help *across*
94+
model types too so that `trees` will be the same argument across
95+
random forest as well as boosting or bagging.
9696

9797
Using the example above, the parsnip approach would be:
9898

@@ -166,18 +166,18 @@ This project is released with a [Contributor Code of
166166
Conduct](https://contributor-covenant.org/version/2/0/CODE_OF_CONDUCT.html).
167167
By contributing to this project, you agree to abide by its terms.
168168

169-
- For questions and discussions about tidymodels packages, modeling,
170-
and machine learning, please [post on RStudio
171-
Community](https://community.rstudio.com/new-topic?category_id=15&tags=tidymodels,question).
169+
- For questions and discussions about tidymodels packages, modeling, and
170+
machine learning, please [post on RStudio
171+
Community](https://community.rstudio.com/new-topic?category_id=15&tags=tidymodels,question).
172172

173-
- If you think you have encountered a bug, please [submit an
174-
issue](https://github.com/tidymodels/parsnip/issues).
173+
- If you think you have encountered a bug, please [submit an
174+
issue](https://github.com/tidymodels/parsnip/issues).
175175

176-
- Either way, learn how to create and share a
177-
[reprex](https://reprex.tidyverse.org/articles/articles/learn-reprex.html)
178-
(a minimal, reproducible example), to clearly communicate about your
179-
code.
176+
- Either way, learn how to create and share a
177+
[reprex](https://reprex.tidyverse.org/articles/articles/learn-reprex.html)
178+
(a minimal, reproducible example), to clearly communicate about your
179+
code.
180180

181-
- Check out further details on [contributing guidelines for tidymodels
182-
packages](https://www.tidymodels.org/contribute/) and [how to get
183-
help](https://www.tidymodels.org/help/).
181+
- Check out further details on [contributing guidelines for tidymodels
182+
packages](https://www.tidymodels.org/contribute/) and [how to get
183+
help](https://www.tidymodels.org/help/).

man/details_boost_tree_h2o.Rd

+2-2
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

man/details_boost_tree_lightgbm.Rd

+2-2
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

man/details_boost_tree_xgboost.Rd

+4-4
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

man/details_rule_fit_xrf.Rd

+2-2
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

man/parsnip-package.Rd

+3-3
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

man/rmd/boost_tree_h2o.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -118,11 +118,11 @@ Non-numeric predictors (i.e., factors) are internally converted to numeric. In t
118118

119119
The `mtry` argument denotes the number of predictors that will be randomly sampled at each split when creating tree models.
120120

121-
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that $[0, 1]$ is always a valid range for that parameter, regardless of input data.
121+
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that `[0, 1]` is always a valid range for that parameter, regardless of input data.
122122

123123
parsnip and its extensions accommodate this parameterization using the `counts` argument: a logical indicating whether `mtry` should be interpreted as the number of predictors that will be randomly sampled at each split. `TRUE` indicates that `mtry` will be interpreted in its sense as a count, `FALSE` indicates that the argument will be interpreted in its sense as a proportion.
124124

125-
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within $[0, 1]$.
125+
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within `[0, 1]`.
126126

127127
## Initializing h2o
128128

man/rmd/boost_tree_lightgbm.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -115,11 +115,11 @@ Non-numeric predictors (i.e., factors) are internally converted to numeric. In t
115115

116116
The `mtry` argument denotes the number of predictors that will be randomly sampled at each split when creating tree models.
117117

118-
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that $[0, 1]$ is always a valid range for that parameter, regardless of input data.
118+
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that `[0, 1]` is always a valid range for that parameter, regardless of input data.
119119

120120
parsnip and its extensions accommodate this parameterization using the `counts` argument: a logical indicating whether `mtry` should be interpreted as the number of predictors that will be randomly sampled at each split. `TRUE` indicates that `mtry` will be interpreted in its sense as a count, `FALSE` indicates that the argument will be interpreted in its sense as a proportion.
121121

122-
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within $[0, 1]$.
122+
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within `[0, 1]`.
123123

124124
### Saving fitted model objects
125125

man/rmd/boost_tree_xgboost.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -121,7 +121,7 @@ boost_tree() %>%
121121
```
122122

123123
```
124-
## Boosted Tree Model Specification (unknown)
124+
## Boosted Tree Model Specification (unknown mode)
125125
##
126126
## Engine-Specific Arguments:
127127
## eval_metric = mae
@@ -139,7 +139,7 @@ boost_tree() %>%
139139
```
140140

141141
```
142-
## Boosted Tree Model Specification (unknown)
142+
## Boosted Tree Model Specification (unknown mode)
143143
##
144144
## Engine-Specific Arguments:
145145
## params = list(eval_metric = "mae")
@@ -162,11 +162,11 @@ By default, the model is trained without parallel processing. This can be change
162162

163163
The `mtry` argument denotes the number of predictors that will be randomly sampled at each split when creating tree models.
164164

165-
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that $[0, 1]$ is always a valid range for that parameter, regardless of input data.
165+
Some engines, such as `"xgboost"`, `"xrf"`, and `"lightgbm"`, interpret their analogue to the `mtry` argument as the _proportion_ of predictors that will be randomly sampled at each split rather than the _count_. In some settings, such as when tuning over preprocessors that influence the number of predictors, this parameterization is quite helpful---interpreting `mtry` as a proportion means that `[0, 1]` is always a valid range for that parameter, regardless of input data.
166166

167167
parsnip and its extensions accommodate this parameterization using the `counts` argument: a logical indicating whether `mtry` should be interpreted as the number of predictors that will be randomly sampled at each split. `TRUE` indicates that `mtry` will be interpreted in its sense as a count, `FALSE` indicates that the argument will be interpreted in its sense as a proportion.
168168

169-
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within $[0, 1]$.
169+
`mtry` is a main model argument for \\code{\\link[=boost_tree]{boost_tree()}} and \\code{\\link[=rand_forest]{rand_forest()}}, and thus should not have an engine-specific interface. So, regardless of engine, `counts` defaults to `TRUE`. For engines that support the proportion interpretation (currently `"xgboost"` and `"xrf"`, via the rules package, and `"lightgbm"` via the bonsai package) the user can pass the `counts = FALSE` argument to `set_engine()` to supply `mtry` values within `[0, 1]`.
170170

171171
### Early stopping
172172

man/rmd/discrim_regularized_klaR.md

+2
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,8 @@ For this engine, there is a single mode: classification
66
## Tuning Parameters
77

88

9+
10+
911
This model has 2 tuning parameter:
1012

1113
- `frac_common_cov`: Fraction of the Common Covariance Matrix (type: double, default: (see below))

man/rmd/mlp_h2o.md

+2
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,8 @@ For this engine, there are multiple modes: classification and regression
55

66
## Tuning Parameters
77

8+
9+
810
This model has 6 tuning parameters:
911

1012
- `hidden_units`: # Hidden Units (type: integer, default: 200L)

0 commit comments

Comments
 (0)