This document describes lazyeval, a package that provides principled tools to perform non-standard evaluation (NSE) in R. You should read this vignette if you want to program with packages like dplyr and ggplot2^{1}, or you want a principled way of working with delayed expressions in your own package. As the name suggests, non-standard evaluation breaks away from the standard evaluation (SE) rules in order to do something special. There are three common uses of NSE:

**Labelling**enhances plots and tables by using the expressions supplied to a function, rather than their values. For example, note the axis labels in this plot:`par(mar = c(4.5, 4.5, 1, 0.5)) grid <- seq(0, 2 * pi, length = 100) plot(grid, sin(grid), type = "l")`

**Non-standard scoping**looks for objects in places other than the current environment. For example, base R has`with()`

,`subset()`

, and`transform()`

that look for objects in a data frame (or list) before the current environment:`df <- data.frame(x = c(1, 5, 4, 2, 3), y = c(2, 1, 5, 4, 3)) with(df, mean(x)) #> [1] 3 subset(df, x == y) #> x y #> 5 3 3 transform(df, z = x + y) #> x y z #> 1 1 2 3 #> 2 5 1 6 #> 3 4 5 9 #> 4 2 4 6 #> 5 3 3 6`

**Metaprogramming**is a catch-all term that covers all other uses of NSE (such as in`bquote()`

and`library()`

). Metaprogramming is so called because it involves computing on the unevaluated code in some way.

This document is broadly organised according to the three types of non-standard evaluation described above. The main difference is that after labelling, we’ll take a detour to learn more about formulas. You’re probably familiar with formulas from linear models (e.g. `lm(mpg ~ displ, data = mtcars)`

) but formulas are more than just a tool for modelling: they are a general way of capturing an unevaluated expression.

The approaches recommended here are quite different to my previous generation of recommendations. I am fairly confident these new approaches are correct, and will not have to change substantially again. The current tools make it easy to solve a number of practical problems that were previously challenging and are rooted in long-standing theory.

In base R, the classic way to turn an argument into a label is to use `deparse(substitute(x))`

:

```
my_label <- function(x) deparse(substitute(x))
my_label(x + y)
#> [1] "x + y"
```

There are two potential problems with this approach:

For long some expressions,

`deparse()`

generates a character vector with length > 1:`my_label({ a + b c + d }) #> [1] "{" " a + b" " c + d" "}"`

`substitute()`

only looks one level up, so you lose the original label if the function isn’t called directly:`my_label2 <- function(x) my_label(x) my_label2(a + b) #> [1] "x"`

Both of these problems are resolved by `lazyeval::expr_text()`

:

```
my_label <- function(x) expr_text(x)
my_label2 <- function(x) my_label(x)
my_label({
a + b
c + d
})
#> [1] "{\n a + b\n c + d\n}"
my_label2(a + b)
#> [1] "a + b"
```

There are two variations on the theme of `expr_text()`

:

`expr_find()`

find the underlying expression. It works similarly to`substitute()`

but will follow a chain of promises back up to the original expression. This is often useful for metaprogramming.`expr_label()`

is a customised version of`expr_text()`

that produces labels designed to be used in messages to the user:`expr_label(x) #> [1] "`x`" expr_label(a + b + c) #> [1] "a + b + c" expr_label(foo({ x + y })) #> [1] "foo(...)"`

`plot()`

uses`deparse(substitute(x))`

to generate labels for the x and y axes. Can you generate input that causes it to display bad labels? Write your own wrapper around`plot()`

that uses`expr_label()`

to compute`xlim`

and`ylim`

.Create a simple implementation of

`mean()`

that stops with an informative error message if the argument is not numeric:`x <- c("a", "b", "c") my_mean(x) #> Error: `x` is a not a numeric vector. my_mean(x == "a") #> Error: `x == "a"` is not a numeric vector. my_mean("a") #> Error: "a" is not a numeric vector.`

Read the source code for

`expr_text()`

. How does it work? What additional arguments to`deparse()`

does it use?

Non-standard scoping is probably the most useful NSE tool, but before we can talk about a solid approach, we need to take a detour to talk about formulas. Formulas are a familiar tool from linear models, but their utility is not limited to models. In fact, formulas are a powerful, general purpose tool, because a formula captures two things:

- An unevaluated expression.
- The context (environment) in which the expression was created.

`~`

is a single character that allows you to say: “I want to capture the meaning of this code, without evaluating it right away”. For that reason, the formula can be thought of as a “quoting” operator.

Technically, a formula is a “language” object (i.e. an unevaluated expression) with a class of “formula” and an attribute that stores the environment:

```
f <- ~ x + y + z
typeof(f)
#> [1] "language"
attributes(f)
#> $class
#> [1] "formula"
#>
#> $.Environment
#> <environment: R_GlobalEnv>
```

The structure of the underlying object is slightly different depending on whether you have a one-sided or two-sided formula:

One-sided formulas have length two:

`length(f) #> [1] 2 # The 1st element is always ~ f[[1]] #> `~` # The 2nd element is the RHS f[[2]] #> x + y + z`

Two-sided formulas have length three:

`g <- y ~ x + z length(g) #> [1] 3 # The 1st element is still ~ g[[1]] #> `~` # But now the 2nd element is the LHS g[[2]] #> y # And the 3rd element is the RHS g[[3]] #> x + z`

To abstract away these differences, lazyeval provides `f_rhs()`

and `f_lhs()`

to access either side of the formula, and `f_env()`

to access its environment:

```
f_rhs(f)
#> x + y + z
f_lhs(f)
#> NULL
f_env(f)
#> <environment: R_GlobalEnv>
f_rhs(g)
#> x + z
f_lhs(g)
#> y
f_env(g)
#> <environment: R_GlobalEnv>
```

A formula captures delays the evaluation of an expression so you can later evaluate it with `f_eval()`

:

```
f <- ~ 1 + 2 + 3
f
#> ~1 + 2 + 3
f_eval(f)
#> [1] 6
```

This allows you to use a formula as a robust way of delaying evaluation, cleanly separating the creation of the formula from its evaluation. Because formulas capture the code and context, you get the correct result even when a formula is created and evaluated in different places. In the following example, note that the value of `x`

inside `add_1000()`

is used:

```
x <- 1
add_1000 <- function(x) {
~ 1000 + x
}
add_1000(3)
#> ~1000 + x
#> <environment: 0x7f8b0e996bb8>
f_eval(add_1000(3))
#> [1] 1003
```

It can be hard to see what’s going on when looking at a formula because important values are stored in the environment, which is largely opaque. You can use `f_unwrap()`

to replace names with their corresponding values:

```
f_unwrap(add_1000(3))
#> ~1000 + 3
```

`f_eval()`

has an optional second argument: a named list (or data frame) that overrides values found in the formula’s environment.

```
y <- 100
f_eval(~ y)
#> [1] 100
f_eval(~ y, data = list(y = 10))
#> [1] 10
# Can mix variables in environment and data argument
f_eval(~ x + y, data = list(x = 10))
#> [1] 110
# Can even supply functions
f_eval(~ f(y), data = list(f = function(x) x * 3))
#> [1] 300
```

This makes it very easy to implement non-standard scoping:

```
f_eval(~ mean(cyl), data = mtcars)
#> [1] 6.1875
```

One challenge with non-standard scoping is that we’ve introduced some ambiguity. For example, in the code below does `x`

come from `mydata`

or the environment?

`f_eval(~ x, data = mydata)`

You can’t tell without knowing whether or not `mydata`

has a variable called `x`

. To overcome this problem, `f_eval()`

provides two pronouns:

`.data`

is bound to the data frame.`.env`

is bound to the formula environment.

They both start with `.`

to minimise the chances of clashing with existing variables.

With these pronouns we can rewrite the previous formula to remove the ambiguity:

```
mydata <- data.frame(x = 100, y = 1)
x <- 10
f_eval(~ .env$x, data = mydata)
#> [1] 10
f_eval(~ .data$x, data = mydata)
#> [1] 100
```

If the variable or object doesn’t exist, you’ll get an informative error:

```
f_eval(~ .env$z, data = mydata)
#> Error: Object 'z' not found in environment
f_eval(~ .data$z, data = mydata)
#> Error: Variable 'z' not found in data
```

`f_eval()`

has one more useful trick up its sleeve: unquoting. Unquoting allows you to write functions where the user supplies part of the formula. For example, the following function allows you to compute the mean of any column (or any function of a column):

```
df_mean <- function(df, variable) {
f_eval(~ mean(uq(variable)), data = df)
}
df_mean(mtcars, ~ cyl)
#> [1] 6.1875
df_mean(mtcars, ~ disp * 0.01638)
#> [1] 3.779224
df_mean(mtcars, ~ sqrt(mpg))
#> [1] 4.43477
```

To see how this works, we can use `f_interp()`

which `f_eval()`

calls internally (you shouldn’t call it in your own code, but it’s useful for debugging). The key is `uq()`

: `uq()`

evaluates its first (and only) argument and inserts the value into the formula:

```
variable <- ~cyl
f_interp(~ mean(uq(variable)))
#> ~mean(cyl)
variable <- ~ disp * 0.01638
f_interp(~ mean(uq(variable)))
#> ~mean(disp * 0.01638)
```

Unquoting allows you to create code “templates”, where you write most of the expression, while still allowing the user to control important components. You can even use `uq()`

to change the function being called:

```
f <- ~ mean
f_interp(~ uq(f)(uq(variable)))
#> ~mean(disp * 0.01638)
```

Note that `uq()`

only takes the RHS of a formula, which makes it difficult to insert literal formulas into a call:

```
formula <- y ~ x
f_interp(~ lm(uq(formula), data = df))
#> ~lm(x, data = df)
```

You can instead use `uqf()`

which uses the whole formula, not just the RHS:

```
f_interp(~ lm(uqf(formula), data = df))
#> ~lm(y ~ x, data = df)
```

Unquoting is powerful, but it only allows you to modify a single argument: it doesn’t allow you to add an arbitrary number of arguments. To do that, you’ll need “unquote-splice”, or `uqs()`

. The first (and only) argument to `uqs()`

should be a list of arguments to be spliced into the call:

```
variable <- ~ x
extra_args <- list(na.rm = TRUE, trim = 0.9)
f_interp(~ mean(uq(variable), uqs(extra_args)))
#> ~mean(x, na.rm = TRUE, trim = 0.9)
```

Create a wrapper around

`lm()`

that allows the user to supply the response and predictors as two separate formulas.Compare and contrast

`f_eval()`

with`with()`

.Why does this code work even though

`f`

is defined in two places? (And one of them is not a function).`f <- function(x) x + 1 f_eval(~ f(10), list(f = "a")) #> [1] 11`

Non-standard scoping (NSS) is an important part of R because it makes it easy to write functions tailored for interactive data exploration. These functions require less typing, at the cost of some ambiguity and “magic”. This is a good trade-off for interactive data exploration because you want to get ideas out of your head and into the computer as quickly as possible. If a function does make a bad guess, you’ll spot it quickly because you’re working interactively.

There are three challenges to implementing non-standard scoping:

You must correctly delay the evaluation of a function argument, capturing both the computation (the expression), and the context (the environment). I recommend making this explicit by requiring the user to “quote” any NSS arguments with

`~`

, and then evaluating explicit with`f_eval()`

.When writing functions that use NSS-functions, you need some way to avoid the automatic lookup and be explicit about where objects should be found.

`f_eval()`

solves this problem with the`.data.`

and`.env`

pronouns.You need some way to allow the user to supply parts of a formula.

`f_eval()`

solves this with unquoting.

To illustrate these challenges, I will implement a `sieve()`

function that works similarly to `base::subset()`

or `dplyr::filter()`

. The goal of `sieve()`

is to make it easy to select observations that match criteria defined by a logical expression. `sieve()`

has three advantages over `[`

:

It is much more compact when the condition uses many variables, because you don’t need to repeat the name of the data frame many times.

It drops rows where the condition evaluates to

`NA`

, rather than filling them with`NA`

s.It always returns a data frame.

The implementation of `sieve()`

is straightforward. First we use `f_eval()`

to perform NSS. Then we then check that we have a logical vector, replace `NA`

s with `FALSE`

, and subset with `[`

.

```
sieve <- function(df, condition) {
rows <- f_eval(condition, df)
if (!is.logical(rows)) {
stop("`condition` must be logical.", call. = FALSE)
}
rows[is.na(rows)] <- FALSE
df[rows, , drop = FALSE]
}
df <- data.frame(x = 1:5, y = 5:1)
sieve(df, ~ x <= 2)
#> x y
#> 1 1 5
#> 2 2 4
sieve(df, ~ x == y)
#> x y
#> 3 3 3
```

`sieve()`

Imagine that you’ve written some code that looks like this:

```
sieve(march, ~ x > 100)
sieve(april, ~ x > 50)
sieve(june, ~ x > 45)
sieve(july, ~ x > 17)
```

(This is a contrived example, but it illustrates all of the important issues you’ll need to consider when writing more useful functions.)

Instead of continuing to copy-and-paste your code, you decide to wrap up the common behaviour in a function:

```
threshold_x <- function(df, threshold) {
sieve(df, ~ x > threshold)
}
threshold_x(df, 3)
#> x y
#> 4 4 2
#> 5 5 1
```

There are two ways that this function might fail:

The data frame might not have a variable called

`x`

. This will fail unless there’s a variable called`x`

hanging around in the global environment:`rm(x) df2 <- data.frame(y = 5:1) # Throws an error threshold_x(df2, 3) #> Error in eval(expr, envir, enclos): object 'x' not found # Silently gives the incorrect result! x <- 5 threshold_x(df2, 3) #> y #> 1 5 #> 2 4 #> 3 3 #> 4 2 #> 5 1`

The data frame might have a variable called

`threshold`

:`df3 <- data.frame(x = 1:5, y = 5:1, threshold = 4) threshold_x(df3, 3) #> x y threshold #> 5 5 1 4`

These failures are partiuclarly pernicious because instead of throwing an error they silently produce the wrong answer. Both failures arise because `f_eval()`

introduces ambiguity by looking in two places for each name: the supplied data and formula environment.

To make `threshold_x()`

more reliable, we need to be more explicit by using the `.data`

and `.env`

pronouns:

```
threshold_x <- function(df, threshold) {
sieve(df, ~ .data$x > .env$threshold)
}
threshold_x(df2, 3)
#> Error: Variable 'x' not found in data
threshold_x(df3, 3)
#> x y threshold
#> 4 4 2 4
#> 5 5 1 4
```

Here `.env`

is bound to the environment where `~`

is evaluated, namely the inside of `threshold_x()`

.

The `threshold_x()`

function is not very useful because it’s bound to a specific variable. It would be more powerful if we could vary both the threshold and the variable it applies to. We can do that by taking an additional argument to specify which variable to use.

One simple approach is to use a string and `[[`

:

```
threshold <- function(df, variable, threshold) {
stopifnot(is.character(variable), length(variable) == 1)
sieve(df, ~ .data[[.env$variable]] > .env$threshold)
}
threshold(df, "x", 4)
#> x y
#> 5 5 1
```

This is a simple and robust solution, but only allows us to use an existing variable, not an arbitrary expression like `sqrt(x)`

.

A more general solution is to allow the user to supply a formula, and use unquoting:

```
threshold <- function(df, variable = ~x, threshold = 0) {
sieve(df, ~ uq(variable) > .env$threshold)
}
threshold(df, ~ x, 4)
#> x y
#> 5 5 1
threshold(df, ~ abs(x - y), 2)
#> x y
#> 1 1 5
#> 5 5 1
```

In this case, it’s the responsibility of the user to ensure the `variable`

is specified unambiguously. `f_eval()`

is designed so that `.data`

and `.env`

work even when evaluated inside of `uq()`

:

```
x <- 3
threshold(df, ~ .data$x - .env$x, 0)
#> x y
#> 4 4 2
#> 5 5 1
```

There is one more tool that you might find useful for functions that take `...`

. For example, the code below implements a function similar to `dplyr::mutate()`

or `base::transform()`

.

```
mogrify <- function(`_df`, ...) {
args <- list(...)
for (nm in names(args)) {
`_df`[[nm]] <- f_eval(args[[nm]], `_df`)
}
`_df`
}
```

(NB: the first argument is a non-syntactic name (i.e. it requires quoting with ```

) so it doesn’t accidentally match one of the names of the new variables.)

`transmogrifty()`

makes it easy to add new variables to a data frame:

```
df <- data.frame(x = 1:5, y = sample(5))
mogrify(df, z = ~ x + y, z2 = ~ z * 2)
#> x y z z2
#> 1 1 2 3 6
#> 2 2 5 7 14
#> 3 3 3 6 12
#> 4 4 1 5 10
#> 5 5 4 9 18
```

One problem with this implementation is that it’s hard to specify the names of the generated variables. Imagine you want a function where the name and expression are in separate variables. This is awkward because the variable name is supplied as an argument name to `mogrify()`

:

```
add_variable <- function(df, name, expr) {
do.call("mogrify", c(list(df), setNames(list(expr), name)))
}
add_variable(df, "z", ~ x + y)
#> x y z
#> 1 1 2 3
#> 2 2 5 7
#> 3 3 3 6
#> 4 4 1 5
#> 5 5 4 9
```

Lazyeval provides the `f_list()`

function to make writing this sort of function a little easier. It takes a list of formulas and evaluates the LHS of each formula (if present) to rename the elements:

```
f_list("x" ~ y, z = ~z)
#> $x
#> ~y
#>
#> $z
#> ~z
```

If we tweak `mutate()`

to use `f_list()`

instead of `list()`

:

```
mogrify <- function(`_df`, ...) {
args <- f_list(...)
for (nm in names(args)) {
`_df`[[nm]] <- f_eval(args[[nm]], `_df`)
}
`_df`
}
```

`add_new()`

becomes much simpler:

```
add_variable <- function(df, name, expr) {
mogrify(df, name ~ uq(expr))
}
add_variable(df, "z", ~ x + y)
#> x y z
#> 1 1 2 3
#> 2 2 5 7
#> 3 3 3 6
#> 4 4 1 5
#> 5 5 4 9
```

Write a function that selects all rows of

`df`

where`variable`

is greater than its mean. Make the function more general by allowing the user to specify a function to use instead of`mean()`

(e.g.`median()`

).Create a version of

`mogrify()`

where the first argument is`x`

? What happens if you try to create a new variable called`x`

?

In some situations you might want to eliminate the formula altogether, and allow the user to type expressions directly. I was once much enamoured with this approach (witness ggplot2, dplyr, …). However, I now think that it should be used sparingly because explict quoting with `~`

leads to simpler code, and makes it more clear to the user that something special is going on.

That said, lazyeval does allow you to eliminate the `~`

if you really want to. In this case, I recommend having both a NSE and SE version of the function. The SE version, which takes formuals, should have suffix `_`

:

```
sieve_ <- function(df, condition) {
rows <- f_eval(condition, df)
if (!is.logical(rows)) {
stop("`condition` must be logical.", call. = FALSE)
}
rows[is.na(rows)] <- FALSE
df[rows, , drop = FALSE]
}
```

Then create the NSE version which doesn’t need the explicit formula. The key is the use of `f_capture()`

which takes an unevaluated argument (a promise) and captures it as a formula:

```
sieve <- function(df, expr) {
sieve_(df, f_capture(expr))
}
sieve(df, x == 1)
#> x y
#> 1 1 2
```

If you’re familiar with `substitute()`

you might expect the same drawbacks to apply. However, `f_capture()`

is smart enough to follow a chain of promises back to the original value, so, for example, this code works fine:

```
scramble <- function(df) {
df[sample(nrow(df)), , drop = FALSE]
}
subscramble <- function(df, expr) {
scramble(sieve(df, expr))
}
subscramble(df, x < 4)
#> x y
#> 3 3 3
#> 2 2 5
#> 1 1 2
```

If you want a `...`

function that doesn’t require formulas, I recommend that the SE version take a list of arguments, and the NSE version uses `dots_capture()`

to capture multiple arguments as a list of formulas.

```
mogrify_ <- function(`_df`, args) {
args <- as_f_list(args)
for (nm in names(args)) {
`_df`[[nm]] <- f_eval(args[[nm]], `_df`)
}
`_df`
}
mogrify <- function(`_df`, ...) {
mogrify_(`_df`, dots_capture(...))
}
```

- Recreate
`subscramble()`

using`base::subset()`

instead of`sieve()`

. Why does it fail?

The final use of non-standard evaluation is to do metaprogramming. This is a catch-all term that encompasses any function that does computation on an unevaluated expression. You can learn about metaprogrgramming in http://adv-r.had.co.nz/Expressions.html, particularly http://adv-r.had.co.nz/Expressions.html#ast-funs. Over time, the goal is to move all useful metaprogramming helper functions into this package, and discuss metaprogramming more here.

Currently neither ggplot2 nor dplyr actually use these tools since I’ve only just figured it out. But I’ll be working hard to make sure all my packages are consistent in the near future.↩