This vignette is current as of
furniture
1.10.0.
furniture
We will first make a fictitious data set:
df <- data.frame(a = rnorm(1000, 1.5, 2),
b = seq(1, 1000, 1),
c = c(rep("control", 400), rep("Other", 70), rep("treatment", 500), rep("None", 30)),
d = c(sample(1:1000, 900, replace=TRUE), rep(-99, 100)))
There are four functions that we’ll demonstrate here:
washer
table1
tableC
tableF
washer
is a great function for quick data cleaning. In
situations where there are placeholders, extra levels in a factor, or
several values need to be changed to another.
Now that the data is “washed” we can start exploring and reporting.
##
## ────────────────────────────────
## Mean/Count (SD/%)
## n = 900
## a
## 1.5 (2.0)
## b
## 450.5 (260.0)
## factor(c)
## control 470 (52.2%)
## treatment 430 (47.8%)
## d
## 492.1 (290.6)
## ────────────────────────────────
The variables must be numeric or factor. Since we use a special type
of evaluation (i.e. Non-Standard Evaluation) we can change the variables
in the function (e.g., factor(c)
). This can be extended to
making a whole new variable in the function as well.
##
## ───────────────────────────────────────
## Mean/Count (SD/%)
## n = 900
## a
## 1.5 (2.0)
## b
## 450.5 (260.0)
## d
## 492.1 (290.6)
## ifelse(a > 1, 1, 0)
## 0.6 (0.5)
## ───────────────────────────────────────
This is just the beginning though. Two powerful things the function can do are shown below:
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
##
## ─────────────────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## a 0.204
## 1.6 (2.0) 1.4 (2.1)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.422
## 499.5 (294.3) 483.9 (286.6)
## ifelse(a > 1, 1, 0) 0.426
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
The splitby = ~factor(c)
stratifies the means and counts
by a factor variable (in this case either control or treatment). When we
use this we can also automatically compute tests of significance using
test=TRUE
.
We can also use it intuitively within the pipe (for more about this, see the “Table 1” vignette):
##
## ─────────────────────────────────────────────────────────
## c
## control treatment P-Value
## n = 470 n = 430
## a 0.204
## 1.6 (2.0) 1.4 (2.1)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.422
## 499.5 (294.3) 483.9 (286.6)
## ifelse(a > 1, 1, 0) 0.426
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
In this case, we used the group_by()
function from
dplyr
(within the tidyverse
) and
table1()
knows to use that as the grouping variable in
place of the splitby
argument.
If the parametric tests (default) are not appropriate, you can set
param = FALSE
.
##
## ─────────────────────────────────────────────────────────
## c
## control treatment P-Value
## n = 470 n = 430
## a 0.23
## 1.6 (2.0) 1.4 (2.1)
## b <.001
## 235.5 (135.8) 685.5 (124.3)
## d 0.43
## 499.5 (294.3) 483.9 (286.6)
## ifelse(a > 1, 1, 0) 0.426
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────────────────
Finally, you can polish it quite a bit using a few other options. For example, you can do the following:
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 1.6 (2.0) 1.4 (2.1) 0.204
## B 235.5 (135.8) 685.5 (124.3) <.001
## D 499.5 (294.3) 483.9 (286.6) 0.422
## New Var 0.6 (0.5) 0.6 (0.5) 0.426
## ─────────────────────────────────────────────
Note that var_names
can be used for more complex naming
(e.g., with spaces, brackets) that otherwise cannot be used with data
frames. Alternatively, for more simple naming, we can name them
directly.
table1(df, A = a, B = b, D = d, A2 = ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 1.6 (2.0) 1.4 (2.1) 0.204
## B 235.5 (135.8) 685.5 (124.3) <.001
## D 499.5 (294.3) 483.9 (286.6) 0.422
## A2 0.6 (0.5) 0.6 (0.5) 0.426
## ────────────────────────────────────────
You can also format the numbers (adding a comma for big numbers such as in 20,000 instead of 20000):
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
format_number = TRUE)
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: B
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## A 0.204
## 1.6 (2.0) 1.4 (2.1)
## B <.001
## 235.5 (135.8) 685.5 (124.3)
## D 0.422
## 499.5 (294.3) 483.9 (286.6)
## New Var 0.426
## 0.6 (0.5) 0.6 (0.5)
## ─────────────────────────────────────────────
The table can be exported directly to a folder in the working
directory called “Table1”. Using export
, we provide it with
a string that will be the name of the CSV containing the formatted
table.
table1(df, a, b, d, ifelse(a > 1, 1, 0),
splitby=~factor(c),
test=TRUE,
var_names = c("A", "B", "D", "New Var"),
format_number = TRUE,
export = "example_table1")
This can also be outputted as a latex, markdown, or pandoc table
(matching all the output types of knitr::kable
). Below
shows how to do a latex table (not using kable
however, but
a built-in function that provides the variable name at the top of the
table):
table1(df, a, b, d, "new var" = ifelse(a > 1, 1, 0),
splitby = ~factor(c),
test = TRUE,
output = "latex2")
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
## \begin{table}[ ht ]
## \centering
## \caption{}\label{}
## \begin{tabular}{ l c c c }
## \toprule
## & \multicolumn{ 2 }{c}{ factor(c) }\\
## & control & treatment & P-Value \\
## & n = 470 & n = 430 & \\
## \midrule
## a & & & 0.204\\
## \hspace{6pt} & 1.6 (2.0) & 1.4 (2.1) & \\
## b & & & <.001\\
## \hspace{6pt} & 235.5 (135.8) & 685.5 (124.3) & \\
## d & & & 0.422\\
## \hspace{6pt} & 499.5 (294.3) & 483.9 (286.6) & \\
## new var & & & 0.426\\
## \hspace{6pt} & 0.6 (0.5) & 0.6 (0.5) & \\
## \bottomrule
##
## \end{tabular}
## \end{table}
Last item to show you regarding table1()
is that it can
be printed in a simplified and condensed form. This instead of reporting
counts and percentages for categorical variables, it reports only
percentages and the table has much less white space.
table1(df, a, b, d, "new var" = ifelse(a > 1, 1, 0),
splitby = ~factor(c),
test = TRUE,
type = c("simple", "condensed"))
## Breusch-Pagan Test of Heteroskedasticity suggests `var.equal = FALSE` in t.test() for: b
##
## ─────────────────────────────────────────────
## factor(c)
## control treatment P-Value
## n = 470 n = 430
## a 1.6 (2.0) 1.4 (2.1) 0.204
## b 235.5 (135.8) 685.5 (124.3) <.001
## d 499.5 (294.3) 483.9 (286.6) 0.422
## new var 0.6 (0.5) 0.6 (0.5) 0.426
## ─────────────────────────────────────────────
This function is to create simple, beautiful correlation tables. The
syntax is just like table1()
in most respects. Below we
include all the numeric variables to see their correlations. Since there
are missing values in d
we will use the natural
na.rm=TRUE
.
## N = 900
## Note: pearson correlation (p-value).
##
## ─────────────────────────────────────────
## [1] [2] [3]
## [1]a 1.00
## [2]b -0.022 (0.507) 1.00
## [3]d -0.008 (0.821) 0.002 (0.944) 1.00
## ─────────────────────────────────────────
All the adjustments that you can make in table1()
can be
done here as well. For example,
## N = 900
## Note: pearson correlation (p-value).
[1] | [2] | [3] | |
---|---|---|---|
[1]A | 1.00 | ||
[2]B | -0.022 (0.507) | 1.00 | |
[3]D | -0.008 (0.821) | 0.002 (0.944) | 1.00 |
This function is to create simple frequency tables. The syntax is
just like table1()
and tableC()
in most
respects, except that it uses only one variable instead of many.
##
## ────────────────────────────────────────────────
## a Freq CumFreq Percent CumPerc
## -5.53209030993269 1 1 0.10% 0.10%
## -5.04322932333854 1 2 0.10% 0.20%
## -4.90581321997993 1 3 0.10% 0.30%
## -4.42199333492323 1 4 0.10% 0.40%
## -4.23175062665298 1 5 0.10% 0.50%
## -3.80952053951653 1 6 0.10% 0.60%
## -3.80827941232605 1 7 0.10% 0.70%
## -3.35868667717424 1 8 0.10% 0.80%
## -3.26334385900036 1 9 0.10% 0.90%
## -3.24914302199172 1 10 0.10% 1.00%
## ... ... ... ... ...
## 5.8951156975105 1 990 0.10% 99.00%
## 5.96547625814913 1 991 0.10% 99.10%
## 6.04558912082311 1 992 0.10% 99.20%
## 6.12342995889386 1 993 0.10% 99.30%
## 6.33056995649048 1 994 0.10% 99.40%
## 6.34872956645027 1 995 0.10% 99.50%
## 6.58469384540867 1 996 0.10% 99.60%
## 6.66462405848191 1 997 0.10% 99.70%
## 7.17623963822854 1 998 0.10% 99.80%
## 7.37106757277121 1 999 0.10% 99.90%
## 7.46025098335742 1 1000 0.10% 100.00%
## ────────────────────────────────────────────────
Similarly to table1()
we can use a splitby
argument (or group_by()
).
## Variable:d
##
## ─────────────────────────────────────────────────────
## control Freq CumFreq Percent CumPerc Valid CumValid
## 1 2 2 0.40% 0.40% 0.43% 0.43%
## 3 1 3 0.20% 0.60% 0.21% 0.64%
## 4 1 4 0.20% 0.80% 0.21% 0.85%
## 10 1 5 0.20% 1.00% 0.21% 1.06%
## 11 3 8 0.60% 1.60% 0.64% 1.70%
## 15 1 9 0.20% 1.80% 0.21% 1.91%
## 17 1 10 0.20% 2.00% 0.21% 2.13%
## 18 2 12 0.40% 2.40% 0.43% 2.55%
## 20 1 13 0.20% 2.60% 0.21% 2.77%
## 21 2 15 0.40% 3.00% 0.43% 3.19%
## ... ... ... ... ... ... ...
## 986 1 460 0.20% 92.00% 0.21% 97.87%
## 987 1 461 0.20% 92.20% 0.21% 98.09%
## 988 1 462 0.20% 92.40% 0.21% 98.30%
## 989 1 463 0.20% 92.60% 0.21% 98.51%
## 990 2 465 0.40% 93.00% 0.43% 98.94%
## 992 1 466 0.20% 93.20% 0.21% 99.15%
## 994 1 467 0.20% 93.40% 0.21% 99.36%
## 995 1 468 0.20% 93.60% 0.21% 99.57%
## 997 1 469 0.20% 93.80% 0.21% 99.79%
## 998 1 470 0.20% 94.00% 0.21% 100.00%
## Missing 30 500 6.00% 100.00%
## ─────────────────────────────────────────────────────
##
## ───────────────────────────────────────────────────────
## treatment Freq CumFreq Percent CumPerc Valid CumValid
## 1 1 1 0.20% 0.20% 0.23% 0.23%
## 2 1 2 0.20% 0.40% 0.23% 0.47%
## 5 1 3 0.20% 0.60% 0.23% 0.70%
## 9 1 4 0.20% 0.80% 0.23% 0.93%
## 13 2 6 0.40% 1.20% 0.47% 1.40%
## 19 1 7 0.20% 1.40% 0.23% 1.63%
## 20 1 8 0.20% 1.60% 0.23% 1.86%
## 21 1 9 0.20% 1.80% 0.23% 2.09%
## 22 1 10 0.20% 2.00% 0.23% 2.33%
## 27 3 13 0.60% 2.60% 0.70% 3.02%
## ... ... ... ... ... ... ...
## 974 1 419 0.20% 83.80% 0.23% 97.44%
## 977 1 420 0.20% 84.00% 0.23% 97.67%
## 979 1 421 0.20% 84.20% 0.23% 97.91%
## 980 1 422 0.20% 84.40% 0.23% 98.14%
## 983 2 424 0.40% 84.80% 0.47% 98.60%
## 984 2 426 0.40% 85.20% 0.47% 99.07%
## 989 1 427 0.20% 85.40% 0.23% 99.30%
## 995 1 428 0.20% 85.60% 0.23% 99.53%
## 996 1 429 0.20% 85.80% 0.23% 99.77%
## 998 1 430 0.20% 86.00% 0.23% 100.00%
## Missing 70 500 14.00% 100.00%
## ───────────────────────────────────────────────────────
## Using dplyr::group_by() groups: c
## Variable:d
##
## ─────────────────────────────────────────────────────
## control Freq CumFreq Percent CumPerc Valid CumValid
## 1 2 2 0.40% 0.40% 0.43% 0.43%
## 3 1 3 0.20% 0.60% 0.21% 0.64%
## 4 1 4 0.20% 0.80% 0.21% 0.85%
## 10 1 5 0.20% 1.00% 0.21% 1.06%
## 11 3 8 0.60% 1.60% 0.64% 1.70%
## 15 1 9 0.20% 1.80% 0.21% 1.91%
## 17 1 10 0.20% 2.00% 0.21% 2.13%
## 18 2 12 0.40% 2.40% 0.43% 2.55%
## 20 1 13 0.20% 2.60% 0.21% 2.77%
## 21 2 15 0.40% 3.00% 0.43% 3.19%
## ... ... ... ... ... ... ...
## 986 1 460 0.20% 92.00% 0.21% 97.87%
## 987 1 461 0.20% 92.20% 0.21% 98.09%
## 988 1 462 0.20% 92.40% 0.21% 98.30%
## 989 1 463 0.20% 92.60% 0.21% 98.51%
## 990 2 465 0.40% 93.00% 0.43% 98.94%
## 992 1 466 0.20% 93.20% 0.21% 99.15%
## 994 1 467 0.20% 93.40% 0.21% 99.36%
## 995 1 468 0.20% 93.60% 0.21% 99.57%
## 997 1 469 0.20% 93.80% 0.21% 99.79%
## 998 1 470 0.20% 94.00% 0.21% 100.00%
## Missing 30 500 6.00% 100.00%
## ─────────────────────────────────────────────────────
##
## ───────────────────────────────────────────────────────
## treatment Freq CumFreq Percent CumPerc Valid CumValid
## 1 1 1 0.20% 0.20% 0.23% 0.23%
## 2 1 2 0.20% 0.40% 0.23% 0.47%
## 5 1 3 0.20% 0.60% 0.23% 0.70%
## 9 1 4 0.20% 0.80% 0.23% 0.93%
## 13 2 6 0.40% 1.20% 0.47% 1.40%
## 19 1 7 0.20% 1.40% 0.23% 1.63%
## 20 1 8 0.20% 1.60% 0.23% 1.86%
## 21 1 9 0.20% 1.80% 0.23% 2.09%
## 22 1 10 0.20% 2.00% 0.23% 2.33%
## 27 3 13 0.60% 2.60% 0.70% 3.02%
## ... ... ... ... ... ... ...
## 974 1 419 0.20% 83.80% 0.23% 97.44%
## 977 1 420 0.20% 84.00% 0.23% 97.67%
## 979 1 421 0.20% 84.20% 0.23% 97.91%
## 980 1 422 0.20% 84.40% 0.23% 98.14%
## 983 2 424 0.40% 84.80% 0.47% 98.60%
## 984 2 426 0.40% 85.20% 0.47% 99.07%
## 989 1 427 0.20% 85.40% 0.23% 99.30%
## 995 1 428 0.20% 85.60% 0.23% 99.53%
## 996 1 429 0.20% 85.80% 0.23% 99.77%
## 998 1 430 0.20% 86.00% 0.23% 100.00%
## Missing 70 500 14.00% 100.00%
## ───────────────────────────────────────────────────────
Lastly, tableX()
is a pipe-able two-way version of
table()
with a similar syntax to that of the rest of the
furniture
functions.
## ifelse(d > 500, 1, 0)
## c 0 1 Missing Total
## control 234 236 30 500
## treatment 224 206 70 500
## Total 458 442 100 1000
By default, it provides the total counts for the rows and columns with flexibility as to what is displayed and where.
The four functions: table1()
, tableC()
,
tableF()
, and washer()
add simplicity to
cleaning up and understanding your data. Use these pieces of furniture
to make your quantitative life a bit easier.