-
Notifications
You must be signed in to change notification settings - Fork 118
/
02_Names_and_values.Rmd
executable file
·486 lines (350 loc) · 14.8 KB
/
02_Names_and_values.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
\mainmatter
```{r, include = FALSE}
source("common.R")
```
# (PART) Foundations {-}
\stepcounter{chapter}
# Names and values
<!-- 2 -->
## Prerequisites {-}
<!-- 2.0 -->
In this chapter we will use the `{lobstr}` package [@lobstr] to help answer questions regarding the internal representation of R objects.
```{r setup, message = FALSE}
library(lobstr)
```
\stepcounter{section}
## Binding basics
<!-- 2.2 -->
__[Q1]{.Q}__: Explain the relationship between `a`, `b`, `c`, and `d` in the following code:
```{r}
a <- 1:10
b <- a
c <- b
d <- 1:10
```
__[A]{.solved}__: `a`, `b`, and `c` point to the same object (with the same address in memory). This object has the value `1:10`. `d` points to a different object with the same value.
```{r}
list_of_names <- list(a, b, c, d)
obj_addrs(list_of_names)
```
__[Q2]{.Q}__: The following code accesses the mean function in multiple ways. Do they all point to the same underlying function object? Verify this with `lobstr::obj_addr()`.
```{r, eval = FALSE}
mean
base::mean
get("mean")
evalq(mean)
match.fun("mean")
```
__[A]{.solved}__: Yes, they point to the same object. We confirm this by inspecting the address of the underlying function object.
```{r}
mean_functions <- list(
mean,
base::mean,
get("mean"),
evalq(mean),
match.fun("mean")
)
unique(obj_addrs(mean_functions))
```
__[Q3]{.Q}__: By default, base R data import functions, like `read.csv()`, will automatically convert non-syntactic names to syntactic ones. Why might this be problematic? What option allows you to suppress this behaviour?
__[A]{.solved}__: Column names are often data, and the underlying `make.names()` transformation is non-invertible, so the default behaviour corrupts data. To avoid this, set `check.names = FALSE`.
__[Q4]{.Q}__: What rules does `make.names()` use to convert non-syntactic names into syntactic ones?
__[A]{.solved}__: A valid name must start with a letter or a dot (not followed by a number) and may further contain numbers and underscores (`"_"`s are allowed since R version 1.9.0).
Three main mechanisms ensure syntactically valid names (see `?make.names`):
1. Names that do not start with a letter or a dot will be prepended with an `"X"`.
```{r}
make.names("") # prepending "x"
```
The same holds for names that begin with a dot followed by a number.
```{r}
make.names(".1") # prepending "X"
```
2. Additionally, non-valid characters are replaced by a dot.
```{r}
make.names("non-valid") # "." replacement
make.names("@") # prepending "X" + "." replacement
make.names(" R") # prepending "X" + ".." replacement
```
3. Reserved R keywords (see `?reserved`) are suffixed by a dot.
```{r}
make.names("if") # "." suffix
```
Interestingly, some of these transformations are influenced by the current locale. From `?make.names`:
> The definition of a letter depends on the current locale, but only ASCII digits are considered to be digits.
__[Q5]{.Q}__: I slightly simplified the rules that govern syntactic names. Why is `.123e1` not a syntactic name? Read `?make.names` for the full details.
__[A]{.solved}__: `.123e1` is not a syntactic name, because it starts with one dot which is followed by a number. This makes it a double, `1.23`.
## Copy-on-modify
<!-- 2.3 -->
__[Q1]{.Q}__: Why is `tracemem(1:10)` not useful?
__[A]{.solved}__: When `1:10` is called an object with an address in memory is created, but it is not bound to a name. Therefore, the object cannot be called or manipulated from R. As no copies will be made, it is not useful to track the object for copying.
```{r}
obj_addr(1:10) # the object exists, but has no name
```
__[Q2]{.Q}__: Explain why `tracemem()` shows two copies when you run this code. Hint: carefully look at the difference between this code and the code shown earlier in the section.
```{r, results = FALSE}
x <- c(1L, 2L, 3L)
tracemem(x)
x[[3]] <- 4
```
__[A]{.solved}__: Initially the vector `x` has integer type. The replacement call assigns a double to the third element of `x`, which triggers copy-on-modify.
```{r, eval = FALSE}
x <- c(1L, 2L, 3L)
tracemem(x)
#> <0x66a4a70>
x[[3]] <- 4
#> tracemem[0x55eec7b3af38 -> 0x55eec774cc18]:
```
We can avoid the copy by sub-assigning an integer instead of a double:
```{r, eval = FALSE}
x <- c(1L, 2L, 3L)
tracemem(x)
#> <0x55eec6940ae0>
x[[3]] <- 4L
```
Please be aware that running this code in RStudio will result in additional copies because of the reference from the environment pane.
__[Q3]{.Q}__: Sketch out the relationship between the following objects:
```{r}
a <- 1:10
b <- list(a, a)
c <- list(b, a, 1:10)
```
__[A]{.solved}__: `a` contains a reference to an address with the value `1:10`. `b` contains a list of two references to the same address as `a`. `c` contains a list of `b` (containing two references to `a`), `a` (containing the same reference again) and a reference pointing to a different address containing the same value (`1:10`).
```{r, echo = FALSE, out.width='180pt'}
knitr::include_graphics("images/names_values/copy_on_modify_fig3.png",
dpi = 300)
```
We can confirm these relationships by inspecting the reference tree in R.
```{r, eval = FALSE}
ref(c)
#> █ [1:0x55erc93cbdd8] <list> # c
#> ├─█ [2:0x55efcb8246e8] <list> # - b
#> │ ├─[3:0x55eac7df4e98] <int> # -- a
#> │ └─[3:0x55eac7df4e98] # -- a
#> ├─[3:0x55eac7df4e98] # - a
#> └─[4:0x55etc7aa6968] <int> # - 1:10
```
__[Q4]{.Q}__: What happens when you run this code:
```{r}
x <- list(1:10)
x[[2]] <- x
```
Draw a picture.
__[A]{.solved}__: The initial reference tree of `x` shows that the name `x` binds to a list object. This object contains a reference to the integer vector `1:10`.
```{r, eval = FALSE}
x <- list(1:10)
ref(x)
#> █ [1:0x55853b74ff40] <list>
#> └─[2:0x534t3abffad8] <int>
```
```{r, echo = FALSE, out.width='135pt'}
knitr::include_graphics("images/names_values/copy_on_modify_fig1.png",
dpi = 300)
```
When `x` is assigned to an element of itself, copy-on-modify takes place and the list is copied to a new address in memory.
```{r, eval = FALSE}
tracemem(x)
x[[2]] <- x
#> tracemem[0x55853b74ff40 -> 0x5d553bacdcd8]:
```
The list object previously bound to `x` is now referenced in the newly created list object. It is no longer bound to a name. The integer vector is referenced twice.
```{r, eval = FALSE}
ref(x)
#> █ [1:0x5d553bacdcd8] <list>
#> ├─[2:0x534t3abffad8] <int>
#> └─█ [3:0x55853b74ff40] <list>
#> └─[2:0x534t3abffad8]
```
```{r, echo = FALSE, out.width = '150pt'}
knitr::include_graphics("images/names_values/copy_on_modify_fig2.png",
dpi = 300)
```
## Object size
<!-- 2.4 -->
__[Q1]{.Q}__: In the following example, why are `object.size(y)` and `obj_size(y)` so radically different? Consult the documentation of `object.size()`.
```{r}
y <- rep(list(runif(1e4)), 100)
object.size(y)
obj_size(y)
```
__[A]{.solved}__: `object.size()` doesn't account for shared elements within lists. Therefore, the results differ by a factor of ~ 100.
__[Q2]{.Q}__: Take the following list. Why is its size somewhat misleading?
```{r}
funs <- list(mean, sd, var)
obj_size(funs)
```
__[A]{.solved}__: All three functions are built-in to R as part of the `{base}` and `{stats}` packages and hence always available. So, what does it mean to measure the size of something that's already included in R?
(There's typically a more general question about what you want to know when you ask for the size of something — do you want to know how much data you'd need to send to communicate the object to someone else (e.g. serialise it), or do you want to know how much memory you'd free if you deleted it?)
Let us look for how many other objects this applies to as well.
The following packages are usually loaded by default.
```{r}
base_pkgs <- c(
"package:stats", "package:graphics", "package:grDevices",
"package:utils", "package:datasets", "package:methods",
"package:base"
)
```
To look up all functions from these packages we iterate over `base_pkgs` and apply `ls()` and `mget()` within each iteration.
```{r}
base_objs <- base_pkgs %>%
lapply(as.environment) %>%
lapply(function(x) mget(ls(x, all.names = TRUE), x)) %>%
setNames(base_pkgs)
```
This gives us more than 2700 objects which are usually available by default.
```{r}
sum(lengths(base_objs))
# We can also show the sizes in MB per package
vapply(base_objs, obj_size, double(1)) / 1024^2
# Check if we've over-counted
as.numeric(obj_size(!!!base_objs)) / 1024^2
```
__[Q3]{.Q}__: Predict the output of the following code:
```{r, eval = FALSE}
a <- runif(1e6)
obj_size(a)
b <- list(a, a)
obj_size(b)
obj_size(a, b)
b[[1]][[1]] <- 10
obj_size(b)
obj_size(a, b)
b[[2]][[1]] <- 10
obj_size(b)
obj_size(a, b)
```
__[A]{.solved}__: In R (on most platforms) a length-0
vector has 48 bytes of overhead.
```{r}
obj_size(list())
obj_size(double())
obj_size(character())
```
A single double takes up an additional 8 bytes of memory.
```{r}
obj_size(double(1))
obj_size(double(2))
```
So, a 1 million double should take up 8,000,048 bytes.
```{r}
a <- runif(1e6)
obj_size(a)
```
(If you look carefully at the amount of memory occupied by short vectors, you will notice that the pattern is actually more complicated. This has to do with how R allocates memory and is not that important. If you want to know the full details, they're discussed in the 1st edition of *Advanced R*: http://adv-r.had.co.nz/memory.html#object-size).
For `b <- list(a, a)` both list elements contain references to the same memory address.
```{r}
b <- list(a, a)
ref(a, b)
```
Therefore, no additional memory is required for the second list element. The list itself requires 64 bytes, 48 bytes for an empty list and 8 bytes for each element (`obj_size(vector("list", 2))`). This lets us predict 8,000,048 B + 64 B = 8,000,112 B.
```{r}
obj_size(b)
```
When we modify the first element of `b[[1]]` copy-on-modify occurs. Both elements will still have the same size (8,000,040 B), but the first one gets a new address in memory. As `b`'s elements don't share references anymore, its object size adds up to the sum of the elements and the length-2 list: 8,000,048 B + 8,000,048 B + 64 B = 16,000,160 B (16 MB).
```{r}
b[[1]][[1]] <- 10
obj_size(b)
```
The second element of `b` still references the same address as `a`, so the combined size of `a` and `b` is the same as `b`.
```{r}
obj_size(a, b)
ref(a, b)
```
When we modify the second element of `b`, this element will also point to a new memory address. This does not affect the size of the list.
```{r}
b[[2]][[1]] <- 10
obj_size(b)
```
However, as `b` doesn't share references with `a` anymore, the memory usage of the combined objects increases.
```{r}
ref(a, b)
obj_size(a, b)
```
## Modify-in-place
<!-- 2.5 -->
__[Q1]{.Q}__: Explain why the following code doesn't create a circular list.
```{r}
x <- list()
x[[1]] <- x
```
__[A]{.solved}__: In this situation copy-on-modify prevents the creation of a circular list. Let us step through the details:
```{r, eval = FALSE}
x <- list() # creates initial object
obj_addr(x)
#> [1] "0x55862f23ab80"
tracemem(x)
#> [1] "<0x55862f23ab80>"
x[[1]] <- x # Copy-on-modify triggers new copy
#> tracemem[0x55862f23ab80 -> 0x55862e8ce028]:
obj_addr(x) # copied object has new memory address
#> [1] "0x55862e8ce028"
obj_addr(x[[1]]) # list element contains old memory address
#> [1] "0x55862f23ab80"
```
__[Q2]{.Q}__: Wrap the two methods for subtracting medians into two functions, then use the `{bench}` package to carefully compare their speeds. How does performance change as the number of columns increase?
__[A]{.solved}__: First, we define a function to create some random data.
```{r}
create_random_df <- function(nrow, ncol) {
random_matrix <- matrix(runif(nrow * ncol), nrow = nrow)
as.data.frame(random_matrix)
}
create_random_df(2, 2)
```
Next, we wrap the two approaches to subtract numerical values (in our case medians) from each column of a data frame in their own function. We name these functions depending on whether the approach operates on a data frame or a list. For a fairer comparison, the second function also contains the overhead code to convert between data frame and list objects.
```{r}
subtract_df <- function(x, medians) {
for (i in seq_along(medians)) {
x[[i]] <- x[[i]] - medians[[i]]
}
x
}
subtract_list <- function(x, medians) {
x <- as.list(x)
x <- subtract_df(x, medians)
list2DF(x)
}
```
This lets us profile the performance, via benchmarks on data frames with differing numbers of columns. Therefore, we create a small helper that creates our random data frame and its medians before it benchmarks the two approaches by employing the `{bench}` package [@bench].
```{r}
benchmark_medians <- function(ncol) {
df <- create_random_df(nrow = 1e4, ncol = ncol)
medians <- vapply(df, median, numeric(1), USE.NAMES = FALSE)
bench::mark(
"data frame" = subtract_df(df, medians),
"list" = subtract_list(df, medians),
time_unit = "ms"
)
}
benchmark_medians(1)
```
The `bench::press()` function allows us to run our helper across a grid of parameters. We will use it to slowly increase the number of data frame columns in our benchmark.
```{r}
results <- bench::press(
ncol = c(1, 10, 50, 100, 250, 300, 400, 500, 750, 1000),
benchmark_medians(ncol)
)
```
Finally, we can plot and interpret our results.
```{r}
library(ggplot2)
ggplot(
results,
aes(ncol, median, col = attr(expression, "description"))
) +
geom_point(size = 2) +
geom_smooth() +
labs(
x = "Number of Columns",
y = "Execution Time (ms)",
colour = "Data Structure"
) +
theme(legend.position = "top")
```
When working directly with the data frame, the execution time grows quadratically with the number of columns in the input data. This is because (e.g.) the first column must be copied n times, the second column n-1 times, and so on. When working with a list, the execution time increases only linearly.
Obviously in the long run, linear growth creates shorter run-times, but there is some cost to this strategy — we have to convert between data structures with `as.list()` and `list2DF()`. Even though this is fast and probably doesn't hurt much, the improved approach doesn't really pay off in this scenario until we get to a data frame that is about 300 columns wide (with the exact value depending on the characteristics of the system running the code).
__[Q3]{.Q}__: What happens if you attempt to use `tracemem()` on an environment?
__[A]{.solved}__: `tracemem()` cannot be used to mark and trace environments.
```{r, error = TRUE}
x <- new.env()
tracemem(x)
```
The error occurs because "it is not useful to trace NULL, environments, promises, weak references, or external pointer objects, as these are not duplicated" (see `?tracemem`). Environments are always modified in place.