Last updated: 2021-09-24

Checks: 7 0

Knit directory: myTidyTuesday/

This reproducible R Markdown analysis was created with workflowr (version 1.6.2). The Checks tab describes the reproducibility checks that were applied when the results were created. The Past versions tab lists the development history.


Great! Since the R Markdown file has been committed to the Git repository, you know the exact version of the code that produced these results.

Great job! The global environment was empty. Objects defined in the global environment can affect the analysis in your R Markdown file in unknown ways. For reproduciblity it’s best to always run the code in an empty environment.

The command set.seed(20210907) was run prior to running the code in the R Markdown file. Setting a seed ensures that any results that rely on randomness, e.g. subsampling or permutations, are reproducible.

Great job! Recording the operating system, R version, and package versions is critical for reproducibility.

Nice! There were no cached chunks for this analysis, so you can be confident that you successfully produced the results during this run.

Great job! Using relative paths to the files within your workflowr project makes it easier to run your code on other machines.

Great! You are using Git for version control. Tracking code development and connecting the code version to the results is critical for reproducibility.

The results in this page were generated with repository version a2779cd. See the Past versions tab to see a history of the changes made to the R Markdown and HTML files.

Note that you need to be careful to ensure that all relevant files for the analysis have been committed to Git prior to generating the results (you can use wflow_publish or wflow_git_commit). workflowr only checks the R Markdown file, but you know if there are other scripts or data files that it depends on. Below is the status of the Git repository when the results were generated:


Ignored files:
    Ignored:    .Rhistory
    Ignored:    .Rproj.user/
    Ignored:    catboost_info/
    Ignored:    data/2021-09-08/
    Ignored:    data/CNHI_Excel_Chart.xlsx
    Ignored:    data/CommunityTreemap.jpeg
    Ignored:    data/Community_Roles.jpeg
    Ignored:    data/YammerDigitalDataScienceMembership.xlsx
    Ignored:    data/acs_poverty.rds
    Ignored:    data/fmhpi.rds
    Ignored:    data/grainstocks.rds
    Ignored:    data/hike_data.rds
    Ignored:    data/us_states.rds
    Ignored:    data/us_states_hexgrid.geojson
    Ignored:    data/weatherstats_toronto_daily.csv

Untracked files:
    Untracked:  code/YammerReach.R
    Untracked:  code/work list batch targets.R

Note that any generated files, e.g. HTML, png, CSS, etc., are not included in this status report because it is ok for generated content to have uncommitted changes.


These are the previous versions of the repository in which changes were made to the R Markdown (analysis/2021_01_12_tidy_tuesday.Rmd) and HTML (docs/2021_01_12_tidy_tuesday.html) files. If you’ve configured a remote Git repository (see ?wflow_git_remote), click on the hyperlinks in the table below to view the files as they were in that past version.

File Version Author Date Message
Rmd a2779cd opus1993 2021-09-24 add Tate art collection TidyTuesday

The data this week comes from the Tate Art Collections. It consists of metadata for around 70,000 artworks that Tate owns or jointly owns with the National Galleries of Scotland as part of ARTIST ROOMS. Metadata for around 3,500 associated artists is also included.

The post will be in two sections. The first is an exploration of the content of the dataset and a submission to Twitter. The second is a machine learning model that helps us better understand the importance of changes in art materials over time.

Load the weekly Data

Download from the R4DS github website and make available in the tt object.

tt <- tt_load("2021-01-12")

    Downloading file 1 of 2: `artists.csv`
    Downloading file 2 of 2: `artwork.csv`
art <- tt$artwork %>%
  rename(art_url = url) %>%
  left_join(tt$artists, by = c("artistId" = "id")) %>%
  mutate(
    gender = factor(gender),
    on = str_sub(medium, start = str_locate(medium, " on ")[, 2])
  )

Visualize

What is the distribution of artwork over creation year?

art %>%
  ggplot(aes(year)) +
  geom_histogram(alpha = 0.8) +
  labs(title = "The Tate Collection acquisitions are bimodally distributed across years")

A treemap of major Tate artists

art %>%
  mutate(birthDecade = yearOfBirth %/% 10 * 10) %>%
  count(artist, birthDecade) %>%
  slice_max(order_by = n, n = 60) %>%
  filter(!is.na(birthDecade), artist != "Turner, Joseph Mallord William") %>%
  ggplot(aes(area = n, fill = birthDecade, label = artist, subgroup = birthDecade)) +
  geom_treemap() +
  geom_treemap_subgroup_border() +
  geom_treemap_text(color = "black", place = "top", reflow = TRUE) +
  geom_treemap_subgroup_text(
    color = "gray40", place = "bottomright",
    fontface = "italic", min.size = 0, alpha = 0.3
  ) +
  scale_fill_fermenter(
    palette = "YlOrRd", direction = -1,
    guide = "colorsteps"
  ) +
  labs(
    fill = NULL, title = "The Tate Collection's Artists by birth year (Non-Turner)",
    subtitle = "Born in 1775, Joseph Mallord William Turner's works account for 57% alone. The size of\nthe box corresponds to the number of pieces of each non-Turner artist in the collection.",
    caption = "Data Source: Tate Art Collection"
  )

And the Tidytuesday tweet:


A machine learning model

Let’s focus on just a couple of variables for modeling, and at least focus on just the art created after 1750.

tate_df <- art %>%
  filter(year > 1750) %>%
  select(year, medium, width, height) %>%
  na.omit() %>%
  arrange(year)

What are the most common words used in describing the media?

tate_df %>%
  unnest_tokens(word, medium) %>%
  count(word, sort = TRUE)

Lots of paper, graphite, oil and watercolour paints.

We can start by splitting our data into training and testing sets, and creating resamples.

art_split <- initial_split(tate_df, strata = year)
art_train <- training(art_split)
art_test <- testing(art_split)

art_folds <- vfold_cv(art_train, strata = year)

Next, let’s preprocess to get it ready for modeling. We can use specialized steps from textrecipes, along with the general recipe steps.

art_rec <- recipe(year ~ ., data = art_train) %>%
  step_tokenize(medium) %>%
  step_stopwords(medium) %>%
  #  step_stem(medium) %>%
  step_ngram(medium, num_tokens = 3, min_num_tokens = 1) %>%
  step_tokenfilter(medium, max_tokens = 700) %>%
  step_tfidf(medium) # a normalized input is required for regularized regression

art_rec
Data Recipe

Inputs:

      role #variables
   outcome          1
 predictor          3

Operations:

Tokenization for medium
Stop word removal for medium
ngramming for medium
Text filtering for medium
Term frequency-inverse document frequency with medium

Let’s walk through the steps in this recipe, which are what are sensible defaults for training a model with text features.

Next, it’s time to specify our model and put it together with our recipe into a workflow. To use a fast, compact sparse data structure, we will create a hardhat blueprint with composition = "dgCMatrix".

sparse_bp <- hardhat::default_recipe_blueprint(composition = "dgCMatrix")

lasso_spec <- linear_reg(penalty = tune(), mixture = 1) %>%
  set_engine("glmnet")

art_wf <- workflow() %>%
  add_recipe(art_rec, blueprint = sparse_bp) %>%
  add_model(lasso_spec)

art_wf
== Workflow ====================================================================
Preprocessor: Recipe
Model: linear_reg()

-- Preprocessor ----------------------------------------------------------------
5 Recipe Steps

* step_tokenize()
* step_stopwords()
* step_ngram()
* step_tokenfilter()
* step_tfidf()

-- Model -----------------------------------------------------------------------
Linear Regression Model Specification (regression)

Main Arguments:
  penalty = tune()
  mixture = 1

Computational engine: glmnet 

The only other piece we need to get ready for model fitting is values for the regularization penalty to try. The default goes down to very tiny penalties and I don’t think we’ll need that, so let’s change the range().

lambda_grid <- grid_regular(penalty(range = c(-3, 0)), levels = 20)

Now let’s tune the lasso model on the resampled datasets we created.

all_cores <- parallelly::availableCores(omit = 1)
all_cores
system 
    11 
future::plan("multisession", workers = all_cores) # on Windows

lasso_rs <- tune_grid(
  art_wf,
  resamples = art_folds,
  grid = lambda_grid
)

lasso_rs

That was quite fast, because it’s a linear model and we used the sparse data structure.

Evaluate model

Now we can do what we really came here for, which is to talk about how we can evaluate a model like this and see if it was a good idea. How do the results look?

autoplot(lasso_rs)

The best $ R^2 $ is around 0.80, and is a measure of how well the model fits the training data. The best RMSE is around 31.5 or so, and is on the scale of the original outcome, i.e. years. What are some of the best penalty values, in terms of RMSE?

show_best(lasso_rs, "rmse")

We can select the best penalty, and finalize the workflow with it.

best_rmse <- select_best(lasso_rs, "rmse")

final_lasso <- finalize_workflow(art_wf, best_rmse)
final_lasso
== Workflow ====================================================================
Preprocessor: Recipe
Model: linear_reg()

-- Preprocessor ----------------------------------------------------------------
5 Recipe Steps

* step_tokenize()
* step_stopwords()
* step_ngram()
* step_tokenfilter()
* step_tfidf()

-- Model -----------------------------------------------------------------------
Linear Regression Model Specification (regression)

Main Arguments:
  penalty = 0.0379269019073225
  mixture = 1

Computational engine: glmnet 

The function last_fit() fits this finalized lasso model one last time to the training data and evaluates one last time on the testing data. The metrics are computed on the testing data.

art_final <- last_fit(final_lasso, art_split)
collect_metrics(art_final)

We can use the fitted workflow in art_final to explore variable importance using the vip package.

art_vip <- pull_workflow_fit(art_final$.workflow[[1]]) %>%
  vi()

another visualization, this time of variable importance in ascertaining the year of the artwork:

art_vip %>%
  group_by(Sign) %>%
  slice_max(abs(Importance), n = 20) %>%
  ungroup() %>%
  mutate(
    Variable = str_remove(Variable, "tfidf_medium_"),
    Importance = abs(Importance),
    Variable = fct_reorder(Variable, Importance),
    Sign = if_else(Sign == "POS", "in later art", "in earlier art")
  ) %>%
  ggplot(aes(Importance, Variable, fill = Sign)) +
  geom_col(show.legend = FALSE) +
  facet_wrap(~Sign, scales = "free") +
  labs(y = NULL)

This tells quite a story about what predictors are most important in pushing the prediction for an observation up or down the most.

How well does the model actually do, though? Let’s plot true and predicted values for years, for the testing data.

collect_predictions(art_final) %>%
  ggplot(aes(year, .pred)) +
  geom_abline(lty = 2, color = "gray50", size = 1.2) +
  geom_point(size = 1.5, alpha = 0.3) +
  coord_fixed()

There are clumps of artwork that are predicted well at the high and low end, but notice that prominent horizontal line of observations that are all predicted to be created at about ~1900. Let’s dig more into misclassifications.

misclassified <- collect_predictions(art_final) %>%
  bind_cols(art_test %>% select(medium)) %>%
  filter(abs(year - .pred) > 100)

misclassified %>%
  arrange(year)

These are pieces of art that were created very early but predicted much later. In fact, notice that “Oil paint on canvas” also predicts to 1886 (with small influence of width and height features); that is about the mean or median of this whole dataset, that was one of the most common media, and this is how a linear model works!

misclassified %>%
  arrange(-year)

These are pieces of art that were created very recently but predicted much earlier. Notice that they have used what we might think of as antique or traditional techniques.

Now, finally, it is time for the residuals. We can compute residuals for the test set with the augment() function.

augment(art_final) %>%
  ggplot(aes(.pred, .resid)) +
  geom_hline(yintercept = 0, lty = 2, color = "gray50", size = 1.2) +
  geom_point(size = 1.5, alpha = 0.3) +
  geom_smooth(color = "midnightblue") +
  labs(title = "Heterscedasticity in Regularized Regression")

This plot exhibits significant heteroscedasticity, with lower variance for recent artwork and higher variance for artwork in other time periods. If the model predicts a recent year, we can be more confident that it is right than if the model predicts an older year, and there is basically no time information in the fact that an artwork was created with a medium like oil on canvas. So is this model bad and not useful? It’s not great, for most goals I can think of, but it’s interesting to notice how much we can learn about our data even from such a model.


sessionInfo()
R version 4.1.1 (2021-08-10)
Platform: x86_64-w64-mingw32/x64 (64-bit)
Running under: Windows 10 x64 (build 19043)

Matrix products: default

locale:
[1] LC_COLLATE=English_United States.1252 
[2] LC_CTYPE=English_United States.1252   
[3] LC_MONETARY=English_United States.1252
[4] LC_NUMERIC=C                          
[5] LC_TIME=English_United States.1252    

attached base packages:
[1] stats     graphics  grDevices utils     datasets  methods   base     

other attached packages:
 [1] glmnet_4.1-2       Matrix_1.3-4       vctrs_0.3.8        rlang_0.4.11      
 [5] stopwords_2.2      vip_0.3.2          textrecipes_0.4.1  yardstick_0.0.8   
 [9] workflowsets_0.1.0 workflows_0.2.3    tune_0.1.6         rsample_0.1.0     
[13] recipes_0.1.16     parsnip_0.1.7.900  modeldata_0.1.1    infer_1.0.0       
[17] dials_0.0.10       scales_1.1.1       broom_0.7.9        tidymodels_0.1.3  
[21] tidytext_0.3.1     tweetrmd_0.0.9     countrycode_1.3.0  hrbrthemes_0.8.0  
[25] treemapify_2.5.5   tidytuesdayR_1.0.1 forcats_0.5.1      stringr_1.4.0     
[29] dplyr_1.0.7        purrr_0.3.4        readr_2.0.1        tidyr_1.1.3       
[33] tibble_3.1.4       ggplot2_3.3.5      tidyverse_1.3.1    workflowr_1.6.2   

loaded via a namespace (and not attached):
  [1] readxl_1.3.1       backports_1.2.1    systemfonts_1.0.2 
  [4] selectr_0.4-2      plyr_1.8.6         splines_4.1.1     
  [7] listenv_0.8.0      SnowballC_0.7.0    usethis_2.0.1     
 [10] digest_0.6.27      foreach_1.5.1      htmltools_0.5.2   
 [13] viridis_0.6.1      fansi_0.5.0        magrittr_2.0.1    
 [16] tzdb_0.1.2         globals_0.14.0     ggfittext_0.9.1   
 [19] modelr_0.1.8       gower_0.2.2        extrafont_0.17    
 [22] vroom_1.5.5        R.utils_2.10.1     extrafontdb_1.0   
 [25] hardhat_0.1.6      colorspace_2.0-2   rvest_1.0.1       
 [28] textshaping_0.3.5  haven_2.4.3        xfun_0.26         
 [31] crayon_1.4.1       jsonlite_1.7.2     survival_3.2-11   
 [34] iterators_1.0.13   glue_1.4.2         gtable_0.3.0      
 [37] ipred_0.9-12       R.cache_0.15.0     Rttf2pt1_1.3.9    
 [40] shape_1.4.6        future.apply_1.8.1 DBI_1.1.1         
 [43] Rcpp_1.0.7         viridisLite_0.4.0  bit_4.0.4         
 [46] GPfit_1.0-8        lava_1.6.10        prodlim_2019.11.13
 [49] httr_1.4.2         RColorBrewer_1.1-2 ellipsis_0.3.2    
 [52] farver_2.1.0       R.methodsS3_1.8.1  pkgconfig_2.0.3   
 [55] nnet_7.3-16        sass_0.4.0         dbplyr_2.1.1      
 [58] utf8_1.2.2         here_1.0.1         labeling_0.4.2    
 [61] tidyselect_1.1.1   DiceDesign_1.9     later_1.3.0       
 [64] munsell_0.5.0      cellranger_1.1.0   tools_4.1.1       
 [67] cachem_1.0.6       cli_3.0.1          generics_0.1.0    
 [70] evaluate_0.14      fastmap_1.1.0      ragg_1.1.3        
 [73] yaml_2.2.1         rematch2_2.1.2     bit64_4.0.5       
 [76] knitr_1.34         fs_1.5.0           nlme_3.1-152      
 [79] future_1.22.1      whisker_0.4        R.oo_1.24.0       
 [82] xml2_1.3.2         tokenizers_0.2.1   compiler_4.1.1    
 [85] rstudioapi_0.13    curl_4.3.2         reprex_2.0.1      
 [88] lhs_1.1.3          bslib_0.3.0        stringi_1.7.4     
 [91] highr_0.9          gdtools_0.2.3      lattice_0.20-44   
 [94] styler_1.6.1       conflicted_1.0.4   pillar_1.6.2      
 [97] lifecycle_1.0.1    furrr_0.2.3        jquerylib_0.1.4   
[100] httpuv_1.6.3       R6_2.5.1           promises_1.2.0.1  
[103] gridExtra_2.3      janeaustenr_0.1.5  parallelly_1.28.1 
[106] codetools_0.2-18   MASS_7.3-54        assertthat_0.2.1  
[109] rprojroot_2.0.2    withr_2.4.2        mgcv_1.8-36       
[112] parallel_4.1.1     hms_1.1.0          grid_4.1.1        
[115] rpart_4.1-15       timeDate_3043.102  class_7.3-19      
[118] rmarkdown_2.11     git2r_0.28.0       pROC_1.18.0       
[121] lubridate_1.7.10  
---
title: "Tate Art Collection"
date: "January 12, 2021"
output:
  workflowr::wflow_html:
    toc: no
    code_folding: hide
    code_download: true
    df_print: paged
editor_options:
  chunk_output_type: console
---

The data this week comes from the [Tate Art Collections](https://github.com/tategallery/collection). It consists of metadata for around 70,000 artworks that Tate owns or jointly owns with the National Galleries of Scotland as part of [ARTIST ROOMS](http://www.tate.org.uk/artist-rooms). Metadata for around 3,500 associated artists is also included.

The post will be in two sections. The first is an exploration of the content of the dataset and a submission to Twitter. The second is a machine learning model that helps us better understand the importance of changes in art materials over time.

```{r setup, include=FALSE}
suppressPackageStartupMessages({
library(tidyverse)
library(tidytuesdayR)
library(treemapify)
library(hrbrthemes)
library(countrycode)
library(tweetrmd)

library(tidytext)
library(tidymodels)
library(textrecipes)
library(vip)
})
extrafont::loadfonts(quiet = TRUE)

source(here::here("code","_common.R"),
       verbose = FALSE,
       local = knitr::knit_global())

theme_set(theme_jim(base_size = 9))

```

## Load the weekly Data

Download from the R4DS github website and make available in the `tt` object.

```{r Load}

tt <- tt_load("2021-01-12")

```


```{r prepare}
art <- tt$artwork %>% 
  rename(art_url = url) %>% 
  left_join(tt$artists, by = c("artistId" = "id")) %>% 
  mutate(gender = factor(gender),
         on = str_sub(medium, start = str_locate(medium, " on ")[,2])) 

```


## Visualize

What is the distribution of artwork over creation year?

```{r}
art %>%
  ggplot(aes(year)) +
  geom_histogram(alpha = 0.8) +
  labs(title = "The Tate Collection acquisitions are bimodally distributed across years")
```

A treemap of major Tate artists

```{r Visualize, fig.asp=1}

art %>% 
  mutate(birthDecade = yearOfBirth %/% 10 * 10) %>% 
  count(artist, birthDecade) %>% 
  slice_max(order_by = n, n = 60) %>% 
  filter(!is.na(birthDecade), artist != "Turner, Joseph Mallord William") %>% 
  ggplot(aes(area = n, fill = birthDecade, label = artist, subgroup = birthDecade)) +
  geom_treemap() +
  geom_treemap_subgroup_border() +
  geom_treemap_text(color = "black", place = "top", reflow = TRUE) +
  geom_treemap_subgroup_text(color = "gray40", place = "bottomright", 
                             fontface = "italic", min.size = 0, alpha = 0.3) +
  scale_fill_fermenter(palette = "YlOrRd", direction = -1,
                       guide = "colorsteps") +
  labs(fill = NULL, title = "The Tate Collection's Artists by birth year (Non-Turner)",
       subtitle = "Born in 1775, Joseph Mallord William Turner's works account for 57% alone. The size of\nthe box corresponds to the number of pieces of each non-Turner artist in the collection.",
       caption = "Data Source: Tate Art Collection")
  
```

And the Tidytuesday tweet:

```{r echo=FALSE}
tweet_embed('https://twitter.com/jim_gruman/status/1349150540610076673')
```

----
# A machine learning model 

Let’s focus on just a couple of variables for modeling, and at least focus on just the art created after 1750.

```{r}
tate_df <- art %>%
  filter(year > 1750) %>%
  select(year, medium, width, height) %>%
  na.omit() %>%
  arrange(year)
```

What are the most common words used in describing the media?

```{r}
tate_df %>% 
  unnest_tokens(word, medium) %>% 
  count(word, sort = TRUE)
```

Lots of paper, graphite, oil and watercolour paints.

We can start by splitting our data into training and testing sets, and creating resamples.

```{r}
art_split <- initial_split(tate_df, strata = year)
art_train <- training(art_split)
art_test <- testing(art_split)

art_folds <- vfold_cv(art_train, strata = year)
```

Next, let’s preprocess to get it ready for modeling. We can use specialized steps from `textrecipes`, along with the general recipe steps.

```{r}
art_rec <- recipe(year ~ ., data = art_train) %>%
  step_tokenize(medium) %>%
  step_stopwords(medium) %>%
#  step_stem(medium) %>%
  step_ngram(medium, num_tokens = 3, min_num_tokens = 1) %>% 
  step_tokenfilter(medium, max_tokens = 700) %>%
  step_tfidf(medium)   # a normalized input is required for regularized regression

art_rec
```

Let’s walk through the steps in this recipe, which are what are sensible defaults for training a model with text features.

- First, we must tell the `recipe()` what our model is going to be (using a formula here) and what data we are using.

- Next, we tokenize our text, with the default tokenization into single words.

- Next, we remove stop words (just the default set, to remove “on” and “and”).

- Next, we stem words to a similar base word. In this case, I am going to leave stemming out, as the medium words are mostly technical materials.

- Next, we form ngram combinations back. In this case, "acrylic paint paper" results in features for *acrylic*, *paint*, *paper*, *acrylic paint*, *paint paper*, *acrylic paper*, and *arylic paint paper*.

- It wouldn’t be practical to keep all the tokens from this whole dataset in our model, so we can filter down to only keep, in this case, the top 700 most-used tokens (after removing stop words).

- We need to decide on some kind of weighting for these tokens next, either something like term frequency or, what we used here, tf-idf.

Next, it’s time to specify our model and put it together with our recipe into a workflow. To use a fast, compact sparse data structure, we will create a hardhat blueprint with `composition = "dgCMatrix"`.

```{r}
sparse_bp <- hardhat::default_recipe_blueprint(composition = "dgCMatrix")

lasso_spec <- linear_reg(penalty = tune(), mixture = 1) %>%
  set_engine("glmnet")

art_wf <- workflow() %>%
  add_recipe(art_rec, blueprint = sparse_bp) %>%
  add_model(lasso_spec)

art_wf
```

The only other piece we need to get ready for model fitting is values for the regularization penalty to try. The default goes down to very tiny penalties and I don’t think we’ll need that, so let’s change the `range()`.

```{r}
lambda_grid <- grid_regular(penalty(range = c(-3, 0)), levels = 20)
```

Now let’s tune the lasso model on the resampled datasets we created.

```{r}
all_cores <- parallelly::availableCores(omit = 1)
all_cores

future::plan("multisession", workers = all_cores) # on Windows

lasso_rs <- tune_grid(
  art_wf,
  resamples = art_folds,
  grid = lambda_grid
)

lasso_rs
```

That was quite fast, because it’s a linear model and we used the sparse data structure.

## Evaluate model

Now we can do what we really came here for, which is to talk about how we can evaluate a model like this and see if it was a good idea. How do the results look?

```{r}
autoplot(lasso_rs)
```

The best $ R^2 $ is around 0.80, and is a measure of how well the model fits the training data. The best RMSE is around 31.5 or so, and is on the scale of the original outcome, i.e. years. What are some of the best penalty values, in terms of RMSE?

```{r}
show_best(lasso_rs, "rmse")
```

We can select the best penalty, and finalize the workflow with it.

```{r}
best_rmse <- select_best(lasso_rs, "rmse")

final_lasso <- finalize_workflow(art_wf, best_rmse)
final_lasso
```

The function `last_fit()` fits this finalized lasso model one last time to the training data and evaluates one last time on the testing data. The metrics are computed on the testing data.

```{r}
art_final <- last_fit(final_lasso, art_split)
collect_metrics(art_final)
```

We can use the fitted workflow in art_final to explore variable importance using the vip package.

```{r}
art_vip <- pull_workflow_fit(art_final$.workflow[[1]]) %>%
  vi()
```

another visualization, this time of variable importance in ascertaining the year of the artwork:

```{r}
art_vip %>%
  group_by(Sign) %>%
  slice_max(abs(Importance), n = 20) %>%
  ungroup() %>%
  mutate(
    Variable = str_remove(Variable, "tfidf_medium_"),
    Importance = abs(Importance),
    Variable = fct_reorder(Variable, Importance),
    Sign = if_else(Sign == "POS", "in later art", "in earlier art")
  ) %>%
  ggplot(aes(Importance, Variable, fill = Sign)) +
  geom_col(show.legend = FALSE) +
  facet_wrap(~Sign, scales = "free") +
  labs(y = NULL)
```

This tells quite a story about what predictors are most important in pushing the prediction for an observation up or down the most.

How well does the model actually do, though? Let’s plot true and predicted values for years, for the testing data.

```{r}
collect_predictions(art_final) %>%
  ggplot(aes(year, .pred)) +
  geom_abline(lty = 2, color = "gray50", size = 1.2) +
  geom_point(size = 1.5, alpha = 0.3) +
  coord_fixed()
```

There are clumps of artwork that are predicted well at the high and low end, but notice that prominent horizontal line of observations that are all predicted to be created at about ~1900. Let’s dig more into misclassifications.

```{r}
misclassified <- collect_predictions(art_final) %>%
  bind_cols(art_test %>% select(medium)) %>%
  filter(abs(year - .pred) > 100)

misclassified %>%
  arrange(year)
```

These are pieces of art that were created very early but predicted much later. In fact, notice that “Oil paint on canvas” also predicts to 1886 (with small influence of width and height features); that is about the mean or median of this whole dataset, that was one of the most common media, and this is how a linear model works!

```{r}
misclassified %>%
  arrange(-year)
```

These are pieces of art that were created very recently but predicted much earlier. Notice that they have used what we might think of as antique or traditional techniques.

Now, finally, it is time for the residuals. We can compute residuals for the test set with the `augment()` function.

```{r}
augment(art_final) %>%
  ggplot(aes(.pred, .resid)) +
  geom_hline(yintercept = 0, lty = 2, color = "gray50", size = 1.2) +
  geom_point(size = 1.5, alpha = 0.3) +
  geom_smooth(color = "midnightblue") +
  labs(title = "Heterscedasticity in Regularized Regression")
```

This plot exhibits significant heteroscedasticity, with lower variance for recent artwork and higher variance for artwork in other time periods. If the model predicts a recent year, we can be more confident that it is right than if the model predicts an older year, and there is basically no time information in the fact that an artwork was created with a medium like oil on canvas. So is this model bad and not useful? It’s not great, for most goals I can think of, but it’s interesting to notice how much we can learn about our data even from such a model.


