11  Importing data in R

R can import data from files in many different formats. For example:

In this chapter, you will learn how to import tabular data from an external text file as a tibble. We will use functions from the package readr which is part of the tidyverse package, so make sure that you have activated it with library(tidyverse).

Before we start

You must be familiar with tibbles and data types (such as numeric, character, etc).

You will need the following packages (in addition to those installed earlier)

  • here

You need to be working in an RStudio project (see Chapter 3).

Exercise

To import some data, you first need to download some.

We have made some files available which you can download with

usethis::use_zip("biostats-r/biostats-data")

When asked, agree to delete the zip file.

This will make a new directory called something like “biostats-r-biostats-data-9db4ebe/”. Rename this directory “data”. You can do this in the RStudio file tab.

11.1 What are tabular data?

Tabular data are data that is organized in the form of a table with rows and columns. A table often has a header, i.e. an additional row that displays variable names. Here is an example of a formatted table with header:

Table 11.1: A simple table with header row
month temperature (Celsius) precipitation (mm) Wind (m/s)
January -3.4 29.4 1.3
February -0.4 15.4 0.9
March -1.7 57.6 1.7
april 3.8 3.2 0.8
May 5.1 25.2 1.4
June 10.6 52.4 1.1
July 13.1 65.0 1.0
August 12.8 67.4 0.7
September 6.9 79.0 1.2
October 2.5 18.2 0.8
November -2.2 7.8 0.8
December 1.5 92.0 1.5

In this example, the table consists of 4 variables (columns) and 12 observations (rows) in addition to the header (top row). All cell contents are clearly delimited.

Below is an example of a non-tabular dataset. This dataset is a list of profiles with recurrent fields (Name, Position, Institution). Each profile contains three lines with a field:value pair.

Name: Aud Halbritter
Position: Researcher
Institution: UiB
-----------------------------
Name: Jonathan Soule
Position: Senior Engineer
Institution: UiB
-----------------------------
Name: Richard J. Telford
Position: Associate Professor
Institution: UiB
-----------------------------

11.2 About tidy data

Tidy data is a standard, consistent way to organize tabular data. Briefly, tidy data follows a short series of rules:

  • each variable in the data set is presented in a specific column,
  • each observation in the data set is presented in a specific row,
  • each cell at the intersection of a row and a column contains a single value.

The following figure illustrates these rules.

Schematic diagram showing that variables should be in columns, observations in rows and each cell should contain a single value

In a tidy dataset, variables are in columns, observations are in rows, and values are in cells. — Source: R for Data Science

The dataset presented in Table 11.1 respects all three rules, and is thus a tidy dataset.

On the contrary, the following dataset is not tidy:

Date 17.06.2016 21.06.2016 23.06.2016 26.06.2016 27.06.2016 28.06.2016 29.06.2016 30.06.2016 02.07.2016 04.07.2016 05.07.2016 06.07.2016 08.07.2016 10.07.2016 11.07.2016 13.07.2016 15.07.2016 17.07.2016 19.07.2016 21.07.2016 23.07.2016 24.07.2016 26.07.2016 28.07.2016 30.07.2016 02.08.2016 04.08.2016 06.08.2016 11.08.2016
Time 10:00 11:00 11:00 13:00 13:00 09:33 13:00 15:00 09:00 09:15 09:30 09:15 09:15 09:15 09:15 09:30 10:00 13:00 17:30 10 10:00 09:30 09:30 09:30 09:15 09:30 15:00 11:00 15:00
Weather Sunny_PartlyCloudy Cloudy_Fog Cloudy_Sunny Cloudy_Rainy_Windy_Fog Cloudy_Windy_Rainy Windy_Rainy_MostlyCloudy Windy_Sunny_Cloudy Windy_Cloudy Windy_Cloudy Cloudy_Sunny Cloudy_Sunny Cloudy_Fog_Rainy Sunny_PartlyCloudy Cloudy_Rainy Cloudy Cloudy_Sunny Cloudy Cloudy_Rainy Cloudy_Rainy Cloudy_Sunny Sunny_Cloudy Cloudy_Sunny Cloudy_Rainy Cloudy_Rainy Cloudy_Rainy Cloudy_Rainy Cloudy_Rainy Cloudy_Rainyn Cloudy_Snowy_Sunny
Observer LV_AH LV LV LV LV_SB LV_SB LV_SB LV LV_SB SB_LV SB_LV LV_SB LV_SB LV_SB LV LV LV_SB_AH SB_LV SB_LV SB_LV SB_LV LV LV LV LV LV LV LV LV
E01a 0 NA 0 0 0 0 1 1 2 2 3 3 6 8 8 11 13 10 9 7 3 6 3 3 1 1 1 2 1
E01b 0 NA 0 0 1 1 1 1 1 1 1 1 3 3 3 8 19 23 26 24 16 16 8 5 2 3 4 3 2
E01c 0 NA 0 0 0 0 0 0 0 0 0 0 0 0 1 5 9 12 10 7 4 3 3 3 2 2 0 1 1
E01d 0 NA 0 0 0 0 0 0 0 0 0 0 1 1 1 2 4 3 2 4 1 1 1 2 2 1 1 1 0
E01e 0 NA 0 0 0 0 0 0 0 0 0 0 0 0 0 1 4 4 4 5 4 3 1 0 0 0 0 0 0

Indeed, at least one of the rules is broken since columns display data matching several variables (Date, time, Weather, etc).

Importing data from a file containing tidy data is a great way to start your work, but it is not a prerequisite to data import. As long as your data is tabular, you will be able to import it in R, and tidy it later (see Chapter 13).

11.3 File formats

Tabular data may be stored in files using various formats, spreadsheets, etc. The most common spreadsheets store data in their own, proprietary file format, e.g. MS Excel which produces .xls and .xlsx files. Such formats may be a limitation to data management in R. Simpler formats such as plain text files with .txt or .csv should always be preferred when saving or exporting data from spreadsheets.

11.4 Delimited files

One of the most common format for storing tabular data is plain text files with a delimiter between the columns. The delimiter is often a comma or a semi-colon, but other delimiters are possible including tabs.

Comma or semi-colon delimited files are known as CSV files, which stands for Comma-Separated Values, and often have a .csv extension. They may also have a .txt extension.

In a CSV-formatted file, the data is stored in a similar manner to this:

Screenshot of a CSV file in notepad.

Figure 11.1: Contents of a CSV file viewed in the text editor Notepad

For information, this file matches the example in Table 11.1. Each line corresponds to a row in the table (including header) and cell contents are separated with a comma ,. Note that the period symbol . is used as decimal separator.

The use of commas in the CSV format is however not universal. Other symbols such as a semi-colon ; may be used as a delimiter. This is the case in several European countries where commas are decimal separator. Here is the same data set as above, this time in the European format:

Screenshot of a CSV file in notepad with columns separated by semi-colons

Contents of another CSV file

11.5 Know your data!

There are several reasons why different symbols are used in CSV files. Among these reasons are:

  • locale, i.e. the parameters that define the language and regional settings (currency, time and date, number format setting) in use on your machine,
  • software-based preferences, the settings which are defined by default by the software that you have used to punch your data,
  • user-based preferences, the settings that you choose when punching or saving data.

It is thus very important to know what your CSV file is made of. We therefore recommend to systematically inspect your data file before importing in R. One way to do it is to open the text file in a text editor (for example RStudio), or to read some of the file into R with read_lines('file.txt') (the n_max argument is useful if there is lots of data) and determine:

  • which symbol is used as decimal separator (, or .)
  • which symbol is used as delimiter (, or ;)
  • any extra lines of data that need removing

Here is our previous example:

read_lines('data/weather.csv')
 [1] "month,temperature (Celsius),precipitation (mm),Wind (m/s)"
 [2] "January,-3.4,29.4,1.3"                                    
 [3] "February,-0.4,15.4,0.9"                                   
 [4] "March,-1.7,57.6,1.7"                                      
 [5] "april,3.8,3.2,0.8"                                        
 [6] "May,5.1,25.2,1.4"                                         
 [7] "June,10.6,52.4,1.1"                                       
 [8] "July,13.1,65.0,1.0"                                       
 [9] "August,12.8,67.4,0.7"                                     
[10] "September,6.9,79.0,1.2"                                   
[11] "October,2.5,18.2,0.8"                                     
[12] "November,-2.2,7.8,0.8"                                    
[13] "December,1.5,92.0,1.5"                                    

In this file, the decimal separator is . and the delimiter is ,.

11.6 Importing delimited files

We can import delimited files such as csv files with readr::read_delim(). read_delim() will import the data as a tibble.

In the following example, we import the file weather.csv located in the subfolder data of the current RStudio project into the object weather.

library(tidyverse)
weather <- read_delim(file = "data/weather.csv")
Rows: 12 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (1): month
dbl (3): temperature (Celsius), precipitation (mm), Wind (m/s)

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

When the function is executed, the console shows a short series of messages. In the frame above, the message Column specification tells you that the content of the column month has been recognized as of data type character, and that the three other columns have been recognized as double (which is similar to numeric).

Now the tibble weather is available in R as you may see in the tab Environment (see Figure 11.2).

Screenshot of the environment tab showing one object "weather" with 12 obsevations and 4 variables.

Figure 11.2: The tibble weather is now stored in R.

You can display the table as follows:

weather
# A tibble: 12 × 4
   month     `temperature (Celsius)` `precipitation (mm)` `Wind (m/s)`
   <chr>                       <dbl>                <dbl>        <dbl>
 1 January                      -3.4                 29.4          1.3
 2 February                     -0.4                 15.4          0.9
 3 March                        -1.7                 57.6          1.7
 4 april                         3.8                  3.2          0.8
 5 May                           5.1                 25.2          1.4
 6 June                         10.6                 52.4          1.1
 7 July                         13.1                 65            1  
 8 August                       12.8                 67.4          0.7
 9 September                     6.9                 79            1.2
10 October                       2.5                 18.2          0.8
11 November                     -2.2                  7.8          0.8
12 December                      1.5                 92            1.5
Remember to assign!

If you don’t assign (<-) the imported data to a name, the tibble will be printed to the screen and then lost.

11.6.1 File paths

The first argument to read_delim() is file. This wants the location of the file on your computer. Before we can import a file, we have to find it!

You might be tempted to write the absolute path to the file, something like:

"c:/user/bad13/documents/project/my_file.csv"

This would be a bad idea. Why? Because this path only works on your current computer (provided you don’t move the file). This code won’t work for anyone else, so you cannot share the code with your supervisor and have them run it, nor archive it with a manuscript. The code is not reproducible.

Much better is to a path relative to your RStudio project (remember to use an RStudio project - for more info about them, read Chapter 3). If the contents of your RStudio project looked like this

.
├── R
│   ├── 01_import.R
│   └── 02_analyse.R
├── data
│   ├── 2022-3-24_lake-tilo-chemistry.csv
│   ├── 2022-3-24_lake-tilo-diatoms.csv
│   └── 2022-8-15_lake-tilo-diatoms.csv
├── my_project.Rproj
├── output
│   └── manuscript.pdf
└── quarto
    ├── manuscript.qmd
    └── references.bib

Now we could use

"data/2022-3-24_lake-tilo-chemistry.csv"

to access the chemistry data. If we move the RStudio project to a different location on your computer, or another computer, the code will still work. You don’t even need to type the full name: just type the quote marks and a few characters and autocomplete will start making suggestions.

There is still a problem though. When we use a quarto document, the working directory is the directory where the quarto document is stored. This can cause some confusion, which we can avoid if we use the here package to find the RStudio project root.

Our import code would now look like this

library("here")
weather <- read_delim(file = here("data/weather.csv"))

Using here is highly recommended.

Forward slashes and backslashes

Remember to use forward slashes / to separate directories in the path. If you use back slashes \, you will get an error as these are used to escape the following character (e.g., \t is a tab). If you really want back slashes, use two of them \\ as the first will escape the second. If you let autocomplete fill in your path, it will do it correctly

11.6.2 Delimiters

The second argument to read_delim() is the column delimiter delim.
It used to be very important to specify the delimiter (or choose a function such as read_csv() that had it set by default), but, after a recent update, read_delim() will guess what the delimiter is and we only need to specify it if it gets it wrong.

Exercise

Import the “weather.csv” data into R and assign it to an name.

11.6.3 Review the imported data

Once you have imported the data, it is important to inspect it to check that everything has worked as expected.

  • does it have the correct number of columns (may indicate problems with the delimiter)
  • are the columns the expected type
  • are the values correct.

To do this, we can print the object or View() it. We can also get details of the column specification

spec(weather)
cols(
  month = col_character(),
  `temperature (Celsius)` = col_double(),
  `precipitation (mm)` = col_double(),
  `Wind (m/s)` = col_double()
)

Any parsing problems (for example if there text in a column that should be numeric) will be shown with problems()

problems(weather)
# A tibble: 0 × 5
# ℹ 5 variables: row <int>, col <int>, expected <chr>, actual <chr>, file <chr>

Here, there are no parsing problems so we get a tibble with zero rows.

Let’s try another example. File weather2.csv has semi-colons as delimiters.

weather2 <- read_delim(file = "data/weather2.csv")
Rows: 12 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ";"
chr (2): month, Wind (m/s)
num (2): temperature (Celsius), precipitation (mm)

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.
spec(weather)
cols(
  month = col_character(),
  `temperature (Celsius)` = col_double(),
  `precipitation (mm)` = col_double(),
  `Wind (m/s)` = col_double()
)

Here, the message Column specification tells you that the content of the columns month and Wind (m/s) has been recognized as of data type character, and that the two other columns have been recognized as number. While read_delim() got things right about the number of variables, something went wrong with the variables as we could expect Wind (m/s) to be recognized as double. To find out about this issue we need to review the imported data in the object weather2 and compare it to the original file weather2.csv.

The example above shows the importance of carefully verifying that the imported data in R matches the original data set. The following figure compares the content of the file weather2.csv (Figure 11.3 left) to the content of the object weather2 (Figure 11.3 right):

Screenshots of the original text file with decimals separated by commas and the imported object in R, mostly without decimal separators

Figure 11.3: The data in the object does not match the data in the file.

Looking at the rows in both pictures, one can understand that all commas have simply been ignored, excepted those in the last column. To solve that issue, you must impose the separator using locale = locale(decimal_mark = ","):

weather3 <- read_delim(file = "data/weather2.csv", 
                       locale = locale(decimal_mark = ","))
Rows: 12 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ";"
chr (1): month
dbl (3): temperature (Celsius), precipitation (mm), Wind (m/s)

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

The last column is now recognized as double and the data in the object weather3 matches the data in the file:

The data in the object now matches the data in the file.
Decimal seperators

While readr can guess the column delimiter, it cannot guess the decimal separator. The English/American standard is to use “.” as the separator. In Norway it is common to use a comma as the decimal separator (but “.” are also common). If you don’t specify the decimal separator, readr assumes that the comma is a thousand separator and ignores it. Then the values can be 10x higher than expected.

11.6.4 read_csv2()

A shortcut for files with a “;” as the column delimiter and commas as the decimal separator is read_csv2().

weather_csv2 <- read_csv2(file = "data/weather2.csv")
ℹ Using "','" as decimal and "'.'" as grouping mark. Use `read_delim()` for more control.
Rows: 12 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ";"
chr (1): month
dbl (3): temperature (Celsius), precipitation (mm), Wind (m/s)

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

Note the message above Column specification that indicates which symbols read_csv2() has considered when importing the data. The resulting tibble looks correct:

weather_csv2
# A tibble: 12 × 4
   month     `temperature (Celsius)` `precipitation (mm)` `Wind (m/s)`
   <chr>                       <dbl>                <dbl>        <dbl>
 1 January                      -3.4                 29.4          1.3
 2 February                     -0.4                 15.4          0.9
 3 March                        -1.7                 57.6          1.7
 4 april                         3.8                  3.2          0.8
 5 May                           5.1                 25.2          1.4
 6 June                         10.6                 52.4          1.1
 7 July                         13.1                 65            1  
 8 August                       12.8                 67.4          0.7
 9 September                     6.9                 79            1.2
10 October                       2.5                 18.2          0.8
11 November                     -2.2                  7.8          0.8
12 December                      1.5                 92            1.5
Exercise

Import the file weather2.csv. Make sure it has imported correctly.

11.6.5 Useful arguments

read_delim() and the other functions discussed above have a multitude of arguments that allow for adjusting the way data is read and displayed. Here we review a handful of them.

11.6.5.1 n_max

n_max sets a limit to the number of observations to be read. Note that n_max does not consider the first row (header) of the data file as an observation.

weather <- read_delim(file = "data/weather.csv", 
                      n_max = 6)
weather
# A tibble: 6 × 4
  month    `temperature (Celsius)` `precipitation (mm)` `Wind (m/s)`
  <chr>                      <dbl>                <dbl>        <dbl>
1 January                     -3.4                 29.4          1.3
2 February                    -0.4                 15.4          0.9
3 March                       -1.7                 57.6          1.7
4 april                        3.8                  3.2          0.8
5 May                          5.1                 25.2          1.4
6 June                        10.6                 52.4          1.1

11.6.5.2 skip

The argument skip may be used to skip rows when reading the data file.
Be aware that:

  • the header in the original data file counts as one row,
  • the first row that comes after those which have been skipped becomes the header for the resulting tibble.

In the following example, we read the data file weather.csv and skip the first row:

weather <- read_delim(file = "data/weather.csv", 
                      skip = 1)
weather
# A tibble: 11 × 4
   January   `-3.4` `29.4` `1.3`
   <chr>      <dbl>  <dbl> <dbl>
 1 February    -0.4   15.4   0.9
 2 March       -1.7   57.6   1.7
 3 april        3.8    3.2   0.8
 4 May          5.1   25.2   1.4
 5 June        10.6   52.4   1.1
 6 July        13.1   65     1  
 7 August      12.8   67.4   0.7
 8 September    6.9   79     1.2
 9 October      2.5   18.2   0.8
10 November    -2.2    7.8   0.8
11 December     1.5   92     1.5

As expected, the first row (with the header) is skipped, and the data from the observation January have become the header of the tibble. The argument col_names() introduced below will be useful for solving this issue.

11.6.5.3 col_names

The argument col_names may be used to define the name of the variables/columns of the tibble.

col_names = may also be followed by the actual variable names that you want to use. In that case, write col_names = c() and list the names between the parentheses. Be aware that the first row of the original data file, which would have been the column names, will become the first row of the resulting tibble. This means that if you want to replace existing names, you should also use the skip argument.

Here is an example:

weather <- read_delim(file = "data/weather.csv", 
                      skip = 1,
                      col_names = c("Month", "Temp", "Precip", "Wind"))
weather
# A tibble: 12 × 4
   Month      Temp Precip  Wind
   <chr>     <dbl>  <dbl> <dbl>
 1 January    -3.4   29.4   1.3
 2 February   -0.4   15.4   0.9
 3 March      -1.7   57.6   1.7
 4 april       3.8    3.2   0.8
 5 May         5.1   25.2   1.4
 6 June       10.6   52.4   1.1
 7 July       13.1   65     1  
 8 August     12.8   67.4   0.7
 9 September   6.9   79     1.2
10 October     2.5   18.2   0.8
11 November   -2.2    7.8   0.8
12 December    1.5   92     1.5

If lots of column names are badly formatted, it may be easier to use janitor::clean_names() to improve them.

weather <- read_delim(file = "data/weather.csv") |> 
  janitor::clean_names()
weather
# A tibble: 12 × 4
   month     temperature_celsius precipitation_mm wind_m_s
   <chr>                   <dbl>            <dbl>    <dbl>
 1 January                  -3.4             29.4      1.3
 2 February                 -0.4             15.4      0.9
 3 March                    -1.7             57.6      1.7
 4 april                     3.8              3.2      0.8
 5 May                       5.1             25.2      1.4
 6 June                     10.6             52.4      1.1
 7 July                     13.1             65        1  
 8 August                   12.8             67.4      0.7
 9 September                 6.9             79        1.2
10 October                   2.5             18.2      0.8
11 November                 -2.2              7.8      0.8
12 December                  1.5             92        1.5

11.6.5.4 col_types

The argument col_types may be used to modify the data type of the variables. This is useful for example when you need to set a variable as factor whereas R has defined it as character. Here is an example with three simple variables.

counts <- read_delim(file = "data/groups.csv")
Rows: 9 Columns: 3
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (1): group
dbl (2): ID, count

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

As the message clearly indicates, the first and last variables are recognized as double (i.e. numeric) while the second one is recognized as character. The tibble in the object counts displays as follows:

counts
# A tibble: 9 × 3
     ID group count
  <dbl> <chr> <dbl>
1     1 A         5
2     2 A         3
3     3 A         2
4     4 B         4
5     5 B         5
6     6 B         8
7     7 C         1
8     8 C         2
9     9 C         9

Let’s use col_types = cols() to manually set the data types to double, factor and integer, respectively.

counts <- read_delim(file = "data/groups.csv",
                     col_types = cols(col_double(), col_factor(), col_integer()))

Now the tibble displays like this, with the correct data types for each column:

counts
# A tibble: 9 × 3
     ID group count
  <dbl> <fct> <int>
1     1 A         5
2     2 A         3
3     3 A         2
4     4 B         4
5     5 B         5
6     6 B         8
7     7 C         1
8     8 C         2
9     9 C         9

If you only need to modify one or a few variables, you must name it/them when writing the code:

counts <- read_delim(file = "data/groups.csv",
                     col_types = cols(group = col_factor()))

The tibble displays like this:

counts
# A tibble: 9 × 3
     ID group count
  <dbl> <fct> <dbl>
1     1 A         5
2     2 A         3
3     3 A         2
4     4 B         4
5     5 B         5
6     6 B         8
7     7 C         1
8     8 C         2
9     9 C         9

Note that col_types is particularly useful when importing tabular data that includes formatted dates. Dates are usually recognized as character when their format does not match the expected format set as default (locale, etc). In the following example, dates entered as yyyy-mm-dd in the last column are recognized as of type date:

countsdates <- read_delim(file = "data/groups-dates1.csv")
Rows: 9 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr  (1): group
dbl  (2): ID, count
date (1): date

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

However, in the example below, dates entered as dd-mm-yyyy are converted to character:

countsdates <- read_delim(file = "data/groups-dates2.csv")
Rows: 9 Columns: 4
── Column specification ────────────────────────────────────────────────────────
Delimiter: ","
chr (2): group, date
dbl (2): ID, count

ℹ Use `spec()` to retrieve the full column specification for this data.
ℹ Specify the column types or set `show_col_types = FALSE` to quiet this message.

To solve this issue, add the function col_date() in cols() to help R understand how dates are formatted and shall be read.

countsdates <- read_delim(file = "data/groups-dates2.csv", 
                     col_types = cols(date = col_date("%d-%m-%Y")))

The result is the following tibble now correctly formatted:

countsdates 
# A tibble: 9 × 4
     ID group count date      
  <dbl> <chr> <dbl> <date>    
1     1 A         5 2021-03-01
2     2 A         3 2021-03-02
3     3 A         2 2021-03-03
4     4 B         4 2021-03-01
5     5 B         5 2021-03-02
6     6 B         8 2021-03-03
7     7 C         1 2021-03-01
8     8 C         2 2021-03-02
9     9 C         9 2021-03-03

You will find the whole list of cols() parameters for col_types here.

Exercise

Import the data from “biomass2015_H.csv”. These data are plant biomass data from an alpine grassland near Mt Gonga, China.

Very large files

readr imports objects into the computers memory. This is fine for most datasets. But very large datasets (gigabytes) - will not fit into memory and another solution is needed. Here we can use apache arrow with the arrow package. This can work with dplyr functions to process large datasets efficiently.

11.6.6 White space delimited files

Some files have white space separating the columns. If a single tab is used then read_delim() will automatically detect it and work correctly. Sometimes multiple spaces are used so the columns line up, making the text file easy for humans to read. read_delim() will not work with these files as it expects a single delimiter. Instead we can use read_table(), shown here with some test data.

read_table(" Species   Value
            Navicula       1
            Cymbella       2")
# A tibble: 2 × 2
  Species  Value
  <chr>    <dbl>
1 Navicula     1
2 Cymbella     2

Another type of data you will sometimes see is the fixed width format. There are no delimiters, but each field has the same position in each row. These files can be imported with read_fwf(). Determining the position of each column can be painful unless you have good metadata.

Exercise

Import part “3a. Clegg et al. 2011 Hudson Lake Midge Temperature Data” from “alaska2011midge-t.txt”

Hint skip and n_max argument can be useful.

11.7 Excel files

Even though we don’t recommend you use Excel files for your own data, you need to know how to import such data as you may be given Excel files for find them in a data repository. To make your analysis reproducible, it is important to keep the files in their original format rather than converting them to CSV so that all the processing is done with code.

Excel files can be imported with the readxl package which is installed with tidyverse but not loaded until you run library(readxl). The function read_excel() will import the data; it can be worth using excel_sheets() first to check the names of the sheets.

library(readxl)
excel_sheets(path = "data/biomass2015.xls")
biomass_l <- read_excel(path = "data/biomass2015.xls", sheet = "Site L")

Many of the arguments for skipping rows at the start of the file, enforcing column types, etc, are similar to those in readr::read_delim()

Exercise

Import the biomass data from Site A from biomass2015.xls

Further Reading

Contributors

  • Jonathan Soulé
  • Aud Halbritter
  • Richard Telford