# Import the pyplot and image modules from the matplotlib library
import matplotlib.pyplot as plt
# Read and display an image from file
'img/laser-cycle.png'))
plt.imshow(plt.imread('off') # Hide axes
plt.axis( plt.show()
A Coding Case Study with Quarto
LASER Orientation Module
0. INTRODUCTION
Welcome to your first LASER Case Study! The case study activities included in each module demonstrate how key Learning Analytics (LA) techniques featured in exemplary STEM education research studies can be implemented with R or Python. Case studies also provide a holistic setting to explore important foundational topics integral to Learning Analytics such as reproducible research, use of APIs, and ethical use of educational data.
This orientation case study will also introduce you to Quarto, which is heavily integrated into each LASER Module. You may have used Quarto before - or you may not have! Either is fine as this task will be designed with the assumption that you have not used Quarto before.
How to use this Quarto document
What you are working in now is an Quarto markdown file as indicated by the .qmd file name extension. Quarto documents are fully reproducible and use a productive notebook interface to combine formatted text and “chunks” of code to produce a range of static and dynamic output formats including: HTML, PDF, Word, HTML5 slides, Tufte-style handouts, books, dashboards, shiny applications, scientific articles, websites, and more.
Quarto docs can include specially formatted callout boxes like this one to draw special attention to provide notes, tips, cautions, warnings, and important information.
Pro tip: Quarto documents also have a handy Outline feature that allow you to easily navigate the entire document. If the outline is not currently visible, click the Outline button located on the right of the toolbar at the top of this document.
Source vs. Visual Editor
Following best practices for reproducible research (Gandrud 2021), Quarto files store information in plain text markdown syntax. You are currently viewing this Quarto document using the visual editor, The visual editor is set as the default view in the Quarto YAML header at the top of this document. Basically, a YAML header is:
a short blob of text that… not only dictates the final file format, but a style and feel for our final document.
The visual editor allows you to view formatted headers, text and code chunks and is a bit more “human readable” than markdown syntax but there will be many occasions where you will want to take a look at the plain text source code underlying this document. This can be viewed at any point by switching to source mode for editing. You can toggle back and forth between these two modes by clicking on Source and Visual in the editor toolbar.
You may have noticed a special kind of link in the text above. Specifically, a link citing Reproducible Research with R and R Studio by Chris Gandrud. The YAML header includes a bibliography option and points to our reference.bib
file in the lit
folder of this project, which produces a nice tooltip for linked references and a bibliography when our doc is rendered and published. Click the following link to learn more about citations in Quarto.
👉 Your Turn ⤵
LASER case studies include many interactive elements in which you are asked to perform an action, answer some questions, or write some code. These are indicated by the 👉 Your Turn ⤵ header. Now it’s your turn to do something.
Take a look at the markdown syntax used to create this document by viewing with the source editor. To do so, click the “Source” button in the toolbar at the top of this file. After you’ve had a look, click back to the visual editor to continue.
Great job! Let’s continue!
Code “Chunks”
In addition to including formatted text hyperlinks, and embedded images like above, Quarto documents can also include a specially formatted text box called a “code chunk.” These chunks allows you to run code from multiple languages including R, Python, and SQL. For example, the code chunk below is intended to run Python code as specified by “python” inside the curly brackets {}
. It also contains a contains some code “comments” as indicted by the # hashtags and several lines of python code. You may have also noticed a set of buttons in the upper right corner of the code chunk which are used to execute the code.
👉 Your Turn ⤵
Click the green arrow icon on the right side of the code chunk to run the Python code and view the image file name laser-cycle.png
stored in the img
folder in your files pane. Quarto will execute the code and its output and any related messages are displayed below the chunk.
Nice work! For this case study, don’t stress too much about understanding the code. We’ll spend a lot of time doing that in the other modules. For now, take a look at the image displayed and answer the question that follows by typing your response directly in this document.
âť“Question
In LASER case studies, you will often see as part of “Your Turns” a ❓ icon that indicates you are being promoted to answer a question. Type your response to the following question by deleting “YOUR RESPONSE HERE” and adding your own response:
What do you think this image is intended to illustrate?
- YOUR RESPONSE HERE
The Data-Intensive Research Workflow
The diagram shown above illustrates a Learning Analytics framework called the Data-Intensive Research workflow and comes from the excellent book, Learning Analytics Goes to School (Krumm, Means, and Bienkowski 2018). You can check that out later, but don’t feel any need to dive deep into it for now - we spend more time unpacking this framework in our Learning Analytics Workflow Modules; just know that this case study and all of the case studies in our LASER curriculum modules are organized around the five main components of this workflow.
In this introductory coding case study, we’ll focus on the following tasks specific to each component of the workflow:
- Prepare. Understand the research context, software packages, and data collected.
- Wrangle. Select and filter variables and “wrangle” them in a tabular (think spreadsheet!) format.
- Explore. Create some basic summary tables and plots to understand our data better.
- Model. Run a basic model - specifically, a simple regression model.
- Communicate. Create a reproducible report of your work that you can share with others.
Now, let’s get started!
1. PREPARE
First and foremost, data-intensive research involves defining and refining a research question and developing an understanding of where your data comes from (Krumm, Means, and Bienkowski 2018). This part of the process also involves setting up a reproducible research environment so your work can be understood and replicated by other researchers (Gandrud 2021). For now, we’ll focus on just a few parts of this process, diving in much more deeply into these components in later learning modules.
Research Question
In this case study, we’ll be working with data come from an unpublished research study by LASER team member, Josh Rosenberg, which utilized a number of different data sources to understand high school students’ motivation within the context of online courses.
These data sets and related research questions are explored in much greater detail in other modules, but for the purpose of this case study, our analysis will be driven by the following research question:
Is there a relationship between the time students spend on a course (as measured through their learning management system) and their final course grade?
Projects & Packages 📦
As highlighted in Chapter 6 of Data Science in Education Using R (Estrellado et al. 2020), one of the first steps of every research workflow should be to set up a “Project” within RStudio.
A Project is the home for all of the files, images, reports, and code that are used in any given project.
We are working in Posit Cloud with an R project cloned from GitHub, so a project has already been set up for you as indicated by the .Rproj
file in the main directory.
👉 Your Turn ⤵
Locate the Files tab lower right hand window pane and see if you can find the file named laser-orientation.Rproj
.
Since a project already set up for us, we will instead focus on loading the required packages we’ll need for analysis.
Packages, sometimes referred to as libraries, are shareable collections of R code that can contain functions, data, and/or documentation and extend the functionality of R.
pandas 📦
One package that we’ll be using extensively is {pandas}. Pandas (McKinney 2010) is a powerful and flexible open source data analysis and wrangling tool for Python that is used widely by the data science community.
NumPy 📦
NumPy is a fundamental package for scientific computing with Python and includes a collection of mathematical algorithms and convenience functions. NumPy offers comprehensive mathematical functions, random number generators, linear algebra routines, Fourier transforms, and more.
Pyplot 📦
Pyplot is a module in the {matplotlib) package, a comprehensive library for creating static, animated, and interactive visualizations in Python. pyplot
provides a MATLAB-like interface for making plots and is particularly suited for interactive plotting and simple cases of programmatic plot generation.
Statsmodels 📦
The statsmodels package provides classes and functions for the estimation of many different statistical models, as well as for conducting statistical tests, and statistical data exploration. An extensive list of result statistics are available for each estimator. The results are tested against existing statistical packages to ensure that they are correct.
scikit-learn 📦
The scikit-learn package is an open source machine learning library that supports supervised and unsupervised learning. It also provides various tools for model fitting, data preprocessing, model selection, model evaluation, and many other utilities.
👉 Your Turn ⤵
Click the arrow to execute the code in the cell below to load the required packages and functions for this case study.
import pandas as pd # for data wrangling
import numpy as np # for descriptive statistics
import matplotlib.pyplot as plt # for data visualization
import statsmodels.api as sm
from sklearn.linear_model import LinearRegression # for data modeling
Loading (or reading in) data
The data we’ll explore in this case study were originally collected for a research study, which utilized a number of different data sources to understand students’ course-related motivation. These courses were designed and taught by instructors through a state-wide online course provider designed to supplement – but not replace – students’ enrollment in their local school.
The data used in this case study has already been “wrangled” quite a bit, but the original datasets included:
A self-report survey assessing three aspects of students’ motivation
Log-trace data, such as data output from the learning management system (LMS)
Discussion board data
Academic achievement data
To know more, see Chapter 7 of Data Science in Education Using R (Estrellado et al. 2020).
👉 Your Turn ⤵
Next, we’ll load our data - specifically, a CSV (comma separated value) text file, the kind that you can export from Microsoft Excel or Google Sheets - into pandas, using the pd.read_csv()
function in the next chunk.
#read sci-online-classes.csv to sci_data and display the output
= pd.read_csv("data/sci-online-classes.csv") sci_data
Nice work! You should now see a new data “object” named sci_data
saved in your Environment pane. Try clicking on it and see what happens!
It’s important to note that by manipulating data with pandas we do not change the original file. Instead, the data is stored in memory and can be viewed in our Environment pane, and can later be exported and saved as a new file is desired.
Viewing and inspecting data
Now let’s learn another way to inspect our data.
👉 Your Turn ⤵
Run the next chunk and look at the results of the data frame you “assigned” to the sci_data
object in the previous code-chunk:
sci_data
student_id | course_id | total_points_possible | total_points_earned | percentage_earned | subject | semester | section | Gradebook_Item | Grade_Category | ... | q7 | q8 | q9 | q10 | TimeSpent | TimeSpent_hours | TimeSpent_std | int | pc | uv | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | 43146 | FrScA-S216-02 | 3280 | 2220 | 0.676829 | FrScA | S216 | 2 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 5.0 | 5.0 | 4.0 | 5.0 | 1555.1667 | 25.919445 | -0.180515 | 5.0 | 4.5 | 4.333333 |
1 | 44638 | OcnA-S116-01 | 3531 | 2672 | 0.756726 | OcnA | S116 | 1 | ATTEMPTED | NaN | ... | 4.0 | 5.0 | 4.0 | 4.0 | 1382.7001 | 23.045002 | -0.307803 | 4.2 | 3.5 | 4.000000 |
2 | 47448 | FrScA-S216-01 | 2870 | 1897 | 0.660976 | FrScA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 4.0 | 5.0 | 3.0 | 5.0 | 860.4335 | 14.340558 | -0.693260 | 5.0 | 4.0 | 3.666667 |
3 | 47979 | OcnA-S216-01 | 4562 | 3090 | 0.677335 | OcnA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 4.0 | 5.0 | 5.0 | 5.0 | 1598.6166 | 26.643610 | -0.148447 | 5.0 | 3.5 | 5.000000 |
4 | 48797 | PhysA-S116-01 | 2207 | 1910 | 0.865428 | PhysA | S116 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 4.0 | 4.0 | NaN | 3.0 | 1481.8000 | 24.696667 | -0.234663 | 3.8 | 3.5 | 3.500000 |
... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... |
598 | 97265 | PhysA-S216-01 | 3101 | 2078 | 0.670106 | PhysA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 4.0 | 4.0 | 4.0 | 4.0 | 817.4501 | 13.624168 | -0.724983 | 3.8 | 3.5 | 4.000000 |
599 | 97272 | OcnA-S216-01 | 2872 | 1733 | 0.603412 | OcnA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 3.0 | 5.0 | 5.0 | 3.0 | 1638.4500 | 27.307500 | -0.119048 | 4.4 | 3.0 | 5.000000 |
600 | 97374 | BioA-S216-01 | 8586 | 6978 | 0.812718 | BioA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | NaN | NaN | NaN | NaN | 470.8000 | 7.846667 | -0.980827 | NaN | NaN | NaN |
601 | 97386 | BioA-S216-01 | 2761 | 1937 | 0.701557 | BioA | S216 | 1 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 3.0 | 4.0 | 3.0 | 3.0 | 71.0166 | 1.183610 | -1.275885 | 3.8 | 3.0 | 3.666667 |
602 | 97441 | FrScA-S216-02 | 2607 | 2205 | 0.845800 | FrScA | S216 | 2 | POINTS EARNED & TOTAL COURSE POINTS | NaN | ... | 5.0 | 5.0 | 2.0 | 4.0 | 208.6664 | 3.477773 | -1.174293 | 4.4 | 4.0 | 2.000000 |
603 rows Ă— 30 columns
You can also enlarge this output by clicking the “Show in New Window” button located in the top right corner of the output.
âť“Question
What do you notice about this data set? What do you wonder? Add one or two observations in the space below:
- YOUR RESPONSE HERE
Data Types
Now, let’s examine our data a little more more systematically. The first step in getting to know your data is to discover the different data types it contains.
There are two general types of data:
Categorical data represent categories or groups that are distinct and separable. It usually consists of names, labels, or attributes and is represented by words or symbols.
Numerical data represents qualities that can be measured and represented as numbers.
👉 Your Turn ⤵
One way to explore the data types is by using info()
function. Complete the following code to take a look at the data types for each column in our sci_data
data frame.
Hint: Type the name of the function after the name of the dataset, using .
before and ()
after it.
sci_data.info()
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 603 entries, 0 to 602
Data columns (total 30 columns):
# Column Non-Null Count Dtype
--- ------ -------------- -----
0 student_id 603 non-null int64
1 course_id 603 non-null object
2 total_points_possible 603 non-null int64
3 total_points_earned 603 non-null int64
4 percentage_earned 603 non-null float64
5 subject 603 non-null object
6 semester 603 non-null object
7 section 603 non-null int64
8 Gradebook_Item 603 non-null object
9 Grade_Category 0 non-null float64
10 FinalGradeCEMS 573 non-null float64
11 Points_Possible 603 non-null int64
12 Points_Earned 511 non-null float64
13 Gender 603 non-null object
14 q1 480 non-null float64
15 q2 477 non-null float64
16 q3 480 non-null float64
17 q4 478 non-null float64
18 q5 476 non-null float64
19 q6 476 non-null float64
20 q7 474 non-null float64
21 q8 474 non-null float64
22 q9 474 non-null float64
23 q10 474 non-null float64
24 TimeSpent 598 non-null float64
25 TimeSpent_hours 598 non-null float64
26 TimeSpent_std 598 non-null float64
27 int 527 non-null float64
28 pc 528 non-null float64
29 uv 528 non-null float64
dtypes: float64(20), int64(5), object(5)
memory usage: 141.5+ KB
Nice work!!
âť“Question
Which of the columns in our dataset contain categorical and which contain numerical data? Name a few.
- YOUR RESPONSE HERE
Which data types do you see? Which ones are numerical and which are categorical? Name a few.
- YOUR RESPONSE HERE
If you look at “Grade_category”, you will notice that all values are NaNs or missing values which means we do not have any information about this variable (or parameter).
What other columns do you think have missing values? Why do you think so?
- YOUR RESPONSE HERE
2. WRANGLE
By wrangle, we refer to the process of cleaning and processing data, and, in some cases, merging (or joining) data from multiple sources. Often, this part of the process is very (surprisingly) time-intensive! Wrangling your data into shape can itself be an important accomplishment! And documenting your code using Python scripts or Quarto files will save yourself and others a great deal of time wrangling data in the future!
Selecting variables
Recall from our Prepare section that we are interested the relationship between the time students spend on a course and their final course grade. Let’s practice selecting these variables!
👉 Your Turn ⤵
Run the following code chunk using sci_data[[]]
and the names of the columns:
FinalGradeCEMS
(i.e., students’ final grades on a 0-100 point scale)TimeSpent
(i.e., the number of minutes they spent in the course’s learning management system)
'FinalGradeCEMS','TimeSpent']] sci_data[[
FinalGradeCEMS | TimeSpent | |
---|---|---|
0 | 93.453725 | 1555.1667 |
1 | 81.701843 | 1382.7001 |
2 | 88.487585 | 860.4335 |
3 | 81.852596 | 1598.6166 |
4 | 84.000000 | 1481.8000 |
... | ... | ... |
598 | 84.569444 | 817.4501 |
599 | 84.239532 | 1638.4500 |
600 | 12.352941 | 470.8000 |
601 | 54.158289 | 71.0166 |
602 | 23.137698 | 208.6664 |
603 rows Ă— 2 columns
Notice how the number of columns (variables) is now different!
It’s important to note that since we haven’t “assigned” this filtered data frame to a new object using the <-
assignment operator, the number of variables in the sci_data
data frame in our environment is still 30, not 2.
Let’s include one additional variable that you think might be a predictor of students’ final course grade or useful in addressing our research question.
First, we need to figure out what variables exist in our dataset (or be reminded of this - it’s very common in Python to be continually checking and inspecting your data)!
Recall that you can use a function named info()
to do this.
sci_data.info()
<class 'pandas.core.frame.DataFrame'>
RangeIndex: 603 entries, 0 to 602
Data columns (total 30 columns):
# Column Non-Null Count Dtype
--- ------ -------------- -----
0 student_id 603 non-null int64
1 course_id 603 non-null object
2 total_points_possible 603 non-null int64
3 total_points_earned 603 non-null int64
4 percentage_earned 603 non-null float64
5 subject 603 non-null object
6 semester 603 non-null object
7 section 603 non-null int64
8 Gradebook_Item 603 non-null object
9 Grade_Category 0 non-null float64
10 FinalGradeCEMS 573 non-null float64
11 Points_Possible 603 non-null int64
12 Points_Earned 511 non-null float64
13 Gender 603 non-null object
14 q1 480 non-null float64
15 q2 477 non-null float64
16 q3 480 non-null float64
17 q4 478 non-null float64
18 q5 476 non-null float64
19 q6 476 non-null float64
20 q7 474 non-null float64
21 q8 474 non-null float64
22 q9 474 non-null float64
23 q10 474 non-null float64
24 TimeSpent 598 non-null float64
25 TimeSpent_hours 598 non-null float64
26 TimeSpent_std 598 non-null float64
27 int 527 non-null float64
28 pc 528 non-null float64
29 uv 528 non-null float64
dtypes: float64(20), int64(5), object(5)
memory usage: 141.5+ KB
👉 Your Turn ⤵
In the code chunk below, add a new variable, being careful to type the new variable name as it appears in the data. We’ve added some code to get you started. Consider how the names of the other variables are separated as you think about how to add an additional variable to this code.
'FinalGradeCEMS','TimeSpent', 'total_points_earned']] sci_data[[
FinalGradeCEMS | TimeSpent | total_points_earned | |
---|---|---|---|
0 | 93.453725 | 1555.1667 | 2220 |
1 | 81.701843 | 1382.7001 | 2672 |
2 | 88.487585 | 860.4335 | 1897 |
3 | 81.852596 | 1598.6166 | 3090 |
4 | 84.000000 | 1481.8000 | 1910 |
... | ... | ... | ... |
598 | 84.569444 | 817.4501 | 2078 |
599 | 84.239532 | 1638.4500 | 1733 |
600 | 12.352941 | 470.8000 | 6978 |
601 | 54.158289 | 71.0166 | 1937 |
602 | 23.137698 | 208.6664 | 2205 |
603 rows Ă— 3 columns
Once added, the output should be different than in the code above - there should now be an additional variable included in the print-out.
Cleaning data
We have already seen that there are missing values in our target columns. There is another way to do that by selecting the column and using the isnull()
function and adding sum()
function in the end to find the number of those values.
Hint: you can use one function and then another function like data.function_1().function_2()
sum() sci_data.isnull().
student_id 0
course_id 0
total_points_possible 0
total_points_earned 0
percentage_earned 0
subject 0
semester 0
section 0
Gradebook_Item 0
Grade_Category 603
FinalGradeCEMS 30
Points_Possible 0
Points_Earned 92
Gender 0
q1 123
q2 126
q3 123
q4 125
q5 127
q6 127
q7 129
q8 129
q9 129
q10 129
TimeSpent 5
TimeSpent_hours 5
TimeSpent_std 5
int 76
pc 75
uv 75
dtype: int64
Handling Missing Values
There are several conventional ways to deal with the missing values.
We can drop those values and not use the entire row in which this element is missing.
We can substitute missing values with the column mean if the variance within a row is not very big.
👉 Your Turn ⤵
Below are the code chunks to execute both ways. Choose one you think is more appropriate as executing one excludes the use of the other and you will add changes to the dataset.
Drop missing values
The following code performs data cleaning and provides an overview of the cleaned dataset. Specifically, the dropna()
function removes any rows from the DataFrame sci_data
that contain missing values (NaN) in the columns FinalGradeCEMS
and TimeSpent
.
# Remove rows with missing values in 'FinalGradeCEMS' and 'TimeSpent' columns
= sci_data.dropna(subset=['FinalGradeCEMS', 'TimeSpent'])
sci_data
# Display a summary of the DataFrame
sci_data.info()
<class 'pandas.core.frame.DataFrame'>
Index: 573 entries, 0 to 602
Data columns (total 30 columns):
# Column Non-Null Count Dtype
--- ------ -------------- -----
0 student_id 573 non-null int64
1 course_id 573 non-null object
2 total_points_possible 573 non-null int64
3 total_points_earned 573 non-null int64
4 percentage_earned 573 non-null float64
5 subject 573 non-null object
6 semester 573 non-null object
7 section 573 non-null int64
8 Gradebook_Item 573 non-null object
9 Grade_Category 0 non-null float64
10 FinalGradeCEMS 573 non-null float64
11 Points_Possible 573 non-null int64
12 Points_Earned 486 non-null float64
13 Gender 573 non-null object
14 q1 465 non-null float64
15 q2 462 non-null float64
16 q3 465 non-null float64
17 q4 463 non-null float64
18 q5 461 non-null float64
19 q6 461 non-null float64
20 q7 459 non-null float64
21 q8 460 non-null float64
22 q9 459 non-null float64
23 q10 460 non-null float64
24 TimeSpent 573 non-null float64
25 TimeSpent_hours 573 non-null float64
26 TimeSpent_std 573 non-null float64
27 int 503 non-null float64
28 pc 504 non-null float64
29 uv 504 non-null float64
dtypes: float64(20), int64(5), object(5)
memory usage: 138.8+ KB
Substitute with column means
The following code snippet addresses missing values in the dataset by replacing them with the mean values of their respective columns. This approach ensures that the dataset remains comprehensive and suitable for analysis, while maintaining the overall statistical properties.
Calculating Mean Values: The
mean()
function calculates the mean (average) values for the FinalGradeCEMS and TimeSpent columns.Filling Missing Values: The
fillna()
function replaces any missing values (NaN) in these columns with their corresponding mean values.
# Calculate the mean value for 'FinalGradeCEMS' column
= sci_data['FinalGradeCEMS'].mean()
mean_value_grade
# Calculate the mean value for 'TimeSpent' column
= sci_data['TimeSpent'].mean()
mean_value_time
# Replace missing values in 'FinalGradeCEMS' column with the mean value
'FinalGradeCEMS'].fillna(value=mean_value_grade, inplace=True)
sci_data[
# Replace missing values in 'TimeSpent' column with the mean value
'TimeSpent'].fillna(value=mean_value_time, inplace=True) sci_data[
/var/folders/n_/y03lvw5130b2ct_8v03d6r840000gq/T/ipykernel_9109/92196522.py:8: FutureWarning: A value is trying to be set on a copy of a DataFrame or Series through chained assignment using an inplace method.
The behavior will change in pandas 3.0. This inplace method will never work because the intermediate object on which we are setting values always behaves as a copy.
For example, when doing 'df[col].method(value, inplace=True)', try using 'df.method({col: value}, inplace=True)' or df[col] = df[col].method(value) instead, to perform the operation inplace on the original object.
sci_data['FinalGradeCEMS'].fillna(value=mean_value_grade, inplace=True)
/var/folders/n_/y03lvw5130b2ct_8v03d6r840000gq/T/ipykernel_9109/92196522.py:8: SettingWithCopyWarning:
A value is trying to be set on a copy of a slice from a DataFrame
See the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy
sci_data['FinalGradeCEMS'].fillna(value=mean_value_grade, inplace=True)
/var/folders/n_/y03lvw5130b2ct_8v03d6r840000gq/T/ipykernel_9109/92196522.py:11: FutureWarning: A value is trying to be set on a copy of a DataFrame or Series through chained assignment using an inplace method.
The behavior will change in pandas 3.0. This inplace method will never work because the intermediate object on which we are setting values always behaves as a copy.
For example, when doing 'df[col].method(value, inplace=True)', try using 'df.method({col: value}, inplace=True)' or df[col] = df[col].method(value) instead, to perform the operation inplace on the original object.
sci_data['TimeSpent'].fillna(value=mean_value_time, inplace=True)
/var/folders/n_/y03lvw5130b2ct_8v03d6r840000gq/T/ipykernel_9109/92196522.py:11: SettingWithCopyWarning:
A value is trying to be set on a copy of a slice from a DataFrame
See the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy
sci_data['TimeSpent'].fillna(value=mean_value_time, inplace=True)
Filtering variables
Finally, let’s explore filtering variables by certain values using single brackets [ ]
.
👉 Your Turn ⤵
Check out and run the next chunk of code, imagining that we wish to filter our data to view only the rows associated with students who earned a final grade (as a percentage) of 70% - or higher and the TimeSpent
associated with it.
'TimeSpent'][sci_data['FinalGradeCEMS']>70] sci_data[
0 1555.1667
1 1382.7001
2 860.4335
3 1598.6166
4 1481.8000
...
592 244.3835
593 2264.4834
595 2676.7501
598 817.4501
599 1638.4500
Name: TimeSpent, Length: 438, dtype: float64
âť“Question
Roughly much time do you think you need to spend to get the grade higher than 70%? Is there a consistent pattern?
- YOUR RESPONSE HERE
3. EXPLORE
Exploratory data analysis, or exploring your data, involves processes of describing your data (such as by calculating the means and standard deviations of numeric variables, or counting the frequency of categorical variables) and, often, visualizing your data. As we’ll learn in later labs, the explore phase can also involve the process of “feature engineering,” or creating new variables within a dataset (Krumm, Means, and Bienkowski 2018).
In this section, we’ll quickly pull together some basic stats and introduce you to a basic data visualization.
Summary Statistics
Let’s repurpose what we learned from our wrangle section to select just a few variables and quickly gather some descriptive statistics to see where the data is centered, its values to identify trends by using describe()
function.
sci_data.describe()
student_id | total_points_possible | total_points_earned | percentage_earned | section | Grade_Category | FinalGradeCEMS | Points_Possible | Points_Earned | q1 | ... | q7 | q8 | q9 | q10 | TimeSpent | TimeSpent_hours | TimeSpent_std | int | pc | uv | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
count | 573.000000 | 573.000000 | 573.000000 | 573.000000 | 573.000000 | 0.0 | 573.000000 | 573.000000 | 486.000000 | 465.000000 | ... | 459.000000 | 460.000000 | 459.000000 | 460.000000 | 573.000000 | 573.000000 | 573.000000 | 503.000000 | 504.000000 | 504.000000 |
mean | 86231.424084 | 4303.863874 | 3267.818499 | 0.758350 | 1.364747 | NaN | 77.202655 | 78.235602 | 69.417905 | 4.290323 | ... | 3.910675 | 4.284783 | 3.492375 | 4.104348 | 1873.849317 | 31.230822 | 0.054687 | 4.215540 | 3.603009 | 3.730974 |
std | 10392.690030 | 2305.719003 | 1824.242324 | 0.089552 | 0.675195 | NaN | 22.225076 | 169.019143 | 145.754144 | 0.685135 | ... | 0.821833 | 0.688678 | 0.985399 | 0.937004 | 1335.608192 | 22.260137 | 0.985739 | 0.595033 | 0.644908 | 0.703211 |
min | 43146.000000 | 840.000000 | 672.000000 | 0.466447 | 1.000000 | NaN | 0.000000 | 5.000000 | 0.000000 | 1.000000 | ... | 1.000000 | 1.000000 | 1.000000 | 1.000000 | 0.700000 | 0.011667 | -1.327782 | 2.000000 | 1.500000 | 1.000000 |
25% | 85644.000000 | 2832.000000 | 2059.000000 | 0.704907 | 1.000000 | NaN | 71.251142 | 10.000000 | 7.600000 | 4.000000 | ... | 3.000000 | 4.000000 | 3.000000 | 4.000000 | 943.066900 | 15.717782 | -0.632272 | 3.900000 | 3.000000 | 3.333333 |
50% | 88340.000000 | 3641.000000 | 2777.000000 | 0.776543 | 1.000000 | NaN | 84.569444 | 10.000000 | 10.000000 | 4.000000 | ... | 4.000000 | 4.000000 | 4.000000 | 4.000000 | 1598.616600 | 26.643610 | -0.148447 | 4.200000 | 3.500000 | 3.791667 |
75% | 92733.000000 | 5083.000000 | 3882.000000 | 0.826181 | 2.000000 | NaN | 92.099323 | 30.000000 | 26.535000 | 5.000000 | ... | 4.500000 | 5.000000 | 4.000000 | 5.000000 | 2454.333200 | 40.905553 | 0.483111 | 4.700000 | 4.000000 | 4.166667 |
max | 97441.000000 | 15552.000000 | 12208.000000 | 0.910640 | 4.000000 | NaN | 100.000000 | 935.000000 | 828.200000 | 5.000000 | ... | 5.000000 | 5.000000 | 5.000000 | 5.000000 | 8870.883300 | 147.848055 | 5.218815 | 5.000000 | 5.000000 | 5.000000 |
8 rows Ă— 25 columns
👉 Your Turn ⤵
âť“ What do you notice about this dataset? Which columns are most important for our research question? What do you wonder?
- YOUR RESPONSE HERE
Data Visualization
Data visualization is an extremely common practice in Learning Analytics, especially in the use of data dashboards. Data visualization involves graphically representing one or more variables with the goal of discovering patterns in data. These patterns may help us to answer research questions or generate new questions about our data, to discover relationships between and among variables, and to create or select features for data modeling.
The Graphing Workflow
At it’s core, you can create some very simple but attractive graphs with just a couple lines of code. Matplotlib follows the common workflow for making graphs. To make a graph, you simply:
Start the graph with
plt
and include type of graph `.hist()’in our case, and add the data as an argument;“Add” elements to the graph using the
bins =
, or changing the color;Select variables to graph on each axis with the
xlabel()
argument.
Let’s give it a try by creating a simple histogram of our FinalGradeCEMS
variable. The code below creates a histogram, or a distribution of the values, in this case for students’ final grades. Go ahead and run it:
# Clear the previous figure to prevent overlapping in Quarto
plt.clf()
# Create a histogram for the 'FinalGradeCEMS' column in the sci_data DataFrame
'FinalGradeCEMS'], bins=30, color="skyblue")
plt.hist(sci_data[
# Label the x-axis as 'FinalGradeCEMS'
'FinalGradeCEMS')
plt.xlabel(
# Display the histogram
plt.show()
👉 Your Turn ⤵
Now use the code chunk below to visualize the distribution of another variable in the data, specifically TimeSpent
. You can do so by swapping out the variable FinalGradeCEMS
with our new variable. Also, change the color to one of your choosing; consider this list of valid color names here: https://matplotlib.org/stable/gallery/color/named_colors.html
Tip: There is no shame in copying and pasting code from above. Remember, reproducible research is also intended to help you save time!
# Clear the previous figure to prevent overlapping in Quarto
plt.clf()
# Create a histogram for the 'TimeSpent' column in the sci_data DataFrame
'TimeSpent'], bins=30, color="green")
plt.hist(sci_data[
# Label the x-axis as 'TimeSpent'
'TimeSpent')
plt.xlabel(
# Display the histogram
plt.show()
Scatterplots
Finally, let’s create a scatter plot for the relationship between these two variables. Scatterplots are most useful for displaying the relationship between two continuous variables. You can change type of graph by using the scatter()
function. You can also choose the size of the marker and color.
👉 Your Turn ⤵
Complete the code chunk below to create a simple scatterplot with TimeSpent
on the x axis and FinalGradeCEMS
on the y axis.
# Clear the previous figure to prevent overlapping in Quarto
plt.clf()
# Create a scatter plot with 'TimeSpent' on the x-axis and 'FinalGradeCEMS' on the y-axis
=sci_data['TimeSpent'], y=sci_data['FinalGradeCEMS'], marker=".", color='#88c999')
plt.scatter(x
# Label the x-axis as 'TimeSpent'
'TimeSpent')
plt.xlabel(
# Label the y-axis as 'FinalGradeCEMS'
'FinalGradeCEMS')
plt.ylabel(
# Display the scatter plot
plt.show()
Well done! As you can see, there appears to be a positive relationship between the time students spend in the online course and their final grade!
4. MODEL
“Model” is one of those terms that has many different meanings. For our purpose, we refer to the process of simplifying and summarizing our data as modeling. Thus, models can take many forms; calculating means represents a legitimate form of modeling data, as does estimating more complex models, including linear regressions, and models and algorithms associated with machine learning tasks. For now, we’ll run a base linear regression model to further examine the relationship between TimeSpent
and FinalGradeCEMS
.
We’ll dive much deeper into modeling in subsequent case studies, but for now let’s see if there is a statistically significant relationship between students’ final grades, FinaGradeCEMS
, and the TimeSpent
on the course.
An Inferential Model
An inferential statistical model is used to understand the relationships between variables and make inferences about the population from a sample. It aims to identify the underlying mechanisms and determine the significance of these relationships.
Key characteristics of an inferential model include:
Focus on Understanding: These models aim to understand the causal or correlational relationships between variables.
Hypothesis Testing: They often involve hypothesis testing to determine if observed patterns are statistically significant.
Parameter Estimates: Inferential models provide estimates of parameters (such as coefficients) that describe the relationship between variables, along with confidence intervals and p-values to assess their significance.
Generalization: The goal is to generalize findings from the sample to the larger population.
👉 Your Turn ⤵
Let’s use {statsmodels} package run a basic linear regression model to further examine the relationship between TimeSpent
and FinalGradeCEMS
:
# Define the independent variable (predictor) and the dependent variable (response)
= sci_data['TimeSpent']
X = sci_data['FinalGradeCEMS']
y
# Add a constant to the independent variable (this adds the intercept term to the model)
= sm.add_constant(X)
X
# Fit the linear regression model
= sm.OLS(y, X).fit()
model
# Print the summary of the model
print(model.summary())
OLS Regression Results
==============================================================================
Dep. Variable: FinalGradeCEMS R-squared: 0.134
Model: OLS Adj. R-squared: 0.132
Method: Least Squares F-statistic: 87.99
Date: Sun, 14 Jul 2024 Prob (F-statistic): 1.53e-19
Time: 13:16:58 Log-Likelihood: -2548.5
No. Observations: 573 AIC: 5101.
Df Residuals: 571 BIC: 5110.
Df Model: 1
Covariance Type: nonrobust
==============================================================================
coef std err t P>|t| [0.025 0.975]
------------------------------------------------------------------------------
const 65.8085 1.491 44.131 0.000 62.880 68.737
TimeSpent 0.0061 0.001 9.380 0.000 0.005 0.007
==============================================================================
Omnibus: 136.292 Durbin-Watson: 1.537
Prob(Omnibus): 0.000 Jarque-Bera (JB): 252.021
Skew: -1.381 Prob(JB): 1.88e-55
Kurtosis: 4.711 Cond. No. 3.97e+03
==============================================================================
Notes:
[1] Standard Errors assume that the covariance matrix of the errors is correctly specified.
[2] The condition number is large, 3.97e+03. This might indicate that there are
strong multicollinearity or other numerical problems.
It looks like TimeSpent
in the online course is indeed positively associated with a higher final grade! That is, students who spent more time in the LMS also earned higher grades. However, before we get too excited, there are likely other factors influencing student performance that are not captured by our model as study time only accounts for a small portion of why students get different grades. There are clearly other variables that likely play a role in determining final grades.
Don’t worry too much for now if interpreting model outputs is relatively new for you, or if it’s been a while. We’ll be working with models in greater depth in our other LASER modules.
A Predictive Model
A predictive model is designed to make accurate predictions about future or unseen data. The primary goal is to maximize the accuracy of predictions rather than understanding the underlying relationships.
Key characteristics of an inferential model include:
Focus on Prediction: These models aim to accurately predict the outcome variable for new data points.
Model Performance: The effectiveness of predictive models is evaluated based on metrics like R-squared, mean squared error, or other prediction accuracy measures.
Complex Algorithms: Predictive modeling often uses more complex algorithms (e.g., machine learning techniques) that might not be easily interpretable but offer higher predictive power.
Validation: Predictive models rely heavily on validation techniques like cross-validation to ensure that the model performs well on unseen data.
Before we can run out model, we’ll need to do some transformations of our two parameters. In our case it is linear regression. We encode the data from the two columns FinaGradeCEMS
and TimeSpent
into the format our scikit-learn linear regression model could understand. We then train (or fit) or model to the data we selected to find a line of best fit.
👉 Your Turn ⤵
Run the following code chunk to fit a linear regression model to predict FinalGradeCEMS
(final grades) based on TimeSpent
(time in LMS) and visualize the data points and the fitted regression line on a scatter plot:
#dependent variable is what you want to predict - y axis; independent(exploratory) variable
= np.array(sci_data['TimeSpent']).reshape(-1, 1)
X = np.array(sci_data['FinalGradeCEMS']).reshape(-1, 1)
y
= LinearRegression().fit(X, y)
reg =".", color = '#88c999')
plt.scatter(X, y, marker='hotpink')
plt.plot(X, reg.predict(X),color'TimeSpent')
plt.xlabel('FinalGradeCEMS')
plt.ylabel( plt.show()
Note that on the y-axis we put a dependent variable, the one we want to predict and x-axis is for and independent, or exploratory variable.
We can now use our model to predict a student’s grade depending on how much time (in minutes) they spent in the course:
# Create a new DataFrame with a column 'TimeSpent' in minutes
= pd.DataFrame({'TimeSpent': [1000, 1500, 2000]})
new_data
# Extracting the 'TimeSpent' values from 'new_data'
= np.array(new_data['TimeSpent']).reshape(-1, 1)
X_new
# Use the predict method to obtain predictions
= reg.predict(X_new)
predicted_grades
# Print the predicted grades
print(predicted_grades)
[[71.88911858]
[74.92942315]
[77.96972773]]
Hmm… Assuming our model is accurate (a big assumption!) even at 2000 minutes the best a student might hope for is a C+!
👉 Your Turn ⤵
Change the amount of time a student is spends in the course and see how the predicted grade changes.
print(reg.predict([[5000]]))
[[96.21155517]]
âť“Question
How much time would a student need to spend in the course to get an A?
- YOUR RESPONSE HERE
It’s important to note that these models are just used for illustrative examples of typical modeling approaches used in learning analytics and should be taken with a grain of salt. In our predictive model, for example, we haven’t attended to the accuracy of our model or how well it performs well on unseen data.
5. COMMUNICATE
The final step in the workflow/process is sharing the results of your analysis with wider audience. Krumm et al. Krumm, Means, and Bienkowski (2018) have outlined the following 3-step process for communicating with education stakeholders findings from an analysis:
Select. Communicating what one has learned involves selecting among those analyses that are most important and most useful to an intended audience, as well as selecting a form for displaying that information, such as a graph or table in static or interactive form, i.e. a “data product.”
Polish. After creating initial versions of data products, research teams often spend time refining or polishing them, by adding or editing titles, labels, and notations and by working with colors and shapes to highlight key points.
Narrate. Writing a narrative to accompany the data products involves, at a minimum, pairing a data product with its related research question, describing how best to interpret the data product, and explaining the ways in which the data product helps answer the research question and might be used to inform new analyses or a “change idea” for improving student learning.
In later modules, you will have an opportunity to create a “data product” designed to illustrate some insights gained from your analysis and ideally highlight an action step or change idea that can be used to improve learning or the contexts in which learning occurs.
For example, imagine that as part of a grant to improve student performance in online courses, your are working with the instructors who taught for this online course provider. As part of some early exploratory work to identify factors influencing performance, you are interested in sharing some of your findings about hours logged in the LMS and their final grades. One way we might communicate these findings is through a simple data dashboard like the one shown below. Dashboards are a very common reporting tool and their use, for better or worse, has become ubiquitous in the field of Learning Analytics.
Render Document
For now, we will wrap up this case study by converting your work to an HTML file that can be published and used to communicate your learning and demonstrate some of your new R skills. To do so, you will need to “render” your document. Rendering a document does two important things, namely when you render a document it:
checks through all your code for any errors; and,
creates a file in your project directory that you can use to share you work.
👉 Your Turn ⤵
Now that you’ve finished your first case study, let’s render this document by clicking the Render button in the toolbar at that the top of this file. Rendering will covert this Quarto document to a HTML web page as specified in our YAML header. Web pages are just one of the many publishing formats you can create with Quarto documents.
If the files rendered correctly, you should now see a new file named orientation-case-study-R.html
in the Files tab located in the bottom right corner of R Studio. If so, congratulations, you just completed the getting started activity! You’re now ready for the unit Case Studies that we will complete during the third week of each unit.
If you encounter errors when you try to render, first check the case study answer key located in the files pane and has the suggested code for the Your Turns. If you are still having difficulties, try copying and pasting the error into Google or ChatGPT to see if you can resolve the issue. Finally, contact your instructor to debug the code together if you’re still having issues.
Publish File
There are a wide variety of ways to publish documents, presentations, and websites created using Quarto. Since content rendered with Quarto uses standard formats (HTML, PDFs, MS Word, etc.) it can be published anywhere. Additionally, there is a quarto publish
command available for easy publishing to various popular services such as Quarto Pub, Posit Cloud, RPubs , GitHub Pages, or other services.
👉 Your Turn ⤵
Choose of of the following methods described below for publishing your completed case study.
Publishing with Quarto Pub
Quarto Pub is a free publishing service for content created with Quarto. Quarto Pub is ideal for blogs, course or project websites, books, reports, presentations, and personal hobby sites.
It’s important to note that all documents and sites published to Quarto Pub are publicly visible. You should only publish content you wish to share publicly.
To publish to Quarto Pub, you will use the quarto publish
command to publish content rendered on your local machine or via Posit Cloud.
Before attempting your first publish, be sure that you have created a free Quarto Pub account.
The quarto publish
command provides a very straightforward way to publish documents to Quarto Pub.
For example, here is the Terminal command to publish a generic Quarto document.qmd
to each of this service:
Terminal
quarto publish quarto-pub document.qmd
You can access your the terminal from directly Terminal Pane in the lower left corner as shown below:
The actual command you will enter into your terminal to publish your orientation case study is:
quarto publish quarto-pub orientation-case-study-R.qmd
When you publish to Quarto Pub using quarto publish
an access token is used to grant permission for publishing to your account. The first time you publish to Quarto Pub the Quarto CLI will automatically launch a browser to authorize one as shown below.
Terminal
$ quarto publish quarto-pub
? Authorize (Y/n) ›
❯ In order to publish to Quarto Pub you need to
authorize your account. Please be sure you are
logged into the correct Quarto Pub account in
your default web browser, then press Enter or
'Y' to authorize.
Authorization will launch your default web browser to confirm that you want to allow publishing from Quarto CLI. An access token will be generated and saved locally by the Quarto CLI.
Once you’ve authorized Quarto Pub and published your case study, it should take you immediately to the published document. See my example Orientation Case Study complete with answer key here: https://sbkellogg.quarto.pub/laser-orientation-case-study-key.
After you’ve published your first document, you can continue adding more documents, slides, books and even publish entire websites!
Publishing with R Pubs
An alternative, and perhaps the easiest way to quickly publish your file online is to publish directly from RStudio using Posit Cloud or RPubs. You can do so by clicking the “Publish” button located in the Viewer Pane after you render your document and as illustrated in the screenshot below.
Similar to Quarto Pub, be sure that you have created a free Posit Cloud or R Pub account before attempting your first publish. You may also need to add your Posit Cloud or R Pub account before being able to publish.
See below for an example of my published Orientation Case Study complete with answer key here using:
Your First LASER Badge!
Congratulations, you’ve completed your first case study!
Once you have shared a link to your published document with your instructor and they have reviewed your work, you will be provided a physical or digital version of the badge pictured below!