Responses.csv
Some pages have items that ask for student input (e.g., multiple choice questions, coding exercises, written response questions). When a student encounters one of these and interacts with it, the interaction is recorded in the “responses.csv” file. This file provides information about the prompt and the student’s response. The response data are organized into a table with a variable number of columns (depending on lrn_option_
Below is a description of each variable in the responses table. You can also download the description of responses.csv (CSV, 3KB)
. Note that the responses.csv file will likely be very large (200 students will yield around 300,000 responses).
Column | Description |
---|---|
institution_id | a unique identifier for the institution the class was taken at |
class_id | a unique identifier for this particular class |
course_name | the GitHub repository for the course |
release | release number (version of the book, e.g., 4.2, 5.1, 5.3) |
book | book name |
branch | the Github branch for the course |
student_id | a unique identifier for each student on CourseKata |
lms_id | student id within LMS (given by LMS) |
item_id | a unique identifier for this particular question (the values are arbitrary — treat them as random identifiers) |
item_type | whether this is a learnosity or coding item |
chapter | the chapter that the item appears in |
page | the page that the item appears on |
response | the value of the response: either the value (for shorttext, plaintext, ratings, datacamp, etc.) or an array of numbers indicating the position of the multiple choice answers chosen and which correspond to the columns lrn_option_ |
prompt | the question prompt for this response |
points_possible | the number of points possible if the completely correct answer is given |
points_earned | the number of points earned |
dt_submitted | a datetime object indicating when the response was submitted (timezone: GMT/UTC) |
competes_page | whether this response completed the page. After completing that item, the student has no more items to complete on that page. Thus, if TRUE occurs for an earlier item, we can infer that the student skipped that item and then went back and did it last. |
attempt | the number of times the question has been attempted, including the current attempt |
user_agent | the browser user agent string for the user (see for details: https://developer.chrome.com/multidevice/user-agent; also, this R package for parsing: uaparserjs) |
lrn_session_id | the unique ID for this user session on Learnosity |
lrn_response_id | the unique ID for this particular response on Learnosity |
lrn_activity_reference | the unique ID for the activity on Learnosity |
lrn_question_reference | the unique ID for the question on Learnosity |
lrn_question_position | for multi-question items, the position of the question in the item |
lrn_type | the Learnosity type of the question (e.g. mcq, shorttext, plaintext, rating, datacamp, etc.) |
lrn_dt_started | a datetime object indicating when the responses was started (timezone: GMT/UTC) |
lrn_dt_saved | a datetime object indicating when the responses was saved (timezone: GMT/UTC) |
lrn_status | the status of the question response on Learnosity (e.g. “Completed”) |
lrn_response_json | the fully-detailed JSON response object (converted to a list object when the data are processed) |
lrn_option_ | for Learnosity items that have multiple answer options (like multiple-choice, Likert, etc.) these are the 0-indexed options the student could choose from. lrn_option_0 is the learnosity answer choice that comes first, lrn_option_1 is the learnosity answer option that comes second etc. |
For more about the distinction between Learnosity Activities, Items, and Questions, see the documentation.