Responses.csv

Some pages have items that ask for student input (e.g., multiple choice questions, coding exercises, written response questions). When a student encounters one of these and interacts with it, the interaction is recorded in the “responses.csv” file. This file provides information about the prompt and the student’s response. The response data are organized into a table with a variable number of columns (depending on lrn_option_, see column description below) and a number of rows equivalent to the number of responses made to questions in the course.

Below is a description of each variable in the responses table. You can also download the description of responses.csv (CSV, 3KB). Note that the responses.csv file will likely be very large (200 students will yield around 300,000 responses).

Column Description
institution_id a unique identifier for the institution the class was taken at
class_id a unique identifier for this particular class
course_name the GitHub repository for the course
release release number (version of the book, e.g., 4.2, 5.1, 5.3)
book book name
branch the Github branch for the course
student_id a unique identifier for each student on CourseKata
lms_id student id within LMS (given by LMS)
item_id a unique identifier for this particular question (the values are arbitrary — treat them as random identifiers)
item_type whether this is a learnosity or coding item
chapter the chapter that the item appears in
page the page that the item appears on
response the value of the response: either the value (for shorttext, plaintext, ratings, datacamp, etc.) or an array of numbers indicating the position of the multiple choice answers chosen and which correspond to the columns lrn_option_.
prompt the question prompt for this response
points_possible the number of points possible if the completely correct answer is given
points_earned the number of points earned
dt_submitted a datetime object indicating when the response was submitted (timezone: GMT/UTC)
competes_page whether this response completed the page. After completing that item, the student has no more items to complete on that page. Thus, if TRUE occurs for an earlier item, we can infer that the student skipped that item and then went back and did it last.
attempt the number of times the question has been attempted, including the current attempt
user_agent the browser user agent string for the user (see for details: https://developer.chrome.com/multidevice/user-agent; also, this R package for parsing: uaparserjs)
lrn_session_id the unique ID for this user session on Learnosity
lrn_response_id the unique ID for this particular response on Learnosity
lrn_activity_reference the unique ID for the activity on Learnosity
lrn_question_reference the unique ID for the question on Learnosity
lrn_question_position for multi-question items, the position of the question in the item
lrn_type the Learnosity type of the question (e.g. mcq, shorttext, plaintext, rating, datacamp, etc.)
lrn_dt_started a datetime object indicating when the responses was started (timezone: GMT/UTC)
lrn_dt_saved a datetime object indicating when the responses was saved (timezone: GMT/UTC)
lrn_status the status of the question response on Learnosity (e.g. “Completed”)
lrn_response_json the fully-detailed JSON response object (converted to a list object when the data are processed)
lrn_option_ for Learnosity items that have multiple answer options (like multiple-choice, Likert, etc.) these are the 0-indexed options the student could choose from. lrn_option_0 is the learnosity answer choice that comes first, lrn_option_1 is the learnosity answer option that comes second etc.

For more about the distinction between Learnosity Activities, Items, and Questions, see the documentation.