Autograding interactive graphics and distributed programs in Submitty
Authors
Maicus, Evan
ORCID
Loading...
Other Contributors
Cutler, Barbara M.
Carothers, Christopher D.
Patterson, Stacy
Carothers, Christopher D.
Patterson, Stacy
Issue Date
2018-12
Keywords
Computer science
Degree
MS
Terms of Use
This electronic version is a licensed copy owned by Rensselaer Polytechnic Institute, Troy, NY. Copyright of original work retained by author.
Full Citation
Abstract
In recent years, Computer Science departments have been faced with rapidly increasing enrollments. As course sizes have ballooned to meet this demand, professors and staff have been met with a proportional increase in student assignment submissions, but not necessarily a proportional increase in teaching or grading staff. In an effort to speed the process of evaluation and ensure consistency in grading, many instructors have turned to automated grading systems. These systems run student code submissions against a battery of tests with known solutions. Students are given full or partial credit based on whether their code is able to match these known outputs. This system of evaluation works very well on assignments with text-based outputs that can be run on a single machine. However, there are very few off-the-shelf autograding solutions which assist instructors who run courses with more complex assignments.
Description
December 2018
School of Science
School of Science
Department
Dept. of Computer Science
Publisher
Rensselaer Polytechnic Institute, Troy, NY
Relationships
Rensselaer Theses and Dissertations Online Collection
Access
Restricted to current Rensselaer faculty, staff and students. Access inquiries may be directed to the Rensselaer Libraries.