410 likes | 826 Views
Reinventing CS Curriculum and Other Projects at The University of Nebraska Leen-Kiat Soh Computer Science and Engineering NCWIT Academic Alliance November Meeting 2007 Introduction Reinventing CS Curriculum Project Placement Exam Learning Objects Closed Labs CS1, CS2 Educational Research
E N D
Reinventing CS Curriculumand Other Projects atThe University of Nebraska Leen-Kiat Soh Computer Science and Engineering NCWIT Academic Alliance November Meeting 2007
Introduction • Reinventing CS Curriculum Project • Placement Exam • Learning Objects • Closed Labs CS1, CS2 • Educational Research • Computer-Aided Education • I-MINDS (Computer-Supported Collaborative Learning) • ILMDA (Intelligent Tutoring System) • Affinity Learning Authoring System
Placement Exam • The primary purpose of the placement test • Place students into one of CS0, CS1, and CS2 • Our approach emphasizes both pedagogical contexts and validation of the test • Placement exams we researched • Not used as a pre- and post-test • Do not explicitly consider pedagogical contexts such as Bloom’s taxonomy • Results not used to improve course instruction • No formative or summative analyses available Reinventing CS Curriculum
Placement Exam • 10 major content areas • based on ACM/IEEE Computing Curricula 2001 • Functions, sets, basic logic, data structures, problem solving, representation of data, etc. • addressed in the CS0 and CS1 courses • students’ knowledge are tested at multiple levels of competency based on Bloom’s Taxonomy • First five (25 questions) address prerequisite skills; second five (25 questions) represent the topics students are expected to know after completion of CS1 Reinventing CS Curriculum
Placement Exam Bloom’s Taxonomy 1. Knowledge/Memory 2. Comprehension 3. Application 4. Analysis 5. Synthesis 6. Evaluation Reinventing CS Curriculum
Placement Exam | Statistics • Degree of difficulty (mean) • The percentage of test takers who answer the question correctly • Too easy or too difficult – not a meaningful discriminator • Targeted mean for each question is between 0.40 and 0.85 • Item-total correlation • Shows the strength of the relationship between the students’ response to a question and their total score • A good question should have a strong positive correlation between the two • 0.3 is generally regarded as a good target, 0.2 is acceptable • Frequency of response for the choices • Unpicked choices are not providing any discrimination and should either be modified or dropped Reinventing CS Curriculum
Placement Exam | Reliability & Validity • Internal Consistency Reliability • A measure of item-to-item consistency of a student’s response within a single test • Cronbach’s alpha statistic [0 – 1] • Results show 0.70 to 0.74, which is acceptable for research purposes • Goal is to obtain 0.80 or higher • Content Validity • Determined by expert opinion by CSE faculty • Predictive Validity • Determined by correlating a student’s total score on the placement test with his/her exam scores in the course • E.g., 0.58 for Spring 2004 Reinventing CS Curriculum
Placement Exam | Implementation • Duration: 1 hour • 50 questions • 10 content areas • 5 questions in each area • Each question is classified into one of the Bloom’s competence level • Students are not informed of the competence levels • The presentation order is by the competence level within each content area • “knowledge” first, then “comprehension”, and so on. • Placement recommendation cutoffs • Greater than or equal to 10/25 CS1 • Greater than or equal to 35/50 CS2 • Otherwise CS0 Reinventing CS Curriculum
Placement Exam | Some Results • Pre-Post comparisons • T(63) = 11.036, p<.001; highly significant • Instructional effectiveness of the CS1 validated • Significant predictor of total test scores in CS1 • Test’s predictive validity * Spring 2004 session Reinventing CS Curriculum
Placement Exam | Some Results • Students who scored 48% or better vs. students who scored less • A one-way ANOVA found a significant difference between these two groups on total course points • F(1,64) = 4.76, p. < 0.5 • Students who scored higher on the placement test received a higher grade in the course • Pre-Post Test • Overall Test: T(68) = 11.81, p < 0.001 • Individual Bloom’s category: All show highly significant results (p < 0.001) • Greatest improvement on “knowledge” questions: t(68) = 8.27, p < 0.001) Reinventing CS Curriculum
Learning Objects • Development of web-based learning objects on “Simple Class” and “Recursion” • Small, stand-along “chunks” of instruction • SCORM – compliant (Shareable Content Object Reference Model) • Operating within Blackboard Course Management System • With extensive tracking for data collection Reinventing CS Curriculum
Learning Objects • Tutorial component Reinventing CS Curriculum
Learning Objects • Tutorial component Reinventing CS Curriculum
Learning Objects • Real-world examples component Reinventing CS Curriculum
Learning Objects • Practice exercises component Reinventing CS Curriculum
Learning Objects • Assessment component Reinventing CS Curriculum
Learning Objects • Self-paced, with learner control of additional practice • Extensive, elaborative feedback for remediation and instruction • Tracking System • Student outcomes and time-spent data captured in real time • Provides data on students’ problems and progress Reinventing CS Curriculum
Learning Objects | Some Results • No significant difference between lab and learning object instruction • Evaluation results showed positive student response to the learning objects • Modular, web-based learning objects can be used successfully for independent learning and are a viable option for distance learning Reinventing CS Curriculum
Closed Labs • Closed labs have multiple advantages • Active learning through goal-oriented problem solving • Promote students’ cognitive activities in comprehension and application • Some evidence that students test performance improves • Facilitates cooperative learning Reinventing CS Curriculum
Closed Labs | Design • Lectures • 2-hour laboratory (16 weeks) • 20 – 30 students per lab • Provide students with structured, hands-on activities • Intended to reinforce and supplement the material covered in the course lectures • Majority of the time allocated to student activities Reinventing CS Curriculum
Closed Labs | Design • A set of core topics are based on • Lecture topics • Modern software engineering practices • Computing Curricula 2001 • We developed 5 components for each laboratory • Pre-Tests • Laboratory Handouts • Activity Worksheets • Instructor Script • Post-Tests Reinventing CS Curriculum
Closed Labs | Design • Pre-Tests • Students are required to pass an on-line test prior to coming to lab • May take it multiple times • Passing score : 80% • Intended to encourage students to prepare for the lab and test their understanding of the lab objectives • Questions are categorized according to Bloom’s Taxonomy Reinventing CS Curriculum
Closed Labs | Design • Laboratory Handouts • Lab objectives • Activities students will perform in the lab (including the source code where appropriate), • Provide references to supplemental materials that should be studied prior to the lab • Additional materials that can be reviewed after the student has completed the lab Reinventing CS Curriculum
Closed Labs | Design • Activity Worksheets • Students are expected to answer a series of questions related to the specific lab activities • Record their answers on a worksheet (paper) • Questions provide the students with an opportunity to regulate their learning • Used to assess the student’s comprehension of the topics practiced in the lab Reinventing CS Curriculum
Closed Labs | Design • Instructor Script • The lab instructor is provided with an instructional script • Includes supplemental material that may not be covered during lecture, special instructions for the lab activities, hints, and resource links • Space for comments and suggestions Reinventing CS Curriculum
Closed Labs | Design • Post-Tests • During the last ten minutes of each lab, students take an on-line test • One-time-only • Another measure of their comprehension of lab topics • Questions are categorized according to Bloom’s Taxonomy Reinventing CS Curriculum
Closed Labs | Some Results • Study 1: To determine the most effective pedagogy for CS1 laboratory achievement • Participants: 68 students in CS1, Fall 2003 • Procedures • Structured cooperative groups had prescribed roles (driver and reviewers) • Unstructured cooperative groups did not have prescribed roles • Direct instruction students work individually • Randomly assigned the pedagogy of each lab section • Used stratified random assignment to assign students to their cooperative groups within each section • Based on ranking of the placement test scores for this course (high, middle, low) Reinventing CS Curriculum
Closed Labs | Some Results • Study 1, Cont’d … • Dependent Measures • Total laboratory grades • Combined worksheet scores and post-test grades for each lab • Although some students work in groups, all students were required to take the post-test individually • Pre-Post-Test measuring self-efficacy and motivation • Taken during the first and last week of the semester • Adapted 8 questions taken from Motivated Strategies for Learning Questionnaire by Pintrich and De Groot (1990) • Returned a reliability measure (Cronbach’s alpha) of .90 with a mean of 3.45 and standard deviation of .09; good reliability Reinventing CS Curriculum
Closed Labs |Some Results • Results of Study 1 • Both cooperative groups performed significantly better than the direct instruction group (F(2,66) = 6.325, p < .05) • Cooperative learning is more effective than direct instruction • No significant difference between the structured cooperative and unstructured cooperative groups • 6 out of 8 questions showed statistically significant changes in student perceived self-efficacy and motivation
Closed Labs |Some Results • Study 2 • Same objective; revised motivation/self-efficacy tool, additional qualitative feedback; revised laboratories • Participants: 66 students in CS1, Spring 2004 • Results • Both cooperative groups performed better than the direct instruction group (F(2,64) = 2.408, p < .05) • Discussion • Similar conclusions
Computer-Aided Education • Studies on the use of Computer-Supported Collaborative Learning (CSCL) tools • I-MINDS • structured cooperative learning (Jigsaw) vs. non-structured cooperative learning • CSCL vs. non-CSCL • Studies on the use of Intelligent Tutoring System (ITS) • ILMDA • ITS vs. Lab • ITS+Lab vs. Lab • Studies on the use of authoring tools • Affinity Learning Authoring System • How authoring tools impact learning • Graphical vs. non-graphical
Ongoing Work • Summer Institute with Center for Math, Science, and Computer Education • Teaching multimedia computing to student-teachers • NSF Advanced Learning Technologies Project • Intelligent Learning Object Guide (iLOG) • Developing SCORM-standard metadata to capture use characteristics of learning objects and student models • Developing software to automatically capture and generate metadata to tag learning objects • Creating SCORM-compliant learning objects for CS0, CS1, CS2
Ongoing Work 2 • Renaissance Computing • Joint curricular programs with other departments • School of Biological Sciences • School of Music • College of Agricultural Sciences and Natural Resources • Digital Humanities • Multi-flavored introductory CS courses • Object first vs. traditional • Multimedia, Engineering, Life Sciences, Arts
NCWIT Academic Alliance Focus • Renaissance Computing • Multi-flavored introductory CS courses in conjunction with joint curricular programs with other departments (that have larger female populations) to promote more female participation in CS • Computer-Aided Education • Online learning objects for K-12 teachers to help them expose their students to computational thinking and real-world IT applications • Collaborative writing (via I-MINDS) for secondary female students on the use of CS paradigms to solve real-world problems • Reinventing CS Curriculum • Use placement exam as pre- and post-tests for future studies on learning performance of female students • Use cooperative learning in labs to recruit and improve retention of female students
People • Rich Sincovec, CSE Department Chair • Reinventing CS Curriculum Project • Leen-Kiat Soh, Ashok Samal, Chuck Riedesel, Gwen Nugent • Computer-Aided Education • Leen-Kiat Soh, Hong Jiang, Dave Fowler, Art Zygielbaum
Others • UNL • College of Education and Human Sciences • Center for Math, Science, and Computer Education • J.D. Edwards Honors Program (CS+Business) • Extended Education and Outreach (AP Courses) • Department of History, School of Biological Sciences, School of Music, etc. • Bellevue University (I-MINDS) • University of Wisconsin-Madison ADL Co-Lab (learning objects)
Publications • Reinventing CS Curriculum • Framework • L.-K. Soh, A. Samal, and G. Nugent (2007). An Integrated Framework for Improved Computer Science Education: Strategies, Implementations, and Results, Computer Science Education, 17(1):59-83 • Learning Objects • G. Nugent, L.-K. Soh, and A. Samal (2006). Design, Development and Validation of Learning Objects, Journal of Educational Technology Systems, 34(3):271-281
Publications 2 • Reinventing CS Curriculum, Cont’d … • Placement Exam • G. Nugent, L.-K. Soh, A. Samal, and J. Lang (2006). A Placement Test for Computer Science: Design, Implementation, and Analysis, Computer Science Education, 16(1):19-36 • Structured Labs & Cooperative Learning • J. Lang , G. Nugent, A. Samal, and L.-K. Soh (2006). Implementing CS1 with Embedded Instructional Research Design in Laboratories, IEEE Transactions on Education, 49(1):157-165 • Soh, L.-K., G. Nugent, and A. Samal (2005). A Framework for CS1 Closed Laboratories, Journal of Educational Resources in Computing, 5(4):1-13
Publications 3 • Computer-Aided Education • Computer-Supported Collaborative Learning • L.-K. Soh, N. Khandaker, and H. Jiang (2007). I-MINDS: A Multiagent System for Intelligent Computer-Supported Cooperative Learning and Classroom Management, to appear in Int. Journal of Artificial Intelligence in Education • Intelligent Tutoring System • L.-K. Soh and T. Blank (2007). Integrating Case-Based Reasoning and Multistrategy Learning for a Self-Improving Intelligent Tutoring System, to appear in Int. Journal of Artificial Intelligence in Education • Affinity Learning Authoring Tool • L.-K. Soh, D. Fowler, and A. I. Zygielbaum (2007). The Impact of the Affinity Learning Authoring Tool on Student Learning, to appear in J. of Educational Technology Systems