Applicability of CMMI to the IS Curriculum: A Panel Discussion Bruce A. White1 School of Business, Quinnipiac University Hamden, Connecticut 06518, USA Herbert E. Longenecker, Jr.2 School of CIS, University of South Alabama Mobile, Alabama 36688, USA Paul M. Leidig3 John H. Reynolds4 CS & IS Department, Grand Valley State University Allendale, Michigan 49401, USA David M. Yarbrough5 Northrop-Grumman Information Technology, Northrop-Grumman, Inc. Pascagoula, Mississippi 39568, USA Abstract Information Systems (IS), as an academic discipline, has grown from its early days as data processing to where we are now, recognized as a true academic discipline, with graduates assuming positions as systems analysts, project managers, applications developers, web designers and more. Business cannot live without quality information. Quality information systems have become the lifeblood that sustains business; but, information systems projects became notorious for being late, over budget and not delivering what was originally promised or worse yet, having no idea about the true nature of the system requirements at all. A major step to remedying this was the promulgation of the Capability Maturity Model (CMM), which assessed how effectively software development groups were performing. Recently, Capability Maturity Model Integrated (CMMI) was advanced by Carnegie Mellon as the extension of the CMM concepts to organizations. This panel will explore the combining of academic discipline specifications of IS 2002 with the controlled and measured approaches of the CMMI to effectively monitor and deal with IS curriculum change and quality enhancement. Keywords: CMMI, IS model curriculum, assessment, quality Panel Overview The panel will address the following topics: 1. The proposition that CMMI can be used in higher education curriculum assessment. 2. The specific implementation issues surrounding the proposition. 3. The interpretation of the Key Process Area mapping to higher education. Intended Audience Educators who are interested in: * Improving the quality of IS education * Assessment of IS curriculum * Preparing for accreditation (regional or program) Panel Background and Rationale 1. Introduction The IS 2002 Model Curriculum (Gorgone, Davis, Valacich, Topi, Feinstein, and Longenecker 2002) represents the extension of IS’97. These Model Information Systems Curricula define a sequence of learning units consisting of goal and objective statements and specified elements of the body of information systems knowledge. IS2002 specifies a set of skills (Landry, Longenecker, Haigood, and Feinstein 2000) that are mapped to the learning units (www.IS2002.org, see reports). These mapped skill-to-learning unit pairs have become the basis for construction of an assessment and certification mechanism for graduating IS students (Landry, Reynolds, and Longenecker 2003; McKell, Reynolds, Longenecker, and Landry 2003; Reynolds, Longenecker, Landry, Pardue, and Applegate 2003). 2. CMM Background CMM was originally developed at the direction of the Secretary of Defense by the Software Engineering Institute (SEI) of Carnegie Mellon University in 1987. The initial CMM model was developed as a mechanism for maturation of software development firms. The idea was that these firms would pass through stages of maturation that reflected on their advancement. The current version of CMMI-SW has five levels (CMMI 2002) as shown in Table 1. Table 1 – Overview of CMM Levels CMM Level Description Level 1 – initial Characterized by individual effort, little control and monitoring and even has been described by some as chaos. Level 2 – repeatable The software development task has some established processes. For CMM-SW level 2, there are six KPAs – Key Process Areas – software configuration management, etc. etc. The emphasis is on establishing and following processes that are considered to be among the best of practices. Level 3 – defined This level introduces five more KPAs – including training – etc. The emphasis now shifts to using metric measurements to help the organization monitor and control the software development process. Level 4 – managed More KPA’s – more metrics, more making the defined processes to be managed processes. Level 5 – optimized At this level software projects are managed and controlled through metric feedback and control. 3. Synthesis In 2002, a new model IS undergraduate curriculum model was released. This model is a well-defined effort building on previous IS curriculum models – with a strong consistency between participants in the model building process. This model had the support and endorsement of three major IS professional organizations – Association for Information Systems (AIS), Association for Computing Machinery (ACM) and Association for Information Technology Professionals (AITP) and has 10 courses that cover the goals and objectives of an IS program. In 2003, a task force was organized to develop an assessment test tied to the IS 2002 model curriculum. IS 2002 contains a set of exit-level skills based on job ad analysis. The skills were summarized within the document text and further expanded on the website (www.is2002.org). These exit level skills are achieved within the curriculum and throughout the curriculum. Aspects of specific skills are contained within the learning units. There is a precise specification of the skill—learning unit mappings on the website. Exit objectives based on these skill—LU maps were determined by a task force meeting in February 2003. These objectives were used as the basis for developing assessment questions. This task force administered a beta test version of their assessment test in the spring of 2003 when over 500 students took the test. The test has been thoroughly analyzed for reliability and validity (Reynolds et al, 2003). Questions were carefully analyzed to determine the effective measurement of the learning objectives of the model curriculum. In June 2003, this task force reconvened to conduct further analysis of the assessment test and process. Using the metrics that were gathered, the task force identified questions that did not effectively measure the learning objectives. 4. Use of the IS 2002 Model by Universities It was intended within the design of IS’97 and IS 2002 that individual institutions could prove their relationship to the national model curriculum by mapping the learning units to their own courses. Such an accounting was done at the University of South Alabama (Daigle et al 2003). As a component of the IS Curriculum Development website (www.is2002.org), software is provided to accomplish this mapping. During the mapping process the interpretation by the university implementation of a learning unit in a given course is specified through a Local Objective (LO). Additional software is provided within the website to show how a sequenced thread of LOs culminates in a task force specified exit objective. Scoring on the task force questions relates to the effectiveness of the LO sequence and specific coverage. Deficiencies provide opportunity for improvement. 5. Proposition Curriculum models and reform can be discussed as a form of capability modeling. In the past institutions have inspected the model curricula and developed courses in an ad hoc manner. Table 2 shows how the CMMI process can be utilized to enable institutions to use the national model and certification exams to attain a desired level of CMMI performance. Also, while past model curriculum design activities may have been described as Level 1 efforts – heroic efforts of IS curriculum experts over the years – the IS 2002 model has become a very stable representation of the IS curriculum. With the assessment test metrics and with the metrics from the analysis of job ads and positions around the world, the CMMI process can be extended further as the basis for careful and meaningful modifications of the model curriculum in the future. 6. References Capability Maturity Model Integration Version 1.1, (2002). Pittsburgh: Carnegie Mellon University CMU/SEI-2002-TR-011. Daigle, R.J., H.E. Longenecker, Jr., J.P. Landry, and J.H. Pardue. (2003). “Using the IS 2002 Model Curriculum for Mapping an IS Curriculum.” Proceedings of ISECON 2003, November 6-9. Gorgone, J.T., G.B. Davis, J.S. Valacich, H. Topi, D.L. Feinstein, and H.E. Longenecker, Jr. (2002). IS 2002 Model Curriculum and Guidelines for Undergraduate Degree Programs in Information Systems. Atlanta: Association for Information Systems. Landry, J.P., H.E. Longenecker, Jr., B. Haigood, and D.L. Feinstein. (2000). “Comparing Entry-Level Skill Depths Across Information Systems Job Types: Perceptions of IS Faculty.” Proceedings of the 2000 Americas Conference on Information Systems, August 10-13. Landry, J.P., J.H. Reynolds, and H.E. Longenecker, Jr. (2003). “Assessing Readiness of IS Majors to Enter the Job Market: An IS competency Exam Based on the Model Curriculum.” Proceedings of the 2003 Americas Conference on Information Systems, August 4-6. McKell, L.J., J.H. Reynolds, H.E. Longenecker, Jr., and J.P. Landry. (2003). “Aligning ICCP Certification with the IS2002 Model Curriculum: A New International Standard.” Proceedings of the European Applied Business Research Conference, June 9-13. Reynolds, J.H., H.E. Longenecker, Jr., J.P. Landry, J. H. Pardue, and B. Applegate. (2003). “Information Systems National Assessment Update: The Results of a Beta Test of a New Information Systems Exit Exam Based on the IS 2002 Model Curriculum,” Proceedings of ISECON 2003, November 6-9. Table 2: Key Process Areas Applied to IS Curriculum Assessment6 Level Characteristic Key Process Area (KPA) Key Process Area Characteristic Interpretation for IS Curriculum Assessment Optimizing (5) Continuous process capability improvement Technology change Management Process Change Management Defect Prevention 1. New technology insertion into courses. 2. A comparative analysis of industry educational & skill requirements for applicants and the exit criteria for the current courses. 3. A comparative analysis of the postgraduate industry performance of a representative student population based on exit objective criterion analysis of Learning Units (LU) Managed (4) Product Quality Planning Tracking of measured IS courseware process IS Course Quality Management Quantitative Process Management 1. Insure that each developed course is consistent with the overall IS curriculum objectives and meets the objectives of the specific course outline. Provided by an analysis based on developing skill threads based on LU local objective development and thread analysis. 2. Maintain quantitative control within the course set. Provided by analysis of student performance in classes of instructors: comparative aggregation based on nation-wide analysis at LU level as aggregated to the industry standard. Defined (3) IS Course development process defined & institutionalized to provide product quality control Peer Reviews Inter-group Coordination Courseware Product Engineering Training Program Organizational Process Definition Organizational Process Focus 1. Intergroup reviews for courseware content & development. Provided by development of Curriculum Exit Objectives compatible with industry skill analysis; and, development of a question database with question validation for LU and exit criterion analysis for skills. 2. Intergroup reviews for courseware interaction between IS sub areas. Provided by development of Curriculum Exit Objectives compatible with industry skill analysis; and, development of a question database with question validation for LU and exit criterion analysis for skills. 3. Training Program for active participation in CMMI & courseware development. 4. Define courseware development process. Provided by the mapping of national LU’s to courses; and, definition of course responsibilities in skill production based on LU mapping and thread analysis. 5. Organization process focus within the IS department. Provided by the mapping of national LU’s to courses; definition of course responsibilities in skill production based on LU mapping and thread analysis. Repeatable (2) Management oversight & project tracking; Stable planning & product baselines Course Configuration Management Course Quality Management Course Project Tracking & Oversight Courseware Project Planning Requirements Management 1. Courseware Configuration Management processes that insure content and presentation. Provided by adjustment of the curriculum based on new skills, revised LU’s and/or thread analysis of local objectives. 2. Quality assurance that provides for outcome based expectation fulfillment. Provided by an analysis of student performance in classes of instructors: comparative aggregation based on nation-wide analysis at LU level as aggregated to the national standard. 3. Course Project Tracking & Oversight that provides the IS department appropriate project management. 4. Course Project Planning that provides for new and continuing courseware development 5. Requirement management, which provides assurance that developed courseware meets the desired objectives. Provided by an analysis of job ads and survey of academics and industry with confirmative factor analysis. Initial (1) Ad hoc “People” Level 1 is characterized by success achieved by ad hoc methods carried out by heroes and exceptionally hard workers. 1 bruce.white@quinnipiac.edu 2 hlongenecker@usouthal.edu 3 leidig@gvsu.edu 4 john.reynolds@csis.gvsu.edu 5 dyarbrough12@comcast.net 6 A process area is defined as a cluster of related practices in an area that, when performed collectively, satisfy a goal set which is considered important for making significant improvements in that area.