a measurement model of building information modelling maturity

24
A measurement model of building information modelling maturity Yunfeng Chen Building Construction Management, Purdue University, West Lafayette, Indiana, USA Hazar Dib Building Construction Management & Computer Graphics Technology, Purdue University, West Lafayette, Indiana, USA, and Robert F. Cox Building Construction Management, Purdue University, West Lafayette, Indiana, USA Abstract Purpose – There is a growing requirement for a rating system of building information modelling maturity (BIMM) to compare the effectiveness of modelling processes in construction projects. The literature related to BIMM contains theoretical proposals and description of their maturity models. However, the research efforts are limited and lacking substantial theoretical and empirical justifications. This paper is a unique attempt to integrate previous models by performing empirical investigations of key factors for measuring BIMM in construction projects. The paper aims to discuss these issues. Design/methodology/approach – A national survey was designed to extract the perception of 124 BIM-related practitioners and academicians about the conceptual model. Then, exploratory and confirmatory factor analyses were employed to identify and test the key factors underlying the 27 areas. Findings – A principal component factor analysis of the collected data had suggested a five-factor model, which explained 69.839 per cent of the variance. The construct validity of the model was further tested by confirmatory factor analysis. The results indicated that all factors were important in measuring BIMM; however, compared with the factors of technology and people, more emphasis was put on the factors of process and information. Originality/value – The key value of the paper is to increase the understanding of multi-dimension nature of BIMM through empirical evidence and to provide practitioners and researchers with the insight regarding particular emphasis on the factors related to modelling process and information. Keywords Quality management, Factor analysis, Process improvement, Maturity model, BIM, Construction management Paper type Research paper Introduction Building information modelling (BIM) has been increasingly adopted and implemented in the architectural, engineering and construction (AEC) industry since its inception in the 1970s, with many professionals expecting it to transform how this business operates (Eastman et al., 2011; Hardin, 2009; Succar, 2010; Wong et al., 2011). The current issue and full text archive of this journal is available at www.emeraldinsight.com/1471-4175.htm The authors would like to thank the ConstrucTech Magazine and the buildingSMART alliance for their support in distributing the questionnaire of this research. Construction Innovation Vol. 14 No. 2, 2014 pp. 186-209 q Emerald Group Publishing Limited 1471-4175 DOI 10.1108/CI-11-2012-0060 CI 14,2 186

Upload: robert-f

Post on 31-Dec-2016

231 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: A measurement model of building information modelling maturity

A measurement model of buildinginformation modelling maturity

Yunfeng ChenBuilding Construction Management, Purdue University,

West Lafayette, Indiana, USA

Hazar DibBuilding Construction Management & Computer Graphics Technology,

Purdue University, West Lafayette, Indiana, USA, and

Robert F. CoxBuilding Construction Management, Purdue University,

West Lafayette, Indiana, USA

Abstract

Purpose – There is a growing requirement for a rating system of building information modellingmaturity (BIMM) to compare the effectiveness of modelling processes in construction projects.The literature related to BIMM contains theoretical proposals and description of their maturity models.However, the research efforts are limited and lacking substantial theoretical and empiricaljustifications. This paper is a unique attempt to integrate previous models by performing empiricalinvestigations of key factors for measuring BIMM in construction projects. The paper aims to discussthese issues.

Design/methodology/approach – A national survey was designed to extract the perception of124 BIM-related practitioners and academicians about the conceptual model. Then, exploratory andconfirmatory factor analyses were employed to identify and test the key factors underlying the27 areas.

Findings – A principal component factor analysis of the collected data had suggested a five-factormodel, which explained 69.839 per cent of the variance. The construct validity of the model was furthertested by confirmatory factor analysis. The results indicated that all factors were important inmeasuring BIMM; however, compared with the factors of technology and people, more emphasis wasput on the factors of process and information.

Originality/value – The key value of the paper is to increase the understanding of multi-dimensionnature of BIMM through empirical evidence and to provide practitioners and researchers with theinsight regarding particular emphasis on the factors related to modelling process and information.

Keywords Quality management, Factor analysis, Process improvement, Maturity model, BIM,Construction management

Paper type Research paper

IntroductionBuilding information modelling (BIM) has been increasingly adopted and implemented inthe architectural, engineering and construction (AEC) industry since its inception in the1970s, with many professionals expecting it to transform how this businessoperates (Eastman et al., 2011; Hardin, 2009; Succar, 2010; Wong et al., 2011).

The current issue and full text archive of this journal is available at

www.emeraldinsight.com/1471-4175.htm

The authors would like to thank the ConstrucTech Magazine and the buildingSMART alliancefor their support in distributing the questionnaire of this research.

Construction InnovationVol. 14 No. 2, 2014pp. 186-209q Emerald Group Publishing Limited1471-4175DOI 10.1108/CI-11-2012-0060

CI14,2

186

Page 2: A measurement model of building information modelling maturity

The 2009 SmartMarket Report showed that there were 48 per cent of the industry claimedemploying BIM or BIM-related tools, which represented a 75 per cent growth in BIM usagecompared to 2007 (McGraw-Hill, 2009). Among those users, some may just employ a fewcopies of BIM-related software occasionally in projects that were dictated by ownersor contacts; some may facilitate BIM implementation in a small scale of projects byequipping staff and infrastructure; others might launch BIM on majority of their projectsand currently focus on developing and optimising internal and external collaborative BIMprocesses. It is obvious that the extent to which BIM is “explicitly defined, managed,integrated and optimised”, which refers to BIM maturity (BIMM), is different among theusers. So while everyone branded themselves as BIM-able, how can one tell the difference?A reliable and valid benchmark of BIMM is needed for meaningful evaluation andcomparison (Chen et al., 2012; NIBS, 2007; Smith and Tardif, 2009; Succar, 2010).

The maturity model specifies the key areas and characteristics a process must possessto qualify as a process of some maturity (Jalote, 2000; Paulk et al., 1995). It originated in thefield of quality management (Cooke-Davies, 2002). Its underlying premise is that thequality of a final product depends largely on the process used to create it (Paulk et al.,1995). It was expected that as the visibility into the process increases, the process“predictability”, process “control”, and business performance improve (Paulk et al.,1995, p. 27). Previous research also showed that the more mature any business process is,the better forecast, control and performance the business can have (Lockamy andMcCormack, 2004; McCormack and Johnson, 2001). Rooted in Walter Shewhart’s principleof statistical quality control in process improvement, maturity models proliferated in themanufacturing and software industries. One of the most well-known and adoptedmaturity models is the capability maturity model (CMM) proposed by the SoftwareEngineering Institute (SEI) (Vaidyanathan and Howell, 2007), which is a framework forsoftware organisations to evaluate and improve their software process (Paulk et al., 1995).Other maturity models extended from CMM include capability maturity modelintegration (CMMI), people-CMM (P-CMM), personal software process (PSP), teamsoftware process (TSP) and trillium (Alshawi, 2007). To adjust to the European market,a process model titled Bootstrap was developed by the European commission forthe software development process improvement soon after CMM (Alshawi, 2007). Theconcept of maturity model was also adopted in other industries, like theProject Management Process Maturity Model (PM)2 in project management industry,manufacturing supply chain maturity model (MSCMM) in the manufacturingindustry and lean enterprise transformation maturity model in the aerospace industry(Vaidyanathan and Howell, 2007). Inspired by the success of maturity model in otherindustries, researchers in the construction industry started to investigate the applicabilityof maturity models. One significant initiative was the Standardised Process Improvementfor Construction Enterprise (SPICE), which concluded that CMM in the software industrycannot be reused directly to the construction industry due to its incapability address thesupply chain issue (Sarshar et al., 2000). The construction supply chain maturity model(CSCMM) was then proposed for supply chain members to improve performance throughoperation excellence (Vaidyanathan and Howell, 2007). This model cannot be used formeasurement of BIMM, because it focuses on the aspect of multi-enterprise supply chainother than BIM. There were four maturity model proposed in the AEC industry thatclaimed the ability to measure BIM when this study was conducted, which was discussedin the section of BIMM.

Measurementmodel of BIMM

187

Page 3: A measurement model of building information modelling maturity

With the proliferation of BIM in the AEC industry, there is a growing awarenessamong the professionals about the need of a BIMM matrix. However, the establishmentof a BIMM matrix can be difficult even not impossible, because of the multi-dimensionalnature of BIM (Smith and Tardif, 2009). BIM is a concept functioning differently todifferent professions. Architects use BIM as “a process and technology” to model“the physical and functional characteristics” of a building (AIA, 2008, A295,Section 1.3.5). Contractors use BIM as a computer software model to improve decisionmaking and facility delivery process (AGC, 2006). Compared with other stakeholders,owners perceive “BIM as more of a collaborative process” (McGraw-Hill, 2009). After all,BIM “is more than a technology” (Ashcraft, 2008); it is a set of interacting process,human, information, and technology issues. According to the technical report by theCenter for Integrated Facility Engineering (CIFE, 2012), the impediments of BIMadoption were shifting from “technical issues such as contractual language andhardware and software to people issues such as training and availability of qualifiedstaff” (Gilligan and Kunz, 2007, p. 1). The adoption of BIM is more than the equipment ofstaff and technology infrastructure; it is a systematic approach to the lifecycleinformation related to a building (Smith and Tardif, 2009). According to SmartMarketReport, the focus of BIM investment changes over time; beginners rated BIM softwareand training as their highest priority investment, while experienced users rankedcollaborative BIM procedures and marketing as their top priority investment(McGraw-Hill, 2009). The employment of any IT solution cannot reach its fullpotential when focusing only on technology (Alshawi, 2007). Unlike other standalonesolution of information technology, the value of BIM can be maximised when it is used asa collaborative platform and process to facilitate communication and interdependenceamong stakeholders (Ashcraft, 2008). Although there have been some attempts topropose metrics for measuring BIM implementation, most studies account only for onedimension of BIM, and focus mainly on the final BIM model other than the process usedto create it. Moreover, because many research efforts lacked substantial theoretical andempirical justifications, the reliability and validity of the models remain questionable.This study was conducted to fill these gaps by integrating previous efforts throughperforming an empirical investigation of key factors for measuring BIMM. Suchresearch is necessary because the results may help non-users to orient their BIMinitiation, as well as current users to locate and improve their BIM implementation.It also offered empirical evidence to the multi-dimensional nature of BIM, and may alsoserve as a starting point for future researches to quantify BIM and its values, as well asthe development of metrics with lower-level items.

Research methodologyEvery project is different, which can result in different requirements for BIM and thereforeBIMM. However, it was believed by the authors through a post-positivism lens that thereshould be a common framework underlying BIMM. To identify the key areas formeasuring BIMM, a two-step quantitative exploratory and confirmatory factor analyticapproach was used, which is a typical methodology for the development of valid andreliable scale (Cudeck and Browne, 1983; Brown, 2006; Hair et al., 2009; Tabachnick andFidell, 2012). A non-experimental design with questionnaire survey was adopted here,because of the factorial analysis’s requirement for large amount of data, the limit of timeand money, as well as the difficulty to find experiment subjects and provide 27 treatments.

CI14,2

188

Page 4: A measurement model of building information modelling maturity

As shown in Figure 1, this study commenced with the development of 27 BIMM areasfor measuring BIMM by reviewing literature. The 27 BIMM areas, their correspondingoriginal areas, as well as their descriptions/definitions were presented in Table AI ofAppendix. The BIM-related academicians’ and practitioners’ perceived importance ofBIMM areas was then collected through a questionnaire survey based on the 27 areas.Qualified responses from 124 participants were received and screened for potentialproblems before formal analysis. Then the exploratory factor analysis (EFA) was usedto identify the key areas and their underlying factors, while confirmatory factor analysis(CFA) was employed to test the construct validity of the factorial structure.

BIM maturityAmong all the maturity models proposed in the AEC industry, only four modelsclaimed the ability to measure BIMM when this study was conducted. They includedthe CMM by the National Institute of Building Science (NIBS, 2007), the BIM

Figure 1.Research methodology

Literature review

Identification of BIMM areas

Assessment of content validity

Phase I:DevelopmentofMeasurementAreas

Design of questionnaire based on theidentified BIMM areas

Review of questionnaire

Development of web-based questionnaire

Phase II:Developmentof surveyinstrument

Exploratory factor analysis (EFA)

Confirmatory Factor Analysis (CFA)

Research conclusions

Phase IV:Dataanalysis andconclusions

Identification of target respondents

Data collection

Data screening

Phase III:Datacollectionandscreening

Measurementmodel of BIMM

189

Page 5: A measurement model of building information modelling maturity

deliverable matrix by the Alliance for Construction Excellence (ACE, 2008), the BIMproficiency matrix by Indiana University (IU, 2009), and the BIM competency sets bySuccar (2010). NIBS’s (2007) CMM was developed for the users to evaluate theirinternal business process and practices. There are two versions of NIBS’s CMM. Thefirst is the static tabular CMM consisting of eleven areas of interest againstten maturity level (NIBS, 2007). Based on the tabular CMM, the second is theinteractive CMM that offered output through interaction with the users (NIBS, 2007).The NIBS’s CMM is “a good first step toward establishing BIM implementationbenchmarks” (Smith and Tardif, 2009, p. 44); a minimum BIM is defined in terms of the11 areas and only the users meet the minimum requirements can claim themselves asBIM-able. CMM was validated by a test bed of BIM Award winners in 2007(McCuen et al., 2012) and was considered as the default standard for the measurementof information management (McGraw-Hill, 2008). Nevertheless, its inability inassessing any other metrics above information management has limited itsapplicability (Succar, 2010). Given that the success of a project team depends on theclear regulation of the project scope and deliverables for each stakeholder, ACE’s (2008)BIM deliverable matrix lists BIM services and deliverables at each phase of a typicalBIM project against three implementation levels, as well as the types of software usedby different stakeholders. This matrix concentrates on the description of the digitalmodels of each phase other than the process to create the models. IU’s (2009) BIMproficiency matrix is used to evaluate the BIM capability and proficiency of a firm, theresult of which serves as a selection criterion for a given project. The matrix contained32 credit areas under eight categories. There are few documents about the developmentprocess of BIM proficiency matrix and the rationale of the credit areas. Based on thelimited examples corresponding to each credit area in the enhanced BIM proficiencymatrix, it was perceived that “the matrix focuses on the accuracy and richness ofdigital model data and has less focus on the process of creating that digital model”(Succar, 2010, p. 102). Succar’s (2010) BIM competency set is intended to evaluate thecapability and maturity of a BIM player. It is a hierarchical set of competency areasunder four granularities. Competency areas with varied breadth and specificity can beused for the different goals of measurement, including discovery, evaluation,certification and auditing. The BIM competency set is comprehensive in covering thekey areas related to technology, process and policy. However, there is limiteddescription about the areas related to information management. What is more, there isredundancy of the areas. For example, the area of “networking solution” overlaps“network deliverables” (Succar, 2010, p. 72). Additionally, although the competency setwas claimed to be generated based on previous frameworks related to performance,excellence, and quality management, there is no explicit explanation or comparison ofthe competency areas with items exposed in previous frameworks. More importantly,the competency set was not validated empirically.

Although there have been some efforts toward BIMM, most research studied BIM asa one-dimensional concept or an end-product. More significantly, most studies werelimited to the theoretical proposal, and there is limited empirical research to test thereliability and validity of the models and frameworks. Key themes were identified andextracted based on a review of previous literature of maturity, quality, and performancemodels. It turned out that the areas included in the CMM of NIBS and Succar’s BIMcompetency set covered most of the key themes. Considering all above, 27 areas were

CI14,2

190

Page 6: A measurement model of building information modelling maturity

extracted and combined from the NIBS’s CMM and Succar’s BIM competency set toestablish the initial pool of areas for measuring BIMM. The content validity of the27 areas was confirmed by aligning with areas/items exposed in previous models ofmaturity, excellence, and quality management. The 27 BIMM areas of this research,their corresponding original areas in NIBS’s CMM and Succar’s BIM competency set andsimilar areas in previous research were listed in Table I of Chen et al. (2012). The 27 areaswere then evaluated by two BIM-related experts with academic and industry experienceas having validity to tap the BIMM. The content validity of the areas was furtherconfirmed by participants, since most of them commented that the questionnaire wascomprehensive and thorough in measuring BIMM.

Data collection and screeningData collectionAn on-line questionnaire survey was conducted for data collection. The questionnairewas designed to solicit the perception of BIM-related experts about the degree ofimportance of the 27 areas in measuring BIMM through a seven-point Likert scale.A sample questionnaire can be found in Dib et al. (2012). To ensure the relevance of theresponses, the questionnaires were only sent to the academicians conducting researchin BIM and practitioners with BIM-related experience. In order to safeguard thereliability of the received responses, both types of professionals were asked about theirBIM-related experience. If the academic respondent replied that s/he had notBIM-related experience of project or research, her/his response was not used for theanalysis. Similarly, if the practical respondent claimed to be involved in less thantwo BIM-implemented projects or had less than one-year direct working experience withBIM, her/his was discarded as well. The survey was conducted to a total of579 BIM-related experts in the USA. 131 of them responded to the survey,which represented a response rate of 22.63 per cent. Based on the criteria listed above,124 qualified responses were filtered out for further data analysis. The detailedinformation of the responses was shown in Table I. The industry sample represented awide spectrum of professional disciplines working as owner, architect or contractor.These profiles were listed in Tables II and III.

Data screeningBefore the data analysis, the raw data should be screened for potential problems,including missing data, departure from normality and collinearity. First, the data profileincluded 0.96 per cent missing values, which was then substituted with the mean ofthe corresponding areas. Second, considering that most responses lied between 4 and 7,the skewness index (SI) and kurtosis index (KI) were used to examine the normality

Questionnaires sentResponses received

(%)Valid responses

(%)Qualified responses

(%)

Academicians 206 52 (25.24%) 50 (24.27%) 49 (23.79%)Practitioners 373 92 (24.66%) 81 (21.72%) 75 (20.11%)Total 579 144 (24.87%) 131(22.63%) 124 (21.42%)

Source: Adapted from Chen et al. (2012)

Table I.Information about

response fromacademicians and

practitioners in the USA

Measurementmodel of BIMM

191

Page 7: A measurement model of building information modelling maturity

of the distribution of each area. Because the absolute value of SI and KI for each area wasless than the cut-off point of 3 and 10 (Kline, 2010), the normality of the distribution ofeach area was not rejected. Finally, in order to use factor analysis, the data matrix shouldhave sufficient correlations greater than 0.30 while less than 0.7 (Hair et al., 2009, p. 104).There are multiple ways with different “rules of thumb” to detect multi-collinearity,including the bivariate correlation (r), the squared multiple correlation (R2

sme ¼ 0:778),and the variance inflation factor (VIF). Commonly, an r above 0.7 for small sample or 0.85for larger sample (Berry and Feldman, 1985), a R2

sme above 0.9 (Kline, 2010), and a VIFmore than 4 or 10 (O’Brien, 2007) may indicate extreme collinearity. Based on therecommended approaches, it was found that there was collinearity between hardwareequipment (HE) (r ¼ 0.846, R2

sme ¼ 0:835, VIF ¼ 6.065) and hardware upgrade (HU)(r ¼ 0.846, R2

sme ¼ 0:778, VIF ¼ 4.505). Considering the nature of the two areas, HE andHU were combined to form a new area of hardware, the score of which was the average ofthe original two areas (Kline, 2010). The further examination of the data set revealed thatthere may be collinearity between the areas of standard operation procedure (SOP)(r ¼ 0.731, R2

sme ¼ 0:648, VIF ¼ 2.841) and documentation and modelling standard(DMS) (r ¼ 0.731, R2

sme ¼ 0:659, VIF ¼ 2.933). Although the related detection indices forthese two areas were not as extreme as the ones for HD and HU, it still posed considerablethreat. Besides, from the theoretical perspective, SOP and DMS aligned on thestandardisation of modelling and operation. To be more conservative, it was decided torun exploratory factor analyses, of the 26 areas with HE and HU combined intohardware and of the 25 areas with SOP and DMS combined into Standardisation,respectively.

Business types Amount (%)

General contractor/construction manager 21 (28.00%)Architect/engineer 19 (25.33%)Consultant 12 (16.00%)Owner/developer 6 (8.00%)Software vendor 6 (8.00%)Subcontractor 5 (6.67%)Others 6 (8.00%)Total 75 (100%)

Table II.Profile of industryrespondents(by business types)

Professions Amount (%)

Company manager (e.g. president/VP/CEP/owner) 18 (24.00%)Project manager (e.g. project/construction/product/program manager) 13 (17.33%)Model director (e.g. VDC/BIM/CAD director) 14 (18.66%)Model designer (e.g. designer/engineer/specialist/coordinator/consultant) 23 (30.67%)Model involver (e.g. estimator/account manager/sales manager) 3 (4.00%)Not specified 4 (5.33%)Total 75 (100%)

Table III.Profile of industryrespondents(by professions)

CI14,2

192

Page 8: A measurement model of building information modelling maturity

Data analysis and resultTo ensure the reliability and validity of a scale development, a two-step exploratory andconfirmatory factor analytic approach was typically applied (Cudeck and Browne, 1983;Brown, 2006; Hair et al., 2009; Tabachnick and Fidell, 2012). The same developmentprocess was used to develop measurement scale for knowledge management maturity(Chen and Fong, 2012), construction project success (Tabish and Jha, 2012) and climatesafety (Hon et al., 2013; Zhou et al., 2011). EFA was a data-driven approach used toidentify the key areas and their underlying factors, which was usually applied in theearly stage of scale development (Brown, 2006). CFA was typically used to test theconstruct validity of the identified factorial solution by evaluating how well itreproduces the sample correlation matrix (Brown, 2006). In this research, step I involvedthe development of BIMM framework based on EFA. Step II used CFA to test theconstruct validity of the first-level and second-level measurement models originatedempirically from EFA result at step I. Throughout this process, a five-factor second-levelmeasurement model was developed to satisfactorily fit the structure of BIMM.

Stage I: measurement model – EFAThe underlying constructs of the initial pool of areas for measuring BIMM wereexplored by the principal component factor analysis (PCFA) with a varimax rotation.Based on the discussion in the section of data screening, the 26 areas with HE and HUcombined was analysed separately from the 25 areas with the additional combinationof SOP and DMS. The two measurement models were then compared from theperspectives of quantitative indices and theory.

EFA of 26 areasThe Kaiser-Meyer-Olkin (KMO) measures of sampling adequacy of 0.857 and thesignificant Bartlett test of sphericity (1,537.952, p , 0.000) indicated that there weremeritoriously significant intercorrelation among the areas and the factor analysis wasappropriate (Hair et al., 2009). In order to obtain a simpler solution with factors that hadclearer interpretation, the areas with factor loadings less than 0.5 (Hair et al., 2009) orthe difference of cross-loadings less than 0.2 (Statwiki, 2012) were removed. Aftertwo-iteration of PCFA, nine inconsistent areas were removed. The acceptance of eacharea explained by the factor solution was confirmed by its correspondingcommunality, all of which were bigger than 0.510. The identified five-factor modelexplained 68.127 per cent of the variance. The structure of the model and its relatedstatistics were listed in Table IV. For the interpretation of the five factors, please alsorefer to the data analysis and interpretation of Chen et al. (2012).

EFA of the 25 areasThe results from the KMO measure of sampling adequacy (0.850) and the Bartlett testof sphericity (1,453.401, p , 0.000) indicated the appropriateness of factor analysis.Following the same criteria about convergent validity and discriminant validity, nineinconsistent areas were discarded and 16 significant areas were retained afterthree-iteration of PCFA. Table V presented the detail about the model structure andstatistical measures.

The five factors accounted for 69.839 per cent of the variance. Factor 1 (processdefinition and management (PDM)) explained 22.670 per cent of the variance and loaded

Measurementmodel of BIMM

193

Page 9: A measurement model of building information modelling maturity

on areas of standardisation, role, strategic planning, senior leadership, quality controland information accuracy (IAc). Factor 2 (13.627 per cent) concerned informationmanagement, including areas of lifecycle process, work flow, and geospatial capability.Factor 3 (11.488 per cent) loaded on items concerning training, involving trainingprogram and training delivery method. Factor 4 (11.083 per cent) included areas ofapplications, hardware and process and technology innovation (PTI), all of whichaligned on technology. Factor 5 (10.971 per cent) focused on information delivery,consisting areas of information delivery method and information assurance (IA).

Comparison of the two identified five-factor measurement modelsThe two models were quite close since the key structure of the measurement models stayedquite the same, as well as the statistical measures. This was understandable because theonly difference between the two initial pools lied in the combination of SOP and DMS.However, compared with the first model accounting for 68.127 per cent of the total variancewith 17 areas, the second model accounted for more variance (69.839 per cent) with fewer

BIMM areas Factor 1 Factor 2 Factor 3 Factor 4 Factor 5 Communalities

Doc. and modellingstandards 0.811 0.213 20.066 20.029 0.032 0.709Standard operatingprocess 0.809 0.159 0.050 20.084 0.095 0.699Role 0.738 0.229 0.067 0.090 0.008 0.610Quality control 0.733 0.113 0.304 0.193 20.054 0.682Strategic planning 0.684 0.098 0.145 0.358 0.063 0.630Senior leadership 0.681 20.084 0.211 0.297 0.213 0.649Work flow 0.102 0.807 0.128 0.128 0.060 0.698Lifecycle process 0.069 0.789 0.046 0.166 0.137 0.675Geospatial capability 0.265 0.683 0.215 20.004 0.045 0.586Real-time data 0.199 0.588 20.003 0.355 0.012 0.512Training program 0.160 0.177 0.848 0.037 0.088 0.786Training deliverymethod 0.151 0.112 0.841 20.067 0.178 0.779Information deliverymethod 0.119 0.214 0.056 0.836 0.072 0.767Information assurance 0.157 0.182 20.087 0.835 0.053 0.776Applications 20.047 20.010 0.137 0.093 0.821 0.704Hardware 0.118 0.035 0.425 0.072 0.702 0.694Process and tech.innovation 0.206 0.371 20.131 20.022 0.662 0.637Eigenvalue 5.369 1.913 1.865 1.321 1.115% variance 21.154 14.711 11.106 10.921 10.235Cronbach’s a 0.848KMO measure ofsampling adequacy 0.797Bartlett’s test ofsphericity: approx.x 2 (sig.) 822.620 (0.000)

Notes: Significant factor loadings are highlighted; the factors represent: 1 – PDM; 2 – informationmanagement; 3 – training; 4 – information delivery; 5 – technologySource: Adapted from Chen et al. (2012)

Table IV.Factor structure andvariance explained basedon responses from theexperts in the USA

CI14,2

194

Page 10: A measurement model of building information modelling maturity

areas (16 areas). In addition, from the theoretical perspective, SOP and DMS wereperceived to measure the different aspects of standardisation: SOP focused more on thestandardisation of process while DMS emphasised on final product. What is more,the author tested both models with CFA, the result of which turned out that the overallfitting of higher-level measurement model (x2

99 2 146:250, CFI ¼ 0.931, RMSEA ¼ 0.062)based on the 16-area structure was better than the one with further modification based onthe 17-area structure (x2

113 2 177:523, CFI ¼ 0.915, RMSEA ¼ 0.068). Considering allabove, the second model extracted from 25 areas was used for further CFA.

Stage II: measurement models: CFAThe result from EFA provided empirically derived hypotheses as to the number andnature of factors of BIMM. The hypothesised measurement model was then tested forits goodness-of-fit and validity using the first-order CFA. Then the common constructof BIMM underlying the five factors were further examined by second-order CFA.

First-order measurement modelBased on the result from EFA, the model to be tested hypothesised a priori that BIMM wasa five-factor structure composed of PDM, information management (Info Management),training, technology and information delivery (Info Delivery), as shown in Figure 2.

Factor 1 Factor 2 Factor 3 Factor 4 Factor 5 Communalities

Standardisation 0.764 0.200 20.002 0.042 20.057 0.629Quality control 0.762 0.123 0.278 20.069 0.138 0.697Role 0.755 0.252 0.029 20.023 20.008 0.636Strategic planning 0.755 0.101 0.094 0.054 0.268 0.664Senior leadership 0.750 20.078 0.157 0.201 0.180 0.667Info accuracy 0.666 0.078 0.134 0.244 0.300 0.617Lifecycle process 0.080 0.832 0.028 0.112 0.222 0.762Work flow 0.098 0.824 0.129 0.039 0.184 0.741Geospatial capability 0.287 0.642 0.218 0.050 20.022 0.545Training program 0.166 0.170 0.857 0.094 0.057 0.803Training deliverymethod 0.180 0.116 0.825 0.189 20.083 0.769Applications 20.050 20.021 0.153 0.820 0.110 0.710Hardware 0.149 0.048 0.407 0.703 0.050 0.687Process and tech.innovation 0.245 0.396 20.174 0.655 20.047 0.678Information assurance 0.203 0.138 20.081 0.047 0.848 0.788Info delivery method 0.174 0.186 0.054 0.064 0.842 0.782Eigenvalue 5.232 1.849 1.684 1.332 1.078% variance 22.67 13.627 11.488 11.083 10.971Cronbach’s a 0.845KMO measure ofsampling adequacy 0.812Bartlett’s test ofsphericity: approx. x 2

(sig.) 748.178 (0.000)

Notes: Significant factor loadings are highlighted; the factors represent: 1 – PDM;2 – information management; 3 – training; 4 – technology; 5 – information delivery

Table V.Factor structure and

variance explained basedon responses from the

experts in the USA

Measurementmodel of BIMM

195

Page 11: A measurement model of building information modelling maturity

The model and its components were specified as following:. Responses to the measuring instrument of BIMM can be explained by

five factors, PDM, Info Management, training, technology and Info Delivery.. The five factors are inter-correlated.. Each observed variable loads on only one factor.. Error terms associated with each observed variable were uncorrelated.

Different goodness-of-fitness (GOF) measured different facets of a model’s ability torepresent the data. Based on the recommendation from previous literature, the absolute fitof the normed x 2, the goodness-of-fit index (GFI) and root mean square error ofapproximation (RMSEA) were reported (Tabachnick and Fidell, 2012), as well as theincremental fit of comparative fit index (CFI) and the parsimony fit of adjusted goodness offit index (AGFI) (Hair et al., 2009). The satisfactory threshold for model fit can be achievedby attaining x 2/df , 2.00 (Tabachnick and Fidell, 2006), GFI . 0.85 (Maruish, 2004),RMSEA , 0.07 (Steiger, 2007), CFI . 0.95 (Hair et al., 2009) and AGFI . 0.8 (Maruish,2004). The fit statistics suggested good support for the model. Specifically, x 2/94 ¼ 1.361,

Figure 2.Hypothesised modelof factorial structurefor the BIMM

CI14,2

196

Page 12: A measurement model of building information modelling maturity

GFI ¼ 0.891, RMSEA ¼ 0.054, CFI ¼ 0.951 and AGFI ¼ 0.842. The standardised outputof the first-order measurement model was presented in Figure 3. The parameter estimatesfor all the factor loadings were significant and in the expected positive direction. Except theparameter estimate for the area of innovation, all loadings were higher than 0.5, whichsuggested that the areas were strongly related to their associated factors and indicated theconstruct validity (Hair et al., 2009). The modification indices (MIs) were examined todetect any possible misspecification (Byrne, 2009). Table VI presented the information ofthe biggest ten MIs. However, given the meaningless and trivialness of these MIs, as wellas the adequate fit of the existing model (Byrne, 2009), no additional parameters wereincluded in the model.

Second-order measurement modelTo facilitate theoretical understanding of multi-dimensional nature of BIMM, thefirst-order measurement model was specified into a second-order measurement model.The difference between the two specifications is that a structure is “imposed on thecorrelational pattern among the first-order factors” (Byrne, 2009, p. 143). In our casehere, BIMM was specified as the construct underlying the five first-order factors,as schematically shown in Figure 4.

Figure 3.Standardised output

of the first-order modelof factorial structure

for the BIMM

Measurementmodel of BIMM

197

Page 13: A measurement model of building information modelling maturity

The CFA model to be tested postulated priorities that:. Responses to the instrument of BIMM can be explained by five first-order factors

(PDM, Info Mgmt, training, technology and Info Delivery) and one second-orderfactor (BIMM).

Path M.I. Par change

Innovation ˆ lifecycle process 10.964 0.215Innovation ˆ Info Mgmt 9.948 0.284e14 $ Info Mgmt 9.526 0.236Geospatial ˆ quality control 7.596 0.239Innovation ˆ work flow 7.336 0.229e1 $ e3 7.107 0.159e9 $ PDM 5.921 0.123e4 $ e10 5.574 20.112e5 $ Info Mgmt 5.483 20.171Innovation ˆ info accuracy 5.474 0.171

Table VI.Ten largestmodification indexes

Figure 4.Hypothesisedsecond-order modelof factorial structurefor the BIMM

CI14,2

198

Page 14: A measurement model of building information modelling maturity

. Each observed variable loads on only one first-order factor.

. Error terms associated with each observed variable were uncorrelated.

. Co-variation among the five first-order factors is fully explained by theirregression on the second-order factor (BIMM).

The adequacy of the parameter estimates and the overall model suggested good supportto the validity of the BIMM measurement models (Byrne, 2009). Specifically, all thefactor loadings were significant and in the expected direction, as indicated by thestandardised output in Figure 5. The adequacy of the model as a whole was evidenced byx 2/99 ¼ 1.477, GFI ¼ 0.877, RMSEA ¼ 0.062, CFI ¼ 0.931 and AGFI ¼ 0.830. Bothabove support that the hypothesised model fitted the data quite well. The biggest tenMIs were listed in Table VII. In light of the nonsense and triviality of the MIs, as well asthe good fit of the model, it was concluded that the second-order model shown in Figure 4was the optimal representation for measuring BIMM for the US BIM-related experts.

The multi-dimensional hypothesis of the BIMM was not rejected by the test result ofparameter estimates. All the five factors were statistically significant, which furthersuggested that all factors were important in measuring BIMM. Among the five factors,

Figure 5.Standardised output

of the second-ordermodel of factorial

structure for the BIMM

Measurementmodel of BIMM

199

Page 15: A measurement model of building information modelling maturity

PDM had the highest standardised second-order factor loading of 0.71, followed by IMof 0.63, technology of 0.576, training of 0.574 and ID of 0.545. It indicated thatcompared with factors of people and technology, the factors of process and informationwere perceived to be more important in measuring BIMM.

DiscussionThe key findings of this research were further compared with other independent BIMand maturity related research.

Theoretically, the proposed maturity framework was considered ascomprehensive because the key factors and content of other BIM and maturityrelated research can be categorised under the four dimensions. Based on the nature ofthe five factors, the five factors of BIMM can be reduced to four dimensions oftechnology (Factor 4: technology), process (Factor 1: PDM), people (Factor 3: training),and information (Factor 2: information management; Factor 5: information delivery).The four dimensions of this research were validated due to their consistencywith the key content of global BIM-related research. As listed in Table AII ofAppendix, the four dimensions of this research covered the areas/content/indices ofBIM standards/guides/protocols/specifications in Australia, UK, China, Finland andSingapore (BCA, 2012). The factorial structure of BIMM was further confirmed bycomplying with the key factors of other maturity models, as listed in Table VIII.

Empirically, the research findings were validated by comparing them with those ofSmartMarket Report by McGraw-Hill, which was the “default gold standard” of BIMimplementation status in North America (Suermann, 2009). First, all the BIMM factorswere found to be important in measuring BIMM. This finding was confirmed giventhat the four dimensions of BIMM were corresponding to the four most importantfactors for increasing BIM benefits, as indicated in the SmartMarket Report byMcGraw-Hill. Specifically, the four factors included “Improved Interoperabilitybetween Software Applications”, “Improved BIM Software Functionality”, “MoreClearly-Defined BIM Deliverables between Parties” and “More Owners Asking forBIM” (2012, p. 22), which corresponded to the four dimensions of information,technology, process and people. Second, it was found that factors related to process andinformation were more important than factors of technology and people. This findingwas consistent with findings of SmartMarket Report, which concluded that the topobstacles to BIM improvement were related to information and process (2009).

Path M.I. Par change

Hardware ˆ training delivery 9.530 0.230Innovation ˆ lifecycle process 9.329 0.195e13 $ F3 8.602 0.226e14 $ F2 8.499 0.220Innovation ˆ Info Mgmt 7.855 0.251Innovation ˆ Info Mgmt 7.855 0.251e11 $ e13 7.504 0.233e11 $ F5 7.177 20.136Applications ˆ strategic planning 6.880 20.210Geospatial ˆ quality control 6.876 0.226e1 $ e3 6.531 0.151

Table VII.Ten largestmodification indexes

CI14,2

200

Page 16: A measurement model of building information modelling maturity

Last, the factorial structure of BIMM was generalisable due to the consistency ofBIMM measurement models based on the perception of USA and non-USA BIM-relatedprofessionals. This research is part of a study of BIMM measurement models based onperception and experience of global BIM-related professionals. Given that the samplesfrom different countries might be heterogeneous due to the inference of issues asculture, industry practice and regulation, infrastructure and BIM adoption status,separate measurement models were developed for the USA and non-USAprofessionals, respectively, (Dib et al., 2012). Similar factors related to information,process, people and technology were recognised. In addition, the factors of informationand process ranked higher than the factors related to technology and people for bothmodels. The key difference of the two measurement models lied in the emphasis ofsimilar factors. For example, for the factors related to people, the USA experts focusedon the training while the non-USA experts emphasised the regulation of teamstructure. This may suggest that the USA experts perceived BIM more as a technology,while the non-USA experts treated BIM more as a process. For more detaileddiscussion, please refer to Chen (2013).

Study\factor Technology Process People Information

Leanenterprise self-assessmenttool (MIT andUW, 2001)

“EnablingInfrastructure”:“must support andimplementation ofLean Principles,practices andbehaviour” (p. 11)

“Life-cycleProcesses”: refers tothe “Implement leanpractices across life-cycle processes” (p. 8)

“LeanTransformation/Leadership”: refers tothe development of“lean implementationplans” (p. 6)

Constructionsupply chainmaturitymodel(Vaidyanathanand Howell,2007)

“TechnologyAssessment”: refersto “the various tools”used to businessprocess efficiency”(p. 175)

“ProcessAssessment”: is “toidentify the currentas-is business processmethodology” (p. 175)

“Strategyassessment”: “is todetermine the currentbusiness strategy”with customersand suppliers (p. 176)“Value assessment”:

is to “assess thecurrent pain points”of “as-is businessprocess” (p. 176)

NIBS (2007) “InformationManagement”(p. 80)

BIM maturitymatrix(Succar, 2010,p. 69)

Technology: this field“clusters a group ofplayers whospecialise indeveloping software,hardware, equipmentand networkingsystems [. . .]”

Process: this field“clusters a group ofplayers who procure,design, construct,manufacture, use,manage and maintainstructures”

Policy: this field“clusters a groupof players focusedon preparingpractitioners,delivering research,distributing benefits,allocating risks andminimising conflicts[. . .]”

Table VIII.Comparison of BIMM

dimensions of thisresearch with other

maturity models

Measurementmodel of BIMM

201

Page 17: A measurement model of building information modelling maturity

ConclusionWith the proliferation of BIM adoption in the AEC industry, there is a need for a scientificrating system to benchmark the different levels of BIM implementation (NIBS, 2007).Based on a substantial theoretical and empirical justification, a valid and reliablemeasurement model of BIMM with five factors under four dimensions was proposed atthe end of this study. There were several theoretical and empirical implications from thisresearch. The key value of this research is to increase the understanding ofmulti-dimensional nature of BIMM through empirical evidence, and to providepractitioners and researchers with the insight regarding particular emphasis on thefactors related to process and information besides the factors of technology and people.In addition, the BIMM measurement model can be used by the industry to benchmarkBIMM across AEC projects, by the current 71 per cent BIM users of the USA AECprojects as a scientific checklist to evaluate their BIM implementation, as well as aguideline for the non-users of the remaining 29 per cent AEC projects to initiate their BIMefforts. What is more, different projects have different requirements for BIM. BIM can beengineered according to project stakeholders’ perceived priority of the BIMM areas tooptimise their investment. Case study can be conducted to examine the efficiency ofthe application of the proposed BIMM model to real-world projects. Finally, thequantification of BIM’s impact had been considered impossible due to themultidimensional nature of BIM and methodological difficulties. The BIMMmeasurement model here can serve as an initial point for quantifying the impact ofBIM on its intangible benefits like trust and collaboration or tangible benefits such asproductivity and construction performance. Specifically, hypothesised relationshipbetween BIMM and its benefits can be established based their respective measurementmatrixes, which can be further tested by data collected from surveys or projects.

The BIMM measurement was considered as generalisable from three perspectives.First, the key dimensions and areas of BIMM measurement models were consistentwith other maturity models and BIM-related standards. Second, the four dimensionsand their ranking in this research aligned with the four most important factorsimpacting BIMM benefits as indicated in the SmartMarket Report by McGraw-Hill.Third, similar factors were perceived as important by both the USA and the non-USABIM-related professionals.

Some limitations of this research need to be acknowledged. One issue is about howwell the sample represented the population of the USA BIM users. There was a widerange of respondents in terms of their experience, education, position and demography.The BIMM measurement model might differ with different composition of participantswith distinct experience and background. Therefore, the measurement model extractedhere might not reflect the perception of the population. A natural extension of this line ofresearch would be to consider the differences in the perception of BIMM across theprofessionals with different backgrounds, such as the difference between the group ofarchitects and the group of constructors or the difference between practitioners withdifferent levels of industry experience. Another limitation is that the findings from thisstudy may not be generalised directly to other specific countries; however, as discussedbefore, similar factors were identified by both the USA and the non-USA experts.Therefore, the research findings here can be adapted for use in other countries andregions. Future studies of the perception of the experts in other countries and globalexperts could also be conducted and compared with the findings in this research.

CI14,2

202

Page 18: A measurement model of building information modelling maturity

References

ACE (2008), Building Information Modeling: An Introduction and Best Methods Approach,Alliance for Construction Excellence, available at: www.ace4aec.com/pdf/Introduction%20and%20Best%20Methods%20Manual%202008.pdf (accessed 29 July 2012).

AEC (UK) Committee (2012), AEC (UK ) BIM Protocol, available at: http://aecuk.files.wordpress.com/2012/09/aecukbimprotocol-v2-0.pdf (accessed 15 August 2013).

AGC (2006), The Contractors’ Guide to BIM, The Associated General Contractors of America,available at: www.ipd-ca.net/images/Integrated%20Project%20Delivery%20Definition.pdf (accessed 8 August 2011).

AIA (2008), AIA Document E202-2008, The American Institute of Architects, available at: www.fm.virginia.edu/fpc/ContractAdmin/ProfSvcs/BIMAIASample.pdf (accessed 25 July 2012).

Alshawi, M. (2007), Rethinking IT in Construction and Engineering: Organisational Readiness,Taylor & Francis, New York, NY.

Ashcraft, H.W. (2008), “Building information modeling: a framework for collaboration”,Construction Lawyer, Vol. 28 No. 3, pp. 1-14.

BCA (2012), Singapore BIM Guide, Building and Construction Authority, available at: http://bimsg.files.wordpress.com/2012/05/singapore-bim-guide-version-1.pdf (accessed 5 August2013).

Berry, W.D. and Feldman, S. (1985), Multiple Regression in Practice, Sage, Newbury Park, CA.

Blanchard, P.N. and Thacker, J. (2006), Effective Training – Systems, Strategies, and Practice,Prentice-Hall, Upper Saddle River, NJ.

Brown, T.A. (2006), Confirmatory Factor Analysis for Applied Research, 1st ed., The GuilfordPress, New York, NY.

BuildingSmart Finland (2012), Common BIM Requirements, available at: www.en.buildingsmart.kotisivukone.com/3 (accessed 10 August 2013).

Businessdictionary (2011), Reward System, available at: www.businessdictionary.com/definition/reward-system.html (accessed 24 December 2011).

Byrne, B.M. (2009), Structural Equation Modeling with AMOS, Taylor & Francis, New York, NY.

Chase, R.B. (2007), Operation Management in Manufacturing and Services, available at:www.ateneonline.it/chase2e/areastudenti.asp#tn (accessed 25 September 2011).

Chen, Y. (2013), “Measurement models of building information modeling maturity”,PhD dissertation, Purdue University, West Lafayette, IN, 2 August.

Chen, Y., Dib, H. and Cox, R.F. (2012), “A framework for measuring building informationmodeling maturity in construction projects”, paper presented at the 14th InternationalConference on Computing in Civil and Building Engineering, Moscow, Russia, 27-29 June.

Chen, L. and Fong, P.S.W. (2012), “Revealing performance heterogeneity through knowledgemanagement maturity evaluation: a capability-based approach”, Expert Systems withApplications, Vol. 39, pp. 13523-13539.

CIFE (2012), CIFE VDC Scorecard and BIM Scorecard, Center for Integrated FacilityEngineering, available at: http://cifevdc.weebly.com/index.html (accessed 28 May 2012).

Cooke-Davies, T. (2002), “Project management maturity models – does it make sense to adoptone?”, Project Manager Today, Vol. 14 No. 5, pp. 16-19.

Cudeck, R. and Browne, M.W. (1983), “Cross-validation of covariance structures”, MultivariateBehavioral Research, Vol. 18 No. 2, pp. 147-167.

Measurementmodel of BIMM

203

Page 19: A measurement model of building information modelling maturity

Dib, H., Chen, Y. and Cox, R. (2012), “A framework for measuring building information modelingmaturity based on perception of practitioners and academics outside the USA”, paperpresented at CIB W78 2012 Conference, Beirut, Lebanon, 17-19 October.

DOT (2012), Specifications, Department of Transportation, available at: www.state.nj.us/transportation/eng/specs/ (access 24 June 2012).

Eastman, C., Teicholz, P., Sacks, R. and Liston, K. (2011), BIM Handbook: A Guide to BuildingInformation Modeling for Owners, Managers, Designers, Engineers, and Contractors,Wiley, Hoboken, NJ.

Encyclopedia (2011), Real-Time Data, available at: http://encyclopedia.thefreedictionary.com/Real-timeþdata#cite_note-0 (accessed 28 December 2011).

Gattol, M. (2012), Hardware and Related Stuff, available at: www.markus-gattol.name/ws/hardware.html (accessed 7 January 2012).

Gilligan, B. and Kunz, J. (2007), VDC Use in 2007: Significant Value, Dramatic Growth, andApparent Business Opportunity, Technical Report, Center for Integrated FacilityEngineering, Stanford University, Stanford, CA, 4 December.

Hair, J.F., Black, W.C., Babin, B.J. and Anderson, R.E. (2009), “Multivariate data analysis”,in Heine, J. (Ed.), Multiple Regression Analysis, Prentice-Hall, Upper Saddle River, NJ,pp. 155-234.

Hardin, B. (2009), BIM and Construction Management: Proven Tools, Methods, and Workflows,Wiley, Indianapolis, IN.

HKI (2011), BIM Project Specification, Hong Kong Institute of Building Information Modeling,available at: www.hkibim.org/?page_id¼1386 (accessed 10 August 2013).

Hon, C.K.H., Chan, A.P.C. and Yam, M.C.H. (2013), “Determining safety climate factors in therepair, maintenance, minor alteration, and addition sector of Hong Kong”, Journal ofConstruction Engineering and Management, Vol. 139, pp. 519-528.

IEEE (1991), IEEE Standard Computer Dictionary: A Compilation of IEEE Standard ComputerGlossaries – 610, Institute of Electrical and Electronics Engineers (IEEE), New York, NY.

ITGI (2007), COBIT 4.1, Governance Institute, available at: www.isaca.org/Knowledge-Center/cobit/Pages/Downloads.aspx (accessed 19 May 2011).

IU (2009), Indiana University BIM Standard, Indiana University, available at: www.indiana.edu/ , uao/iubim.html (accessed 15 May 2011).

Ivanov, K. (1972), “Quality-control of information: on the concept of accuracy of informationdata-banks and in management information systems”, PhD dissertation, The RoyalInstitute of Technology KTH, Stockholm, 11 December.

Jalote, P. (2000), CMM in Practice: Processes for Executing Software Projects at Infosys,Addison-Wesley Professional, Boston, MA.

Kapur, P.K., Tandon, A. and Kaur, G. (2010), “Multi up-gradation software reliability model”,The 2nd International Conference on Reliability, Safety and Hazard, Institute of Electricaland Electronics Engineers (IEEE), Mumbai, India, 14-16 December, pp. 468-474.

Kline, R.B. (2010), Principles and Practice of Structural Equation Modeling, The Guilford Press,New York, NY.

Lockamy, A. and McCormack, K. (2004), “The development of a supply chain managementprocess maturity model using the concepts of business process orientation”, Supply ChainManagement: An International Journal, Vol. 9 No. 4, pp. 272-278.

McCormack, K.P. and Johnson, W.C. (2001), Business Process Orientation: Gaining the E-BusinessCompetitive Advantage, CRC Press LLC, Boca Raton, FL.

CI14,2

204

Page 20: A measurement model of building information modelling maturity

McCuen, T., Suermann, P.C. and Krogulecki, M.J. (2012), “Evaluating award-winning BIMprojects using the national building information model standard capability maturitymodel”, Journal of Management in Engineering, Vol. 28 No. 2, pp. 224-230.

McGraw-Hill (2008), Building Information Modeling: Transforming Design and Construction toAchieve Greater Industry Productivity, available at: www.dbia.org/NR/rdonlyres/1631EDF1-8040-410D-BE1A-CF96AB1E9F84/0/McGrawHillConstructionBIMSmartMarketReportDecember2008.pdf (accessed 6 June 2011).

McGraw-Hill (2009), SmartMarket Report – The Business Value of BIM: Getting BuildingInformation Modeling to the Bottom Line, available at: www.bim.construction.com/research/ (accessed 28 April 2011).

Maruish, M.E. (2004), The Use of Psychological Testing for Treatment Planning and OutcomesAssessment: Volume 2: Instruments for Children and Adolescents, Lawrence ErlbaumAssociates, Inc., Mahwah, NJ.

NATSPEC (2011), NATSPEC National BIM Guide, available at: http://bim.natspec.org/index.php/natspec-bim-documents/national-bim-guide (accessed 15 August 2013).

NIBS (2007), United States National Building Information Modeling Standard, National Instituteof Building Science, available at: www.wbdg.org/pdfs/NBIMSv1_p1.pdf (accessed 29 July2012).

Nightingale, D.J. and Mize, J.H. (2002), “Development of a lean enterprise transformationmaturity model”, Information Knowledge Systems Management, Vol. 3 No. 1, pp. 15-30.

NIST (2011), 2011-2012 Criteria for Performance Excellence, National Institute of Standards andTechnology, available at: www.nist.gov/baldrige/publications/upload/2011_2012_Business_Nonprofit_Criteria.pdf (accessed 25 July 2012).

O’Brien, R.M. (2007), “A caution regarding rules of thumb for variance inflation factors”, Qualityand Quantity, Vol. 41, pp. 673-690.

Paulk, M.C., Weber, C.V. and Curtis, B. (1995), The Capability Maturity Model: Guidelines forImproving the Software Process, Addison-Wesley Professional, Reading, MA.

Sarshar, M., Haigh, R., Finnemore, M., Aouad, G., Barrett, P. and Baldry, D. (2000), “SPICE:a business process diagnostics tool for construction projects”, Engineering, Constructionand Architectural Management, Vol. 7 No. 3, pp. 241-250.

SEI (1994), The Capability Maturity Model: Guidelines for Improving the Software Process,Software Engineering Institute, Addison-Wesley Longman, Inc., Boston, MA.

SEI (2008), People Capability Maturity Model (P-CMM) Version 2.0, 2nd ed., SoftwareEngineering Institute, available at: www.sei.cmu.edu/library/abstracts/reports/09tr003.cfm (accessed 15 May 2011).

Seddigh, N., Pieda, P., Matrawy, A., Nandy, B., Lambadaris, J. and Hatfield, A. (2004), “Currenttrends and advances in information assurance metrics”, Proceedings of the 2nd AnnualConference on Privacy, Security and Trust, Fredericton, NB, pp. 197-205.

Smith, D.K. and Tardif, M. (2009), Building Information Modeling: A Strategic ImplementationGuide, Wiley, Hoboken, NJ.

Statwiki (2012), Exploratory Factor Analysis, available at: http://statwiki.kolobkreations.com/wiki/Exploratory_Factor_Analysis (accessed 7 January 2012).

Steele, P.K. and Kirkpatrick, S. (2011), Developing Competency Profiles, available at:www.regislearning.com/html/Publications/Article_steele_kirkpatrick_competency.pdf(accessed 28 December 2011).

Steiger, J.H. (2007), “Understanding the limitations of global fit assessment in structural equationmodeling”, Personality and Individual Differences, Vol. 42 No. 5, pp. 893-898.

Measurementmodel of BIMM

205

Page 21: A measurement model of building information modelling maturity

Succar, B. (2010), “Building information modeling maturity matrix”, in Underwood, J. andIsikdag, U. (Eds), Handbook of Research on Building Information Modeling andConstruction Informatics: Concepts and Technologies, Information Science Publishing,Hershey, PA, pp. 65-103.

Suermann, P.C. (2009), “Evaluating the impact of building information modeling (BIM) onconstruction”, PhD dissertation, University of Florida, Gainesville, FL, May.

Tabachnick, B.G. and Fidell, L.S. (2006), Using Multivariate Statistics, Ally & Bacon, New York, NY.

Tabachnick, B.G. and Fidell, L.S. (2012), Using Multivariate Statistics, 6th ed., Pearson Education,Upper Saddle River, NJ.

Tabish, S.Z.S. and Jha, K.N. (2012), “Success traits for a construction project”, Journal ofConstruction Engineering and Management, Vol. 138, pp. 1131-1138.

TFD (2011), Graphics, The Free Dictionary, available at: www.thefreedictionary.com/graphics(accessed 23 December 2011).

(The) US EPA (2007), Guidance for Preparing Standard Operating Procedures (SOPs), The USEnvironmental Protection Agency.

Vaidyanathan, K. and Howell, G. (2007), “Construction supply chain maturity model – conceptualframework”, Proceedings of the International Group for Lean Construction, Lansing, MI,USA, pp. 170-180.

Wong, A.K.D., Wong, F.K.W. and Nadeem, A. (2011), “Government roles in implementing buildinginformation modelling systems: comparison between Hong Kong and the United States”,Construction Innovation: Information, Process, Management, Vol. 11 No. 1, pp. 61-76.

Zhou, Q., Fang, D. and Mohamed, S. (2011), “Safety climate improvement: case study in a Chineseconstruction company”, Journal of Construction Engineering and Management, Vol. 137,pp. 86-95.

Further reading

FMI and Construction Management Association of America (CMAA) (2007), The Eighth AnnualSurvey of Owners, available at: http://images.autodesk.com/flashassets/company/BIM/downloads/FMI_CMAA_report.pdf (accessed 28 June 2011).

IBM (2011), Workflow Process, available at: http://publib.boulder.ibm.com/infocenter/cmgmt/v8r5m0/index.jsp?topic¼%2Fcom.ibm.administeringcm.doc%2Fmwf00008.htm(accessed 8 December 2011).

About the authorsYunfeng Chen is a graduate student in the Department of Building Construction Management withthe research focus on BIM, maturity and process models, project management, collaboration andtrust. Yunfeng Chen is the corresponding author and can be contacted at: [email protected]

Dr Hazar Dib is an Assistant Professor in a joint appointment with the Department ofComputer Graphics Technology and the Department of Building Construction Management withthe research focus on BIM, object-oriented-modelling (OOM), construction delivery, visualization,and sustainability.

Dr Robert F. Cox is the Associate Dean for Globalization and Engagement in the College ofTechnology and the Interim Department Head and Professor of the Department of BuildingConstruction Management with the research focus on construction management, partnering,trust, and technique application.

CI14,2

206

To purchase reprints of this article please e-mail: [email protected] visit our web site for further details: www.emeraldinsight.com/reprints

Page 22: A measurement model of building information modelling maturity

Appendix

BIM

Mar

eas

Sou

rce

for

BIM

Mar

eas

Des

crip

tion

s/d

efin

itio

ns

1a.

Sof

twar

eap

pli

cati

ons

Sof

twar

e“a

pp

lica

tion

s”(S

ucc

ar,

2010

)“A

pp

lica

tion

sar

eth

eau

tom

ated

use

rsy

stem

san

dm

anu

alp

roce

du

res

that

pro

cess

the

info

rmat

ion

”(I

TG

I,20

07,

p.

12)

3.H

ard

war

eeq

uip

men

tH

ard

war

e“e

qu

ipm

ent”

(Su

ccar

)B

IMH

Ere

fers

toth

ep

hy

sica

lar

tifa

cts

ofB

IMte

chn

olog

y(G

atto

l,20

12,

inp

roce

ss)

4.H

ard

war

eu

pg

rad

eH

ard

war

e“m

obil

ity

”(S

ucc

ar)

HU

refe

rsto

the

“rep

lace

men

t”of

har

dw

are

wit

ha

new

erv

ersi

onto

imp

rov

eit

sch

arac

teri

stic

s(K

apu

ret

al.,

2010

)6.

Info

rmat

ion

assu

ran

ce“S

ecu

rity

and

acce

ssco

ntr

ol”

(Su

ccar

)IA

isth

ep

ract

ice

tofa

cili

tate

the

tran

sfer

ofin

form

atio

nb

etw

een

par

ties

inan

accu

rate

and

secu

refa

shio

n(S

edd

igh

etal.,

2004

)7.

Pro

cess

and

tech

nol

ogy

inn

ovat

ion

“In

nov

atio

nan

dre

new

al”

(Su

ccar

)P

TI

refe

rsto

pro

cess

and

tech

nol

ogy

imp

rov

emen

tsan

din

nov

atio

ns

that

wou

ldm

easu

rab

lyim

pro

ve

ap

roje

ct’s

soft

war

ep

roce

sses

(SE

I,19

94)

8.S

trat

egic

pla

nn

ing

“Str

ateg

icat

trib

ute

s”(S

ucc

ar)

Str

ateg

icp

lan

nin

gre

fers

toh

owa

pro

ject

defi

nes

its

stra

teg

ies

and

acti

onp

lan

s(N

IST

,20

11)

9.S

enio

rle

ader

ship

“Org

anis

atio

nal

attr

ibu

tes”

(Su

ccar

)S

enio

rle

ader

ship

des

crib

esh

owa

com

mon

vis

ion

isco

mm

un

icat

edan

den

cou

rag

edb

yse

nio

rle

ader

s(N

IST

,20

11)

19.

Rew

ard

syst

em“R

ewar

ds”

(Su

ccar

)R

ewar

dsy

stem

refe

rsto

“pro

ced

ure

san

dst

and

ard

s”as

soci

ated

wit

hb

enefi

tsal

loca

tion

tore

late

dp

arti

es(B

usi

nes

sdic

tion

ary

,20

11)

20.

Ris

km

anag

emen

t“R

isk

s”(S

ucc

ar)

“Ris

km

anag

emen

tis

asy

stem

atic

pro

cess

tom

anag

eri

skb

yid

enti

fyin

g,

anal

ysi

ng

,an

dco

ntr

olli

ng

the

even

tsth

atm

ayca

use

un

des

irab

leou

tcom

e”(I

TG

I,20

07,

p.

68)

21.

Sta

nd

ard

oper

atin

gp

roce

ss(S

OP

)“G

uid

elin

es”

(Su

ccar

)S

OP

isa

det

aile

d“i

nst

ruct

ion

”ab

out

all

“act

ivit

ies”

ofa

pro

cess

”(T

he

US

EP

A,

2007

)22

.D

ocu

men

tati

onan

dm

odel

ing

stan

dar

ds

“Sta

nd

ard

san

dcl

assi

fica

tion

s”(S

ucc

ar)

Doc

um

enta

tion

and

mod

elin

gst

and

ard

sre

fer

toa

set

ofru

les

for

pre

par

ing

and

eval

uat

ing

doc

um

ents

and

mod

els

for

tech

nic

alp

roje

cts

23.

Qu

alit

yco

ntr

ol“B

ench

mar

ks”

(Su

ccar

)Q

ual

ity

con

trol

isa

pro

cess

tom

onit

oran

dm

anag

eth

eq

ual

ity

of“a

llfa

ctor

s”d

uri

ng

pro

du

ctio

n(C

has

e,20

07)

24.

Sp

ecifi

cati

on“S

erv

ice

spec

ifica

tion

”/”p

rod

uct

spec

ifica

tion

”(S

ucc

ar)

Sp

ecifi

cati

ons

are

“req

uir

emen

tsto

be

sati

sfied

by

ap

rod

uct

orse

rvic

e”(D

OT

,20

12)

25.

Com

pet

ency

pro

file

“Ex

per

ien

ce”,

“kn

owle

dg

e”,

“sk

ills

”,“d

yn

amic

s”(S

ucc

ar)

Com

pet

ency

pro

file

refe

rsto

ase

tof

kn

owle

dg

e,sk

ills

and

abil

itie

s(K

SA

s)th

ata

pos

itio

nre

qu

ires

(Ste

ele

and

Kir

kp

atri

ck,

2011

)26

.T

rain

ing

pro

gra

m“T

rain

ing

pro

gra

mm

es[S

IC]”

(Su

ccar

)T

rain

ing

pro

gra

mre

fers

toa

set

ofre

late

del

emen

tsth

atfo

cus

onad

dre

ssin

ga

pro

ject

’str

ain

ing

nee

ds

(SE

I,20

08)

(con

tinued

)

Table AI.BIMM areas and theircorresponding sources

and descriptions

Measurementmodel of BIMM

207

Page 23: A measurement model of building information modelling maturity

BIM

Mar

eas

Sou

rce

for

BIM

Mar

eas

Des

crip

tion

s/d

efin

itio

ns

27.

Tra

inin

gd

eliv

ery

met

hod

“Res

earc

hef

fort

s,ed

uca

tion

alp

rog

ram

mes

[SIC

]”/

”del

iver

able

s”(S

ucc

ar)

Tra

inin

gd

eliv

ery

met

hod

refe

rsto

the

met

hod

chos

ento

trai

np

arti

cip

ants

inu

seof

BIM

(Bla

nch

ard

and

Th

ack

er,

2006

)2.

Inte

rop

erab

ilit

y“I

nte

rop

erab

ilit

y/I

FC

sup

por

t”(N

IBS

,200

7)“I

nte

rop

erab

ilit

y”

isth

eab

ilit

yof

two

orm

ore

syst

ems

toex

chan

ge

info

rmat

ion

that

can

be

use

d(I

EE

E,

1991

)5.

Info

rmat

ion

del

iver

ym

eth

od“D

eliv

ery

met

hod

”(N

IBS

)In

form

atio

nd

eliv

ery

met

hod

refe

rsto

asy

stem

orap

pro

ach

use

db

ya

pro

ject

tob

rin

gin

form

atio

nto

use

rs(N

IBS

,20

07)

10.

Dat

ari

chn

ess

“Dat

ari

chn

ess”

(NIB

S)

“Dat

ari

chn

ess”

refe

rsto

the

com

ple

ten

ess

and

val

ue

ofd

ata

orin

form

atio

nw

ith

ina

syst

em(N

IBS

,20

07)

11.

Rea

l-ti

me

dat

a“T

imel

ines

s/re

spon

se”

(NIB

S)

“Rea

l-ti

me

dat

ad

enot

esin

form

atio

nth

atis

del

iver

ed”

oru

pd

ated

“im

med

iate

lyaf

ter

coll

ecti

on”

orch

ang

e(E

ncy

clop

edia

,20

11)

12.

Info

rmat

ion

accu

racy

“In

form

atio

nac

cura

cy”

(NIB

S)

IAc

defi

nes

the

clos

enes

sof

the

info

rmat

ion

rece

ived

toth

etr

uth

(Iv

anov

,19

72)

13.

Gra

ph

ics

“Gra

ph

ical

info

rmat

ion

”(N

IBS

)“G

rap

hic

s”ar

e“v

isu

al”

pre

sen

tati

ons

onco

mp

ute

rsc

reen

orp

aper

toin

form

oril

lust

rate

.(T

FD

,20

11)

14.

Geo

spat

ial

cap

abil

ity

“Sp

atia

lca

pab

ilit

y”

(NIB

S)

Geo

spat

ial

cap

abil

ity

refe

rsto

the

cap

abil

ity

ofa

syst

emto

cap

ture

,st

ore,

anal

yse

,m

anag

e,an

dp

rese

nt

dat

aw

ith

refe

ren

ceto

geo

gra

ph

iclo

cati

ond

ata

(NIB

S,

2007

)15

.L

ifec

ycl

ep

roce

sses

“Lif

e-cy

cle

vie

ws”

(NIB

S)

Lif

ecy

cle

pro

cess

esar

eth

ep

roce

sses

resp

onsi

ble

for

afa

cili

tyfr

omco

nce

pti

onth

rou

gh

oper

atio

nan

dm

ain

ten

ance

(Nig

hti

ng

ale

and

Miz

e,20

02)

17.

Ch

ang

em

anag

emen

t“C

han

ge

man

agem

ent”

(NIB

S)

“Ch

ang

em

anag

emen

t”id

enti

fies

,ev

alu

ates

and

imp

lem

ents

imp

rov

emen

tto

the

pro

ject

s’d

efin

edB

IMp

roce

sses

ona

con

tin

uou

sb

asis

(IT

GI,

2007

)18

.R

ole

“Rol

e”(N

IBS

)“R

ole”

refe

rsto

au

nit

ofd

efin

edre

spon

sib

ilit

ies

that

may

be

assu

med

by

one

orm

ore

par

ties

(SE

I,19

94)

Notes:

aT

he

area

sw

ere

nu

mb

ered

corr

esp

ond

ing

toth

eor

der

ofth

eit

ems

inth

eq

ues

tion

nai

re;f

orth

ed

etai

lof

the

qu

esti

onn

aire

,ple

ase

refe

rto

app

end

ixof

the

arti

cle

wri

tten

by

Dib

etal.

(201

2);t

he

area

su

nd

erth

eco

lum

ns

of“B

IMM

area

s”an

d“S

ourc

efo

rB

IMM

area

s”w

ere

rep

rod

uce

dd

irec

tly

from

Tab

leI

ofth

ear

ticl

ew

ritt

enb

yC

hen

etal.

(201

2)

Table AI.

CI14,2

208

Page 24: A measurement model of building information modelling maturity

Stu

dy

Tec

hn

olog

yP

roce

ssP

eop

leIn

form

atio

n

NA

TS

PE

CN

ati

onalB

IMG

uid

e(N

AT

SP

EC

,20

11,

pp

.iv

-v)

“Tec

hn

olog

yP

latf

orm

and

Sof

twar

e”;“

BIM

man

agem

ent

Pla

n(B

MP

)”;

“Req

uir

emen

tsfo

rU

sin

gB

IM”

“Col

lab

orat

ion

Pro

ced

ure

s”“B

IMro

les

and

Res

pon

sib

ilit

ies”

“Mod

elS

har

ing

”;“3

DM

odel

s,F

orm

ats,

and

Mod

elS

tru

ctu

res”

;“F

ile

Sto

rag

ean

dS

ecu

rity

”;“R

equ

irem

ents

for

2DD

raw

ing

s”;

“Mod

elli

ng

Req

uir

emen

ts”

AE

C(U

K)

BIM

Pro

toco

l(A

EC

(UK

)C

omm

itte

e,20

12)

“Res

ourc

es”

(p.

43);

“In

tero

per

abil

ity

”(p

.19

)“P

roje

ctB

IME

xec

uti

onP

lan

”:“P

roje

ctE

xec

uti

onP

lan

”(p

.12)

and

“Pro

ject

BIM

mee

tin

gs”

(p.

13)

“Pro

ject

BIM

Ex

ecu

tion

Pla

n”:

“Rol

esan

dR

esp

onsi

bil

itie

s”(p

.10

)

“Col

lab

orat

ion

BIM

Wor

kin

g”

(p.

14);

“Dat

aS

egre

gat

ion

”(p

.21

)

“Mod

elli

ng

Met

hod

olog

y”

(P.

25);

“Fol

der

Str

uct

ure

and

Nam

ing

Con

ven

tion

s”(p

.32)

;“P

rese

nta

tion

Sty

les”

(p.

39)

Hon

gK

ong

BIM

Pro

ject

Spe

cifica

tion

(HK

I,20

11,

p.

2)

“Cla

shA

nal

ysi

sP

roce

ss”;

“Har

dw

are

Sp

ecifi

cati

ons”

;“S

oftw

are

Sp

ecifi

cati

ons”

“BIM

pro

ject

Ob

ject

ive”

;“B

uil

din

gIn

form

atio

nM

odel

Sp

ecifi

cati

on”;

“BIM

Met

hod

olog

ies

and

Pro

cess

es”;

“Pro

ject

Del

iver

able

sfr

omB

IMP

roce

ss”

“BIM

Man

agem

ent

and

Sta

ffR

esou

rces

”“M

odel

Dat

aan

dL

evel

ofD

etai

l”

Com

mon

BIM

requir

emen

ts(B

uil

din

gS

mar

tF

inla

nd

,20

12)

“Sof

twar

e”(p

.7)

;“M

odel

ing

Too

ls”

(p.

9)“B

IMS

pec

ifica

tion

”(p

.10

)“P

ub

lish

ing

ofM

odel

s”(p

.11

)“R

ole

ofth

eB

IMC

oord

inat

or”

(p.

11)

“BIM

Acc

ura

cy”

(p.

8);

“Th

eB

uil

din

gs,

Flo

orL

evel

san

dD

ivis

ion

s”(p

.9)

;“N

amin

gan

dA

rch

ivin

gof

the

BIM

”(p

.10

);“R

elea

seof

the

BIM

”(p

.7)

;“C

oord

inat

ors

and

Un

its”

(p.

8);

“Wor

kin

gM

odel

s”(p

.11)

;“Q

ual

ity

Ass

ura

nce

ofB

IMs”

(p.

12)

Sin

gapo

reB

IMG

uid

e(B

CA

,20

12,

p.

vii

)

“BIM

Mod

elin

gan

dC

olla

bor

atio

nP

roce

du

res”

:“I

nd

ivid

ual

Dis

cip

lin

eM

odel

ing

”,“C

ross

-Dis

cip

lin

ary

Mod

elC

oord

inat

ion

“BIM

Sp

ecifi

cati

ons”

:“B

IMO

bje

ctiv

ean

dR

esp

onsi

bil

ity

Mat

rix

”,“C

omp

ensa

tion

Ex

pec

tati

ons”

“BIM

Sp

ecifi

cati

ons”

:“B

IMD

eliv

erab

les”

,“L

evel

ofD

etai

lan

dP

roje

ctS

tag

esin

the

Sin

gap

ore

BIM

Gu

ide”

;“B

IMM

odel

ing

and

Col

lab

orat

ion

Pro

ced

ure

s”:“

Mod

elan

dD

ocu

men

tati

onP

rod

uct

ion

Table AII.Comparison of BIMM

dimensions of thisresearchwith content and

areas of globalBIM-related

standards/guides/protocols/specifications

Measurementmodel of BIMM

209