|buildingSMART alliance January 2013 Challenge: AutoDesk|
January 2013 bSa Challenge: AutoDesk
by E. William East, PhD, PE, F.ASCE, Dr. Chris Bogen, PhD, and Mariangelica Carrasquillo, Engineer Research and Development Center, U.S. Army Corps of Engineers
The January 2013 buildingSMART alliance Challenge was the first buildingSMART Challenge event that expanded the set of contracted information exchanges beyond the Construction-Operations Building information exchange (COBie). This event introduced standards for the exchange of building programming information, at the early stages of project planning, and standards for the exchange of building systems, from design through the production of construction shop drawings. Software systems supporting COBie version 2.4, the Building Programming information exchange (BPie), and the Heating, Ventilating and Air Conditioning information exchange (HVACie) formats were demonstrated.
The detailed requirements for this Challenge were published by the buildingSMART alliance in February 2012 and discussed among the participants during monthly software developer meetings between February and October 2012. Meeting notes were provided to all potential Challenge participants. An example of one of the extensive sets of meeting from one of the meetings can be seen here. A new tool for the automated checking of COBie files, the COBie Toolkit was developed and used for this challenge. The rules used by the Toolkit were documented in the COBie Responsibility Matrix and developed with direct feedback by the software developers and programmers. Another important contribution of these monthly discussions was the development of, and agreement to, rules governing the order of precedence when merging multiple model files at the Construction Documents stage of design. Final submissions were due 15 December 2012. This page provides information about the performance of one specific commercial software product.
Files tested in the bSa Challenge are produced from software that can be commercially purchased on, or before, to the date of the Challenge. The table below provides information about the product tested and links to user guides and support hotlines.
Software companies were required to provide a set of files to demonstrate their products compliance with the requirements of the bSa Challenge. The table below provides a link to these files. As appropriate, notes are provided below the table.
Test Procedure: Design Phase
The procedure for testing COBie design models required that the submitted design data must match the information content found on the Medical Clinic drawings. To that end, a presentation documented the detailed asset information found on the Clinic drawings for two typical rooms. This presentation shows the complexity of incorporating the complete set of architectural, mechanical, electrical, plumbing, fire protection, communication, and other systems assets identified in a detailed cross-referenced review of the clinic drawings. This presentation was used to explain the level of detail required to establish if a COBie file contains the same information as printed drawings. Once the file was received an additional room, the first floor mechanical room was also added to the evaluation. To summarize the presentation, evaluation criteria for typical design COBie file should include, at minimum, the following types of reviews.
To support the integration of scheduled information into the model an updated example COBie file based on a manual takeoff of equipment/product schedules. This updated Clinic COBie file will be publically released through the Common BIM File page as "version 12" of the Clinic COBie example files on or about March 2013.
While significant effort was spent to develop the underlying redacted drawings, initial models, and updated COBie models not one design software vendor met the complete COBie Challenge to have the model provided match the provided design drawings or example COBie file. As a result, the following criteria were established to evaluate design model submissions to establish an estimate of the quality of the submitted data file.
The table below provides the results of these tests and evaluations. The first column, "COBie Sheet," identifies each worksheet in a COBie file. Those worksheets not required, must be present, but may be left empty. The second column, "In Scope," indicates if that a given worksheet (or its equivalent IFC formatted) data was required for the related challenge. If the information was in scope, the column contains the value of "Yes." If the information was not in scope, the column contains a value of "No" and while present in the file should be empty (or contain only the first row containing the column headings).
Design Software Results
The three right-most columns in the table below provide the results of the testing team's evaluation of the accuracy of the vendors design model submissions. Quality Control Errors identify technical errors found in the vendor's file. In some cases the COBie Tool Kit reported errors that required additional explanation, in these cases notes are provided. The next column, "Record (Count/Error)," identifies total count of COBie rows found for the required row and the number of rows containing errors when evaluated against information found on design drawings. The "Attribute (Count/Errors)" column identifies the attributes or properties found for all of the records of the given type. Errors related to these attributes were identified if the related schedule data found on the drawings did not, at a minimum match the information found on the related design drawings.
1 There was a discrepancy between the units of measure used for usable height, but based on the unit conversion the values are accurate and no penalty is applied.
2 In the newest version of COBie (2.4) the standard convention for Zone.SpaceNames is to provide only one space instead of comma-delimited lists of space; now a Zone row is uniquely identified by Zone.Name+Zone.Category+Zone.SpaceNames. However, no penalty is applied because the provided data is accurate and is still an acceptable representation for Zones.
3 The description fields for many rows contain the Autodesk family name for the Component or Type. Though this doesn't match the original model, it is a valid use of Description.
4 Based on the inspected data, Voltage values (Cold Water Pump, e.g.) did not match the values on the original model. This incurs a 4-minute penalty for the time it takes to manually fix these errors.
5 This error is due to one component row referencing a Space name that does not exist in the Space worksheet. This incurs a one minute penalty.
This company successfully completed the construction COBie challenge by producing the handover COBie file of the Medical Clinic model. Based on the quality control report, there was only one small error that incurred a one minute penalty with respect to the internal consistency of the output format that would require an estimated 5 minutes to correct.
The vendor did not produce the 100% of the sample data; however the produced data was very accurate to expected sample data found on the project drawings. Based on the content quality criteria a 4 minutes penalty was applied as described in the session above. In sum a total of 5 minutes was applied for the submitted file. This means that a user utilizing Autodesk Revit 2013 software is estimated to have to spend 9 minutes cleaning/fixing the COBie file for a facility of comparable size and complexity.
Previous June 2010 Revit Challenge Results (for COBie Output)