DDC_Skills_for_AI_Agents_in_Construction bim-validation-pipeline
Build automated BIM validation pipelines for IFC/Revit data. Continuous validation against IDS, LOD requirements, COBie, and project-specific BEP standards.
install
source · Clone the upstream repo
git clone https://github.com/datadrivenconstruction/DDC_Skills_for_AI_Agents_in_Construction
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/datadrivenconstruction/DDC_Skills_for_AI_Agents_in_Construction "$T" && mkdir -p ~/.claude/skills && cp -r "$T/2_DDC_Book/4.3-BIM-Validation-Pipeline/bim-validation-pipeline" ~/.claude/skills/datadrivenconstruction-ddc-skills-for-ai-agents-in-construction-bim-validation-p && rm -rf "$T"
manifest:
2_DDC_Book/4.3-BIM-Validation-Pipeline/bim-validation-pipeline/SKILL.mdsource content
BIM Validation Pipeline
Overview
Based on DDC methodology (Chapter 4.3), this skill provides automated BIM data validation pipelines. Validate BIM models against Information Delivery Specification (IDS), Level of Development (LOD) requirements, and project standards.
Book Reference: "Автоматический ETL конвейер для валидации данных" / "Automated ETL Pipeline for Data Validation"
"Автоматизированная валидация BIM-данных позволяет выявлять ошибки на ранних стадиях и обеспечивать соответствие требованиям BEP." — DDC Book, Chapter 4.3
Quick Start
import ifcopenshell import pandas as pd # Load IFC model ifc_model = ifcopenshell.open("model.ifc") # Quick validation checks walls = ifc_model.by_type("IfcWall") print(f"Total walls: {len(walls)}") # Check for required properties issues = [] for wall in walls: # Check if wall has material if not wall.HasAssociations: issues.append(f"Wall {wall.GlobalId}: No material assigned") print(f"Issues found: {len(issues)}")
BIM Validation Framework
Core Validator Class
import ifcopenshell import ifcopenshell.util.element as element_util import pandas as pd from dataclasses import dataclass from typing import List, Dict, Optional from enum import Enum class Severity(Enum): ERROR = "error" WARNING = "warning" INFO = "info" @dataclass class ValidationIssue: element_id: str element_type: str rule_id: str severity: Severity message: str location: Optional[str] = None class BIMValidator: """Comprehensive BIM model validator""" def __init__(self, ifc_path: str): self.model = ifcopenshell.open(ifc_path) self.issues: List[ValidationIssue] = [] self.stats = {} def validate_all(self): """Run all validation checks""" self.validate_geometry() self.validate_properties() self.validate_relationships() self.validate_naming() self.validate_classification() return self.get_report() def validate_geometry(self): """Check geometry validity""" elements_with_geometry = [ e for e in self.model.by_type("IfcProduct") if e.Representation ] for element in elements_with_geometry: # Check for zero volume try: settings = ifcopenshell.geom.settings() shape = ifcopenshell.geom.create_shape(settings, element) # Volume check would go here except: self.issues.append(ValidationIssue( element_id=element.GlobalId, element_type=element.is_a(), rule_id="GEO-001", severity=Severity.ERROR, message="Invalid or missing geometry" )) self.stats['elements_with_geometry'] = len(elements_with_geometry) def validate_properties(self, required_psets: Dict[str, List[str]] = None): """Check required property sets and properties""" if required_psets is None: required_psets = { 'IfcWall': ['Pset_WallCommon', 'BaseQuantities'], 'IfcSlab': ['Pset_SlabCommon', 'BaseQuantities'], 'IfcColumn': ['Pset_ColumnCommon', 'BaseQuantities'], 'IfcBeam': ['Pset_BeamCommon', 'BaseQuantities'] } for ifc_type, psets in required_psets.items(): elements = self.model.by_type(ifc_type) for element in elements: element_psets = element_util.get_psets(element) for required_pset in psets: if required_pset not in element_psets: self.issues.append(ValidationIssue( element_id=element.GlobalId, element_type=ifc_type, rule_id="PROP-001", severity=Severity.WARNING, message=f"Missing PropertySet: {required_pset}" )) def validate_relationships(self): """Check spatial containment and relationships""" products = self.model.by_type("IfcProduct") for product in products: # Check spatial containment if hasattr(product, 'ContainedInStructure'): if not product.ContainedInStructure: self.issues.append(ValidationIssue( element_id=product.GlobalId, element_type=product.is_a(), rule_id="REL-001", severity=Severity.WARNING, message="Element not assigned to building storey" )) # Check material assignment if hasattr(product, 'HasAssociations'): has_material = any( rel.is_a('IfcRelAssociatesMaterial') for rel in (product.HasAssociations or []) ) if not has_material and product.is_a() in ['IfcWall', 'IfcSlab', 'IfcColumn']: self.issues.append(ValidationIssue( element_id=product.GlobalId, element_type=product.is_a(), rule_id="MAT-001", severity=Severity.WARNING, message="No material assigned" )) def validate_naming(self, patterns: Dict[str, str] = None): """Validate element naming conventions""" import re if patterns is None: patterns = { 'IfcBuildingStorey': r'^(Level|L|Floor|Уровень)\s*[-]?\d+', 'IfcWall': r'^W[-_]?\d{3,}|^Wall[-_]', 'IfcColumn': r'^C[-_]?\d{3,}|^Column[-_]', 'IfcSpace': r'^Room[-_]|^Space[-_]' } for ifc_type, pattern in patterns.items(): elements = self.model.by_type(ifc_type) for element in elements: name = element.Name or "" if not re.match(pattern, name, re.IGNORECASE): self.issues.append(ValidationIssue( element_id=element.GlobalId, element_type=ifc_type, rule_id="NAME-001", severity=Severity.INFO, message=f"Name '{name}' doesn't match convention" )) def validate_classification(self, required_systems: List[str] = None): """Check classification system assignments""" if required_systems is None: required_systems = ['Uniclass', 'OmniClass', 'Uniformat'] elements = self.model.by_type("IfcProduct") for element in elements: if hasattr(element, 'HasAssociations'): has_classification = any( rel.is_a('IfcRelAssociatesClassification') for rel in (element.HasAssociations or []) ) if not has_classification: self.issues.append(ValidationIssue( element_id=element.GlobalId, element_type=element.is_a(), rule_id="CLASS-001", severity=Severity.INFO, message="No classification assigned" )) def get_report(self): """Generate validation report""" by_severity = {s: [] for s in Severity} by_type = {} by_rule = {} for issue in self.issues: by_severity[issue.severity].append(issue) if issue.element_type not in by_type: by_type[issue.element_type] = [] by_type[issue.element_type].append(issue) if issue.rule_id not in by_rule: by_rule[issue.rule_id] = [] by_rule[issue.rule_id].append(issue) return { 'total_issues': len(self.issues), 'errors': len(by_severity[Severity.ERROR]), 'warnings': len(by_severity[Severity.WARNING]), 'info': len(by_severity[Severity.INFO]), 'by_type': {k: len(v) for k, v in by_type.items()}, 'by_rule': {k: len(v) for k, v in by_rule.items()}, 'issues': self.issues, 'stats': self.stats }
LOD Validation
Level of Development Checker
class LODValidator: """Validate Level of Development (LOD) requirements""" # LOD requirements by element type LOD_REQUIREMENTS = { 'LOD100': { 'geometry': False, 'properties': [], 'description': 'Conceptual' }, 'LOD200': { 'geometry': True, 'approximate_size': True, 'properties': ['Category'], 'description': 'Schematic Design' }, 'LOD300': { 'geometry': True, 'exact_size': True, 'properties': ['Category', 'Material', 'Type'], 'quantities': ['Length', 'Area', 'Volume'], 'description': 'Design Development' }, 'LOD350': { 'geometry': True, 'exact_size': True, 'properties': ['Category', 'Material', 'Type', 'Manufacturer'], 'quantities': ['Length', 'Area', 'Volume', 'Weight'], 'connections': True, 'description': 'Construction Documentation' }, 'LOD400': { 'geometry': True, 'fabrication_ready': True, 'properties': ['Category', 'Material', 'Type', 'Manufacturer', 'Model', 'Serial', 'InstallationDate'], 'quantities': ['All'], 'connections': True, 'description': 'Fabrication & Assembly' } } def __init__(self, model, target_lod='LOD300'): self.model = model self.target_lod = target_lod self.requirements = self.LOD_REQUIREMENTS.get(target_lod, {}) self.results = [] def validate_element(self, element): """Validate single element against LOD requirements""" issues = [] element_guid = element.GlobalId psets = element_util.get_psets(element) # Check geometry if self.requirements.get('geometry'): if not element.Representation: issues.append({ 'element': element_guid, 'issue': 'Missing geometry', 'required_for': self.target_lod }) # Check required properties required_props = self.requirements.get('properties', []) all_props = {} for pset_name, props in psets.items(): all_props.update(props) for prop in required_props: if prop not in all_props or all_props[prop] is None: issues.append({ 'element': element_guid, 'issue': f'Missing property: {prop}', 'required_for': self.target_lod }) # Check quantities required_quantities = self.requirements.get('quantities', []) if required_quantities != ['All']: qsets = psets.get('BaseQuantities', {}) for qty in required_quantities: if qty not in qsets: issues.append({ 'element': element_guid, 'issue': f'Missing quantity: {qty}', 'required_for': self.target_lod }) return issues def validate_model(self, element_types=None): """Validate entire model""" if element_types is None: element_types = ['IfcWall', 'IfcSlab', 'IfcColumn', 'IfcBeam', 'IfcDoor', 'IfcWindow', 'IfcStair'] all_issues = [] summary = {} for ifc_type in element_types: elements = self.model.by_type(ifc_type) type_issues = [] for element in elements: issues = self.validate_element(element) type_issues.extend(issues) summary[ifc_type] = { 'total': len(elements), 'issues': len(type_issues), 'compliance': ((len(elements) - len(type_issues)) / len(elements) * 100) if elements else 100 } all_issues.extend(type_issues) return { 'target_lod': self.target_lod, 'total_issues': len(all_issues), 'summary': summary, 'issues': all_issues }
IDS Validation
Information Delivery Specification
import xml.etree.ElementTree as ET class IDSValidator: """Validate against IDS (Information Delivery Specification)""" def __init__(self, ids_path: str): self.ids = self._parse_ids(ids_path) def _parse_ids(self, path): """Parse IDS XML file""" tree = ET.parse(path) root = tree.getroot() specifications = [] for spec in root.findall('.//specification'): specifications.append({ 'name': spec.get('name'), 'applicability': self._parse_facets(spec.find('applicability')), 'requirements': self._parse_facets(spec.find('requirements')) }) return specifications def _parse_facets(self, element): """Parse IDS facets""" if element is None: return [] facets = [] for child in element: facet = { 'type': child.tag, 'constraints': {} } for attr, value in child.attrib.items(): facet['constraints'][attr] = value facets.append(facet) return facets def validate(self, model): """Validate IFC model against IDS""" results = [] for spec in self.ids: applicable_elements = self._find_applicable_elements( model, spec['applicability'] ) for element in applicable_elements: issues = self._check_requirements(element, spec['requirements']) if issues: results.append({ 'specification': spec['name'], 'element': element.GlobalId, 'issues': issues }) return results def _find_applicable_elements(self, model, applicability): """Find elements matching applicability criteria""" elements = [] for facet in applicability: if facet['type'] == 'entity': ifc_type = facet['constraints'].get('name') if ifc_type: elements.extend(model.by_type(ifc_type)) return elements def _check_requirements(self, element, requirements): """Check element against requirements""" issues = [] psets = element_util.get_psets(element) for req in requirements: if req['type'] == 'property': pset_name = req['constraints'].get('propertySet') prop_name = req['constraints'].get('name') if pset_name and prop_name: pset = psets.get(pset_name, {}) if prop_name not in pset: issues.append(f"Missing property: {pset_name}.{prop_name}") return issues
Pipeline Automation
Automated Validation Pipeline
import os from datetime import datetime import json class BIMValidationPipeline: """Automated BIM validation pipeline""" def __init__(self, config_path=None): self.config = self._load_config(config_path) self.results_history = [] def _load_config(self, path): if path and os.path.exists(path): with open(path) as f: return json.load(f) return { 'lod_target': 'LOD300', 'required_psets': { 'IfcWall': ['Pset_WallCommon'], 'IfcSlab': ['Pset_SlabCommon'] }, 'naming_patterns': {}, 'fail_on_errors': True, 'warn_threshold': 50 } def run(self, ifc_path): """Run complete validation pipeline""" start_time = datetime.now() # Initialize validators validator = BIMValidator(ifc_path) lod_validator = LODValidator( validator.model, self.config['lod_target'] ) # Run validations bim_report = validator.validate_all() lod_report = lod_validator.validate_model() # Compile results result = { 'file': ifc_path, 'timestamp': start_time.isoformat(), 'duration_seconds': (datetime.now() - start_time).total_seconds(), 'bim_validation': bim_report, 'lod_validation': lod_report, 'passed': self._evaluate_pass(bim_report, lod_report) } self.results_history.append(result) return result def _evaluate_pass(self, bim_report, lod_report): """Determine if validation passed""" if self.config['fail_on_errors'] and bim_report['errors'] > 0: return False if bim_report['warnings'] > self.config['warn_threshold']: return False return True def run_batch(self, ifc_paths): """Run validation on multiple files""" results = [] for path in ifc_paths: try: result = self.run(path) results.append(result) except Exception as e: results.append({ 'file': path, 'error': str(e), 'passed': False }) return { 'total': len(results), 'passed': sum(1 for r in results if r.get('passed', False)), 'failed': sum(1 for r in results if not r.get('passed', True)), 'results': results } def export_report(self, output_path): """Export validation results to Excel""" if not self.results_history: return None latest = self.results_history[-1] with pd.ExcelWriter(output_path, engine='openpyxl') as writer: # Summary summary = pd.DataFrame({ 'Metric': ['File', 'Timestamp', 'Passed', 'Errors', 'Warnings'], 'Value': [ latest['file'], latest['timestamp'], latest['passed'], latest['bim_validation']['errors'], latest['bim_validation']['warnings'] ] }) summary.to_excel(writer, sheet_name='Summary', index=False) # Issues if latest['bim_validation']['issues']: issues_df = pd.DataFrame([ { 'Element': i.element_id, 'Type': i.element_type, 'Rule': i.rule_id, 'Severity': i.severity.value, 'Message': i.message } for i in latest['bim_validation']['issues'] ]) issues_df.to_excel(writer, sheet_name='Issues', index=False) return output_path
Quick Reference
| Rule ID | Description | Severity |
|---|---|---|
| GEO-001 | Invalid/missing geometry | ERROR |
| PROP-001 | Missing PropertySet | WARNING |
| REL-001 | No spatial containment | WARNING |
| MAT-001 | No material assigned | WARNING |
| NAME-001 | Invalid naming convention | INFO |
| CLASS-001 | No classification | INFO |
LOD Requirements Summary
| LOD | Geometry | Properties | Quantities |
|---|---|---|---|
| 100 | No | - | - |
| 200 | Approximate | Category | - |
| 300 | Exact | Material, Type | L, A, V |
| 350 | Exact + connections | Manufacturer | All |
| 400 | Fabrication-ready | All details | All |
Resources
- Book: "Data-Driven Construction" by Artem Boiko, Chapter 4.3
- Website: https://datadrivenconstruction.io
- IDS Standard: https://technical.buildingsmart.org/projects/information-delivery-specification-ids/
- IfcOpenShell: https://ifcopenshell.org
Next Steps
- See
for extracting data from IFCifc-data-extraction - See
for general data validationdata-quality-check - See
for quantity take-off from validated modelsqto-report