Bedrock
insightsconstruction AIdrawing intelligenceconstruction technology

The Last Mile Problem: Why AI Agents Can't Do Construction Work Yet

AI agents are automating every industry end-to-end. Construction is stuck because LLMs cannot read the documents that matter most: drawings. That is about to change.

Stan Liu
Stan Liu · Co-Founder
·8 min read
Share

TL;DR

AI agents are automating end-to-end workflows across legal, finance, healthcare, and software. Construction is stuck. Not because the workflows are too complex, but because the most important documents in construction are invisible to AI. LLMs cannot read construction drawings. That one gap blocks everything downstream: estimating, scheduling, procurement, project controls, change management. Solving it unlocks the entire industry.

AI Agents Are Eating Every Industry. Except One.

In legal, AI agents draft contracts, review discovery, and flag regulatory risk. In finance, they reconcile transactions, generate reports, and monitor compliance. In healthcare, they summarize patient records, cross-reference lab results, and surface treatment options. In software, they write code, review PRs, and deploy to production.

The pattern is the same everywhere: give an AI agent access to the documents and data that drive a workflow, and it automates the workflow end-to-end.

Construction should be next. The industry spent $2.1 trillion in 2024. Its workflows are highly structured, repetitive, and document-driven. Change order preparation, quantity takeoffs, bid leveling, RFI generation, schedule updates: these are all processes where humans extract information from documents and feed it into decisions.

So why isn't construction automated yet?

The Document That Breaks Everything

Every construction workflow traces back to a drawing. Which wall goes where. What gauge conduit feeds which panel. Where the mechanical diffuser sits relative to the beam above it. How many fire dampers appear on the mechanical plan. What changed between Rev 3 and Rev 4.

Drawings are the source of truth. They determine scope, cost, schedule, and risk. And LLMs cannot read them.

This isn't a minor limitation. Ask GPT-4V to compare two revisions of a structural plan and tell you what changed. Accuracy drops to 40-55% on spatial detection tasks. It might catch a deleted room label but miss that a wall shifted 6 inches. Ask it to count fire alarm devices across 80 sheets of electrical drawings. It hallucinates counts. Ask it to identify room boundaries from a floor plan. It produces approximate regions that miss corridors, closets, and irregular spaces.

Construction drawings are dense, multi-layered technical documents where spatial relationships, symbol conventions, and dimensional precision all carry meaning. This is fundamentally different from the text, code, and natural images that LLMs handle well. Pure LLM approaches to drawing interpretation achieve 40-55% accuracy on spatial tasks, which is not reliable enough to automate anything.

The result: every AI agent built for AEC work hits the same wall. It can draft an email, summarize a spec, or format a spreadsheet. But the moment a workflow requires understanding what's on a drawing, the agent is blind.

This is the last-mile problem. LLMs can do everything except the one thing that matters most in construction.

What "Reading" a Drawing Actually Requires

Humans read drawings through a combination of spatial reasoning, domain knowledge, and pattern recognition that took them years to develop.

When a project engineer compares two revisions, they're doing several things simultaneously: aligning sheets that may have shifted in layout, filtering irrelevant changes like title block updates and watermarks, detecting pixel-level differences in linework, classifying those differences as additions, deletions, or modifications, and contextualizing them against the project scope.

When an estimator counts symbols, they're matching visual patterns from a legend across dozens of sheets, accounting for variations in scale, rotation, and rendering quality.

When a PM reviews a floor plan, they're segmenting space into rooms, associating labels, understanding adjacencies, and cross-referencing against a finish schedule.

None of these tasks are about "seeing" the drawing. They're about interpreting it. And interpretation requires understanding spatial relationships, symbol conventions, dimensional precision, and domain context simultaneously.

LLMs can see drawings. They cannot interpret them. That distinction is everything.

What This Blocks

Without drawing interpretation, here's what AI agents cannot do in construction today:

Automated change management. An agent could prepare a change order if it knew what changed between two revisions. It can't, because it can't read the drawings. That manual comparison step is the bottleneck that blocks the entire downstream workflow from being automated.

Intelligent estimating. An agent could generate quantity takeoffs if it could count symbols and extract spatial data from drawings. It can't. So estimators still count fire alarm devices by hand across 80 sheets.

Proactive schedule updates. An agent could flag timeline risk when scope changes hit a drawing revision. It can't detect those scope changes. So PMs find out about problems in OAC meetings, weeks after the drawings landed.

End-to-end procurement. An agent could match spec changes to material orders. But spec changes originate in drawings, and the agent can't read those drawings. So procurement stays manual and reactive.

Every one of these workflows is structured, repetitive, and high-value. The only thing missing is the ability to read the source documents.

According to Autodesk/FMI research, 52% of construction rework stems from poor project data and miscommunication. The data exists in the drawings. The problem is that neither humans nor AI agents extract it systematically.

Solving the Last Mile

The approach that works is not asking an LLM to read a drawing. It's building a specialized system that reads drawings the way humans do, then returning structured data that any LLM, agent, or software platform can act on.

This means:

Computer vision for detection. Pixel-level comparison for change detection. Template matching for symbol identification. Boundary extraction for room segmentation. These are spatial tasks that require spatial methods.

AI for classification. Once changes, symbols, and rooms are detected, AI classifies them: What type of change? What kind of symbol? What room label applies? This is where language models add value, working on structured data rather than raw pixels.

Structured output for agents. The result isn't a PDF or a screenshot. It's JSON: categorized change lists, symbol counts with locations, room polygons with areas. Data that an AI agent can reason about, filter, aggregate, and feed into downstream workflows.

This hybrid approach is what bridges the gap. Give AI agents structured access to drawing data, and the entire AEC automation stack unlocks.

The Reading Layer That Was Missing

The construction tech stack has evolved in layers. First came document management: Procore, PlanGrid, and Bluebeam made drawings accessible, mobile, and markable. That layer solved distribution.

The next layer is reading. Not viewing, not storing, not annotating. Actually extracting structured data from drawings so that software, agents, and people can act on it programmatically.

APIs that let any platform query what's on a drawing, what changed between revisions, how many of a given symbol appear on a sheet, turn drawings from static files into live data sources. A scheduling agent that knows when scope changes hit can flag timeline risk automatically. An estimating agent that reads symbol counts can generate quantity takeoffs without human input. A project controls agent that tracks changes across revisions can catch rework before it starts.

This is the layer Bedrock provides.

Bedrock's Role

We build the AI that reads construction drawings. Comparison, symbol detection, room segmentation: these are the core reading capabilities that make drawing data accessible to agents and platforms.

We started with comparison because that's where the pain was sharpest. Building it confirmed a larger thesis: the real opportunity isn't any single feature. It's solving the last-mile problem so that the entire AEC industry can be automated by AI agents that actually understand the documents driving every workflow.

The Construction Management Association of America (CMAA) has noted that data-driven project delivery is the industry's next frontier. That frontier opens when AI agents can read the industry's most important documents.

The Question for Builders

If you're building AI agents for construction, you've already hit this wall. Your agent can draft emails, summarize specs, and generate reports. But the moment it needs to know what changed in a drawing, what symbols appear on a sheet, or how a floor plan is laid out, it's stuck.

That last mile is solvable now. The question is what you'll build once your agent can actually read.

FAQ

Why can't LLMs read construction drawings?

LLMs process text and natural images well, but construction drawings are dense, multi-layered technical documents where spatial relationships, symbol conventions, and dimensional precision carry meaning. Pure LLM approaches achieve 40-55% accuracy on spatial detection tasks in construction drawings. They miss wall shifts, relocated equipment, dimensional changes, and symbol counts. The spatial reasoning required is fundamentally different from text comprehension.

How is Bedrock's approach different from using GPT-4V or Gemini on drawings?

Bedrock uses a hybrid approach: computer vision for pixel-level detection (changes, symbols, boundaries) combined with AI for classification and interpretation. This is how humans read drawings too: spatial detection first, then contextual interpretation. Pure LLM approaches try to do both at once and fail on the spatial component.

What can AI agents do once they can read drawings?

Automated change management, quantity takeoffs, schedule risk detection, procurement matching, bid leveling, RFI generation: any workflow that starts with extracting information from drawings becomes automatable when the agent has structured access to that data via API.

Can AI replace the project team's review of drawings?

No. AI handles detection and extraction so that both AI agents and project teams can act on structured drawing data rather than starting from raw PDFs. The project team still applies judgment, context, and domain expertise. What changes is that the scanning bottleneck is removed.

What types of drawings does this work on?

All major construction disciplines: architectural, structural, mechanical, electrical, plumbing, fire protection, and civil. The core capabilities work on any technical drawing delivered as a PDF.

Stan Liu
Stan Liu · Co-Founder
·8 min read
Share