Thought Leadership

Topics in AI Systems: Part II — GNN and LLM Types

In Part I, we introduced the idea that solving complex digital-engineering problems requires combining structural reasoning (GNNs) with semantic understanding (LLMs). We used examples from electro-mechanical systems, manufacturing, and service lifecycle to show why this fusion is powerful. In this article, we move from why to what. Specifically, we introduce the major types of GNNs and LLMs so developers and architects can select the right building blocks when designing fused AI solutions for digital industries (see Figure 1). For ease of comparison, the current top-twelve GNNs and LLMs are presented in two tables.

Figure 1: Illustrating what are the major GNN and LLM model families available to AI system developers.

I. The GNN Table with Example Application Use Cases

The legend in the last two columns of the table for the current top-twelve GNNs are defined as:

  • Complexity/effort = engineering + needed infrastructure + tuning cost
  • Maturity/accuracy = reliability + predictability in real deployments

Table 1: Graph Neural Network (GNN) Model Types

GNN TypeBest Applied Problem TypesImplementation ComplexityMaturity / Typical Accuracy
GCN (Graph Convolutional Network)Node classification in product graphs, part similarity, baseline digital-twin analytics, baseline molecular property predictionLowHigh
GraphSAGEInductive embeddings for evolving product structures, supply chains, and asset fleetsLowHigh
GAT (Graph Attention Network)Systems where some connections matter more than others (failure propagation, reliability analysis)MediumHigh
GIN (Graph Isomorphism Network)Graph classification, molecular fingerprints, tasks requiring strong structural discriminationMediumHigh
MPNN (Message Passing Neural Network)Molecular property prediction, chemistry/material science, physics-inspired problemsMediumHigh
R-GCN (Relational GCN)Enterprise knowledge graphs linking requirements, parts, simulations, suppliers, and service dataMediumHigh
Heterogeneous GNN (HAN, HGT)Multi-type nodes/edges: recommender systems, enterprise knowledge graphsHighHigh
Graph Autoencoders (GAE/VGAE)Link prediction and anomaly detection in supply chains, manufacturing networks, and service fleetsMediumMedium
DiffPool / Hierarchical GNNsGraph classification requiring hierarchical structure (proteins, materials, circuits)HighMedium
Graph TransformersLong-range dependencies, large molecular graphs, protein folding, materials discoveryHighMedium
SE(3)/E(n) Equivariant GNNs3D molecular modeling, quantum chemistry, physics simulationsHighHigh (scientific domain)
SchNet / PhysNetAtomistic simulations, energy/force predictionHighHigh (scientific domain)

II. Choosing GNNs for Digital Industry Use Cases

For most digital-engineering applications, the starting point is usually a heterogeneous product or lifecycle graph. A practical selection pattern often looks like:

  • Start with GCN / GraphSAGE for early experimentation and baseline value.
  • Move to R-GCN or Heterogeneous GNNs once multiple data types (requirements, parts, suppliers, simulations) are included.
  • Consider GAT when uncertain about which components play the critical roles in system.
  • Use physics-aware GNNs (SE(3), SchNet) when geometry and simulation fidelity become critical.

III. The LLM Table with Example Application Use Cases

For the LLM table, including the current top-twelve models, we have a similar legend as above:

  • Complexity = engineering + infrastructure + tuning pain
  • Maturity/accuracy = reliability + predictability in real deployments

Table 2: Large Language Model (LLM) Types

LLM TypeBest Applied Problem TypesImplementation ComplexityMaturity / Typical Accuracy
Vanilla Transformer (GPT/BERT-style)General NLP, chat, summarization, coding, classificationMediumVery High
Encoder-only (BERT/RoBERTa)Semantic search across requirements, documentation, simulation reports, and service recordsLowVery High
Decoder-only (GPT-style)Engineering copilots, requirements assistants, code generation, workflow automationMediumVery High
Encoder–Decoder (T5/BART)Translation, summarization, structured text transformationMediumHigh
Instruction-tuned LLMsAssistants, enterprise copilots, task-followingMediumVery High
RLHF / RLAIF tuned LLMsSafe assistants, customer-facing chatbotsHighVery High
Retrieval-Augmented Generation (RAG)Enterprise knowledge copilots grounded in PLM, ALM, MES, and service documentationMediumVery High
Tool-using / Agent LLMsAutomated engineering workflows interacting with CAD, simulation, and data platformsHighHigh
Mixture-of-Experts (MoE)Massive-scale models, cost-efficient inference at scaleVery HighHigh
Long-context TransformersAnalysis of long engineering specifications, certification documents, and design historiesHighHigh
Multimodal LLMs (Vision-Language)Engineering drawings, CAD screenshots, inspection images, and technical documentationVery HighHigh
Domain-Finetuned LLMsEngineering, manufacturing, aerospace, automotive, and industrial automation domainsMedium–HighVery High (in-domain)

IV. Choosing LLMs for Digital Industry Use Cases

In digital industries, the “right” LLM is less about raw benchmark scores and more about how the model will be grounded, governed, and integrated into engineering workflows. A practical selection pattern often looks like:

  • Start with Instruction-tuned, decoder-only LLMs for developer productivity and copilots.
    This approach would be best when we need: Q&A, summarization, explanation, and workflow assistance for engineers.
  • Use Encoder-only models (embeddings) when retrieval is the real product. This would be best when we need: semantic search across requirements, specifications, simulation reports, service notes, and lessons learned.
  • Adopt RAG early for enterprise use, not as an afterthought. Best when we must ground outputs in authoritative and up-to-date sources: PLM/ALM artifacts, engineering standards, test reports, maintenance manuals, and internal wikis.
  • Choose Long-context Transformers when engineers work with “giant documents.” Best when the key inputs are large: certification documents, multi-hundred-page specifications, and system-level design histories.
  • Use Tool-using / Agent LLMs when the output must change the configuration (CAD/CAE/MES), not just talk about it. Best when the model needs to: query databases, run checks, generate scripts, call simulation workflows, open tickets, and produce traceable artifacts.
  • Reserve Domain-Finetuned LLMs for repeatable, high-value tasks with stable data. Best when we have: consistent internal corpora and recurring tasks (requirements normalization, defect triage, service summary classification).
  • Use Multimodal LLMs only when visual inputs are central. Best when the workflow involves: drawings, inspection photos, UI screenshots, and diagram interpretation.

V. Putting the Pieces Together for Fusion

Selecting models for fusion is not about choosing the most advanced architecture. It is about matching system structure with knowledge sources and regulatory/safety requirement. Here are a few examples of the typical problems in digital industries, and choice of GNN and LLM to provide a GNN+LLM solution:

Digital Industry ProblemLikely GNN ChoiceLikely LLM Choice
Digital twin of complex productR-GCN / Heterogeneous GNNRAG + Instruction-tuned LLM
Manufacturing planningGraphSAGE / Temporal GNNTool-using LLM + RAG
Service lifecycle intelligenceTemporal GNNLong-context LLM + RAG
Engineering knowledge assistantsKnowledge-graph GNNDomain-finetuned LLM

Obviously, specific GNN and LLM choices shown in these examples could differ based on the specific industry use case and evolution of AI methods.

VI. Final Remarks and Next Steps

In Part I, we introduced the idea that fusing GNNs and LLMs enables AI systems that understand both the structure of engineered systems and the knowledge surrounding them. In this article, we explored the major model families that developers can use to build these fused solutions. A key takeaway is that most real-world implementations do not require the most complex models immediately. Successful projects typically start with mature, production-ready approaches and evolve toward more specialized architectures as value is proven. Across digital industries, the opportunity is clear:

  • GNNs provide the structural backbone of the digital thread and digital twin.
  • LLMs provide the interface to human knowledge and engineering workflows.
  • Fusion enables a new class of engineering copilots and intelligent automation.

In Part III of this series, we will explore practical architecture and implementation patterns for building GNN+LLM fusion systems in real industrial environments. Note: This blog was created with the help of AI.

Mohsen Rezayat
Chief Solutions Architect / Technical Fellow

Dr. Rezayat has over 40 years of industrial experience at Siemens, with over 80 technical publications and a large number of patents. Furthermore, he was an Adjunct Full Professor at University of Cincinnati and currently serves on the Industrial Advisory Board of the College of Engineering and Applied Sciences there. Dr. Rezayat is also the founder and member of the Board of Directors (BOD) of poverty-alleviation non-profit OMID organization. His current research interests include Generative AI and its ethical use, impact of Industrial Metaverse on productivity and workers, solutions for sustainable growth in developing countries, and global green engineering.

More from this author

Leave a Reply

This article first appeared on the Siemens Digital Industries Software blog at https://blogs.sw.siemens.com/thought-leadership/topics-in-ai-systems-part-ii-gnn-and-llm-types/