Back

Triplex — SOTA LLM for Knowledge Graph Construction

Author photo
Owen Colegrove
Jul 9, 2024
4
min read
On this page

Triplex is an innovative new model that allows you to convert large amounts of unstructured data into knowledge graphs. Triplex exceeds the performance of gpt-4o at knowledge graph construction for less than one-tenth the cost.

Triplex is open source and available on HuggingFace here and ollama here.

Knowledge Graph Sketch
Illustration of a knowledge graph structure

Knowledge graphs excel at answering queries that traditional search methods often struggle with, particularly population-level relational queries. For instance, "Provide a list of AI employees who attended technology schools." Interest in knowledge graphs has intensified after Microsoft's recent GraphRAG paper [1].

However, the process of knowledge graph construction has traditionally been complex and resource-intensive [2], limiting its widespread adoption. Recent estimates suggest that Microsoft's GraphRAG procedure is particularly costly, requiring at least one generated output token for every ingested input token. This high cost makes it impractical for most applications.

Triplex aims to radically disrupt this paradigm by reducing the generation cost of knowledge graphs tenfold. This cost reduction is achieved through Triplex's efficient method of converting unstructured text into "semantic triples" - the building blocks of knowledge graphs. To demonstrate how Triplex creates these graphs, let's observe how it processes simple sentences:

# Inputs
# Entity Types <- CITY, COUNTRY
# Relationships <- CAPITAL_OF, LOCATED_IN
# Text <- Paris is the capital of France

# Output -> (subject > predicate > object)
CITY: Paris > CAPITAL_OF > COUNTRY: France
CITY: Paris > LOCATED_IN > COUNTRY: France

And on a more complex input:

# Inputs
# Entity Types <- ARTIST, ARTWORK, ART_MOVEMENT
# Relationships <- CREATED_BY, BELONGS_TO_MOVEMENT
# Text <- Vincent van Gogh, a post-impressionist painter, created "The Starry Night" in 1889. This iconic artwork, with its swirling clouds, brilliant stars, and crescent moon, exemplifies the artist's unique style and emotional intensity. Van Gogh's bold use of color and expressive brushstrokes influenced many subsequent art movements, including Expressionism and Fauvism.

# Output -> (subject > predicate > object)
ARTIST:Vincent van Gogh > BELONGS_TO_MOVEMENT > ART_MOVEMENT:post-impressionist
ARTWORK:The Starry Night > CREATED_BY > ARTIST:Vincent van Gogh
ARTIST:Vincent van Gogh > BELONGS_TO_MOVEMENT > ART_MOVEMENT:Expressionism
ARTIST:Vincent van Gogh > BELONGS_TO_MOVEMENT > ART_MOVEMENT:Fauvism

Our performance measurements revealed that Triplex significantly outperformed gpt-4o in cost and performance.

Accuracy Comparison

Price Comparison

The triple extraction model achieves results comparable to GPT-4, but at a fraction of the cost. This significant cost reduction is made possible by Triplex's smaller model size and its ability to operate without the need for few-shot context.

Building upon the SFT model, we generated additional preference-based dataset using majority voting and topological sorting to further train Triplex using DPO and KTO. These additional training steps yielded substantial improvements in model performance. To accurately assess these nuanced enhancements, we conducted a rigorous evaluation using Claude-3.5 Sonnet. Our evaluation involved head-to-head comparisons between three models: triplex-base, triplex-kto, and GPT-4o. The results are presented in the table below:

Model 1 Model 2 Model 1 Win Model 2 Win Tie
triplex-base gpt-4o 54% 43% 3%
triplex-kto triplex-base 66% 26% 8%
triplex-kto gpt-4o 56% 40% 4%

The exceptional performance of Triplex results from extensive training on a diverse and comprehensive dataset. Our model leverages proprietary datasets generated from authoritative sources such as DBPedia and Wikidata, as well as web-based text sources and synthetically generated datasets. This broad foundation ensures Triplex's versatility and robustness across a wide range of applications.

Usage

We have designed the R2R RAG engine, together with Neo4J, to immediately leverage Triplex for local knowledge graph construction, a use case that has only now become more viable due to this work. Please read more about how to get started in the documentation here, or try out triplex directly here.

For additional information or to explore how Triplex can transform your data insights, please reach out to us at founders@sciphi.ai. Join us in shaping the future of knowledge representation and extraction!

References

[1] Zhong, L., Wu, J., Li, Q., Peng, H. and Wu, X., 2023. A comprehensive survey on automatic knowledge graph construction. ACM Computing Surveys, 56(4), pp.1-62. [Link]

[2] Edge, Darren, et al. "From local to global: A graph rag approach to query-focused summarization." arXiv preprint arXiv:2404.16130 (2024). [Link]

Read also
Advanced RAG Techniques in R2R
CASE STUDY
Advanced rag
Hybrid search
Knowledge graphs
Rag

Advanced RAG Techniques in R2R

Covering the advanced RAG options in R2R

Author photo
Owen Colegrove
Oct 9, 2024
GraphRAG in R2R
CASE STUDY
Knowledge graphs
Rag

GraphRAG in R2R

Seamlessly design and deploy GraphRAG using R2R

Author photo
Owen Colegrove
Aug 9, 2024
Benchmarking R2R
CASE STUDY
Rag
Devops

Benchmarking R2R

Comparing Performance of Popular RAG Solutions

Author photo
Owen Colegrove
Jul 15, 2024