Hetzerk: AGI Reasoning and Decentralized Science

chevron-icon
Back
Top
chevron-icon
project-presentation-img
user-profile-img
Justin Diamond
Project Owner

Hetzerk: AGI Reasoning and Decentralized Science

Status

  • Overall Status

    ⏳ Contract Pending

  • Funding Transfered

    $0 USD

  • Max Funding Amount

    $50,000 USD

Funding Schedule

View Milestones
Milestone Release 1
$10,000 USD Pending TBD
Milestone Release 2
$20,000 USD Pending TBD
Milestone Release 3
$20,000 USD Pending TBD

Project AI Services

No Service Available

Overview

Hetzerk seeks $50 k to build a modular, open-source toolkit that turns raw scientific data into high-quality MeTTa/MORK knowledge graphs and benchmarks their impact on Hyperon reasoning. The system ingests, refines, deduplicates, and scores graph data, exposing APIs for AGI agents and DeSci apps. Leveraging prior work with SingularityNET (Deep Funding R1) and Cardano Catalyst, the project connects decentralized molecular & physics datasets to Hyperon’s neuro-symbolic engine—boosting analogical reasoning, fact-checking, and hypothesis generation across the ecosystem.

RFP Guidelines

Advanced knowledge graph tooling for AGI systems

Complete & Awarded
  • Type SingularityNET RFP
  • Total RFP Funding $350,000 USD
  • Proposals 39
  • Awarded Projects 5
author-img
SingularityNET
Apr. 16, 2025

This RFP seeks the development of advanced tools and techniques for interfacing with, refining, and evaluating knowledge graphs that support reasoning in AGI systems. Projects may target any part of the graph lifecycle — from extraction to refinement to benchmarking — and should optionally support symbolic reasoning within the OpenCog Hyperon framework, including compatibility with the MeTTa language and MORK knowledge graph. Bids are expected to range from $10,000 - $200,000.

Proposal Description

Our Team

 

Justin Diamond – PhD candidate in Computer Science (ML focus, finishing June).brings cutting-edge machine learning expertise.

Ziyu She – PhD student in AI with publications

Ryan Diamond, Advisor – guides financial strategy and business planning.

Floriane Le Floch, Advisor – Web3 consultant and Auxetic.ai founder

 

Company Name (if applicable)

Hetzerk

Project details

Hetzerk Knowledge Graph Toolkit for AGI Systems

Summary

Hetzerk proposes a $50,000 project to develop a knowledge graph toolkit that advances symbolic reasoning for Artificial General Intelligence (AGI), tightly integrating with the OpenCog Hyperon framework. The toolkit will ingest and structure large-scale scientific datasets (molecular and physics) into a dynamic knowledge graph, then refine and evaluate it for use by Hyperon-based AGI agents. By leveraging MeTTa (the Hyperon language) and the MORK hypergraph database, our system will enable Hyperon agents to perform complex reasoning (analogical problem-solving, multi-hop question answering) on decentralized scientific knowledge. The design is modular and emphasizes open science, while full open-sourcing will be delayed to maintain a competitive edge. This balanced approach ensures that decentralized science (DeSci) data and AGI development reinforce each other, accelerating progress in both domains.

Motivation

For cognitive architectures like Hyperon, symbolic knowledge graphs serve as a backbone for advanced reasoning capabilities such as analogical inference, causal reasoning, and long-term memory. However, constructing and maintaining high-quality knowledge graphs is difficult – especially in dynamic, noisy domains like cutting-edge scientific research. Our experience in decentralized molecular and physics simulations underscores these challenges: scientific data are vast, distributed, and constantly evolving, which makes consistent extraction of structured knowledge very challenging. We often see critical information scattered across publications, experiment logs, and databases. Bridging these into a coherent graph is non-trivial, yet doing so is essential for any AI that needs deep understanding of the domain.

There is a clear gap between raw scientific data and the structured knowledge that AGI systems require for higher-order reasoning. Large language models can parse text, but without a structured knowledge base they struggle with multi-step logical queries and consistency. Traditional databases also fall short for AGI because they cannot flexibly represent complex relationships. A well-structured knowledge graph, by contrast, lets an AI navigate interlinked facts and discover new connections. Hyperon’s Atomspace + MeTTa architecture is designed to exploit such structured knowledge, but it currently lacks automated tools to build and maintain these graphs.

By focusing on decentralized science data as our use case, we address the RFP’s goal of a domain-agnostic knowledge graph tool. Scientific domains like materials science and physics are extreme in complexity and distribution of knowledge, so solving these cases will produce solutions generalizable to other fields. Furthermore, aligning our toolkit with Hyperon means improvements in scientific knowledge management immediately benefit AGI agents’ reasoning abilities. In turn, smarter AGI agents can autonomously curate and expand the knowledge graph, creating a positive feedback loop at the heart of our proposal.

Objectives

Our objectives are to deliver a comprehensive toolkit and demonstrate its impact on AGI reasoning:

  • Data Ingestion: Create pipelines to extract and unify raw scientific data (from distributed molecular and physics sources) into a cohesive knowledge graph with a well-defined schema.

  • Graph Refinement: Implement algorithms for cleaning and optimizing the graph – deduplicating entities, resolving contradictions or inconsistencies, and condensing noisy subgraphs into meaningful, concise representations.

  • Hyperon Integration: Ensure compatibility with the Hyperon framework. The toolkit will export or interface the knowledge graph as MeTTa expressions or directly into the MORK hypergraph store, enabling fast symbolic queries and updates by AGI agents.

  • Benchmark Reasoning: Evaluate the system on tasks requiring symbolic multi-step reasoning. Key benchmarks will include analogical reasoning tests and multi-hop question answering that leverage the knowledge graph, comparing performance with vs. without our toolkit.

  • Scalability & Performance: Leverage MORK’s high-performance engine to handle large knowledge bases (millions of nodes/edges) with interactive query speeds. Optimize data structures and query strategies so the system can support real-time reasoning.

  • Open Science & Modular Design: Follow open science practices by documenting and modularizing each component (ingestion, refinement, integration, evaluation). We will release modules and findings openly to the community, but note that full open-sourcing of the entire codebase will be deferred until we establish initial traction, balancing community benefit with sustainable competitive advantage.

Technical Approach

Data Ingestion and Graph Construction: We will build connectors to pull data from decentralized sources (e.g. chemistry databases, physics simulations) and transform it into a unified knowledge graph. Domain-specific parsers and ontologies will map raw inputs to a common schema of nodes (e.g. compounds, materials, conditions) and labeled relations (e.g. "reacts_with", "has_property"). The result is an initial knowledge graph of scientific facts, ready for refinement.

Symbolic Refinement: Next, we apply refinement algorithms to improve quality: merge duplicates, resolve or flag contradictions, and infer new relationships where possible. We will also attach provenance and confidence metadata to nodes and edges. The refined graph will be more compact, consistent, and enriched than the initial version.

Integration with Hyperon (MeTTa + MORK): We will develop an adapter to import the refined graph into the OpenCog Hyperon environment (as Atomspace atoms via MeTTa). The toolkit will use MORK as the high-performance backend for storage and retrieval, leveraging its optimized hypergraph operations. This integration allows an AGI agent to query the graph via MeTTa and get rapid results even at scale, and also to insert new learned facts which the toolkit will validate and incorporate.

Reasoning Benchmark Evaluation: To validate the toolkit, we will implement two categories of reasoning tasks: analogical reasoning (finding analogous relationships across domains) and multi-hop question answering (answering complex queries by chaining multiple facts). These will evaluate the AGI’s accuracy and efficiency with our knowledge graph, demonstrating improvements over a baseline without structured knowledge.

Milestones and Deliverables

  1. Milestone 1 (Month 1-2): Data Ingestion Prototype – Deliver a basic ingestion pipeline handling at least one molecular and one physics dataset, outputting an initial integrated knowledge graph.

  2. Milestone 2 (Month 3): Graph Refinement – Deliver the graph refinement module (duplicate consolidation, contradiction detection) and produce a cleaned knowledge graph, with a brief summary of improvements.

  3. Milestone 3 (Month 4-5): Hyperon Integration – Deliver integration code enabling MeTTa/MORK to interface with the knowledge graph, and demonstrate a Hyperon agent executing a multi-hop query on the graph.

  4. Milestone 4 (Month 6): Final Evaluation & Dissemination – Deliver comprehensive benchmark results for analogical reasoning and multi-hop Q&A tasks using the toolkit, and deliver documentation plus a phased open-source release plan.

Each milestone provides a concrete step toward the final system, reducing risk progressively. By the project’s end, we will have a validated toolkit and an initial community engagement around it.

Conclusion

This proposal outlines a focused effort to build a symbolic knowledge graph toolkit that strengthens both decentralized scientific research and AGI development. By grounding our work in real-world scientific datasets, we tackle the most demanding knowledge integration challenges and ensure the results are applicable to many domains. Integrating directly with Hyperon means that our outputs will immediately enhance an existing AGI platform, enabling it to reason over structured knowledge far more effectively. Conversely, the AGI’s usage of the knowledge graph can drive automated curation and discovery in the scientific data itself – a two-way synergy. As noted by SingularityNET’s leadership, linking specialized knowledge graphs with an AGI’s meta-knowledge can yield emergent synergies that boost intelligence. Our project embodies this vision: decentralized science and AGI will mutually reinforce one another.

We affirm our commitment to open science and modular design. We will engage the community with interim results and gradually open-source the toolkit, ensuring short-term competitive advantage and long-term public benefit. In doing so, Hetzerk will deliver a powerful new asset for the pursuit of beneficial AGI and collaborative scientific innovation.

Background & Experience

Our team merges deep AI research expertise with business and startup acumen to build advanced knowledge graph AGI tools. We combine academic excellence with industry savvy:

  • Justin Diamond – PhD candidate in Computer Science (ML focus, finishing June) with publications at ICANN, NeurIPS, ICML, ICLR; brings cutting-edge machine learning expertise.

    • google scholar: https://scholar.google.com/citations?user=O-chxmAAAAAJ&hl=en
  • Ziyu She – PhD student in AI (strong mathematical background) with publications in IJCAI, Complex & Intelligent Systems, IEEE ICARM; drives knowledge representation and algorithm design.

    • google scholar: https://scholar.google.com/citations?user=V0KN5esAAAAJ&hl=en

  • Ryan Diamond – Former senior financial analyst; guides financial strategy and business planning.

  • Floriane Le Floch – Web3 consultant and Auxetic.ai founder; brings startup experience and decentralized tech insight.

Links and references

  • SingularityNET. “Advanced knowledge graph tooling for AGI systems.” DeepFunding RFP (2025).

  • TrueAGI. “MeTTa Optimal Reduction Kernel (MORK) – GitHub README.” (2023).

  • SingularityNET & Trace Labs. “Advancing Decentralized Knowledge Graphs – Joint Press Release.” (May 30, 2024).

Additional videos

 

https://youtu.be/jdyoPr5mNZ0?t=532

https://www.youtube.com/watch?v=F575gBjOCr4

Describe the particulars.

Previous work with DeepFunding.

Proposal Video

Not Avaliable Yet

Check back later during the Feedback & Selection period for the RFP that is proposal is applied to.

Reviews & Rating

New reviews and ratings are disabled for Awarded Projects

Overall Community

0

from 0 reviews
  • 5
    0
  • 4
    0
  • 3
    0
  • 2
    0
  • 1
    0

Feasibility

0

from 0 reviews

Viability

0

from 0 reviews

Desirabilty

0

from 0 reviews

Usefulness

0

from 0 reviews

Sort by

0 ratings

    No Reviews Avaliable

    Check back later by refreshing the page.

  • Total Milestones

    3

  • Total Budget

    $50,000 USD

  • Last Updated

    28 Aug 2025

Milestone 1 - Initialization and Knowledge Graph Design

Status
😐 Not Started
Description

This first milestone lays the groundwork for the Hetzerk project by assembling the core team and defining the system architecture with a focus on knowledge graph integration. The team will establish communication channels and outline a comprehensive project blueprint. This blueprint will detail how a knowledge graph will be used to integrate AI (including AGI-oriented algorithms) with large-scale scientific simulation data. Key activities include brainstorming requirements, identifying data sources (e.g., molecular simulation outputs, quantum chemistry results), and designing the knowledge graph schema to represent physical systems, experimental results, and algorithmic insights. By the end of this phase, the project will have a clear technical roadmap that emphasizes how scientific data will be structured and refined in the knowledge graph and how AI components will interact with this knowledge base. This milestone ensures everyone on the team understands the objectives and technical approach, setting a strong foundation for subsequent development.

Deliverables

A detailed Project Blueprint document and initial system setup. This blueprint will include: Architecture Specification: Diagrams and descriptions of the platform’s architecture, highlighting the knowledge graph’s role in storing and connecting scientific data with AI modules. Knowledge Graph Schema & Ontology: A defined schema or ontology outlining key entities (e.g. molecules, simulation parameters, results) and relationships for the knowledge graph, ensuring it can accommodate various scientific datasets and support reasoning tasks. Data & Integration Plan: Documentation of how diverse scientific data (from simulations, experiments, etc.) will be collected, transformed, and loaded into the knowledge graph (i.e., a plan for the data refinement pipeline). This includes identifying file formats and preliminary parsing strategies for each data type. AGI Integration Plan: A design section detailing how AI/AGI components will interface with the knowledge graph – for example, how a machine learning model or cognitive engine will query/update the graph for enhanced reasoning. Team & Resource Allocation: A roster of the project team with roles and a development timeline, ensuring that the expertise (AI, data engineering, domain science) is aligned with each component of the plan. Initial Prototype (if feasible): Optionally, a rudimentary prototype or proof-of-concept demonstrating a small subset of the knowledge graph to validate technology choices.

Budget

$10,000 USD

Success Criterion

This milestone is successful when the project has a well-defined plan and initial setup that guides all future work. Specific success criteria include: Completed Project Blueprint: A comprehensive blueprint is finalized and reviewed by key stakeholders. It should clearly capture all major components (knowledge graph, data pipeline, AI integration) and outline how they fit together. Reviewers should agree that the design is technically sound and addresses the project goals of AGI-aligned knowledge integration and data refinement. Knowledge Graph Design Approval: The knowledge graph schema/ontology is validated through peer review or expert feedback to ensure it appropriately models the necessary scientific information and can be extended as new data types are added. Any required revisions from the review are incorporated. Infrastructure Ready: Basic project infrastructure is set up (e.g., version control repositories, development environments, perhaps a preliminary graph database instance) so that development can begin immediately in the next milestone. Milestone Review Sign-off: Project advisors or funders have reviewed the milestone outcomes (documentation and plans) and given a go-ahead to proceed, indicating confidence that the project is well-planned and ready for implementation.

Link URL

Milestone 2 - Computational Framework and Data Pipeline

Status
😐 Not Started
Description

In the second milestone, the project moves from planning into building the core system – integrating the computational framework with the knowledge graph and implementing the scientific data refinement pipeline. This is a development-intensive phase where the team will create the backbone of the platform. Key tasks include setting up the knowledge graph database and populating it with sample data, developing tools to ingest and refine data from various scientific computations, and integrating AI algorithms to both produce and consume knowledge graph data. For example, simulation outputs (from molecular dynamics, quantum chemistry, etc.) will be parsed by custom scripts or converters and then stored as structured entries in the knowledge graph. Simultaneously, the platform’s AI components (e.g., machine learning models or reasoning engines) will be connected so they can query the knowledge graph for insights or update it with new results. Throughout this milestone, iterative testing will be performed to ensure the system can handle complex computational tasks and data flows. By milestone completion, the Hetzerk platform’s core functionality – a loop that takes raw scientific data, refines it into a knowledge graph, and leverages it for AI-driven analysis – will be implemented and demonstrated on a pilot scale.

Deliverables

A prototype of the integrated platform along with associated documentation and test results including: A deployed graph database containing representative data. For instance, a running instance of the knowledge graph with several sample datasets (e.g., a molecular dataset, a materials science dataset) ingested. The schema from Milestone 1 is realized in this database, and one can query it for stored knowledge. A suite of data processing scripts or modules capable of converting raw output files from at least two types of scientific algorithms into the structured format required by the knowledge graph. Example deliverables may be parsers for molecular simulation trajectories and quantum calculation outputs that import results (e.g., energy values, molecular structures, metadata) into the graph. The data pipeline should also perform refinement steps such as cleaning inconsistent units or merging duplicate entries, so the knowledge graph information is consistent and reliable. The core computational engines (simulation or AI algorithms) integrated with the system. For instance, if a molecular dynamics simulation code is part of the project, it is configured to output results that feed directly into the knowledge graph via the pipeline. Additionally, an AI module is set up to pull data from the knowledge graph. A concrete example could be a neural network that reads simulation results from the graph to predict an outcome or suggest new simulation parameters.

Budget

$20,000 USD

Success Criterion

This milestone is considered achieved when the integrated platform can handle real data and computations in a controlled test environment, proving out the concept of coupling a knowledge graph with scientific computing and AI. The success criteria are: The system successfully processes raw scientific data through to the knowledge graph and subsequent AI analysis without manual intervention. For instance, given a new simulation result file, the pipeline ingests it into the knowledge graph, and an AI query on the graph yields meaningful information. The graph is populated with refined data that is accurate and useful. This can be verified by checking that the ingested data in the graph matches original data sources (ground truth) and is organized correctly (e.g., units converted where necessary, linked to relevant entities without errors). A successful criterion could be that 100% of test dataset entries appear correctly in the graph and can be retrieved via queries that demonstrate the linking of related information . Basic performance benchmarks are met. For example, the system can ingest a typical dataset within a reasonable time, and query responses from the knowledge graph for test queries are within acceptable latency. Success might be defined as ingesting a dataset of size X in Y minutes, and answering a complex query (spanning multiple linked records in the graph) in Z seconds. These metrics ensure the solution is scalable to larger problems in the next phase.

Link URL

Milestone 3 - MVP Deployment and AGI Integration Validation

Status
😐 Not Started
Description

The final milestone focuses on deploying the Hetzerk platform as a Minimum Viable Product (MVP) and validating its performance and integration with advanced AI (AGI) components in a real-world context. In this phase, the system transitions from a prototype in a test environment to a robust platform accessible to end-users or stakeholders (for example, via a user interface or API on a cloud or decentralized network). The knowledge graph will be fully populated with a larger volume of refined scientific data, and the pipeline will be refined for reliability. Crucially, this stage will integrate an AGI-oriented element: for instance, incorporating a cognitive engine or large language model that uses the knowledge graph to perform complex reasoning or to assist users in drawing conclusions from the data. The team will also engage initial users (such as an industrial partner or domain experts) to test the platform on real use cases – e.g., running a new molecular simulation on the platform and using the AI assistant to interpret the results via the knowledge graph. Throughout this milestone, feedback is gathered and the system is tuned for accuracy, security, and scalability. By the end, the MVP will demonstrate the platform’s value in accelerating scientific discovery through AI and knowledge graph integration, meeting both technical benchmarks and user expectations.

Deliverables

A deployed MVP system along with validation reports and user feedback. The key deliverables are: The Hetzerk platform is deployed in an accessible environment. This could be a cloud deployment or a decentralized network deployment. The deployment includes a front-end interface or console for users to submit computational jobs and view results, and the back-end comprising the knowledge graph database and integrated AI services. A working integration with an advanced AI system is delivered as part of the MVP. For example, an AI assistant or agent service that can be queried in natural language or logical queries about the data in the knowledge graph. This could be implemented via a large language model connected to the knowledge graph , or an OpenCog Hyperon module using the graph for reasoning. The deliverable would include the AI component itself and a description of how it connects to the knowledge graph. A demo execution using the MVP on a realistic scenario. For instance, Industrial-scale Molecular Simulation Demo: the team runs a complex molecular simulation through the platform, the data automatically populates the knowledge graph, and then the integrated AI analyzes the results to produce insights (like identifying promising compounds or explaining phenomena). The outputs of this demonstration – including the knowledge graph entries created, the AI’s conclusions or responses, and performance metrics – are compiled.

Budget

$20,000 USD

Success Criterion

The MVP deployment and testing will be deemed successful when the platform operates in a real-world setting and delivers tangible results through its AI-enhanced knowledge graph approach. Success criteria include: The system successfully handles a substantial, real-world dataset. For instance, at least one large-scale simulation or dataset is ingested and managed in the knowledge graph. The success can be measured by data volume and complexity (multiple types of data linked in the graph). The refined knowledge graph should show that even with increased scale, data remains consistent and queryable. If applicable, integration with blockchain for data integrity (as hinted by Cardano integration) is demonstrated without performance loss. The integrated AI/AGI component provides a successful demonstration of advanced insight or decision support. A clear success criterion would be that the AI agent correctly answers complex questions or solves a problem that would be hard to do without the knowledge graph. For example, the AI might identify a non-obvious correlation in the simulation data that leads to a new hypothesis or optimization. Achieving a meaningful result like this – confirmed by domain experts or experimental validation – would illustrate the power of combining knowledge graphs with AGI techniques. Even in absence of a dramatic discovery, the criterion could be that the AI assistant can interact with users about the data as per the project requirements.

Link URL

Join the Discussion (0)

Expert Ratings

Reviews & Ratings

New reviews and ratings are disabled for Awarded Projects

    No Reviews Avaliable

    Check back later by refreshing the page.

Welcome to our website!

Nice to meet you! If you have any question about our services, feel free to contact us.