The word Klug means smart, wise or clever in German and Dossier, which means a file or a folder of Data/Records, comes from the French language. Hence, “KlugDossier - Smart Data Dossier”.
KlugDossier’s need for existence comes from the problems that many firms face concerning raw data, especially in the form of fetching clear and refined data needed for complex businesses involved in algorithmic trading, predictive analysis, cognitive analysis, and other activities that depend heavily on such ready-to-use data.
As an AI-integrated oracle platform that fetches, queries, validates and provides data from external sources to smart, on-chain contracts, it also fetches data from smart contracts themselves. It transports it to other external data sources, regardless of whether they’re on or off-chain, like IT systems and others connected to the KlugDossier platform.
Targeting and solving complex problems is not the sole purpose of KlugDossier—its other goal is to fulfil normal wants from mainstream users like generating NFTs, fetching validated data from data providers and natively linking with the existing Polkadot ecosystem. Our functional testnet is live on polkadotJS, PolkaJS/KlugDossier.
KlugDossier communicates with smart contracts deployed on Ethereum using a Moonbeam / Ethereum Virtual Machine bridge bypassing data from different Ethereum platforms.
KlugDossier also includes TendermintBFT Consensus using Cosmos with its Hub and Zone model. In addition, it integrates the Cosmos chain with the Polkadot Relay Chain as one of the zones that passes information from one data provider to another.
As a cross-chain platform, KlugDossier interacts with different blockchain frameworks like Solana using ChainLink. It does this by using Proof of History Consensus and the Proof of Stake consensus using the Solana blockchain, which delivers high speed, low latency, and inexpensive transactions without sharding. Contracts deployed on Solana will be WASM compatible as well.
Data Providers of different domains send raw, unprocessed and unfiltered data, e.g., with no clear labels, unstructured formats, not refined etc.;
Data sent is not compatible with AI firms' use in the application of machine learning models on top of data and hence are unable to perform predictive analysis using their AI;
Lack of validation of data;
Inability to consolidate data acquired from different sources;
Constantly changing/fluctuating data;
Lack of consistency in data resulting in the suffering of the Prediction Market.
Data validation based on consumer requirements;
Implementation of a decentralised storage mechanism to store and consolidate data received from different sources;
Processing and refining data for consumer and or AI firm use based on the expected formats to process ML models; continuous on-chain/off-chain feeding of data to consumers;
The prediction Market will benefit massively as the converted data will be directly applicable to consumer-based models like ANN, CNN, RNN, LTSM.
KlugDossier's oracle node will collect data from different producers (data providers) and validate it based on the request provided by consumers.
A reward-and-penalty mechanism is provided to every oracle node.
Rewards producers if the validation is successful based on the request.
Penalises producers if the validation is a failure.
The input data feed can be from an Excel file or an API supporting JSON structure. A producer (if needed) can manipulate or apply logic on the uploaded data using WASM contracts to make the format compatible.
Send data to the consumer side using smart contracts or non-smart contract apps like legacy IT systems based on the adapter available from KlugDossier.
Proof of Stake is the consensus mechanism to be used by the oracle, so producers and consumers equally benefit.
WebAssembly contracts are developed using the ink language and KlugDossier's IDE, which will be available to producers/consumers for applying rules to data
KlugDossier IDE will have a syntax validation mechanism and also will be able to develop contracts compatible with parachain/parathreads available in KlugDossier for data processing
Decentralised local storage/IPFS using the storage module in FRAME will be accessible to WASM contracts for storing data and for conversion into the numerical/time series format
Some currently existing data projects can be used for our benefit, and we can complement each other by using each other's protocols.
To use Ocean Protocol as an example: Ocean Protocol is a data exchange platform that's also a producer. So we could utilise Ocean Protocol's data to get relevant and required inputs, with us being an oracle platform. This makes KlugDossier, not a competitor but a processor/consumer of Ocean Protocol. The data from Ocean Protocol can then be used for further processing and refining by KlugDossier.
KlugDossier Architecture gives an overview of different kinds of producers and consumers that use this system; it aims to serve other business domains and uses Polkadot to transfer data between different chains and legacy systems using relay chains. Communication between Polkadot and Cosmos happens by using IBC (Inter-Blockchain Communication), and Polkadot (Relay chain) works as Peg Zones for Cosmos hub.
Substrate chains, once developed based on GRANDPA / AURA / BABE Consensus, are converted into Parachains using Cumulus, data is stored in IPFS / Decentralized Data Storage. WASM Contract IDE can be used to add contracts by a producer to refine data by adding business logic, and later the validator validates the agreement. It is converted into Parachain, which is deployed into Relaychain.
Producers can upload data into the KlugDossier system using Excel/JSON format after validation. It can communicate with legacy IT systems apart from the On-Chain systems; consumers can also consume data from KlugDossier using Defi apps, legacy IT systems, and ML Models. KlugDossier serves Datascience / AI firms apart from the other firms.