Page cover

MEF-AI Architecture – A Technical Deep Dive

Abstract

This document describes the technical details of the proprietary Quantum-Fusion Architecture, which lies at the heart of the MEF-AI project and combines high-accuracy predictions, dynamic risk management, and the principles of Explainable Artificial Intelligence (XAI). Unlike classic single-model approaches, our architecture is a multi-layered, integrated system of systems with specialized components at every stage, from data ingestion to final signal generation. This whitepaper will thoroughly detail the layered structure of the architecture, the core algorithms used (Temporal Fusion Transformer, LightGBM, Deep Reinforcement Learning), the data flow, and the system's performance metrics.


1. Introduction: Why "Quantum-Fusion"?

Financial markets are, by nature, chaotic, non-linear, and highly noisy systems. Classical technical analysis methods and traditional machine learning models are often insufficient to model this multidimensional complexity on their own. The "blind spot" of one model can be the area of expertise for another.

The core philosophy and assertion of MEF-AI is as follows: "Instead of a single "all-knowing" model, we can capture both the quantitative (technical) and qualitative (psychological) dynamics of the market simultaneously by using an 'integrated intelligence network' composed of specialized AI systems that complement, verify, and check one another."

The name "Quantum-Fusion" is derived from the architecture's ability to take information from multiple, different data universes (market data, news, social media) and fuse it into a single, coherent, and probabilistic outcome, much like superposition in quantum physics collapses into a single, definite state. A Note on Terminology: The name "Quantum-Fusion" is the proprietary designation for our architecture. It is used metaphorically to reflect the system's ability to synthesize vast and disparate data universes (market data, sentiment, etc.) into a single, coherent probabilistic outcome, akin to how quantum superposition resolves into a definite state upon observation. This architecture does not imply the current use of quantum computing hardware.

2. The Layered Architecture: Anatomy of a Signal

The unique architecture of MEF-AI represents a methodical and intelligent journey that a signal undertakes from its inception to the end-user. Each layer feeds the next with more refined and enriched data, thereby increasing the overall accuracy and reliability of the system.

2.1. Layer 1: Data Ingestion and Feature Engineering ("The Crude Oil Refinery")

Objective: To transform the ocean of raw, meaningless data into high-octane, meaningful fuel for our machine learning models.

  • Data Sources and Technologies:

    • Real-Time Market Data: Instantaneous OHLCV (Open, High, Low, Close, Volume), order book depth, and trade data from exchanges flow into our system with sub-millisecond latency via Apache Kafka and WebSocket streams.

    • Alternative Data: Financial news feeds (Reuters, Bloomberg API) and social media platforms (Twitter API) are processed in real-time by our finance-oriented, transformer-based Natural Language Processing (NLP) models, such as FinBERT. VADER Sentiment scores extracted from the text quantitatively measure market sentiment (fear, greed, uncertainty).

    • Historical Data: Past price movements, correlation analyses, and macroeconomic data (interest rates, inflation expectations) are stored on Snowflake and PostgreSQL for analysis.

  • Feature Engineering and Feature Store: Over 250 meaningful features are derived from this raw data. Furthermore, a portion of these can be directly downloaded from our platform as an Excel file within seconds. These include standard technical indicators (RSI, MACD, Ichimoku) as well as proprietary metrics that analyze market microstructure (e.g., order book imbalance, spread-to-volatility ratio). All derived features are versioned and stored in our Feature Store, built with the Feast Framework. This ensures that all our models are fed from a consistent, up-to-date, and standardized data set, eliminating the risk of "model-data drift."

  • The 250+ features undergo iterative selection via:

    1. Mutual Information Scoring to filter low-predictiveness features

    2. Clustering Analysis to remove multicollinearity (e.g., retaining only one correlated volatility metric)

    3. SHAP-based Pruning, dropping features contributing <1% to model outputs quarterly

2.2. Layer 2: Core Prediction Models ("The Parallel Experts")

Objective: To send the enriched data to two parallel and independent AI models that analyze the market from different areas of expertise, providing a multi-faceted analysis.

  • Model A: Temporal Fusion Transformer (TFT) – "The Time Traveler"

    • What it Does: This is a state-of-the-art deep learning architecture developed by Google AI that has revolutionized time-series forecasting. It combines past data with known future events (e.g., FED meeting calendar) and asset-specific static data (e.g., sector, market cap) to generate multi-horizon probabilistic scenarios for the future.

    • Technical Superiority: Unlike traditional LSTM/GRU models, its "Attention" mechanisms allow it to automatically learn and weigh which past time windows and which features are more important for future prediction.

  • Model B: LightGBM – "The Market Psychologist"

    • What it Does: This model does not attempt to predict the price itself, but rather classifies the current "regime" or "character" of the market with high accuracy. Using hundreds of features from the Feature Store, it categorizes the market into states such as: "High-Volatility Bull Market," "Low-Volume Consolidation Zone," or "Panic Sell-off Mode."

    • Explainability: Every decision made by the model is analyzed with SHAP (SHapley Additive exPlanations) values. This allows us to see clearly which features (e.g., a rising VIX index, negative news flow) were most influential in making a "Panic Sell-off Mode" decision.

2.3. Layer 3: Hybrid Signal and Dynamic Risk Assessment ("The Strategy Room")

Objective: To convert the raw predictions from the two expert models into an actionable, reliable final signal with adjusted risk parameters.

  • Component A: Meta-Learner (Ensemble Model)

    • How it Works: This layer acts as a "council of experts." It takes the future scenarios from the TFT and the market regime classification from LightGBM as inputs. It combines the opinions of these two experts using advanced ensemble techniques like Bayesian Model Averaging and produces a single, final Confidence Score (0%-100%). If both experts provide strong signals in the same direction, the confidence score increases.

  • Component B: Dynamic Risk Management (Deep Reinforcement Learning Agent)

    • What it Does: This is the most innovative part of our architecture. It completely rejects static rules like "set stop-loss at 2%." A DQN (Deep Q-Network) based Reinforcement Learning agent uses the confidence score from the Meta-Learner and the market's real-time volatility data to dynamically determine the optimal position size and risk parameters (stop-loss, take-profit levels) for each signal.

    • Reward Function: The agent has been trained on millions of past data points to maximize the Sharpe Ratio and minimize the Maximum Drawdown. It learns to behave more cautiously during high volatility and to make bolder decisions in stable trends.

  • Component C: Explainability (XAI) and Signal Packaging

    • The risk parameters determined by the DRL agent are combined with the confidence score from the Meta-Learner. The "reasons for the decision" from the SHAP module are also added to create the "Final Signal Package."

    • Example Signal Package:

      • Signal Direction: BUY

      • Confidence Score: 92%

      • Hash is a signed and unchangeable date and time stamp.

      • Risk Parameters: (Dynamically adjusted SL/TP)

      • Explanation: “A 40% contribution from increasing volume, a 30% contribution from positive news sentiment, and a 20% contribution from the RSI indicator turning up from an oversold region were influential in forming this signal.”

      DRL Implementation Details

    • The DRL agent operates with a 50ms latency window, enabled by:

      • Model quantization

      • FPGA-accelerated inference

      • Co-located servers in NY4/LD4 datacenters (ping <1ms to exchanges)


3. System Architecture & Data Flow Diagram

The Quantum-Fusion architecture’s layered data flow is visualized below:

4. Performance Metrics and Backtest Results

The architecture has been subjected to extensive backtesting. Below are some of the key metrics obtained over 5 years of S&P 500 and major crypto-asset data. (These metrics are presented to demonstrate the system's potential, and it should be noted that past performance does not guarantee future results.)

Metric

Value

Description

Accuracy

85%

The accuracy of the Buy/Sell direction prediction over 5 years of data.

Sharpe Ratio

2.1

The industry-standard measure of risk-adjusted return.

Maximum Drawdown

12%

The largest portfolio value drop according to worst-case scenario (stress test) results.

Signal Confidence Range

70-95%

The typical confidence score range for signals produced by the Meta-Learner.

New XAI Limitations Disclaimer Interpretability Notes:

  • SHAP values provide approximate, not absolute explanations

  • Complex TFT attention mechanisms may resist full decomposition

  • Explanatory fidelity varies by market regime (higher in trending markets)


5. Conclusion and Future Vision

The MEF-AI Quantum-Fusion architecture has a theoretically sound, flexible structure that can explain its own decisions. Unlike "black box" AI models, this system places transparency and dynamic risk management at its core.

Keywords: Temporal Fusion Transformer (TFT), LightGBM, Meta-Learning (Ensemble Learning), Deep Reinforcement Learning, Explainable AI (SHAP), Real-Time Data Processing (Kafka, WebSocket), Feature Store.

© 2025 MEF-AI. All rights reserved.

Last updated