Cybernetic Intelligence

An open exploration of viable human-AI systems.

View the Project on GitHub algoplexity/cybernetic-intelligence

đź§Ş Research Hypothesis:

Autonomous Generation of Decorrelated Alpha Expressions via Self-Supervised Reasoning with Compact Language Models


1. Motivation

Quantitative asset managers rely on vast libraries of “alpha expressions” — mathematical signals derived from financial and alternative data — to inform portfolio construction. However, the marginal value of new alphas has declined as model complexity increases, often resulting in redundant or non-robust factors that fail to generalize out-of-sample (Kelly, Malamud & Zhou, 2023; Buncic, 2024). At the same time, recent advances in reinforcement learning and self-supervised reasoning (Chen et al., 2024) suggest that small language models (≤1B parameters) can acquire advanced reasoning capabilities through interaction with verifiable environments — without relying on pretrained weights or curated data. This raises the possibility of building a compact agent that can autonomously generate novel, profitable, and decorrelated alpha expressions, guided solely by simulator feedback such as Sharpe ratio, turnover, and correlation to production alphas.


2. Hypothesis

A compact language model (≤1B parameters), trained via a self-supervised, reward-driven learning loop without any pretraining or labeled data, can autonomously generate syntactically valid and semantically novel alpha expressions in the Fast Expression language used by the WorldQuant BRAIN platform. These expressions will exhibit positive performance metrics (e.g., Sharpe > 1.25, Fitness > 1.0) and low correlation with existing production alphas, when evaluated over historical market data via the BRAIN simulator.


3. Methodology

This hypothesis will be tested by constructing an autonomous alpha generation agent composed of the following modules:


4. Expected Contributions


5. References