AlphaEvolve: A Learning Framework to Discover Novel Alphas in Quantitative Investment
Abstract: Alphas are stock prediction models capturing trading signals in a stock market. A set of effective alphas can generate weakly correlated high returns to diversify the risk. Existing alphas can be categorized into two classes: Formulaic alphas are simple algebraic expressions of scalar features, and thus can generalize well and be mined into a weakly correlated set. Machine learning alphas are data-driven models over vector and matrix features. They are more predictive than formulaic alphas, but are too complex to mine into a weakly correlated set. In this paper, we introduce a new class of alphas to model scalar, vector, and matrix features which possess the strengths of these two existing classes. The new alphas predict returns with high accuracy and can be mined into a weakly correlated set. In addition, we propose a novel alpha mining framework based on AutoML, called AlphaEvolve, to generate the new alphas. To this end, we first propose operators for generating the new alphas and selectively injecting relational domain knowledge to model the relations between stocks. We then accelerate the alpha mining by proposing a pruning technique for redundant alphas. Experiments show that AlphaEvolve can evolve initial alphas into the new alphas with high returns and weak correlations.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Explain it Like I'm 14
What this paper is about
This paper is about “alphas,” which are rules or small computer programs that try to predict which stocks will go up or down next. The goal is to build a collection of alphas that each make good money but don’t all win or lose at the same time. The authors introduce a new way to automatically create better alphas, called AlphaEvolve, that mixes the best parts of simple formulas and machine learning.
The questions the researchers asked
In simple terms, the paper asks:
- Can we design alphas that are both smart (accurate) and safe (not too risky when combined)?
- Can we automatically “evolve” better alphas from a starting point, instead of hand-crafting them or relying only on very complex machine learning models?
- Can we use information about how stocks are related (like being in the same industry) in a flexible, helpful way?
- Can we make the search for good alphas faster by avoiding wasted work?
How they approached the problem
To explain their approach, imagine alphas as recipes for making a prediction each day. AlphaEvolve is like a kitchen where many recipe variations are tried, improved, and filtered over time.
Here are the key ideas:
- Mixing strengths: Traditional alphas are simple formulas (easy to combine and keep uncorrelated), while machine learning alphas are powerful but complicated (hard to manage and de-correlate). The authors propose a new kind of alpha that uses both simple numbers and richer data (like whole rows or tables of features), keeping things understandable while still learning from more information.
- Evolving alphas: They use an evolutionary algorithm (like natural selection). Start with an initial alpha, make small random changes (mutations), test how well each version works, keep the better ones, and repeat.
- Special “operators” to help the search:
- ExtractionOps: Tools that pull useful pieces out of the big data table (for example, a single feature or a slice of time), so the alpha can use both simple values and longer-term patterns without turning into an overly complex neural network.
- RelationOps: Tools that let an alpha use relationships between stocks (like “rank this stock compared to others in the same industry”) without forcing a heavy, fixed model structure. This keeps the domain knowledge helpful but flexible.
- Pruning (clean-up) for speed: Before testing a candidate alpha, they remove dead ends and repeated steps—like crossing out lines in a recipe that don’t affect the final dish. This avoids wasting time evaluating alphas that won’t matter and speeds up the whole search.
How they judged performance:
- Information Coefficient (IC): A score that checks whether the alpha’s predictions line up with what really happened, mainly by comparing rankings (did we put likely winners near the top?).
- Sharpe ratio: A common finance measure of “return per unit of risk” for a trading strategy built from the alpha.
- Low correlation: They want multiple good alphas that don’t behave the same way. “Weakly correlated” means they don’t all move together, which helps reduce overall risk.
Data and setup:
- They tested on five years of NASDAQ stock data (2013–2017), using over a thousand stocks.
- Features included things like recent averages of prices, price volatility, and daily prices/volume.
- They compared AlphaEvolve to a genetic algorithm (another auto-search method) and to deep learning models (like LSTMs and a graph-based model that uses industry links).
What they found
- AlphaEvolve discovered alphas with higher Sharpe ratios and better IC than the baselines. In plain terms, its alphas were both more accurate and more profitable for the amount of risk taken.
- It was able to produce a set of alphas that were weakly correlated with each other—so the combined portfolio was safer.
- The flexible use of “relationship knowledge” (industry/sector info) helped when appropriate, and didn’t hurt when markets were too noisy. This is better than forcing that structure all the time.
- The pruning (clean-up) step made the search much more efficient, allowing AlphaEvolve to explore many more promising candidates in the same time.
- An “ablation study” (turning off parts to see their value) showed the parameter-updating part (which stores learned info from the past) generally improved prediction quality.
Why this matters:
- Compared to hand-built formulas or very complex machine learning models, AlphaEvolve found a middle path: powerful enough to use richer data, but simple enough to combine into a safe, diverse set.
- It outperformed both a genetic algorithm and deep models (LSTM and a graph model) on this NASDAQ dataset.
Why it matters
- Better tools for building alphas: This framework can help quants (people who design trading strategies) automatically find effective, diverse signals without getting stuck with overly similar or overly complex models.
- Risk control through diversification: Finding several good but weakly correlated alphas helps build portfolios that are steadier over time.
- Smarter use of knowledge: The method can selectively use relationships between stocks without locking into rigid assumptions that may fail in fast-changing markets.
- Practical efficiency: By pruning away wasted work, the system can search more ideas faster, which is valuable in real-world, time-limited research.
Takeaway
AlphaEvolve is like an evolution lab for stock-picking rules. It starts from a simple idea, tries many careful variations, uses helpful shortcuts to read the data and compare stocks, and cleans up junk before spending time on tests. The result is a set of smarter, safer alphas that work well together—promising a better way to build quantitative investment strategies.
Knowledge Gaps
Knowledge gaps, limitations, and open questions
The following list captures what remains missing, uncertain, or unexplored, framed to enable concrete follow-up research:
- Potential data leakage in preprocessing: each feature type is “normalized by its maximum value across all time steps for each stock,” which uses future information relative to the training/validation/test splits. Validate results with strictly forward-only, rolling, or train-only normalization to eliminate look-ahead bias.
- Transaction costs and execution frictions are ignored. Incorporate realistic assumptions on commissions, slippage, bid–ask spreads, taxes, borrow fees, and shorting constraints to assess whether reported Sharpe ratios persist net of costs.
- Survivorship and selection bias: the universe excludes illiquid or “too low price” stocks and (implicitly) delisted names. Quantify the impact of these filters and re-run on survivorship-corrected data to bound bias.
- Single market and regime: all results are on NASDAQ (2013–2017). Test generalization across markets (e.g., NYSE, global equities) and regimes (e.g., 2008 crisis, 2020 pandemic, 2022–2023 inflationary period) using walk-forward evaluation.
- Overfitting risk from repeated model selection on one validation window (116 days). Use nested walk-forward cross-validation, multiple non-overlapping validation/test windows, and out-of-time holdouts; apply multiple-hypothesis corrections and report statistical significance (e.g., t-stats, p-values for IC/SR).
- Portfolio construction specifics are underdefined. Provide and test sensitivity to:
- Weighting scheme (equal-weight vs. rank-weight vs. risk-parity),
- The size of top/bottom buckets (e.g., 25/25, 100/100),
- Holding period and rebalance frequency,
- Leverage and net exposure constraints.
- Quantify how these choices affect IC and Sharpe.
- Turnover, capacity, and market impact are not reported. Measure turnover, average holding period, capacity under liquidity constraints, and how performance degrades with AUM and more realistic fill assumptions.
- Correlation cutoff ambiguities: clarify whether the 15% cutoff is on absolute Pearson correlation and assess stability of correlations over time and regimes. Explore non-linear dependence (e.g., tail dependence, Spearman, distance correlation) and the implications for “weakly correlated” alpha sets.
- Combined portfolio performance of the mined alpha set is missing. Evaluate the ensemble portfolio built from all mined alphas (with realistic weighting and constraints) to demonstrate diversification benefits and incremental information ratio.
- Baseline fairness and breadth: Rank_LSTM and RSR baselines are fed only simple moving-average inputs (and show high variance). Compare against stronger, modern baselines (e.g., Transformers, Temporal Fusion Transformer, N-BEATS, DeepAR), state-of-the-art symbolic regression, gradient-based feature learners, and AutoML-Zero on the same feature space, with comparable hyperparameter search budgets.
- Relational domain knowledge is static (sector/industry). Test dynamic relations (e.g., time-varying learned graphs, correlation/clustering-based neighbors, news-based links) and quantify robustness to misclassification or changing sector compositions.
- The “new class of alphas” remains structurally complex. Assess interpretability and economic plausibility (e.g., economic rationale, factor exposures), and propose methods to simplify, canonicalize, or regularize discovered expressions without degrading performance.
- Training procedure for parameter-updating functions is unclear. Clarify what “train our alpha by one epoch” entails, especially since updates appear rule-based rather than learned via gradient descent. Study the effect of longer training, different optimization schemes, and explicit learning dynamics on out-of-sample performance.
- Pruning and fingerprinting correctness is unproven. Provide formal guarantees or empirical evidence that:
- Pruning never removes operations essential for Update() or relational computations,
- Fingerprinting via operation strings reliably deduplicates functionally equivalent alphas (addressing operator commutativity/associativity and algebraic simplification),
- Cache collision rates are negligible.
- Report wall-clock speedups and memory usage, not just counts of “searched alphas.”
- Sensitivity to AlphaEvolve hyperparameters is not explored. Systematically study mutation rate, population size, tournament size, allowable OP sets, operand limits, and time budgets; report variability across seeds and provide confidence intervals for performance.
- Risk exposures to known factors (market, size, value, momentum, quality, low-vol) are unmeasured. Decompose alpha returns with a multi-factor model to confirm genuine, orthogonal alpha and avoid rediscovering established factors.
- Feature space is limited to price/volume-derived technicals. Evaluate the incremental value of fundamentals (e.g., earnings, balance sheet), macroeconomic series, alternative data (news, sentiment, ESG), and learned embeddings; integrate ExtractionOps for richer modalities.
- Execution timing and signal availability assumptions are not explicit. Confirm that all features used at time t are available before trading decisions for day t (no use of same-day highs/lows/volume). Provide a precise data timestamping and signal generation protocol.
- Reproducibility and code/hardware transparency are lacking. Share code, random seeds, hardware specs, and exact search budgets to enable replication; report compute cost per discovered alpha and scaling behavior with K (number of stocks/tasks).
- Comparison to AutoML-Zero is qualitative. Run head-to-head experiments to quantify sample efficiency, search efficiency, and resulting alpha quality under matched budgets, highlighting when AlphaEvolve’s operators/pruning provide advantages.
- Canonicalization/simplification of discovered formulas is missing. Develop algebraic reduction and equivalence-checking to minimize redundant forms and aid interpretability, maintenance, and deployment.
- Deployment considerations are not discussed. Specify latency, daily compute requirements for feature extraction and inference, monitoring, and model governance needed to run AlphaEvolve alphas in production.
- Robustness to data quality issues (splits, dividends, missing values, corporate actions) is not addressed. Provide preprocessing protocols and stress tests to show stability under realistic data imperfections.
- Ethical and compliance aspects (e.g., short-selling constraints, market manipulation concerns) are unexamined. Outline constraints and compliance checks for real-world use.
These gaps suggest concrete avenues: eliminate leakage; incorporate costs and constraints; broaden baselines and features; conduct walk-forward, multi-regime tests; measure turnover/capacity/risk exposures; formalize pruning; and evaluate ensemble performance of the mined alpha set.
Practical Applications
Immediate Applications
The following applications can be deployed now by adapting the paper’s methods, operators, and workflows to existing data, research pipelines, and compliance practices.
- Quant “alpha factory” to discover and maintain weakly correlated signals
- Sector: finance (hedge funds, asset management)
- What: Integrate AlphaEvolve to evolve existing formulaic or ML alphas into a library of high-IC, low-correlation signals; enforce a correlation gate (e.g., ≤15% Pearson on portfolio returns) during search.
- Tools/workflows:
AlphaEvolve SDKfor evolutionary search;Correlation Gateto filter candidates;Pruning Cachefor redundancy pruning; IC/Sharpe evaluation with a long–short backtest; alpha registry and promotion criteria. - Assumptions/dependencies: High-quality, survivorship-bias–controlled historical data; accurate sector/industry taxonomy; compute budget for large populations; correlation threshold calibration; risk and compliance approvals.
- Long–short portfolio construction with evolved alphas
- Sector: finance
- What: Use top/bottom-ranked signals (e.g., long top-50 and short bottom-50) to construct balanced portfolios; monitor IC and Sharpe; deploy multiple weakly correlated alphas for diversification.
- Tools/workflows:
Ranking Enginefor daily signal sorting;IC/Sharpe Monitor; position rebalancer; cash balancing logic. - Assumptions/dependencies: Liquidity and borrow availability; transaction costs, slippage, and short borrow fees (not modeled in the paper) will affect realized Sharpe; risk-free rate set to 0 in evaluation; market impact modeling needed for production.
- Correlation-aware signal library management
- Sector: finance, software
- What: Maintain an internal catalog of signals with measured cross-correlation of portfolio returns; schedule periodic re-evaluation and pruning of signals whose correlations drift higher.
- Tools/workflows:
Alpha Registry;Correlation Dashboard; scheduled backtests and gatekeeping; archive/evidence of IC histories. - Assumptions/dependencies: Regime changes affect correlation; robust statistics for rolling windows; governance to retire or neutralize overly correlated signals.
- Selective domain-knowledge injection via RelationOps
- Sector: finance
- What: Apply
RelationOps(RankOp, RelationRankOp, RelationDemeanOp) when sector/industry relations are informative; disable or soften injection in highly noisy markets (e.g., NASDAQ) where relations are less stable. - Tools/workflows:
RelationOpsmodule parameterized by sector mapping; toggles for per-market/domain injection. - Assumptions/dependencies: Accurate, timely sector/industry classification; market-specific tuning (relations may hurt on noisier markets).
- AutoML acceleration with redundancy pruning in quant R&D
- Sector: software, data science
- What: Adopt the paper’s redundancy pruning (operation and alpha-level) and fingerprinting-by-structure to reduce repeated evaluations in program-evolution pipelines.
- Tools/workflows:
Program-Pruning Cache Service; structural hashing of operation graphs; early elimination pre-evaluation. - Assumptions/dependencies: Correct graph representation of alphas; collision-resistant hashing; careful caching to avoid stale reuse.
- Academic benchmarking and reproducible experimentation
- Sector: academia
- What: Use the NASDAQ dataset split and evaluation metrics (IC, Sharpe with long–short strategy) to benchmark AutoML-evolved alphas versus GA and deep models; study correlation-aware mining protocols.
- Tools/workflows:
AlphaEvolve Notebooks; open baselines (GA, LSTM, RSR) and operator libraries; reproducible seeds and splits. - Assumptions/dependencies: Data licensing for OHLCV and volumes; compute time budgets (e.g., 60 hours per round); transparent preprocessing (e.g., low-price stock filtering).
- Compliance and audit trail enhancement with alpha fingerprints
- Sector: policy/compliance in finance
- What: Use structural fingerprints of alphas and cached fitness scores to maintain audit trails of model lineage and deployment decisions; support model risk management reviews.
- Tools/workflows:
Alpha Fingerprint Ledger; change-management records; correlation gate decisions logged. - Assumptions/dependencies: IP protection; regulator acceptance of structural fingerprints; integration with MRM frameworks.
- Retail quant education and paper trading sandbox
- Sector: education, consumer finance
- What: Offer a simplified version of the evolutionary search to retail users in a paper-trading environment to learn signal diversification and correlation management.
- Tools/workflows:
Alpha Playground;Paper-Trading Sandbox; broker API (read-only) for practicing. - Assumptions/dependencies: Prominent risk disclosures; constraints to prevent live deployment by unqualified users; simplified operator set and compute limits.
Long-Term Applications
These applications require further research, engineering, scaling, or regulatory development before broad deployment.
- Autonomous alpha discovery and maintenance (“AlphaOps”)
- Sector: finance, software
- What: Always-on system that continuously mines, evaluates, de-correlates, and deploys alphas; integrates execution costs, dynamic risk constraints, and real-time data.
- Tools/workflows:
AlphaOpsorchestration;Live Backtesting Engine; CI/CD for signals; guardrails and kill-switches. - Assumptions/dependencies: Real-time market data ingestion; robust model risk management; operational resiliency; continuous correlation surveillance.
- Cross-asset and multi-market extension
- Sector: finance (FX, futures, commodities, credit, crypto)
- What: Generalize operators and features to multiple asset classes; add RelationOps for cross-asset hierarchies (e.g., sector-to-commodity links, currency blocs).
- Tools/workflows: Asset-specific feature connectors;
RelationOpsfor new taxonomies; multi-market backtesting harness. - Assumptions/dependencies: Data availability and quality across assets; differing microstructure (e.g., 24/7 crypto, futures roll); cost and carry models.
- Cost-aware fitness and execution simulation
- Sector: finance
- What: Extend the IC/Sharpe fitness to include transaction costs, slippage, borrow fees, and market impact; run execution-aware simulations during evolution.
- Tools/workflows:
Cost Modelcomponents;Execution Simulator; broker/venue analytics feeds. - Assumptions/dependencies: Reliable cost estimates; parameterization per venue and liquidity regime; calibration to realized performance.
- Regime-aware correlation management
- Sector: finance
- What: Dynamically adjust correlation cutoffs and alpha selection based on detected market regimes; prevent hidden co-movements during stress periods.
- Tools/workflows:
Regime Detector(change-point/volatility clustering);Correlation Scheduler; robust rolling metrics. - Assumptions/dependencies: Accurate regime identification; robust statistics under non-stationarity; governance to adapt thresholds.
- Explainability and distillation of evolved alphas
- Sector: finance, policy/compliance
- What: Convert evolved operation sequences into compact, human-readable formulas with narratives (e.g., “bounded trend on high prices”); support risk committee reviews.
- Tools/workflows:
Alpha Distiller; operator-to-story mapping; documentation generator. - Assumptions/dependencies: Stable mapping from operators to interpretable constructs; acceptance by oversight bodies.
- Generalized relational AutoML beyond finance
- Sector: healthcare, energy, manufacturing/IoT, software
- What: Apply RelationOps and weak-correlation model mining to multi-task domains where entities share relations:
- Healthcare: patient risk prediction across hospitals/departments; build diverse, weakly correlated risk models.
- Energy: grid fault prediction across nodes/regions; ensemble of low-correlation predictors improves reliability.
- Manufacturing/IoT: anomaly detection across machine families; diverse signals reduce false positives.
- Tools/workflows: Domain-specific
RelationOps(e.g., patient-group, grid-topology, machine-class);Correlation Gatefor ensemble diversity. - Assumptions/dependencies: High-quality relational data; privacy/compliance (e.g., HIPAA); streaming data infrastructure; domain-specific fitness metrics.
- Education: curriculum and toolkits for program-evolution AutoML
- Sector: academia, education
- What: Develop coursework and labs on evolving models with scalar/vector/matrix operators, pruning techniques, and correlation-aware ensemble design.
- Tools/workflows: Teaching datasets;
Relational AutoMLlibrary; lab guides and graded exercises. - Assumptions/dependencies: Access to compute; institutional buy-in; open-source licensing.
- Policy and standards for AutoML-based trading
- Sector: policy/regulation
- What: Create guidelines for backtesting rigor, correlation-risk controls, auditability of evolved models, and stress testing before deployment.
- Tools/workflows: Standardized validation checklists; regulator-facing reporting templates; model lineage requirements.
- Assumptions/dependencies: Industry consensus; regulator capacity to review program-evolved models; cooperation on data sharing.
- Productized relational AutoML library
- Sector: software
- What: Package operators (ExtractionOps, RelationOps), pruning, and structural fingerprinting into a general-purpose library for program-evolution AutoML (beyond AutoML-Zero).
- Tools/workflows:
Relational AutoMLSDK; performance benchmarks across domains; plugin operator catalogs. - Assumptions/dependencies: Broad operator coverage; community contributions; maintenance and performance tuning.
- Human-in-the-loop interactive evolution
- Sector: finance, ESG investing
- What: Allow researchers to steer evolution with constraints (e.g., sector neutrality, ESG screens, factor exposures), and inspect trade-offs between IC, Sharpe, and correlation.
- Tools/workflows:
Interactive Evolution UI; constraint DSL; integrations with ESG/alternative data providers. - Assumptions/dependencies: Responsive compute backends; well-defined constraint solvers; data licensing for constraints (e.g., ESG).
Glossary
- Alpha (trading): A stock prediction model that generates buy/sell signals in quantitative finance. "Alphas are stock prediction models generating trading signals (i.e., triggers to buy or sell stocks)."
- AlphaEvolve: The paper’s AutoML-based evolutionary framework for generating new alphas with high returns and low correlation. "we propose a novel alpha mining framework based on AutoML, called AlphaEvolve, to generate the new alphas."
- AutoML: Automated machine learning used here to search over operators and structures to build alphas. "we propose a novel alpha mining framework based on AutoML, called AlphaEvolve"
- AutoML-Zero: A system that discovers neural networks from scratch, considered as a broader search approach for alpha mining. "A framework called AutoML-Zero \cite{real2020automl} was recently proposed to discover a neural network from scratch"
- Backtesting: Evaluating a trading strategy or alpha on historical data to assess performance before deployment. "and then backtested on the markets to ensure weakly correlated returns."
- ExtractionOps: Operator family that extracts scalar or vector features from the input matrix, aiding search. "We define OPs extracting a scalar feature from and OPs extracting a vector feature from as GetScalarOps and GetVectorOps respectively, or called ExtractionOps in general."
- Frobenius norm: A matrix norm equal to the square root of the sum of squared entries, used as a feature operation. "where calculates the Frobenius norm of a matrix."
- Genetic algorithm: An evolutionary search method previously used to mine formulaic alphas. "or using the genetic algorithm to automatically mine a set of formulaic alphas"
- Graph neural network (GNN): A neural architecture for graph-structured data; cited as a complex component hard to evolve from scratch. "e.g., graph neural network \cite{feng2019temporal}, attention mechanism and LSTM neural networks \cite{8622541}"
- Hedge funds: Institutional investors prominent in markets, often employing long-short strategies and correlation controls. "Hedge funds are institutional investors and among the most critical players in the stock markets"
- Heaviside (step function): A step function used as an operator within alpha equations. "{heaviside}(S 1_{t-2})"
- Information Coefficient (IC): The average cross-sectional Pearson correlation between predicted and actual returns, used as fitness. "We use the Information Coefficient (IC) as the fitness score"
- Long position: Buying assets expected to rise in price. "The long position is built by buying the stocks with the top 50 predicted returns."
- Long-short trading strategy: A portfolio strategy that goes long top-ranked stocks and short bottom-ranked stocks to evaluate alphas. "We use the long-short trading strategy, a popular hedge fund strategy \cite{RePEc:arx:papers:(Kakushadze, 2016)}, to build a portfolio to evaluate an alphaâs investment performance."
- Machine learning alphas: Data-driven models over vector/matrix features that generate trading signals. "Machine learning alphas are data-driven models over vector and matrix features."
- Net Asset Value (NAV): The total value of positions minus cash in the evaluation portfolio. "The net asset value (NAV) is defined as "
- Pearson correlation: A standard correlation measure used to define weak correlation between alpha portfolios. "the sample Pearson correlation of 15\% between portfolio returns of different alphas"
- Pruning technique (redundancy pruning): Optimization that removes redundant operations/alphas and fingerprints before evaluation to speed search. "We thus propose an optimization technique by pruning redundant operations and alphas as well as fingerprinting without evaluation."
- RankOp: A RelationOp that outputs the rank of a scalar among tasks. "RankOp outputs the ranking of the input operand"
- RelationDemeanOp: A RelationOp that de-means a scalar by subtracting the sector/industry mean. "RelationDemeanOp calculates the difference between the input operand calculated on and the mean of those calculated on $\mathbf{s}^{\mathcal{F}_{I}$."
- RelationOps: Operator family that models relations among stocks (e.g., within sector/industry) across tasks. "We design a set of OPs, called RelationOps, to model such relations"
- RelationRankOp: A RelationOp that ranks a scalar within its sector/industry group. "RelationRankOp outputs the ranking of the input operand calculated on among those calculated on $\mathbf{s}^{\mathcal{F}_{I}$"
- Relational domain knowledge: Market structure knowledge (e.g., sector/industry) injected into alphas to leverage stock relations. "selectively injecting relational domain knowledge to model the relations between stocks."
- Risk-adjusted returns: Returns normalized for risk, typically measured via Sharpe ratio. "we use the Sharpe ratio to measure risk-adjusted returns of a portfolio"
- Risk-free rate: The return of a theoretically riskless asset used in Sharpe ratio computation. " is the risk-free rate"
- Sector-industry relations: Groupings of stocks that share sector/industry, used to inform RelationOps. "operators to selectively inject sector-industry relations of all stocks into an alpha."
- Sharpe ratio: A measure of risk-adjusted performance defined as mean excess return over volatility. "The Sharpe ratio is defined as "
- Short position: Borrowing and selling assets expected to fall in price. "The short position is built by borrowing the stocks with the bottom 50 predicted returns and selling them for cash."
- Tournament (selection): Evolutionary selection method choosing the best alpha from a random subset. "called the tournament"
- Volatility: The standard deviation of returns, used in Sharpe ratio and as a feature. "and is the volatility of the portfolio calculated as the standard deviation of ."
- Weakly correlated returns: A low-correlation standard (e.g., 15%) among alpha portfolios to reduce risk. "A set of effective alphas can generate weakly correlated high returns to diversify the risk."
Collections
Sign up for free to add this paper to one or more collections.