top of page

Hybrid AI: How a Model Outperformed Venture Capital Firms at Predicting Startup Success


Large language models (LLMs) are powerful, but variability in accuracy limits their use in high-stakes decisions. Few executives will commit millions based on an opaque model. A recent paper introduces LLM-AR, a framework that pairs LLM-generated rules with probabilistic reasoning to predict startup success at the idea stage using only founders’ professional and behavioral traits (1). The result is a hybrid approach that seeks not just higher precision but also transparent, auditable decisions.

 


Hybrid AI for predicting success

1. The Experiment: Can startup success be predicted?

The LLM-AR research (1) tackles one of the highest-risk decisions in the business world: venture capital investment at the "idea" stage. At this stage, information is sparse and the base rate for identifying outliers is roughly 1.9% (2).


The Experiment's Objective: The researchers wanted to see if an AI could reliably predict a startup's success by exclusively analyzing the professional and behavioral traits of its founders.


Dataset

First, the researchers built a dataset based on real historical data (from LinkedIn and Crunchbase) to define the outcomes of 6,000 founders. They established concrete financial criteria to label the results:

  • Success (real world): A startup with an IPO, an acquisition above $500 million, or more than $500 million raised.

  • Failure (real world): A startup that stalled at a minor funding round (between $100k and $4M) (1).


Only after establishing these real-world outcomes did they extract the 6,000 founder profiles (using only data available before the startup was founded) and convert them into 52 anonymized numerical features. These features represent the founder's "profile traits," such as education_level, vc_experience (experience at other VC firms), and even implicit qualities like perseverance and risk_tolerance (1).


The Contenders (The Test)

With the cleaned dataset, the team ran a head-to-head evaluation to see which approach was best:

  1. The "Pure" Contenders: They tested "pure" LLMs (like GPT-4o-mini, GPT-4o, and DeepSeek-V3) (1, 11, 19). They asked these models to predict success based on the profile traits.

  2. The New "Hybrid" Challenger: They tested their new framework, LLM-AR, on the exact same data.


From an operator’s perspective, the hardest call is betting on people. Understanding who will succeed—whether it's a startup founder to fund, a manager to hire, or a strategic partner—is the highest-risk, highest-return bet a company can make. For a business leader, AI isn't a stylistic exercise; it's a tool to mitigate this risk. The strategic question, therefore, is: can a model help us make this decision more reliably than we do today?

The LLM-AR hybrid approach is designed to do just that: not just to win the precision race (as we'll see in Sec. 2), but to do it transparently and reliably. This article analyzes the results of this comparison, the technology that made it possible (Sec. 3), and the governance implications (Sec. 10).


2. The Results: The precision of the models

In business, P&L discipline sets the bar. Communication to leadership must reflect an obsession with results, focusing on measurable, concrete competitive advantages, not just the technology itself.


As we saw in Section 1, researchers put pure LLMs in a head-to-head comparison against the new LLM-AR hybrid framework. The test's objective was to measure precision in predicting which founders would succeed, based solely on their professional and behavioral traits.

To establish a "human benchmark," they computed average precision for Tier-1 seed funds and rescaled it to match the 10% prevalence used in the study (1, 2).


The results of this head-to-head evaluation are telling:

LLM-AR (hybrid AI)

Average Precision (%): 59.5%


GPT-4o mini (pure LLM)

Average Precision (%): 49.5%


GPT-4o (pure LLM)

Average Precision (%): 32.3%


DeepSeek-V3 (pure LLM)

Average Precision (%): 31.0%


Tier-1 VC funds (human benchmark)

Average Precision (%): 29.5%


o3-mini (pure LLM)

Average Precision (%): 21.6%


Indexing strategy (baseline)

Average Precision (%): 10.0%


Note: The human benchmark (29.5%) is a scaled industry figure used to allow a fair comparison against the models in the 10% prevalence dataset (1).


In the validation dataset, 6,000 founders were analyzed, with success defined as IPO or M&A > $500M, or funding > $500M; "failures" included fundraising between $100K and $4M. The prevalence was set at 10% for the experiment, whereas the real-world "market index" is ~1.9% (2). Figure 1 of the LLM-AR paper shows, at the same 10% prevalence, LLM-AR achieving 59.5% precision versus the 29.5% scaled human benchmark (1).


The LLM-AR framework doesn't just edge out the other models (including GPT-4o mini) (11); it outperforms the Tier-1 venture capital benchmark by a wide margin: it is twice as precise (59.5% vs 29.5%). That is a 5.9× lift over the 10% baseline. This isn't just a minor academic refinement; it's a pragmatic result that, in the venture capital context, translates to minimizing investment in false positives—a mission-critical objective for financial sustainability.


These values are derived from the original paper, which uses ProbLog to formalize rules (e.g., education_level, industries) and an F-score tuned for precision (F0.25) to reduce false positives in high-cost-of-error contexts (1).


3. Beyond LLMs: What is neurosymbolic AI and ProbLog?

The primary problem with today's LLMs, especially in structured corporate environments, isn't just precision—it's their "black box" nature. An executive cannot make a strategic decision based on an output they cannot understand or verify.


A tech-agnostic approach is more resilient. Instead of just searching for the "best" LLM, LLM-AR adopts a hybrid approach inspired by neurosymbolic AI (18, 19). This paradigm integrates the statistical pattern-matching of neural networks (LLMs) with the reasoning power and interpretability of symbolic logic.


This model's strength isn't promoting one tool; it's identifying the right solution. Specifically, LLM-AR combines an LLM (the study used Deepseek-V3) with an automated reasoning engine called ProbLog.


Why ProbLog? Traditional logic systems (like Prolog) struggle with the ambiguity of human language. Phrases like "most," "usually," or "a strong indicator" can't be translated into binary True/False rules. ProbLog solves this by introducing probabilities. It allows "weights" or confidence scores to be assigned to both facts (e.g., "education: 0.7") and the rules themselves (e.g., "0.6::success <= education, experience"). ProbLog descends from probabilistic logic programming and has been integrated with neural nets in DeepProbLog, where neural predicates allow outputs from deep models to be used as probabilistic facts within business rules (16). This enables auditability and what-if analysis on domain conditions, unlike end-to-end models.


The result is a system that doesn't just "guess" a pattern; it reasons transparently and reproducibly while managing real-world uncertainty. Real-world adoption of neurosymbolic/automated reasoning is already visible: AWS (Amazon Web Services), for example, uses automated reasoning techniques to reduce hallucinations and improve the verifiability of its conversational and robotic systems (14). It's the same logic of "external verification" that LLM-AR applies to its rule sets, building trust and positioning the technology as an impartial, verifiable arbitrator.


4. Opening the "Black Box": Interpretability and rule set transparency

Transparency and intellectual honesty are fundamental. A strategic partner, human or otherwise, must allow its work to be verified. The most significant advantage of the LLM-AR framework isn't just its precision, but its interpretability.

Unlike a standard LLM, which returns an answer without explaining how it got there, LLM-AR outputs a human-readable rule set (1). Every single decision path is exposed for human inspection.


The process, in short, works like this: the LLM is used to generate and refine the rules, but the final prediction is executed by the automated reasoning engine (ProbLog). This decouples pattern recognition (where LLMs excel) from logical reasoning (where symbolic systems are transparent).


Let's look at a practical example. Instead of an obscure answer, LLM-AR might operate on rules like these (examples from the study):

  • Success Rules:

    • IF num_acquisitions AND career_growth THEN success (p=0.40)

    • IF perseverance AND vision THEN success (p=0.32)

  • Failure Rules:

    • IF NOT career_growth AND NOT num_acquisitions THEN failure (p=0.96)

    • IF NOT education_level AND NOT education_institution THEN failure (p=0.89)


This transparency is critical. It allows executives to understand why the system recommended an investment or flagged a risk. It allows them to debate, refine, and even correct the system's logic, transforming AI from an oracle into a true strategic assistant. The symbolic layer (ProbLog rules) supports post-hoc traceability and compliant "instructions for use" under the EU AI Act (Reg. (UE) 2024/1689) (4). It simplifies both the "instructions for use" for deployers and the post-market monitoring mandated by the regulation (4).


5. From Agile Adoption to Strategic Integration: A Phased Roadmap

Implementing AI in a company is not a single event; it's a journey. Many initiatives stall when launched as monolithic, high-risk "big-bang" programs. I have analyzed in-depth why 85% of AI projects fail. My conclusion, based on field experience, is that the cause isn't the technology; it's a leadership that passively delegates strategy to the IT department instead of guiding it with a clear vision from the top—a cornerstone of my Rhythm Blues AI method. The often-cited "85%" figure traces back to a 2018 Gartner prediction about erroneous outcomes by 2022; it is a cautionary data point, not a production-failure rate (15). In parallel, 2024-2025 surveys confirm that many companies struggle to scale value from AI (5, 6). This is all the more reason to adopt phased, measurable cycles.


Interestingly, the LLM-AR framework's design mirrors this philosophy. The model isn't "born" complete; it evolves through an "iterative policy-evolution loop" (1). The system is trained on small batches of data, generates an initial rule set, statistically analyzes its own mistakes, and then "reflects" on those errors to produce a better rule set in the next iteration.


This iterative model is the same one businesses should adopt. Instead of investing millions in an all-or-nothing project, it's wiser to start with targeted, low-risk interventions. A consulting approach like that offered by Rhythm Blues AI is built on this logic: service packages (like Starter, Advanced, and Executive) that guide the company on this journey.

It starts with an audit to map real needs and identify high-potential automation processes (as in the Starter package). KPIs are defined, and the necessary "company culture" is built. Only after validating initial results and building internal trust do you proceed to more complex implementations, like generative agents or advanced governance. This phased approach manages risk and ensures every tech investment is tied to a tangible business outcome.


6. How AI Learns to Decide: The iterative loop explained for managers

One of a strategic consultant's hardest jobs is to translate technological complexity into business language. Avoiding jargon isn't about "dumbing it down"; it's about demystifying tech to empower leaders to make informed decisions.

So, how does LLM-AR's iterative training work in practice? Think of it as mentoring a junior analyst through four passes:


Example: After one iteration, the statistical analysis might flag that the pair career_growth ∧ num_acquisitions has a high "lift" and consistent "confidence." In the subsequent reflection, the LLM increases the weight of that rule in ProbLog and downgrades "education_level," which showed low statistical support (1, 3).

  1. Initial Generation (Observation): For each "founder" in the data batch, the LLM is prompted like a VC analyst: "This founder succeeded. In your opinion, what were the most important reasons?" The LLM produces a text analysis (e.g., "Deep industry experience," "Leadership skills").

  2. Synthesis (From observation to rule writing): After analyzing a full batch, the LLM is asked to summarize these individual insights into general, logical rules. For example: IF (ceo_experience) AND (num_acquisitions) THEN success. The LLM also assigns a probability (confidence score) to each rule.

  3. Statistical Analysis (Senior Review): Here, the "adult supervision" kicks in. This initial rule set, based on the LLM's intuition, is statistically vetted. A technique called "association-rule mining" (3) is used to verify if the feature combinations (e.g., ceo_experience and acquisitions) are truly statistically associated with success in the real data. When discussing rules, it's useful to invoke the classic metrics of association-rule mining—support for statistical relevance, confidence for conditional reliability, and lift for informative value—to separate broad correlations from truly insightful clues (3). In practice, rules with a lift > 1 and sufficient support are promoted to the rule set.

  4. Reflection (Learning the Ropes): The LLM is presented with the statistical report. It's told: "Your intuition about 'vision' was correct, but you overestimated the importance of 'education_level.' The data also suggests a strong correlation between X and Y that you missed." The LLM "reflects" (1) on this feedback and produces a new, updated rule set, removing rules with low statistical support and incorporating the new insights.


This cycle repeats, refining the rule set with each pass. It's a perfect example of how intuition (LLM) and rigorous analysis (statistics) can collaborate to produce a superior outcome.


7. Expert-in-the-Loop: Is AI replacing the strategic analyst?

AI adoption presents challenges that are primarily human and organizational. Fear of replacement, the need for continuous training (upskilling), and change management are the real hurdles.

A well-designed AI system doesn't seek to replace the human expert; it seeks to augment them. The LLM-AR framework is explicitly built for an "expert-in-the-loop" (1). This is a fundamental advantage over black-box systems.


Thanks to the rule set transparency (as seen in Sec. 4), human experts (VC analysts, doctors, managers) can interpret the model's reasoning. But the advantage doesn't stop there: the system is designed to be modifiable. A manager, based on their own domain knowledge or contextual information the AI lacks, can directly modify the rules or adjust the probabilities.


This human-in-the-loop design solves two problems. First, it improves the model's accuracy by combining the best of the AI's statistical analysis with human intuition and experience. Second, it addresses the risk of "cognitive debt"—the erosion of critical human skills from over-reliance on technology.


By keeping the human in the loop, the AI becomes a co-pilot that handles large-scale data analysis, while the human retains strategic control and responsibility for the final decision. The AI handles the computational complexity; the human handles the context and strategy. The need for strong leadership to guide the AI revolution is a pillar of my method.


8. Beyond Precision: How to measure AI ROI (Precision vs. Recall and F-score)

Any strategic initiative must be tied to clear business metrics and a demonstrable ROI. One of the main problems companies face in AI adoption is the difficulty in quantifying this return (5, 6).


The LLM-AR framework addresses this by making measurement not just an outcome, but a "tunable" feature of the model itself (1).

In forecasting, two key metrics are often in conflict:

  1. Precision: Of all the times the model predicted "Success," how often was it right? (Goal: Minimize false positives).

  2. Recall: Of all the actual "Successes" in the data, how many did the model find? (Goal: Minimize false negatives).

In the venture capital study, the objective was to maximize precision to avoid wasting capital on startups destined to fail. Accordingly, the team optimized for F0.25 (1). This is a formula (F-beta score) that weights precision and recall. By using F0.25, precision is weighted four times more heavily than recall.


F(beta) = (1 + beta^2) (Precision Recall) / ( (beta^2 * Precision) + Recall)


The strategic point is that this parameter is tunable. Since precision depends on prevalence (the base success rate), expectations must be recalibrated when moving from the 10% experimental dataset to the ≈ 1.9% real market (1, 2). Operationally, raising the rule-activation threshold trades recall for precision (e.g., 100% precision at ~2% recall vs. ~92% recall at ~12.5% precision), keeping the architecture unchanged and adjusting F(β). For an investment committee, this translates into an explicit lever on the cost of false positives (1).

This flexibility allows executives to define their measurable objectives upfront and "tune" the AI to serve that specific business strategy. The F0.25 choice prioritizes precision 4x over recall; in healthcare, one could pivot to F(2) to prioritize sensitivity (1). It's a "business" control, not just a technical one.


9. Advanced Positioning: The future of AI in complex decisions and hybrid systems

Market differentiation requires moving beyond chatbots and basic automation. You must demonstrate expertise in cutting-edge topics like generative agents or hybrid reasoning systems.


LLM-AR sits exactly in this advanced space. It's not just a "tool" (like a standard LLM); it's a strategic "actor" that requires a new management paradigm. It's a concrete example of neurosymbolic AI (18, 19), a field seeking to overcome the limitations of purely neural (LLM) and purely symbolic approaches.


This framework is inspired by influential systems like NS-VQA (Neural-Symbolic Visual Question Answering), which disentangles visual perception (handled by neural networks) from the execution of deterministic symbolic programs to answer questions. The neurosymbolic field has already shown near-perfect accuracy in compositional reasoning tasks (e.g., 99.8% on NS-VQA for CLEVR) precisely because it explicitly executes program traces over symbolic representations (13). This is the same principle we apply to text/tabular data when we derive verifiable rules from LLM-extracted patterns.


The future of high-performance enterprise AI likely lies not in ever-larger LLMs, but in intelligent, hybrid architectures. The LLM-AR research points to several future directions (1):

  • LLM-Powered Feature Selection: Allowing the LLM to propose new features to analyze, which human engineers might not have considered.

  • Alternative Statistical Methods: Exploring the use of Bayesian Networks instead of simple rule association. This would allow for encoding multi-step reasoning, such as "professional athlete implies perseverance," and "perseverance implies success."

  • Alternative AR Implementations: LLM-AR is a framework. ProbLog could be replaced with other symbolic AI systems to adapt the model for different domains.


Speaking this language—of hybrid systems, interpretability, and multi-step reasoning—is what distinguishes a strategic approach to AI from a purely tactical one. Models like o3 and o4-mini add stronger tool use and visual reasoning at department-friendly cost/latency (10); GPT-4o-mini offers ~82% MMLU at a price point drastically lower than high-end models, useful for the "intuition" layer of the pipeline (11). DeepSeek-V3 shows open-source progress on hard benchmarks and multi-token prediction (19), but it doesn't replace the logic-probabilistic traceability required in regulated decisions.


A complementary branch is Random Rule Forest (RRF): an ensemble of YES/NO questions generated by LLMs and voted on by threshold. On a 10% prevalence, it reports ~50-54% precision with full traceability, making it a useful reference when the priority is immediate heuristic explainability (12).


10. Governance and Security: Managing bias, data contamination, and the AI Act

Finally, a strategic partner must demonstrate a holistic understanding of AI's implications, including data security, ethical frameworks, and governance. This is fundamental to building trust with enterprise clients.


The LLM-AR research explicitly addresses two of these risks:

  1. Data Contamination: This is a material risk. It occurs when an LLM has already "seen" the test data during its training (e.g., it read the founders' profiles on the internet). In that case, the model isn't predicting success; it's simply recalling a fact it already knows. The study actively mitigated this risk. Instead of feeding the LLM the founders' names, the text profiles were converted into anonymized, structured numerical features (1, 2). This prevents the LLM from "cheating" by remembering specific people and forces it to reason only about the profile traits (e.g., education_level=3, vc_experience=true). The literature documents the possibility of training data extraction from LLMs (7) and proposes taxonomies and methods for contamination detection (8, 9, 17); designing robust datasets and evaluations is critical.

  2. Transparency on Limitations (Bias and Prevalence Shift): Intellectual honesty requires stating a model's limits. The authors are clear: their dataset was curated to have a 10% success rate, while the real-world "market index" is 1.9% (1, 2). This "prevalence shift" means the performance (like 59.5% precision) (1) cannot be linearly transposed to the real world without caution.


This focus on data governance is crucial. With the European AI Act now in its implementation phase (approved in 2024 with 2025-2026 application deadlines), companies can no longer afford to treat AI as an unregulated experiment. For high-risk cases, the EU AI Act (Reg. (UE) 2024/1689) requires native AI system logging (Art. 12), meaningful documentation for deployers, and registration in an EU database for certain categories (4). The symbolic layer (ProbLog rules) supports post-hoc traceability and compliant "instructions for use" under the EU AI Act (4). Regulatory compliance, risk management, and the ability to explain why a model made a certain decision will become non-negotiable business requirements.


Conclusion: From "Bigger" to "Smarter"

The analysis of the LLM-AR framework offers a realistic, strategic perspective on the future of AI in the enterprise. It teaches us that the race toward ever-larger LLMs may not be the answer for the most complex business decisions. The real business opportunity lies not in brute force, but in architectural intelligence.

For entrepreneurs and executives, this means shifting the focus. Instead of asking, "Which LLM should I buy?" the strategic question becomes, "How can I build a hybrid system that integrates the intuition of generative models with the logic, transparency, and rigor of symbolic reasoning?"


State-of-the-art enterprise work is trending toward hybrids like LLM-AR: interpretable, human-modifiable, and tunable to business KPIs (1). Competing technologies, like traditional expert systems, were transparent but brittle, unable to handle real-world ambiguity. Pure LLMs, conversely, handle ambiguity but are opaque and unreliable.

The neurosymbolic approach is not a compromise; it's a superior synthesis (18, 19). For a manager, investing in this direction means investing in governance. It means building systems you can trust, audit, and defend in front of a board of directors or a regulator. Enterprise AI will not be "magic"—it will be engineered, measurable, and defensible.


Frequently Asked Questions (FAQ)

1. What is neurosymbolic AI and why does it matter to my company?

Neurosymbolic AI is a hybrid approach that combines the statistical pattern-matching of neural networks (like LLMs) with the transparent logic of Symbolic AI (which is good at reasoning) (18). It matters because it creates models, like LLM-AR, that are not only powerful but also interpretable, reliable, and verifiable (1, 14)—all critical requirements for high-stakes business decisions.

2. What is LLM-AR and how is it different from GPT-4?

LLM-AR (LLM-powered Automated Reasoning) is a framework, not a single model. It uses an LLM (like GPT-4 or DeepSeek) as an "intuition engine" to generate rules, but then uses a separate system (like ProbLog) to execute logical, probabilistic reasoning (1). GPT-4 is a "pure" LLM: it gives you an answer, but its internal decision process is an opaque "black box."

3. What does VCBench measure?

The VCBench benchmarks (2) quantify the "market index" (the baseline success rate in venture capital, around 1.9%) and show that the best human VCs improve on that index by a factor of 1.7-2.9x. Hybrid approaches like LLM-AR (1) push precision even further when the prevalence is fixed at 10% for a comparative evaluation.

4. Why does the LLM-AR study focus on "precision" over general accuracy?

In high-risk contexts like venture capital, the cost of a false positive (investing in a company that fails) is extremely high. "Precision" (how many of your "yes" bets were correct) is more important than general accuracy. The study prioritized precision (59.5%) (1) to minimize wasted resources.

5. What does it mean that the model is "tunable"?

It means that without retraining the entire system, you can adjust hyperparameters to change the model's behavior based on business goals. For example, you can "tune" LLM-AR to favor precision (for finance) or recall (for medical diagnostics) by optimizing the F-beta score (1).

6. What is ProbLog? Do I need to understand it to use AI?

ProbLog (Probabilistic Prolog) is a logic programming language that incorporates probability. It lets the system manage uncertainty (e.g., "there's a 70% chance X is true"). Executives don't need to program in ProbLog, but they need to understand why it matters: it's the engine that makes the AI's decisions transparent and based on verifiable rules and probabilities (1, 16), not on an incomprehensible "feeling."

7. What is "data contamination" and why is it a risk?

It's a serious methodological problem where an AI model is tested on data it has already "seen" during its training (7). This leads to inflated, unrealistic results (the model is "memorizing," not "reasoning"). It's a risk for businesses because a poorly tested model will fail dramatically on new, real-world data. Recent taxonomies help classify this risk (8, 9, 17).

8. What does "expert-in-the-loop" mean?

It's the opposite of total automation. It's a system design where the AI acts as a powerful assistant, but the human expert remains at the center of the process. The expert can read, understand, and even modify the rules and decisions the AI proposes (1). LLM-AR is designed for this, making AI an augmentation tool, not a replacement tool.

9. Does the European AI Act affect models like LLM-AR?

Yes. The AI Act (Reg. (UE) 2024/1689) (4), now in its implementation phase, places strict requirements on "high-risk" AI systems (used in finance, HR, medicine, etc.). It requires transparency, traceability, and robustness—for example, Article 12 on logging. An opaque "black box" system will struggle with compliance. An interpretable framework like LLM-AR, which exposes every decision path, is intrinsically better positioned to meet these regulatory demands.

10. Why did LLM-AR beat the human (VC Fund) benchmark in the study?

Hybrid systems like LLM-AR excel at large-scale, unbiased statistical analysis. They can identify statistically significant correlations (via "association-rule mining" (3)) that even an expert human might miss due to cognitive bias or a limited sample of experience. The AI doesn't get "tired" and analyzes all data with the same rigor.

11. What is the first step to implementing strategic AI in my company?

The first step isn't technology; it's strategy. It begins with an audit (like that proposed in the Rhythm Blues AI packages) to map business processes, identify high-value decisions, and define KPIs. You start with a low-risk, high-impact pilot project (as proven by BCG and McKinsey reports (5, 6)) to build competence and trust before scaling adoption.


How We Can Help

Adopting artificial intelligence is not a question of "if," but "how." A strategic approach—based on measurability, governance, and an agile roadmap—is the only way to turn hype into a competitive advantage.


If you want a direct discussion to examine your company's specific needs and identify the most valuable path for AI adoption, Rhythm Blues AI offers an exchange to evaluate opportunities and build a custom action plan.


To book a free 30-minute video call and explore how AI can make a concrete contribution to your business projects, please schedule an appointment at the following link:


Navigating the regulatory complexity of the AI Act and building a robust governance framework is a challenge that requires both technological and business expertise. If you feel the need for a guide to protect your company and turn compliance into a strategic asset, we can analyze your specific situation together and chart a clear, safe path forward.


Sources and References

  1. Chen R. et al. (2025) – LLM-AR: LLM-powered Automated Reasoning Framework, arXiv, 24/10/2025. https://arxiv.org/abs/2510.22034

  2. Chen R. et al. (2025) – VCBench: Benchmarking LLMs in Venture Capital, arXiv. https://arxiv.org/abs/2509.14448

  3. Agrawal R., Imieliński T., Swami A. (1993) – Mining Association Rules between Sets of Items in Large Databases, SIGMOD ’93. https://dl.acm.org/doi/10.1145/170035.170072

  4. Regolamento (UE) 2024/1689 (AI Act) – Official Text EUR-Lex. https://eur-lex.europa.eu/eli/reg/2024/1689/oj/eng

  5. BCG, “Where’s the Value in AI?” (2024 report). https://media-publications.bcg.com/BCG-Wheres-the-Value-in-AI.pdf

  6. McKinsey, “The State of AI: Global Survey 2025.” https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai

  7. Carlini N. et al. (2021) – Extracting Training Data from Large Language Models, USENIX Security. https://www.usenix.org/system/files/sec21-carlini-extracting.pdf

  8. Palavalli M. et al. (2024) – A Taxonomy for Data Contamination in Large Language Models, arXiv. https://arxiv.org/abs/2407.08716

  9. Cheng Y. et al. (2025) – A Survey on Data Contamination for LLMs, arXiv. https://arxiv.org/abs/2502.14425

  10. OpenAI, “Introducing o3 and o4-mini,” 16 Apr 2025. https://openai.com/index/introducing-o3-and-o4-mini/

  11. OpenAI, “GPT-4o mini: advancing cost-efficient intelligence,” 18 Jul 2024. https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/

  12. Griffin B. et al. (2025) – Random Rule Forest: Interpretable Ensembles of LLM-Generated Questions for Predicting Startup Success, arXiv. https://arxiv.org/abs/2505.24622

  13. Yi K. et al. (2018) – “Neural-Symbolic VQA,” NeurIPS. https://arxiv.org/abs/1810.02338

  14. WSJ (2025) – Why Amazon is Betting on “Automated Reasoning” to Reduce AI’s Hallucinations. https://www.wsj.com/articles/why-amazon-is-betting-on-automated-reasoning-to-reduce-ais-hallucinations-b838849e

  15. Gartner, Press Release: “Through 2022, 85% of AI projects will deliver erroneous outcomes,” 13 Feb 2018. https://www.gartner.com/en/newsroom/press-releases/2018-02-13-gartner-says-nearly-half-of-cios-are-planning-to-deploy-artificial-intelligence

  16. R. Manhaeve et al. (2019) – “DeepProbLog,”. https://arxiv.org/abs/1907.08194

  17. J. Chang et al. (2025) – “Challenging Common LLM Contamination Detection Assumptions,”. https://arxiv.org/abs/2502.14200

  18. C. Colelough, W. Regli (2025) – “A Systematic Review of Neuro-Symbolic AI…,” arXiv. https://arxiv.org/abs/2501.05435

  19. DeepSeek-AI (2025) – “DeepSeek-V3 Technical Report,” arXiv. https://arxiv.org/abs/2412.19437


Andrea Viliotti is an AI Strategy Consultant who acts as a "translator" between technology and the C-suite for CEOs, entrepreneurs, and executives. With 20+ years as an entrepreneur, his perspective blends a deep understanding of emerging technologies with a pragmatic, P&L-focused approach centered on measurable results and ROI. Through his proprietary "Rhythm Blues AI" method, he helps companies govern digital transformation, turning the complexity of AI into a sustainable competitive advantage. He is a contributor to leading publications like Agenda Digitale and AI4Business and shares his analysis via his blog(https://www.andreaviliotti.it/blog), YouTube channel(https://www.youtube.com/@Andrea-Viliotti), and podcast(https://podcasts.apple.com/us/podcast/andrea-viliotti/id1770291025). Connect with him on LinkedIn.

Commenti

Valutazione 0 stelle su 5.
Non ci sono ancora valutazioni

Aggiungi una valutazione
bottom of page