Types of Modelling - Statistical vs Discretionary

[ quant ]

So this recently occurred to me while trying to apply statistical modelling to a problem (in my opinion) that was clearly not suited to it. Rather, my prior is that the problem could be solved via discretionary modelling, aka having a hypothesis or thesis about how a series of events in the future would play out.

This got me thinking, what are the different types of modelling approaches? What even is a definition of model?

Modelling - What is it Anyway?

I guess this is a fairly safe definition. To model is to make some prediction about the future, given past data. I suppose there are 3 main types of models: statistical, discretionary, and numerical. Statistical modelling refers to exactly what it says on the tin: your regressions, ML, econometrics, whatever. Discretionary modelling refers to constructing an argument or thesis about what someone will do in the future. And numerical refers to simulating a system with known rules/laws forward in time and generating a result.

With that in mind, we can come up with a nice table:

A Classification of Models

Name Data Mechanism Models
Statistical Modelling (Backwards) Numerical, tabular records of past observations Uncovering statistical relationships (correlations) in past data with the assumption that they can be used for pr ediction in future Regression, ML, etc
Discretionary Modelling (Forwards) Current information/events, past history (what actors did in the past given current point-in-time info) Having a view on what several actors will do in the future and the outcome associated with that Thesis, view on situation evolving
Numerical Modelling Measurements Knowing the physical laws of a system and simulating them forward in time No idea

And:

Name Skill When to Use Examples
Statistical Modelling Understanding underlying mathematical theory, implementing it in code, nature of data and thing you are modelling Many participants in a system (e.g millions), individual actor contribution is small, cannot individually predict each of their behavior, BUT. on aggregate, behave in a manner where relationships can be extracted by a statistical model. Relationships are more stationary, aka past predicts future . Predicting returns
Discretionary Modelling Understanding the individual entities and how they will behave based on new/current information and their past behavior. Relatively fewer participants, individual actors have massive contribution. Actors or events cannot be predicted based on past behavior, e.g lots of random shocks. Past data cannot predict future events or behavior Thesis or view, say in commodities, macro, politics, etc
Numerical Modelling Being able to simulate these things in a compute intensive fashion Behaviors are fixed Weather forecasting

Statistical vs Discretioanry

A clear example of the former vs latter and their interplay would be in finance; say quant vs discretionary LS equity. The former might rebalance daily. The latter might rebalance quarterly.

The former uses statistical modelling to model human behavior over a shorter term horizon. The latter uses discretionary modelling or a thesis of how markets would behave to make a decision. But fundamentally, both lead to the same type of action (aka buying or selling an asset).

A physical commodity trader told me “We don’t rely too much on historical data as markets evolve too quickly”. To me, that is clear evidence that discretionary modelling is prevalent in most physical commodities. All the more so for commodities that have high exposure to the geopolitical factor - a very human-centric thing. What kind of statistical model could predict where Israel is going to hit Iran in response to the missile strikes?

Numerical

Numerical modelling is what we see in applied math/physics, when we build out a set of laws then derive some model (by simulation or theory) to see how a situation or measurements evolve across a time horizon.

How might we deduce these laws? From what I see on X, physics-trained quants are especially good at this. First, they observe the situation. They make assumptions (e.g this phenomenon seems to follow this law) without being too concerned about pinpoint accuracy. They do a quick and dirty estimation, simulating some laws/parameters and them developing a toy model. In this case, there’s no statistical element, just some ‘rules’ that are then applied numerically.

Integrating Discretionary Views in Statistical Modelling

Is it possible to blend or combine discretionary views in statistical modelling in a sensible fashion? Perhaps. One could, get a discretionary analyst to create a feature representing his view of some event as a numerical scale (e.g 1-10), then incorporate it into a statistical model.

In a regression, for example, this would let us quantify the impact in terms of the beta of the feature, as well as the correlation to other features.

But of course, by definition, a discretionary view is something that is difficult to quantify, hence discretionary, so the scale would be inherently imprecise. But that’s about the best way I could see it being done.