11

In regression in general and in linear regression in particular causal interpretation about parameters is sometimes permitted. At least in econometrics literature, but not only, when causal interpretation is permitted is not so clear; for a discussion you can see: Regression and Causation: A Critical Examination of Six Econometrics Textbooks - Chen and Pearl (2013).

For a proper handling of causality in statistical model the best way probably is to use Structural Causal Model as explained, for example (shortly), in: Trygve Haavelmo and the Emergence of Causal Calculus – Pearl 2012 feb.

However, currently, these are not the standard method in basics econometrics model (classic multiple linear regression). Indeed is frequently used the concept of “true model” or “data generating process” that sometime have explicit causal meaning. In any case I want to consider only causal sense. Therefore if we estimate the sample counterpart of “true model” we achieve causal interpretation about parameters.

Keeping in mind the above consideration, my attempt is to grasp

  • the link between the concept of “true model” (of current econometrics textbooks) and structural causal model (of Pearl) … if any.

  • The link between the previous point and the concept of randomized controlled experiment, as used in laboratory, that sometimes is the reference point in econometrics observational study (as good as it). For example Stock and Watson (2013) spend a lot of discussion about that (particularly cap 13). Moreover in Pearl 2012feb pag 14 there is a debate review between “structuralists” and “experimentalists” that is strongly related with this point.

Can you explain me something about these two point in the simplest as possible scenario?

markowitz
  • 3,964
  • 1
  • 13
  • 28
  • 2
    There are multiple schools of thought about causality in econometrics. See e.g., Heckman and Pinto (2015) commenting on Pearl, or the applications of the Rubin potential outcomes model (for example in Mostly Harmless Econometrics by Angrist and Pischke). – Frank Nov 14 '18 at 18:25
  • there are no "true" models in economtrics – Aksakal Nov 14 '18 at 22:31
  • 1
    @Frank: thank you for the paper suggested, they will be surely useful. Regarding the Angrist and Pischke book, I read it. I surveyed about ten econometrics textbooks and Mostly Harmless Econometrics represent the best about causality (my opinion but probably not only). I learned a lot from that. However, at least in my opinion, even this book is exhaustive. For example it says nothing about structural causal model and link between it and potential outcome language. Finally it does not offer an answer a my two question above. Or, at least, I "don't see" it. – markowitz Nov 15 '18 at 08:36
  • 1
    @Aksakal: answer like your says both, everything or nothing. In my opinion says nothing because it don't says me nothing about: firstly, why in many econometric books "true model" is used (and seems very useful if not essential); what are the links among your answer and my questions? If links do not exist the presence of "true model" became very hard to explain. In any case I known that "true model" is sometimes criticized concept and that "real world" is something different and links between them can be artificial, but this is the theory. My question is about econometric theory. – markowitz Nov 15 '18 at 08:52
  • @markowitz: Angrist and Pischke do discuss this, but their language might be confusing you. See section 3.2.1: "the function $f_i(s)$ tells us what $i$ would earn for _any_ value of schooling, $s$. In other words, $f_i(s)$ answers causal 'what if' questions." See also the discussion of the linear constant effects causal model (at the end of the section). – Frank Nov 15 '18 at 14:01
  • @markowitz, it was a comment, not an answer. The subject matter is too unstable to have a "true" model in sense that we have models in natural sciences. Instead, economic "models" are "stories" and vignettes. That's the reason why economists love "stories", and always talk about them when you discuss the papers. This still just a comment, not an answer. – Aksakal Nov 15 '18 at 15:19
  • @markowitz, my views on econometric models is similar to Rubinstein's description of microeconomics, see the first part of the Introduction in his lecture notes here http://assets.press.princeton.edu/rubinstein/front.pdf See also his review of the book http://arielrubinstein.tau.ac.il/papers/100.pdf , He says this about the models: "A story is not meant to be “useful” in the sense that most people use the word... During my forty years in the profession, I have not encountered a single model that could persuade me that economic theory might have direct practical use." – Aksakal Nov 15 '18 at 15:33
  • @Aksakal, you are wrong. Have you ever opened an econometrics textbook? Start with the classics such as Greene "Econometric Analysis", Kennedy "A Guide to Econometrics" or Wooldridge "Introductory Econometrics" - all of them include the notion of the "true model". – Richard Hardy Nov 24 '18 at 12:05
  • @RichardHardy, I know they talk about them "true model", but it's delusional. None of their models are true in conventional sense. – Aksakal Nov 24 '18 at 16:13
  • @Aksakal, who defines the conventional sense? Or where can such a definition be found? My first bet would be on a classical textbook in the field. I presume you are not an econometrician by education and what seems conventional to you might be specific to you alone, or perhaps to your field. It is the same as if I was going to define what is conventional in, say, chemistry (I am not educated as a chemist but an econometrician). Even using the word "delusional" in this context is not only ignorant but also inappropriate, IMHO. – Richard Hardy Nov 24 '18 at 16:27

3 Answers3

6

I will start with the second part of your question, which pertains to the difference between randomized control studies and observational studies, and will wrap it up with the part of your question pertaining to "true model" vs. "structural causal model".

I will use one of Pearl's examples, which is an easy one to grasp. You notice that when the ice cream sales are highest (in the summer), the crime rate is highest (in the summer), and when the ice cream sales are lowest (in the winter), the crime rate is lowest. This makes you wonder whether the level of ice cream sales is CAUSING the level of crime.

If you could perform a randomized control experiment, you would take many days, suppose 100 days, and on each of these days randomly assign the level of sales of ice cream. Key to this randomization, given the causal structure depicted in the graph below, is that the assignment of the level of ice cream sales is independent of the level of temperature. If such a hypothetical experiment could be performed you should find that on the days when the sales were randomly assigned to be high, the average crime rate is not statistically different than on days when the sales were assigned to be low. If you had your hands on such data, you'd be all set. Most of us, however, have to work with observational data, where randomization did not do the magic it did in the above example. Crucially, in observational data, we do not know whether the level of Ice Cream Sales was determined independently of Temperature or whether it depends on temperature. As a result, we'd have to somehow untangle the causal effect from the merely correlative.

Pearl's claim is that statistics doesn't have a way of representing E[Y|We Set X to equal a particular value], as opposed to E[Y|Conditioning on the values of X as given by the joint distribution of X and Y]. This is why he uses the notation E[Y|do(X=x)] to refer to the expectation of Y, when we intervene on X and set its value equal to x, as opposed to E[Y|X=x], which refers to conditioning on the value of X, and taking it as given.

What exactly does it mean to intervene on variable X or to set X equal to a particular value? And how is it different than conditioning on the value of X?

Intervention is best explained with the graph below, in which Temperature has a causal effect on both Ice Cream Sales and Crime Rate, and Ice Cream Sales has a causal effect on Crime Rate, and the U variables stand for unmeasured factors that affect the variables but we do not care to model these factors. Our interest is in the causal effect of Ice Cream Sales on Crime Rate and suppose that our causal depiction is accurate and complete. See the graph below.

enter image description here

Now suppose that we could set the level of ice cream sales very high and observe whether that would be translated into higher crime rates. To do so we would intervene on Ice Cream Sales, meaning that we do not allow Ice Cream Sales to naturally respond to Temperature, in fact this amounts to us performing what Pearl calls "surgery" on the graph by removing all the edges directed into that variable. In our case, since we're intervening on Ice Cream Sales, we would remove the edge from Temperature to Ice Cream sales, as depicted below. We set the level of Ice Cream Sales to whatever we want, rather than allow it be determined by Temperature. Then imagine that we performed two such experiments, one in which we intervened and set the level of ice cream sales very high and one in which we intervened and set the level of ice cream sales very low, and then observe how Crime Rate responds in each case. Then we'll start to get a sense of whether there is a causal effect between Ice Cream Sales and Crime Rate or not.

enter image description here

Pearl distinguished between intervention and conditioning. Conditioning here refers merely to a filtering of a dataset. Think of conditioning on Temperature as looking in our observational dataset only at cases when the Temperature was the same. Conditioning does not always give us the causal effect we're looking for (it doesn't give us the causal effect most of the time). It happens that conditioning would give us the causal effect in the simplistic picture drawn above, but we can easily modify the graph to illustrate an example when conditioning on Temperature would not give us the causal effect, whereas intervening on Ice Cream Sales would. Imagine that there is another variable which causes Ice Cream Sales, call it Variable X. In the graph is would be represented with an arrow into Ice Cream Sales. In that case, conditioning on Temperature would not give us the causal effect of Ice Cream Sales on Crime Rate because it would leave untouched the path: Variable X -> Ice Cream Sales -> Crime Rate. In contrast, intervening on Ice Cream Sales would, by definition, mean that we remove all arrows into Ice Cream, and that would give us the causal effect of Ice Cream Sales on Crime Rate.

I will just mention that one Pearl's greatest contributions, in my opinion, is the concept of colliders and how conditioning on colliders will cause independent variables to be likely dependent.

Pearl would call a model with causal coefficients (direct effect) as given by E[Y|do(X=x)] the structural causal model. And regressions in which the coefficients are given by E[Y|X] is what he says authors mistakenly call "true model", mistakenly that is, when they are looking to estimate the causal effect of X on Y and not merely to forecast Y.

So, what's the link between the structural models and what we can do empirically? Suppose you wanted to understand the causal effect of variable A on variable B. Pearl suggests 2 ways to do so: Backdoor criterion and Front-door criterion. I will expand on the former.

Backdoor Criterion: First, you need to correctly map out all the causes of each variable and using the Backdoor criterion identify the set of variables that you'd need to condition on (and just as importantly the set of variables you need to make sure you do not condition on - i.e. colliders) in order to isolate the causal effect of A on B. As Pearl points out, this is testable. You can test whether or not you've correctly mapped out the causal model. In practice, this is easier said than done and in my opinion the biggest challenge with Pearl's Backdoor criterion. Second, run the regression, as usual. Now you know what to condition on. The coefficients you'll get would be the direct effects, as mapped out in your causal map. Note that this approach is fundamentally different from the traditional approach used in estimating causality in econometrics - Instrumental Variable regressions.

ColorStatistics
  • 2,699
  • 1
  • 10
  • 26
  • I understand your explanation but I fear that it does not give an answer about my question. Except, maybe, the final phrase but I partially disagree with it. You firstly explain the usefulness of an experiment. I agree with you about that. After you said that (shortly) in observational context we have to “somehow untangle the causal effect from the merely correlative” without experimental manipulations. I known. – markowitz Nov 15 '18 at 11:14
  • After, since the conceptual difference between conditioning and intervention, you introduce Pearl interventional notation’s. I understand the usefulness of interventional notation. I’m not confident with causal graph but surely it is useful too. Pearl’s contributions are important. Finally you said that: “Pearl would call a model with causal coefficients (direct effect) as given by $E[Y|do(X=x)]$ the structural causal model.”. Ok, I understand. – markowitz Nov 15 '18 at 11:15
  • 1
    “regressions in which the coefficients are given by $E[Y|X]$ is what he says authors mistakenly call "true model", mistakenly that is, when they are looking to estimate the causal effect of $X$ on $Y$ and not merely to forecast $Y$.” This seems me not strictly true because $E[Y|X]$ is not always the true model. Any OLS regression is a conditional expectation but not every of these are counterpart of the true model frequently used in econometric textbooks. – markowitz Nov 15 '18 at 11:15
  • Let me use your/Pearl example: $X1$= temperature, $X2$= ice cream quantity, $Y$= crime rate. It’s possible to define the true (causal) model as: $Y = beta_0 + beta_1 *X1 + beta_2 * X2 + u$ but we estimate the underspecified model $Y = alfa_0 + alfa_1 *X1 + e$. It’s possible demonstrate that alfas parameters are biased estimates of the true causal parameters (betas); however the latter remain a conditional expectation. This is omitted variables problem. Indeed if we estimate the correct specified form, or a “longer”, the estimates parm. are unbiased and its causal interpretation is permitted. – markowitz Nov 15 '18 at 11:16
  • The explanation, at least in simple case like above, can be complete also without Pearl’s (and related) contributions. Exactly for reasons like the above I looking for the relations among: Structural Causal Model (from Pearl), experimental language (from Rubin); usual true model used in its causal sense (from many econometrics textbooks). I’m convinced that links exist but I’m doubt about the form of them. – markowitz Nov 15 '18 at 11:17
  • This answer is a good exposition of the Pearl framework, but it doesn't shed any light on OP's question: what's the relationship to the methods used in econometrics? – Frank Nov 15 '18 at 16:03
  • Good feedback. Thank you. I added 2 paragraphs to the end to address the link to econometrics practice and pointed out that it is different than Instrumental Variables. Maybe somebody else can directly attack the difference between Pearl's Backdoor Criterion and Instrumental Variables, as I believe this is the implicit heart of this not-so-surgically stated question. – ColorStatistics Nov 15 '18 at 16:32
  • Thanks @ColorStatistics. To your recent addition: econometrics sometimes uses selection on observables (Conditional Independence Assumption), which appears to be similar to (but less rigorous than) the backdoor criterion. – Frank Nov 15 '18 at 16:58
  • Do I understand correctly that “Variable X -> Ice Cream Sales -> Crime Rate” is a problem only when Variable X affects also Crime Rate through a path that does not go via Ice Cream Sales, just like it is with Temperature? Otherwise, controlling for Temperature would still be enough. – Ivan Nov 18 '20 at 06:31
5

In the context of the Pearl paper you've given, what most econometricians would call a true model is input I-1 to the Structural Causal Model: a set of assumptions $A$ and a model $M_A$ that encodes these assumptions, written as a system of structural equations (as in Models 1 and 2) and a list of statistical assumptions relating the variables. In general, the true model need not be recursive, so the corresponding graph can have cycles.

What's an example of a true model? Consider the relationship between schooling and earnings, described in Angrist and Pischke (2009), section 3.2. For individual $i$, what econometricians would call the true model is an assumed function mapping any level of schooling $s$ to an outcome $y_{si}$: $$ y_{si} = f_i(s). $$ This is exactly the potential outcome. One could go further and assume a parametric functional form for $f_i(s)$. For example, the linear constant effects causal model: $$ f_i(s) = \alpha + \rho s + \eta_i. $$ Here, $\alpha$ and $\rho$ are unobserved parameters. By writing it this way, we assume that $\eta_i$ does not depend on $s$. In Pearl's language, this tells us what happens to expected earnings if we fix an individual's schooling at $s_i = s_0$, but we don't observe $\eta_i$: $$ E[y_{si} \mid do(s_i = s_0)] = E[f_i(s_0)] = \alpha + \rho s_0 + E[\eta_i]. $$ We haven't said what queries we're interested in, or what data we have. So the "true model" is not a full SCM. (This is generally true, not just in this example.)

What's the connection between a true model and a randomized experiment? Suppose an econometrician wants to estimate $\rho$. Just observing $(s_i, y_i)$ for a bunch of individuals isn't sufficient. This is identical to Pearl's point about statistical conditioning. Here $$ E[y_{si} \mid s_i = s_0] = E[f_i(s_0) \mid s_i = s_0] = \alpha + \rho s_0 + E[\eta_i \mid s_i = s_0]. $$ As Angrist and Pischke point out, $\eta_i$ may be correlated with $s_i$ in observational data, due to selection bias: an individual's decision about schooling might depend on her value of $\eta_i$.

Randomized experiments are one way to correct for this correlation. Using Pearl's notation loosely here, if we randomly assign our subjects to $do(s_i = s_0)$ and $do(s_i = s_1)$ then we can estimate $E[y_{si} \mid do(s_i = s_1)]$ and $E[y_{si} \mid do(s_i = s_0)]$. Then $\rho$ is given by: $$ E[y_{si} \mid do(s_i = s_1)] - E[y_{si} \mid do(s_i = s_0)] = \rho(s_1 - s_0). $$

With additional assumptions and data, there are other ways to correct for the correlation. A randomized experiment is only considered the "best" because we may not believe the other assumptions. For example, with the Conditional Independence Assumption and additional data, we could estimate $\rho$ by OLS; or we could bring in instrumental variables.

Edit 2 (CIA): This is mainly a philosophical point, and Angrist and Pischke may disagree with my presentation here. The Conditional Independence Assumption (selection on observables) lets us correct for selection bias. It adds an assumption about joint distributions: that $$ f_i(s) \perp\!\!\!\perp s_i \mid X_i $$ for all $s$. Using just conditional expectation algebra (see the derivation in Angrist and Pischke) it follows that we can write $$ y_i = f_i(s_i) = \alpha + \rho s_i + X_i' \gamma + v_i $$ with $E[v_i \mid X_i, s_i] = 0$. This equation allows us to estimate $\rho$ in the data using OLS.

Neither randomization nor the CIA goes into the system of equations that defines the true model. They are statistical assumptions that give us ways to estimate parameters of a model we've already defined, using the data we have. Econometricians wouldn't typically consider the CIA part of the true model, but Pearl would include it in $A$.

Frank
  • 291
  • 1
  • 8
  • I already see the Angrist and Pischke example and I thought about it. That was the best example/explanation that I never found. Your addition represent what I was looking for. Thank you. – markowitz Nov 15 '18 at 22:31
  • However let me add some points and specific questions. Angrist and Pischke call the model as “linear constat effect causal model” but my doubt was exactly if that model was interpretable as “true model” that I usually see in many econometrics textbooks (for simple example see my comment on ColorStatistics answer). My favourite answer was no, because in the usual “true model” is explicitly valid some form of exogeneity that is indeed precisely violated in the present case. This is exactly your final consideration about my phrase but I was aware of this problem. – markowitz Nov 15 '18 at 22:31
  • Now, Angrist and Pischke never speak about “true model” but you suggest this interpretation for it. Probably you are right. In any case you write: “So the "true model" is just one part of the SCM.”. Ok! Your is a well posed form of one guess that I already had in mind. This conclusion hold even in the case of the “true model” that I was remember before? – markowitz Nov 15 '18 at 22:31
  • In the Angrist and Pischke's exposition the CIA plays a key role, and actually play a key role in causality. This is essential for eliminating selection bias. However in the example the causal assumptions explicitly preceded the CIA intervention. I think that if we take the general model $y = a + bx + cZ + e$, where $Z$ is a set of controls, the CIA condition (on $Z$) is never enough. Causal assumption must precede $y$,$x$ relationship shown in the equation. It’s correct? I think yes otherwise the saying “no causal assumpions no causal conclusions” is violated. – markowitz Nov 15 '18 at 22:32
  • Finally in the case of randomized controlled experiment, is correct to says that randomization (about sampling and treatment assignment) is enough in the last case because even if causality assumptions are not explicit are nested in randomization? – markowitz Nov 15 '18 at 22:32
  • @markowitz I added a section on the CIA. Does this address your question? – Frank Nov 16 '18 at 00:19
  • Frank, there are some imprecise statements here. CIA is a causal assumption (constraints on the causal model is a causal assumption, and CIA imposes constraints on the potential outcome---a causal quantity). Also, it's $E[Y|do(s)] = E[Y_{s}]$, when you $do(s)$ that's already the potential outcome had $s$ been $s$. – Carlos Cinelli Nov 16 '18 at 03:11
  • So this sentence is mathematically incorrect: "Neither randomization nor the CIA adds any causal content – nothing there goes into the system of equations that defines the true model. They are only statistical relationships that give us ways to estimate parameters of a model we've already defined, using the data we have." – Carlos Cinelli Nov 16 '18 at 03:14
  • As to SCMs, it seems you are conflating partial specification with the definition of an SCM. In your model, you don't need to give all inputs of an SCM, you give the assumptions you can defend. That is, usually your give only a partial specification of the model. You can give a partial specification of an SCM with a DAG, for instance. Or, in your example, with the CIA, $Y_{s} \perp S$, which defines a class of SCMs. – Carlos Cinelli Nov 16 '18 at 03:17
  • @CarlosCinelli I've removed the statement that the CIA isn't causal, but the CIA is entirely separate from the statement that $E[Y \mid do(s)] = E[Y_s]$ (that came much earlier). I don't get your second point – the system of equations (as given here) is one of the inputs of an SCM, isn't it? – Frank Nov 16 '18 at 21:54
  • @Frank no, the SCM (just like Rubin's "truth table", a table with all potential outcomes filled in) is the complete causal model. If you have that, then you don't need to estimate anything. The inputs in practice are partial specifications of the causal model. The CIA, for instance, is one such partial specification (you are stating that potential responses do not depend on treatment assignment, given a set of covariates). A DAG, for instance, is another such partial specification (you are stating qualitative functional relationships among variables). – Carlos Cinelli Nov 16 '18 at 21:59
  • 1) I think you misunderstand the CIA, see above (I see you fixed this) 2) I'm having trouble connecting your definition of the SCM with the definition given in "Trygve Haavelmo and the Emergence of Causal Calculus" – Frank Nov 16 '18 at 22:00
  • 1
    @Frank you can give any partial specification of a structural causal model, the set of assumptions $A$ that Pearl refers to. Traditionally, this will be qualitative statements of the functional relationships that you can defend on scientific grounds. The CIA is a partial specification of an SCM as well, you are asserting that the potential response $Y_{s}$ does not depend on $S$ given a set of covariates, this imposes constraints on the possible systems of equations. – Carlos Cinelli Nov 16 '18 at 22:07
  • @Frank: about the definitional question of “true model” and particularly to your first Edit e my comment let me add that the model used in Angrsit and Pischke (2009), essentially the above example, is not the usual “true model”. In fact they speak about “linear causal model”. This position is clarified in the article Undergraduate Econometrics Instruction: Through Our Classes, Darkly; Angrist and Pischke (2017). – markowitz Feb 04 '19 at 17:10
  • The authors do not have a great opinion about the so called “true model” and in fact affirm: “The pedagogical emphasis on statistical efficiency and functional form, along with the sophomoric narrative that sets students off in search of “true models” as defined by a seemingly precise statistical fit, is ready for retirement” pag 126. The critics is exactly about the search of the true model that is the usual (old) econometrics goal. In the old sense, what I had in mind, my sentence that you underscore is correct. – markowitz Feb 04 '19 at 17:10
  • @markowitz I don't understand your objection -- Angrist and Pischke describe common practice in their book, and then criticize it in their article. It's still common practice. – Frank Feb 06 '19 at 16:28
  • @Frank: No Frank. Angrist and Pischke, in the 2017 article, criticize the “old style econometrics” that is still common practice, or at least is so in many currently econometrics books. In his 2009 book the same authors use the “new style econometric”. In the old style econometric the “true model” is a key concept and in it control variables is absent or no important. In the new style econometric the “true model”, at least in the usual sense understood, is absent and the controls variables is very important. – markowitz Feb 07 '19 at 11:04
  • Nonetheless, the exposition in AP 2009, section 3.2, is faithful to old style econometric practice, and what it calls the "true model" is consistent with old style econometrics. – Frank Feb 07 '19 at 15:30
  • No Frank, I'm sorry but it's not like that. AP 3.2 is exactly the point of the book where the distinction in argument become more clear. AP 3.2 is completely inside new style econometrics. In this section the “true model” is never mention, is mention “linear causal model” that is something else. In this section the difference between causal variable (of interest) and controls variables is crucial and clear. CIA is used primarily for this distinction. – markowitz Feb 07 '19 at 16:43
  • In the “old style” (past and also present books) so called “true model”, always with exogenous error, is used. Distinction between causal and controls variables is absent or not clear and, in fact, CIA is not used. – markowitz Feb 07 '19 at 16:43
  • Apparently comments are not for extended discussion; we can take this up in chat. I'll just point you to AP 2009 pages 58-59, the 1981 paper cited there, and every "old style" econometrics paper that uses instrumental variables (of which there are many). – Frank Feb 07 '19 at 23:19
  • Frank, you're right, the comments are not for extended discussions. Let me take up my point again. I appreciate a lot your answer, this is what I looked for. The “objection” the I made is primarily terminological. The sentence " Therefore if we estimate the sample counterpart of “true model” we achieve causal interpretation about parameters." seems me not wrong, rather it represent one "mantra" of old style econometrics. In my opinion for benefit of other readers is better to delete or modify your Edit. That’s it. – markowitz Feb 08 '19 at 10:49
  • Others argument as Matching and/or IV estimators are related but different question that is better do not insert here. – markowitz Feb 08 '19 at 10:49
0

The use of ‘causal’ in such regression/correlation based approaches is misleading, in my opinion. Path analysis, structural equation modeling, Granger causality, etc attempts to license causal inferences by imposing some fairly tenuous assumptions. In the case of Structural equation modeling for instance, the paths are directional and A appears to ‘cause’ B, but this simply means that the model as structured is ‘plausible’ in that it reproduces an observed covariance matrix (in fact, the direction of the paths don’t even matter much - just the constraints).

HEITZ
  • 1,682
  • 7
  • 15
  • You write: "In the case of Structural equation modeling for instance, the paths are directional and A appears to ‘cause’ B, but this simply means that the model as structured is ‘plausible’ in that it reproduces an observed covariance matrix" I disagree. Structural equation do not have the role of "reproduces" any statistical quantity. Moreover this question is quite old. Main points are no more obscure for me. I summarized main points here: https://stats.stackexchange.com/questions/493211/under-which-assumptions-a-regression-can-be-interpreted-causally/493905#493905 – markowitz Dec 03 '21 at 11:36