20

I am interested in FDA (data perceived as functions), as someone from a pure mathematics background and I think it can help provide solutions to some major challenges in data analysis (also data science), compared to the ordinary perspective in which data is perceived, as vectors. However, I have not found so much literature about it on the web. Could someone help me understand why this is so? Also, where can I get some good reads in FDA?

kjetil b halvorsen
  • 63,378
  • 26
  • 142
  • 467
KaRJ XEN
  • 303
  • 1
  • 8
  • 2
    I am of the same opinion as you and I do think this has great potential for practical impact. I liked this lecture series on you tube: https://www.youtube.com/watch?v=SUp_Nq8NwfE&list=PLOdMinGElTQenPYWhochrqgAa4M3wXdEP I have some research on this area and I am interested in exploring applications of it. https://pypi.org/project/scikit-fda/ is a python package for the same. R has some really nice packages. If you are looking for collaborators for research, I am interested. Ramsay and Silverman's book is the go to reference and a fantastic resource for it. – Rajiv Sambasivan Feb 16 '22 at 12:28
  • 2
    This recent paper got a decent amount of interest on Twitter: https://arxiv.org/abs/2201.12204. I think NERFs and SIREN networks are somewhat in the direction of FDA, but maybe I'm misunderstanding what it is. – Davis Yoshida Feb 16 '22 at 19:59

2 Answers2

12

As someone with almost zero knowledge in FDA but who has recently started reading and thinking about it, here are some of my thoughts on why FDA is not so popular these days. Please take them with a grain of salt as I am far from being an expert :

  • Most of the Data Science problems people are interested in solving "do not care" about the nature of the data. What I mean is, when trying to do a task such as e.g. regression, classification, or clustering, statisticians will go for the methods that yield the best models for a minimal computational cost. Depending on context, what the definition of "best" is, the available information etc, one may choose different methods which themselves may rely on different possible representations of the data, such as vectors, matrices, graphs, smooth manifolds, functions... Therefore, when a dataset can be nicely represented as a tensor that you can plug into a CNN with solid guarantees on model performance, why would you bother picking the functional representation ?
    On the other hand, there are problems where you are trying to infer information on the functions you sample from themselves, in that specific case there is no way around using FDA (see here for some examples).
  • That brings me to my next point : Functional Data, in practice, is always stored as a (high-dimensional) finite-dimensional object, so the FDA paradigm is never really needed. Indeed, even if the data is in practice a set of realisations of a function depending on continuous parameter(s), which typically would be space and/or time, what you actually have stored in your computer is a discretised version of it. Although it is true that when the "mesh size" is sufficiently small, you are close to dealing with "real" functional data such as e.g. Random Fields, Stocks prices, Video recordings etc... It works fine in practice to consider that data as high (but finite) dimensional objects.
  • And here is, imo, the most crucial point : There are plenty of non-FDA specific algorithms that perform very well on functional data. There are indeed countless examples of successful processing of all kinds of functional data such as Video recordings, audio recordings, satellite imagery and many more. On the other hand, there hasn't been, as far as I know, many (or any) big breakthroughs results justifying the superiority of FDA-specific methods for functional dataover more conventional ones outside of specific context. All hope is definitely not lost, as there definitely are a few theoretical arguments here and there suggesting that the FDA framework can be far superior to the finite-dimensional one in some cases, such as this paper by Delaigle and Hall : Achieving near perfect classification for functional data (2012) which shows that "perfect classification" is possible for functional data under very mild assumptions whereas it is definitely not the case in the finite-dimensional setting. In practice however, it seems that dimensionality reduction + classical methods works just fine.
  • Lastly, I think another factor is that the mathematical knowledge required to be able to contribute to research in FDA tends to be outside of the expertise of most statisticians. Indeed, the proposed algorithms in the literature often rely on rather deep results in Functional Analysis that many statisticians might not be too familiar with, and there are other issues such as, e.g., defining a meaningful measure on a space of functions which are more profound and less likely to interest statisticians who are generally more experts in stuff like Linear Algebra, Optimization, Concentration Inequalities, VC-Theory etc... I think it is kind of the same with Topological Data Analysis (although it seems to be gaining a lot of traction recently) : the ideas are very promising, but require a deep knowledge in pure maths concepts such as homology, Betti numbers or Euler characteristic... to be applied and further improved, which is a knowledge many statisticians do not have.

Even though I said all that, I do believe there is plenty of interesting things to do in that subfield and its full potential hasn't been reached at all. I have mostly read papers related to the problems I am interested in, so I don't have much to recommend, but I have read some of Hsing and Eubank and find it pretty great so far. I also found this review by Wang, Chiou and Müller to be pretty comprehensive to get a rough idea of the current state-of-the-art.

StratosFair
  • 353
  • 2
  • 10
  • 2
    I was going to comment that "needing to know functional analysis" is a reason, but you beat me to it! – hasManyStupidQuestions Feb 16 '22 at 16:05
  • 3
    +1, (almost) all good points. Few comments: 1. H&E book is *not* a book on FDA but rather on the *foundations necessary* for FDA. (I was lucky to have Prof. Hsing make this point in person...) 2. The W, C & M paper is excellent. 3. FDA is at times *just* non-parametric statistics. Especially when working with sparse data. 4. I don't agree with the second point fully, e.g. from Signal processing: we still care for CFT results even if DFT is the main work-horse behind the DSP. 5. "why FDA is not so popular these days" << I would say FDA is the most popular it has been ever nowadays... – usεr11852 Feb 16 '22 at 16:17
  • 1
    I agree with your point that the knowledge burden of FDA is high, but I disagree with the statement that there are plenty of non-FDA specific algorithms that work well on functional data. When you're working with problems that are naturally posed over Sobolev spaces, which is almost always the case in the geophysical sciences, ignoring the functional analysis issues results in algorithms whose convergence properties strongly depend on the discretization and this is *very* bad. – Daniel Shapero Feb 16 '22 at 18:51
  • “Therefore, when a dataset can be nicely represented as a tensor that you can plug into a CNN” – ah, but the CNN architecture has a data=function ($\mathcal{L}^2([0,1]^2)$) assumption all baked into it! — Similarly, “Functional Data, in practice, is always stored as a (high-dimensional) finite-dimensional object” – e.g. image data is actually stored in compressed files which, again, very much have assumption about the function space in them. (But storage arguably doesn't really matter.) — What I'm getting at: the methods people use in practice actually suggest FDA _should be more_ popular. – leftaroundabout Feb 17 '22 at 18:36
7

FDA was late to the party and benefits against established applications are at times incremental. In many cases, "standard longitudinal data techniques" check most boxes already. FDA does provide some unique advantages on specific use cases (e.g. working with density functions or covariance surfaces as the unit of analysis) but those are specialised and rather recent applications.

usεr11852
  • 33,608
  • 2
  • 75
  • 117