ТОП просматриваемых книг сайта:
Multiblock Data Fusion in Statistics and Machine Learning. Tormod Næs
Читать онлайн.Название Multiblock Data Fusion in Statistics and Machine Learning
Год выпуска 0
isbn 9781119600992
Автор произведения Tormod Næs
Жанр Химия
Издательство John Wiley & Sons Limited
1.6 Some History
The history of multiblock data analysis methods goes back a long time. One of the starting points was principal component analysis (PCA, Pearson (1901)). Another early method in statistics was canonical correlation analysis (Hotelling, 1936a) which led to development of many related methods. In the social sciences, path-models were developed, such as LISREL and PLS (Jöreskog and Wold, 1982) which are also used in consumer science and marketing research. In psychometrics, inter-battery factor analysis (Tucker, 1958) was developed and later simultaneous component analysis methods (Ten Berge et al., 1992) which have also sparked many alternatives. In parallel in chemometrics, methods such as consensus PCA and hierarchical PCA were developed (Westerhuis et al., 1998). Ideas on the latter two methods already started in the 1970s at conferences. In the French data analysis school, multiple correspondence analysis was developed (Benzécri, 1980) and many other methods. In multiway data analysis (which can be regarded as a subset of multiblock data analysis methods) the earliest developments started with the work of Cattell (1944), Carroll and Chang (1970) and Harshman (1970). For a history of the latter, see Smilde et al. (2004). Figure 1.8 tries to systematically show many multiblock data analysis methods and how they relate to basic statistical methodology and subspace projection methods from various fields of applied data analysis.
Figure 1.8 Phylogeny of some multiblock methods and relations to basic data analysis methods used in this book.
1.7 Fundamental Choices
In any sort of multiblock data analysis, choices have to be made such as which method to use and what kind of pre-processing to apply. Two fundamental questions which always should be considered (and dealt with) are highlighted below.
Variation explained:Do we only want to explain variation between blocks or also within blocks?Fairness:Should all blocks play a role in the final solution or can we allow some of the blocks to be dominant in this respect?
The first concept – variation explained – pertains to the choice that we can model the variation within the blocks and/or the variation between the blocks. Each multiblock data analysis method makes a different choice in this respect and thus it is up to the user to decide what aspect is the most important: between- or within-variation. A more detailed account is given in Section 2.4.
The concept of fairness relates to the notion that each block should participate in the final solution to a certain degree. Multiblock data analysis methods also differ in this respect: some methods are fair and some are ‘block selectors’ (Smilde et al., 2003; Tenenhaus et al., 2017). To some extent, fairness can be influenced by block-scaling (see Section 2.6), but some methods are invariant to this type of scaling. The method ROSA (see Section 7.5) builds on the principle of fairness: each block is allowed to enter the solution in competition with the other blocks: if it is important then it will be included. The fairness concept has some relation to the concept of invariance discussed in Chapter 7.
1.8 Common and Distinct Components
A crucial concept that plays a role in almost all methods in this book is the idea of common and distinct subspaces and components (Smilde et al., 2017). A schematic illustration of these concepts is shown in Figure 1.9 and a more detailed exposure of these concepts is given in Chapter 2.
Figure 1.9 The idea of common and distinct components. Legend: blue is common variation; dark yellow and dark red are distinct variation and shaded areas are noise (unsystematic variation).
Suppose there are two data blocks X1 and X2 sharing the same samples, i.e., different variables are measured on the same set of samples (see Chapter 3). Then these two blocks can have variation in common (the blue part). This common variation spans a subspace and the common components are then a basis for this subspace.
There is also a part in each block that contains still systematic variation (the dark yellow and dark red parts). These have nothing in common and are, therefore, called distinct parts. These also represent subspaces and the distinct components (two sets; one set for each block) are the bases for these subspaces. What is left in the matrices is unsystematic variation or noise (shaded parts).
The division of each data block in common, distinct, and unsystematic variation should not be read in terms of the individual variables being in common or being distinct but in terms of subspaces. Hence, a part of the variation of a variable in block 1 may be in common with variation of some variables in block 2 whereas the other part of that variable may be distinct, see Elaboration 1.8.
ELABORATION 1.8
Common and distinct in spectroscopy
Suppose that the same set of samples is measured in the UV-Vis regime (block X1) and with near-infrared (NIR, block X2). Also assume that this set of samples contains three chemical components (A,B,C): A absorbs both in UV-Vis and NIR; B only absorbs in the UV-Vis regime and C absorbs only in NIR. Then the common part is the absorption of A in both data blocks; the distinct parts are B in block 1 and C in block 2. However, at a particular wavelength in the NIR region there may be a contribution from both A and C. Hence, this wavelength, i.e., variable, has a common and a distinct part. The same can happen in block 1.
1.9 Overview and Links
In this book we will consider a multitude of methods. To streamline this a bit, we are going to give a summary, at the beginning of each chapter, of the methods and aspects which will be discussed. That will be done in the format of a table. We will specify the following aspects of the methods:
1 A method for unsupervised (U), supervised (S) or complex (C) data structures.
2 The method can deal with heterogeneous data (HET, i.e., different measurement scales) or can only deal with homogeneous data (HOM).
3 A method that uses a sequential (SEQ) or simultaneous (SIM) approach.
4 The method is defined in terms of a model (MOD) or in terms of an algorithm (ALG).
5 A method for finding common (C); common and distinct (CD); or finding common, local and distinct components (CLD).
6 Estimation of the model parameters is based on least squares (LS), maximum likelihood