首页 | 官方网站   微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 46 毫秒
1.
To interpret the biplot, it is necessary to know which points—usually variables—are the ones that are important contributors to the solution, especially when there are many variables involved. This information can be calculated separately as part of the biplot's numerical results, but this means that a table has to be consulted along with the graphical display. We propose a new scaling of the display, called the contribution biplot, which incorporates this diagnostic information directly into the display itself, showing visually the important contributors and thus facilitating the biplot interpretation and often simplifying the graphical representation considerably. The contribution biplot can be applied to a wide variety of analyses, such as correspondence analysis, principal component analysis, log-ratio analysis, and various forms of discriminant analysis, and, in fact, to any method based on dimension reduction through the singular value decomposition. In the contribution biplot, one set of points, usually the rows of a data matrix, optimally represents the spatial positions of the cases or sample units, according to an appropriate distance measure. The other set of points, usually the columns of the data matrix, is represented by vectors that are related to their contributions to the low-dimensional solution. A fringe benefit is that often only one common scale for the row and column points is needed on the principal axes, thus avoiding the problem of enlarging or contracting the scale of one set of points to make the biplot legible. Furthermore, the contribution biplot also solves the problem in correspondence analysis and log-ratio analysis of low-frequency categories that are located on the periphery of the map, giving the false impression that they are important, when they are in fact contributing minimally to the solution. This article has supplementary materials online.  相似文献   

2.
The theory of sparse stochastic processes offers a broad class of statistical models to study signals, far beyond the more classical class of Gaussian processes. In this framework, signals are represented as realizations of random processes that are solution of linear stochastic differential equations driven by Lévy white noises. Among these processes, generalized Poisson processes based on compound-Poisson noises admit an interpretation as random L-splines with random knots and weights. We demonstrate that every generalized Lévy process—from Gaussian to sparse—can be understood as the limit in law of a sequence of generalized Poisson processes. This enables a new conceptual understanding of sparse processes and suggests simple algorithms for the numerical generation of such objects.  相似文献   

3.
Classical biplot methods allow for the simultaneous representation of individuals (rows) and variables (columns) of a data matrix. For binary data, logistic biplots have been recently developed. When data are nominal, both classical and binary logistic biplots are not adequate and techniques such as multiple correspondence analysis (MCA), latent trait analysis (LTA) or item response theory (IRT) for nominal items should be used instead. In this paper we extend the binary logistic biplot to nominal data. The resulting method is termed “nominal logistic biplot”(NLB), although the variables are represented as convex prediction regions rather than vectors. Using the methods from computational geometry, the set of prediction regions is converted to a set of points in such a way that the prediction for each individual is established by its closest “category point”. Then interpretation is based on distances rather than on projections. We study the geometry of such a representation and construct computational algorithms for the estimation of parameters and the calculation of prediction regions. Nominal logistic biplots extend both MCA and LTA in the sense that they give a graphical representation for LTA similar to the one obtained in MCA.  相似文献   

4.
Abstract. In this paper we extend Martindale's result by showing that the symmetric elements or the skew elements of a prime ring with involution of characteristic not 2 do not satisfy certain linear generalized polynomial identities. As applications, we determine the centralizing additive maps and the commuting traces of biadditive maps on the symmetric elements of a prime ring.  相似文献   

5.
A number of classical approaches to nonparametric regression have recently been extended to the case of functional predictors. This article introduces a new method of this type, which extends intermediate-rank penalized smoothing to scalar-on-function regression. In the proposed method, which we call principal coordinate ridge regression, one regresses the response on leading principal coordinates defined by a relevant distance among the functional predictors, while applying a ridge penalty. Our publicly available implementation, based on generalized additive modeling software, allows for fast optimal tuning parameter selection and for extensions to multiple functional predictors, exponential family-valued responses, and mixed-effects models. In an application to signature verification data, principal coordinate ridge regression, with dynamic time warping distance used to define the principal coordinates, is shown to outperform a functional generalized linear model. Supplementary materials for this article are available online.  相似文献   

6.
Social networks have become an important part of agent-based models, and their structure may have remarkable impact on simulation results. We propose a simple and efficient but empirically based approach for spatial agent-based models which explicitly takes into account restrictions and opportunities imposed by effects of baseline homophily, i.e. the influence of local socio-demography on the composition of one’s social network. Furthermore, the algorithm considers the probability of links that depends on geographical distance between potential partners. The resulting network reflects social settings and furthermore allows the modeller to influence network properties by adjusting agent type specific parameters. Especially the parameter for distance dependence and the probability of distant links allow for control of clustering and agent type distribution of personal networks.  相似文献   

7.
Large-scale generalized linear array models (GLAMs) can be challenging to fit. Computation and storage of its tensor product design matrix can be impossible due to time and memory constraints, and previously considered design matrix free algorithms do not scale well with the dimension of the parameter vector. A new design matrix free algorithm is proposed for computing the penalized maximum likelihood estimate for GLAMs, which, in particular, handles nondifferentiable penalty functions. The proposed algorithm is implemented and available via the R package glamlasso. It combines several ideas—previously considered separately—to obtain sparse estimates while at the same time efficiently exploiting the GLAM structure. In this article, the convergence of the algorithm is treated and the performance of its implementation is investigated and compared to that of glmnet on simulated as well as real data. It is shown that the computation time for glamlasso scales favorably with the size of the problem when compared to glmnet. Supplementary materials, in the form of R code, data and visualizations of results, are available online.  相似文献   

8.
This paper aims at analysing the existence of a formal correspondence between spatial interaction models emanating from entropy theory and micro-economic discrete choice theory (in particular, multinomial logit models.). After a concise review of the literature on this issue, the emphasis is placed on an interpretation of formal analogies between both classes of models in a dynamic context. A simple dynamic spatial interaction model—based on optimal control theory—is proposed, and it is shown that the results confirm also the existence of a formal analogy between (macro) dynamic interaction models and (micro) choice models. Similar results are also derived for Alonso's general theory of movement in a spatial system.  相似文献   

9.
Exponential family random graph models (ERGMs) can be understood in terms of a set of structural biases that act on an underlying reference distribution. This distribution determines many aspects of the behavior and interpretation of the ERGM families incorporating it. One important innovation in this area has been the development of an ERGM reference model that produces realistic behavior when generalized to sparse networks of varying sizes. Here, we show that this model can be derived from a latent dynamic process in which tie formation takes place within small local settings between which individuals move. This derivation provides one possible micro-process interpretation of the sparse ERGM reference model and sheds light on the conditions under which constant mean degree scaling can emerge.  相似文献   

10.
In a recent paper, the authors have proved results characterizing convexity-preserving maps defined on a subset of a not-necessarily finite dimensional real vector space as projective maps. The purpose of this note is three-fold. First, we state a theorem characterizing continuous, injective, convexity-preserving maps from a relatively open, connected subset of an affine subspace of ℝ m into ℝ n as projective maps. This result follows from the more general results stated and proved in a coordinate-free manner in the above paper, and is intended to be more accessible to researchers interested in optimization algorithms. Second, based on that characterization theorem, we offer a characterization theorem for collinear scalings first introduced by Davidon in 1977 for deriving certain algorithms for nonlinear optimization, and a characterization theorem for projective transformations used by Karmarkar in 1984 in his linear programming algorithm. These latter two theorems indicate that Davidon’s collinear scalings and Karmarkar’s projective transformations are the only continuous, injective, convexity-preserving maps possessing certain features that Davidon and Karmarkar respectively desired in the derivation of their algorithms. The proofs of these latter two theorems utilize our characterization of continuous, injective, convexity-preserving maps in a way that has implications to the choice of scalings and transformations in the derivation of optimization algorithms in general. The third purpose of this note is to point this out. Received: January 2000 / Accepted: November 2000?Published online January 17, 2001  相似文献   

11.
When using a model-based approach to geostatistical problems, often, due to the complexity of the models, inference relies on Markov chain Monte Carlo methods. This article focuses on the generalized linear spatial models, and demonstrates that parameter estimation and model selection using Markov chain Monte Carlo maximum likelihood is a feasible and very useful technique. A dataset of radionuclide concentrations on Rongelap Island is used to illustrate the techniques. For this dataset we demonstrate that the log-link function is not a good choice, and that there exists additional nonspatial variation which cannot be attributed to the Poisson error distribution. We also show that the interpretation of this additional variation as either micro-scale variation or measurement error has a significant impact on predictions. The techniques presented in this article would also be useful for other types of geostatistical models.  相似文献   

12.
Our paper presents an empirical analysis of the association between firm attributes in electronic retailing and the adoption of information initiatives in mobile retailing. In our attempt to analyze the collected data, we find that the count of information initiatives exhibits underdispersion. Also, zero‐truncation arises from our study design. To tackle the two issues, we test four zero‐truncated (ZT) count data models—binomial, Poisson, Conway–Maxwell–Poisson, and Consul's generalized Poisson. We observe that the ZT Poisson model has a much inferior fit when compared with the other three models. Interestingly, even though the ZT binomial distribution is the only model that explicitly takes into account the finite range of our count variable, it is still outperformed by the other two Poisson mixtures that turn out to be good approximations. Further, despite the rising popularity of the Conway–Maxwell–Poisson distribution in recent literature, the ZT Consul's generalized Poisson distribution shows the best fit among all candidate models and suggests support for one hypothesis. Because underdispersion is rarely addressed in IT and electronic commerce research, our study aims to encourage empirical researchers to adopt a flexible regression model in order to make a robust assessment on the impact of explanatory variables. Copyright © 2014 John Wiley & Sons, Ltd.  相似文献   

13.
A biplot, which is the multivariate generalization of the two-variable scatterplot, can be used to visualize the results of many multivariate techniques, especially those that are based on the singular value decomposition. We consider data sets consisting of continuous-scale measurements, their fuzzy coding and the biplots that visualize them, using a fuzzy version of multiple correspondence analysis. Of special interest is the way quality of fit of the biplot is measured, since it is well known that regular (i.e., crisp) multiple correspondence analysis seriously under-estimates this measure. We show how the results of fuzzy multiple correspondence analysis can be defuzzified to obtain estimated values of the original data, and prove that this implies an orthogonal decomposition of variance. This permits a measure-of-fit to be calculated in the familiar form of a percentage of explained variance, which is directly comparable to the corresponding fit measure used in principal component analysis of the original data. The approach is motivated initially by its application to a simulated data set, showing how the fuzzy approach can lead to diagnosing nonlinear relationships, and finally it is applied to a real set of meteorological data.  相似文献   

14.
We propose a geometric method to parameterize inequivalent vacua by dynamical data. Introducing quantum Clifford algebras with arbitrary bilinear forms we distinguish isomorphic algebras—as Clifford algebras—by different filtrations (resp. induced gradings). The idea of a vacuum is introduced as the unique algebraic projection on the base field embedded in the Clifford algebra, which is however equivalent to the term vacuum in axiomatic quantum field theory and the GNS construction in C*‐algebras. This approach is shown to be equivalent to the usual picture which fixes one product but employs a variety of GNS states. The most striking novelty of the geometric approach is the fact that dynamical data fix uniquely the vacuum and that positivity is not required. The usual concept of a statistical quantum state can be generalized to geometric meaningful but non‐statistical, non‐definite, situations. Furthermore, an algebraization of states takes place. An application to physics is provided by an U (2)‐symmetry producing a gap equation which governs a phase transition. The parameterization of all vacua is explicitly calculated from propagator matrix elements. A discussion of the relation to BCS theory and Bogoliubov–Valatin transformations is given. Copyright © 2001 John Wiley & Sons, Ltd.  相似文献   

15.
Penalized estimation has become an established tool for regularization and model selection in regression models. A variety of penalties with specific features are available and effective algorithms for specific penalties have been proposed. But not much is available to fit models with a combination of different penalties. When modeling the rent data of Munich as in our application, various types of predictors call for a combination of a Ridge, a group Lasso and a Lasso-type penalty within one model. We propose to approximate penalties that are (semi-)norms of scalar linear transformations of the coefficient vector in generalized structured models—such that penalties of various kinds can be combined in one model. The approach is very general such that the Lasso, the fused Lasso, the Ridge, the smoothly clipped absolute deviation penalty, the elastic net and many more penalties are embedded. The computation is based on conventional penalized iteratively re-weighted least squares algorithms and hence, easy to implement. New penalties can be incorporated quickly. The approach is extended to penalties with vector based arguments. There are several possibilities to choose the penalty parameter(s). A software implementation is available. Some illustrative examples show promising results.  相似文献   

16.
以往关于广义博弈Nash平衡的稳定性的研究,均利用可行策略映射之间的一致度量.现考虑在更弱的度量下,利用可行策略映射图像之间的Hausdorff距离定义度量.在此弱图像拓扑下,证明了广义博弈空间的完备性,以及Nash平衡映射的上半连续性和紧性,进而得到广义博弈Nash平衡的通有稳定性.即在Baire分类的意义下,大多数的广义博弈都是本质的.  相似文献   

17.
Reduced-rank restrictions can add useful parsimony to coefficient matrices of multivariate models, but their use is limited by the daunting complexity of the methods and their theory. The present work takes the easy road, focusing on unifying themes and simplified methods. For Gaussian and non-Gaussian (GLM, GAM, mixed normal, etc.) multivariate models, the present work gives a unified, explicit theory for the general asymptotic (normal) distribution of maximum likelihood estimators (MLE). MLE can be complex and computationally hard, but we show a strong asymptotic equivalence between MLE and a relatively simple minimum (Mahalanobis) distance estimator. The latter method yields particularly simple tests of rank, and we describe its asymptotic behavior in detail. We also examine the method's performance in simulation and via analytical and empirical examples.  相似文献   

18.
19.
Correspondence analysis, a data analytic technique used to study two‐way cross‐classifications, is applied to social relational data. Such data are frequently termed “sociometric” or “network” data. The method allows one to model forms of relational data and types of empirical relationships not easily analyzed using either standard social network methods or common scaling or clustering techniques. In particular, correspondence analysis allows one to model:

—two‐mode networks (rows and columns of a sociomatrix refer to different objects)

—valued relations (e.g. counts, ratings, or frequencies).

In general, the technique provides scale values for row and column units, visual presentation of relationships among rows and columns, and criteria for assessing “dimensionality” or graphical complexity of the data and goodness‐of‐fit to particular models. Correspondence analysis has recently been the subject of research by Goodman, Haberman, and Gilula, who have termed their approach to the problem “canonical analysis” to reflect its similarity to canonical correlation analysis of continuous multivariate data. This generalization links the technique to more standard categorical data analysis models, and provides a much‐needed statistical justificatioa

We review both correspondence and canonical analysis, and present these ideas by analyzing relational data on the 1980 monetary donations from corporations to nonprofit organizations in the Minneapolis St. Paul metropolitan area. We also show how these techniques are related to dyadic independence models, first introduced by Holland, Leinhardt, Fienberg, and Wasserman in the early 1980's. The highlight of this paper is the relationship between correspondence and canonical analysis, and these dyadic independence models, which are designed specifically for relational data. The paper concludes with a discussion of this relationship, and some data analyses that illustrate the fart that correspondence analysis models can be used as approximate dyadic independence models.  相似文献   

20.
Profit,Directional Distance Functions,and Nerlovian Efficiency   总被引:26,自引:0,他引:26  
The directional technology distance function is introduced, given an interpretation as a min-max, and compared with other functional representations of the technology including the Shephard input and output distance functions and the McFadden gauge function. A dual correspondence is developed between the directional technology distance function and the profit function, and it is shown that all previous dual correspondences are special cases of this correspondence. We then show how Nerlovian (profit-based) efficiency measures can be computed using the directional technology distance function.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号