To request a blog written on a specific topic, please email James@StatisticsSolutions.com with your suggestion. Thank you!

Thursday, April 9, 2009

Statistical Data Analysis

Statistics is basically a science that involves data collection, data interpretation and finally, data validation. Statistical data analysis is a procedure of performing various statistical operations. Statistical data analysis is a kind of quantitative research, which seeks to quantify the data, and typically, applies some form of statistical analysis. Quantitative data in statistical data analysis basically involves descriptive data, such as survey data and observational data.

Statistical data analysis generally involves some form of statistical tools, which a layman cannot perform without having any statistical knowledge. There are various software packages to perform statistical data analysis. This software includes Statistical Analysis System (SAS), Statistical Package for the Social Sciences (SPSS), Stat soft, etc.

Data in statistical data analysis consists of variable(s). Sometimes the data in statistical data analysis is univariate or multivariate. Depending upon the number of variables in statistical data analysis, the researcher performs different statistical techniques.

If the data in statistical data analysis is multiple in numbers, then several multivariate statistical data analysis can be performed. The multivariate statistical data analyses are factor statistical data analysis, discriminant statistical data analysis, etc. Similarly, if the data in statistical data analysis is singular in number, then the univariate statistical data analysis is performed. This includes t test for significance, z test, f test, ANOVA one way, etc.

The data in statistical data analysis is basically of 2 types, namely, continuous data and discreet data. The continuous data in statistical data analysis is the one that cannot be counted. For example, intensity of a light can be measured but cannot be counted. The discreet data in statistical data analysis is the one that can be counted. For example, the number of bulbs can be counted.

The continuous data in statistical data analysis is distributed under continuous distribution function, which can also be called the probability density function, or simply pdf.

The discreet data in statistical data analysis is distributed under discreet distribution function, which can also be called the probability mass function or simple pmf.

We use the word ‘density’ in continuous data of statistical data analysis because density cannot be counted, but can be measured. We use the word ‘mass’ in discreet data of statistical data analysis because mass cannot be counted.

There are various pdf’s and pmf’s in statistical data analysis. For example, Poisson distribution is the commonly known pmf, and normal distribution is the commonly known pdf in statistical data analysis.

These distributions in statistical data analysis help us to understand which data falls under which distribution. If the data in statistical data analysis is about the intensity of a bulb, then the data would be falling in Poisson distribution.

There is a major task in statistical data analysis, which comprises of statistical inference. The statistical inference in statistical data analysis is mainly comprised of two parts: estimation and tests of hypothesis.

Estimation in statistical data analysis mainly involves parametric data—the data that consists of parameters. On the other hand, tests of hypothesis in statistical data analysis mainly involve non parametric data— the data that consists of no parameters.

For more information on statistical consulting, click here.

Monday, April 6, 2009

Data Analysis

Data analysis is a procedure of collecting and analyzing raw data by interpreting the inference out of raw data. Data analysis is one of the important aspects of the analyst’s work. Data analysis plays a crucial role in deciding whether or not the retrieved data is reliable.

Data analysis is basically a two-step procedure that involves collecting and analyzing data. Data analysis can be explained with the help of the following example:

Suppose a researcher has conducted a survey in order to know if the manufacturing of auto parts in an auto industry is more in Pune or in Chennai. The first step of data analysis is to collect the data through primary or secondary research. The next step of data analysis is to make an inference about the collected data. The second step of data analysis in this case will involve SWOT Analysis. SWOT Analysis stands for Strength, Weakness, Opportunity and Threat of the data under study.

Primary research in data analysis is the one that involves collection of data through questionnaires or telephone interviews. Secondary research in data analysis is the one that involves collection of data using the internet.

There are basically two types of data analysis. These two types are as follows:

Qualitative data analysis:
This kind of data analysis is the one that consists of an unstructured, exploratory research methodology based on small samples intended to provide an insight into the problem being solved.

Quantitative data analysis: On the other hand, this kind of data analysis seeks to quantify the data and typically involves some form of statistical data analysis.

Quantitative data analysis can be performed in those cases when one needs to get statistical inferences about the data. In such cases, data analysis is done by using some statistical techniques. These statistical techniques include Factor Analysis, Discriminate Analysis, etc.

A technical analyst performs data analysis by interpreting the charts using a time series technique, and he/she forecasts the price trends of a particular commodity or share. Thus, data analysis can be used to forecast about the data as well.

Data analysis is an integral part of every research work. The validity of data can be known only through data analysis.

In statistics, data analysis is done on quantitative data. Data analysis in relation to quantitative data analysis can be divided into descriptive statistics, exploratory data analysis and confirmatory data analysis.

Descriptive Statistics in data analysis involves techniques like mean, median, mode, variance, standard deviation, etc.

Exploratory data analysis involves the following steps:

· Formulation of a problem in data analysis.
· Identifying alternative courses of action in data analysis.
· Developing hypotheses in data analysis.
· Isolating key variables and relationships for further examination in data analysis.
· Gaining insights for developing an approach to the formulated problem in data analysis.

Sometimes, qualitative data analysis is undertaken to explain the findings obtained from quantitative data analysis. Thus, one can say that both qualitative data analysis and quantitative data analysis are interrelated with each other.

Data analysis is also synonymous to data modeling. Data modeling is a process in which a perfect model (which represents the data as a whole) is being fitted during the data analysis.

For information on statistical consulting, click here.

Wednesday, April 1, 2009

Statistics Help

Fact and verity go-hand-in-hand in today’s world of constant change and practicality. Statistics has become a necessary facet in daily activity. With statistics as a core subject towards the proper functioning of organizations and firms, help and assistance is made accessible to all those in need. Statistics help has been a compulsory need for businesses as it ensures functionality and efficiency within organizations. Statistical help is used for guiding and assisting clients like students, researchers and members of the business or government communities as it analyzes complicated statistical problems. Backed up by strong statistical facts, organizations with statistical help use these surveys and analyses to create constructive findings and conclusions.

Statistics form a core element in the proper execution of activities and functions of organizations. Statistics form a crucial part in determining business activities and behavior. In the line of business, applications such as risk assessment, data analysis, data mining and decision support can all be carried out through statistics. Statistical help and analysis go a long way at determining business processes. It also helps set a proper course through its research and findings.

Statistics is also applied by students when they write theses, dissertations, reports and term papers. Statistical help and validity may be required for the report or paper. Statistics help also aids organizations at expediting the growth rate and progress. Statistics help should originate from a well-trained work force that is highly skilled at statistics and is experienced and knowledgeable in the field. Such statistics help can be acquired from experts like professors, business consultants, researchers and specialized statistical consultants. The consultants should have good communication skills for interacting with clients, a good scientific and analytical brain, statistical understanding and should be computer proficient. Statisticians should be able to comprehend the needs of the client and fulfill them as per the clients’ requirements. Statistics help is necessarily centered on the needs of the clients, be it analysis, research, survey, etc. Hence, budget should be taken into consideration as quality is emphasized.

When it comes to statistics, the bounds are endless. More than half of the world’s population today depends on statistics help and assistance. Almost every domain of life relies on statistics. The influence of statistics help in the present day is highly credited. While statistics help gains milestones in the field of business, it also achieves goals in the lines of medicine. Statistics help has become such a prevalent feature that the need for firms providing statistics help is on the rise. While some people have skills and qualifications in fields other than that of statistics, to ensure completion of their work or reports, statistics is required. Consequently, such persons fall back and depend on statistics help to guide them at achieving their desired results.

Given that the competition between firms today is on the rise, statistics help assists organizations in achieving more prospects and thus they gain leverage against other contenders. Some organizations, especially the smaller firms who do not have the required skill and ability to perform statistical analyses, rely on statistical help to further their benefits.

Statistics help is a very relevant instrument in today’s world. It warrants efficiency and accuracy and is applicable to almost every sphere of life. The possibilities that statistics help provide goes beyond measure. Putting an end to ball-park figures and estimates, statistics help has taken the world by storm with its precision and exactness and it remains an indispensable part of everyday activities.

For help with your statistical analysis, click here.

Exploratory Factor Analysis

Factor Analysis is a general name denoting a class of procedures primarily used for data reduction and summarization. In research, there are a large number of variables which are extensively correlated and must be reduced to a manageable level. Relationships among sets of many interrelated variables are examined and represented in terms of a few underlying factors.
There are basically 2 approaches to Factor Analysis:

· Exploratory Factor Analysis (EFA) seeks to uncover the underlying structure of a relatively large set of variables. The researcher has a priori assumption that any indicator may be associated with any factor. This is the most common form of factor analysis. There is no prior theory and one uses factor loadings to intuit the factor structure of the data.

· Confirmatory Factor Analysis (CFA) seeks to determine if the number of factors and the loadings of measured (indicator) variables on them conform to what is expected on the basis of pre-established theory. Indicator variables are selected on the basis of prior theory, and factor analysis is used to see if they loaded, as predicted, on the expected number of factors.

The basic difference between Exploratory Factor Analysis and CFA is that in CFA, a researcher’s a priori assumption is that each factor (the number and labels of which may be specified a priori) is associated with a specified subset of indicator variables. The major limitation behind Exploratory Factor Analysis is its simplicity. Hence, the researcher will not get a reliable inference. Therefore, Exploratory Factor Analysis is used less as compared to Confirmatory Factor Analysis.

The following techniques are used in both the approaches—both Exploratory Factor Analysis and CFA:

· Principal Component Technique: This technique is used in Exploratory Factor Analysis, where the total variance in the data is considered. The diagonal of the correlation matrix consists of unities, and full variance is brought into the factor matrix. Principal technique is recommended when the primary concern is to determine the minimum number of factors that will account for maximum variance in the data for use in subsequent multivariate analysis.

There are some techniques, in addition to Principal Component Technique, that are used in Exploratory factor analysis and Confirmatory factor analysis and that are complex. These techniques are also called Extraction Methods. These techniques are as follows:

· Image factoring: This technique in Exploratory Factor Analysis is based on the correlation matrix of predicted or dependent variables rather than actual variables. In this, we predict each variable from the others by using multiple regressions.

· Maximum likelihood factoring(MLF): This technique in Exploratory Factor Analysis is based on a linear combination of variables to form factors, where the parameter estimates are such that they are most likely to have resulted in the observed correlation matrix, by using Maximum Likelihood Estimation (MLE) methods and assuming multivariate normality. Correlations are weighted by each variable's uniqueness. Here, uniqueness refers to the difference between the variability of a variable and its communality. MLF generates a chi-square goodness-of-fit test. The researcher can increase the number of factors one at a time until a satisfactory goodness-of-fit is obtained.

· Alpha factoring: This technique in Exploratory Factor Analysis is based on the maximization of the reliability of factors, assuming that the variables are randomly sampled from a very large set of variables. Unlike other methods, this method does not assume sampled cases and fixed variables.

· Unweighted least squares (ULS) factoring: This technique in Exploratory Factor Analysis is based upon minimizing the sum of squared differences between the observed and estimated correlation matrices, without counting the diagonal.

· Generalized least squares (GLS) factoring: This technique in Exploratory Factor Analysis is based on adjusting ULS by measuring the correlations, which are inversely proportional to their uniqueness (more unique variables weight less). Like MLF, GLS also generates a chi-square goodness-of-fit test. The researcher can increase the number of factors one at a time until a satisfactory goodness-of-fit is obtained.

The major disadvantage of using these techniques in Exploratory Factor Analysis is that they are quiet complex and are not recommended for an inexperienced user. Hence, these methods are usually not used in extraction methods. For help with these techniques, click here.