Data Science: Using Python to Modify Data Distributions
Python allows data scientists to modify data distributions as part of the EDA approach. As a byproduct of data exploration, in an EDA phase you can do the following things:

Obtain new feature creation from the combination of different but related variables

Spot hidden groups or strange values lurking in your data

Try some useful modifications of your data distributions by binning (or other discretizations such as binary variables)
When performing EDA, you need to consider the importance of data transformation in preparation for the learning phase, which also means using certain mathematical formulas. The following information provides an overview of the most common mathematical formulas used for EDA. The data transformation you choose depends on the distribution of your data, with a normal distribution being the most common. In addition, you see the need to match the transformation process to the mathematical formula you use.
Using the normal distribution
The normal, or Gaussian, distribution is the most useful distribution in statistics thanks to its frequent recurrence and particular mathematical properties. It’s essentially the foundation of many statistical tests and models, with some of them, such as the linear regression, widely used in data science.
During data science practice, you’ll meet with a wide range of different distributions — with some of them named by probabilistic theory, others not. For some distributions, the assumption that they should behave as a normal distribution may hold, but for others, it may not, and that could be a problem depending on what algorithms you use for the learning process.
As a general rule, if your model is a linear regression or part of the linear model family because it boils down to a summation of coefficients, then both variable standardization and distribution transformation should be considered.
Creating a Zscore standardization
In your EDA process, you may have realized that your variables have different scales and are heterogeneous in their distributions. As a consequence of your analysis, you need to transform the variables in a way that makes them easily comparable:
from sklearn.preprocessing import scale stand_sepal_width = scale(iris_dataframe[‘sepal width (cm)’])
Transforming other notable distributions
When you check variables with high skewness and kurtosis for their correlation, the results may disappoint you. Using a nonparametric measure of correlation, such as Spearman’s, may tell you more about two variables than Pearson’s r may tell you. In this case, you should transform your insight into a new, transformed feature:
tranformations = {‘x’: lambda x: x, ‘1/x’: lambda x: 1/x, ‘x**2’: lambda x: x**2, ‘x**3’: lambda x: x**3, ‘log(x)’: lambda x: np.log(x)} for transformation in tranformations: pearsonr_coef, pearsonr_p = pearsonr(iris_dataframe[‘sepal length (cm)’], tranformations[transformation](iris_dataframe[‘sepal width (cm)’])) print ‘Transformation: %s t Pearson’s r: %0.3f’ % (transformation, pearsonr_coef) Transformation: x Pearson’s r: 0.109 Transformation: x**2 Pearson’s r: 0.122 Transformation: x**3 Pearson’s r: 0.131 Transformation: log(x) Pearson’s r: 0.093 Transformation: 1/x Pearson’s r: 0.073
In exploring various possible transformations, using a for loop may tell you that a power transformation will increase the correlation between the two variables, thus increasing the performance of a linear machinelearning algorithm. You may also try other, further transformations such as square root np.sqrt(x), exponential np.exp(x), and various combinations of all the transformations, such as log inverse np.log(1/x).