Yi=β0+β1xi+β2zi+ϵi
i=1,…,nx∗i=(xi−x¯)/sd(x)z∗i=(zi−z¯)/sd(z)
and instead fit the model (using ordinary least squares)
Yi=β∗0+β∗1x∗i+β∗2z∗i+ϵi
Then the fitted parameters (betas) will change, but they change in a way that you can calculate by simple algebra from the transformation applied. So if we call the estimated betas from the model using transformed predictors for β∗1,2 and the denote the betas from the untransformed model with β^1,2, we can calculate one set of betas from the other one, knowing the means and standard deviations of the predictors. The realtionship between the transformed and untransformed parameters is the same as between their estimates, when based on OLS. Some algebra will give the relationship as
β0=β∗0−β∗1x¯sd(x)−β∗2z¯sd(z),β1=β∗1sd(x),β2=β∗2sd(z)
So standardization is not a necessary part of modelling. (It might still be done for other reasons, which we do not cover here). This answer depends also upon us using ordinary least squares. For some other fitting methods, such as ridge or lasso, standardization is important, because we loose this invariance we have with least squares. This is easy to see: both lasso and ridge do regularization based on the size of the betas, so any transformation which change the relative sizes of the betas will change the result!
And this discussion for the case of linear regression tells you what you should look after in other cases: Is there invariance, or is it not? Generally, methods which depends on distance measures among the predictors will not show invariance, so standardization is important. Another example will be clustering.