Some basics about correlation analysis based on code, including resampling, standard error, correlation, lsqfit etc.
If you use VS code, install the extension "Better Comments" to see my comments in different colors, they are meaningful and beautiful.
Resampling is a method of estimating the precision of sample statistics (medians, variances, percentiles) by using subsets of available data (bootstrapping) or drawing randomly with replacement from a set of data points (permutation).
Resampling is used to estimate the precision of sample statistics (medians, variances, percentiles) by using subsets of available data (bootstrapping) or drawing randomly with replacement from a set of data points (permutation).
The jackknife is a resampling technique especially useful for variance and bias estimation. The jackknife predates other common resampling methods such as the bootstrap. The jackknife estimator of a parameter is found by systematically leaving out each observation from a dataset and calculating the estimate and then finding the average of these calculations.
The bootstrap is a resampling technique used to estimate statistics on a population by sampling a dataset with replacement. It can be used to estimate summary statistics such as the mean or standard deviation. It is used in applied machine learning to estimate the skill of machine learning models when making predictions on data not included in the training data.
The standard error (SE) of a statistic (usually an estimate of a parameter) is the standard deviation of its sampling distribution or an estimate of that standard deviation. If the parameter or the statistic is the mean, it is called the standard error of the mean (SEM).
In statistics, the method of least squares is a standard approach to the approximate solution of overdetermined systems, i.e., sets of equations in which there are more equations than unknowns. "Least squares" means that the overall solution minimizes the sum of the squares of the errors made in solving every single equation.