1. Trang chủ
  2. » Tất cả

Statistics, data mining, and machine learning in astronomy

2 3 0

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 2
Dung lượng 149,31 KB

Nội dung

Statistics, Data Mining, and Machine Learning in Astronomy 340 • Chapter 8 Regression and Model Fitting A common form for K (x) is the tricubic kernel, K (xi , x) = ( 1− ( |x − xi | h )3 )3 (8 51) for[.]

340 • Chapter Regression and Model Fitting A common form for K (x) is the tricubic kernel,  K (xi , x) = 1−  |x − xi | h 3 3 (8.51) for |xi − x| < h, which is often referred to as lowess (or loess; locally weighted scatter plot smoothing); see [3] There are further extensions possible for local linear regression: Local polynomial regression We can consider any polynomial order However there is a bias–variance (complexity) trade-off, as usual The general consensus is that going past linear increases variance without decreasing bias much, since local linear regression captures most of the boundary bias • Variable-bandwidth kernels Let the bandwidth for each training point be inversely proportional to its kth nearest neighbor’s distance Generally a good idea in practice, though there is less theoretical consensus on how to choose the parameters in this framework • None of these modifications improves the convergence rate 8.7 Nonlinear Regression Forcing data to correspond to a linear model through the use of coordinate transformations is a well-used trick in astronomy (e.g., the extensive use of logarithms in the astronomical literature to linearize complex relations between attributes; fitting y = A exp(B x) becomes a linear problem with z = K + B x, where z = log y and K = log A) These simplifications, while often effective, introduce other complications (including the non-Gaussian nature of the uncertainties for low signal-to-noise data) We must, eventually, consider the case of nonlinear regression and model fitting In the cosmological examples described previously we have fit a series of parametric and nonparametric models to the supernova data Given that we know the theoretical form of the underlying cosmological model, these models are somewhat ad hoc (e.g., using a series of polynomials to parameterize the dependence of distance modulus on cosmological redshift) In the following we consider directly fitting the cosmological model described in eq 8.4 Solving for m and  is a nonlinear optimization problem requiring that we maximize the posterior, p(m ,  |z, I ) ∝ n  i =1 √ 2π σi  exp −(µi − µ(zi |m ,  ))2 2σi2  p(m ,  ) (8.52) with µi the distance modulus for the supernova and zi the redshift In § 5.8 we introduced Markov chain Monte Carlo as a sampling technique that can be used for searching through parameter space Figure 8.5 shows the resulting likelihood contours for our cosmological model after applying the Metropolis– Hastings algorithm to generate the MCMC chains and integrating the chains over the parameter space 8.7 Nonlinear Regression 46 • 341 1.1 100 observations 1.0 44 0.9 42 µ ΩΛ 0.8 0.7 40 0.6 38 0.5 36 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 z 0.4 0.1 0.2 0.3 0.4 0.5 ΩM 0.6 0.7 Figure 8.5 Cosmology fit to the standard cosmological integral Errors in µ are a factor of ten smaller than for the sample used in figure 8.2 Contours are 1σ , 2σ , and 3σ for the posterior (uniform prior in  M and  ) The dashed line shows flat cosmology The dotted lines show the input values An alternate approach is to use the Levenberg–Marquardt (LM) algorithm [15, 16] to optimize the maximum likelihood estimation LM searches for the sum-of-squares minima of a multivariate distribution through a combination of gradient descent and Gauss–Newton optimization Assuming that we can express our regression function as a Taylor series expansion then, to first order, we can write f (xi |θ) = f (xi |θ ) + J dθ , (8.53) where θ is an initial guess for the regression parameters, J is the Jacobian about this point (J = ∂ f (xi |θ)/∂θ ), and dθ is a small change in the regression parameters LM minimizes the sum of square errors,  (yi − f (xi |θ ) − J i dθ )2 (8.54) i for the perturbation dθ This results in an update relation for dθ of (J T C −1 J + λ diag(J T C −1 J )) dθ = J T C −1 (Y − f (X|θ )), (8.55) with C the standard covariance matrix introduced in eq 8.18 In this expression the λ term acts as a damping parameter (in a manner similar to ridge regression regularization discussed in § 8.3.1) For small λ the relation approximates a Gauss–Newton method (i.e., it minimizes the parameters assuming the function is quadratic) For large λ the perturbation dθ follows the direction of steepest descent The diag(J T C −1 J ) term, as opposed to the identity matrix used in ridge regression, ensures that the update of dθ is largest along directions where the gradient is smallest (which improves convergence) LM is an iterative process At each iteration LM searches for the step dθ that minimizes eq 8.54 and then updates the regression model The iterations cease when ... the standard cosmological integral Errors in µ are a factor of ten smaller than for the sample used in figure 8.2 Contours are 1σ , 2σ , and 3σ for the posterior (uniform prior in  M and ... (8.53) where θ is an initial guess for the regression parameters, J is the Jacobian about this point (J = ∂ f (xi |θ)/∂θ ), and dθ is a small change in the regression parameters LM minimizes the sum... This results in an update relation for dθ of (J T C −1 J + λ diag(J T C −1 J )) dθ = J T C −1 (Y − f (X|θ )), (8.55) with C the standard covariance matrix introduced in eq 8.18 In this expression

Ngày đăng: 20/11/2022, 11:17