Show simple item record

dc.contributor.advisorKarasözen, Bülent
dc.contributor.advisorWeber, Gerhard Wilhelm
dc.contributor.authorKayhan, Belgin
dc.date.accessioned2020-12-10T09:06:59Z
dc.date.available2020-12-10T09:06:59Z
dc.date.submitted2010
dc.date.issued2018-08-06
dc.identifier.urihttps://acikbilim.yok.gov.tr/handle/20.500.12812/223885
dc.description.abstract
dc.description.abstractRegression analysis refers to techniques for modeling and analyzing several variablesin statistical learning. There are various types of regression models. In our study,we analyzed Generalized Partial Linear Models (GPLMs), which decomposes inputvariables into two sets, and additively combines classical linear models with nonlinearmodel part. By separating linear models from nonlinear ones, an inverse problemmethod Tikhonov regularization was applied for the nonlinear submodels separately,within the entire GPLM. Such a particular representation of submodels provides botha better accuracy and a better stability (regularity) under noise in the data.We aim to smooth the nonparametric part of GPLM by using a modied form of Mul-tiple Adaptive Regression Spline (MARS) which is very useful for high-dimensionalproblems and does not impose any specic relationship between the predictor anddependent variables. Instead, it can estimate the contribution of the basis functionsso that both the additive and interaction eects of the predictors are allowed to de-termine the dependent variable. The MARS algorithm has two steps: the forward and backward stepwise algorithms. In the rst one, the model is built by adding basisfunctions until a maximum level of complexity is reached. On the other hand, thebackward stepwise algorithm starts with removing the least signicant basis functionsfrom the model.In this study, we propose to use a penalized residual sum of squares (PRSS) insteadof the backward stepwise algorithm and construct PRSS for MARS as a Tikhonovregularization problem. Besides, we provide numeric example with two data sets; onehas interaction and the other one does not have. As well as studying the regular-ization of the nonparametric part, we also mention theoretically the regularizationof the parametric part. Furthermore, we make a comparison between Innite KernelLearning (IKL) and Tikhonov regularization by using two data sets, with the dier-ence consisting in the (non-)homogeneity of the data set. The thesis concludes withan outlook on future research.en_US
dc.languageEnglish
dc.language.isoen
dc.rightsinfo:eu-repo/semantics/openAccess
dc.rightsAttribution 4.0 United Statestr_TR
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectMatematiktr_TR
dc.subjectMathematicsen_US
dc.subjectİstatistiktr_TR
dc.subjectStatisticsen_US
dc.titleParameter estimation in generalized partial linear models with tikhanov regularization
dc.title.alternativeGenelleştirilmiş parçalı doğrusal modellerde tikhanov düzenleme ile parametre tahmini
dc.typemasterThesis
dc.date.updated2018-08-06
dc.contributor.departmentBilimsel Hesaplama Anabilim Dalı
dc.subject.ytmParameter estimation
dc.subject.ytmOptimization problem
dc.subject.ytmGeneralized linear models
dc.subject.ytmOptimization
dc.identifier.yokid384234
dc.publisher.instituteUygulamalı Matematik Enstitüsü
dc.publisher.universityORTA DOĞU TEKNİK ÜNİVERSİTESİ
dc.identifier.thesisid275854
dc.description.pages115
dc.publisher.disciplineDiğer


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

info:eu-repo/semantics/openAccess
Except where otherwise noted, this item's license is described as info:eu-repo/semantics/openAccess