The diagnostic of outliers is very essential since of their responsibility for producing large interpretative problems in linear regression analysis and nonlinear regression analysis. There has been a lot of work accomplished in identifying outliers in linear but not in nonlinear regression. In practice, it is often the case that the assumption of linear regression is violated, such as when highly influential outliers exist in the dataset, which will adversely impact the validity of the statistical analysis. Finding outliers is important because they are responsible for invalid inferences and inaccurate predictions as they have a larger impact on the computed values of various estimations. The outliers must be divided into vertical outliers (VO), good leverage points (GLP), and bad leverage points (BLP) since only the vertical outliers and bad leverage have an undue effect on parameter estimations. We compare several outlier detection techniques using a robust diagnostic plot to correctly classify good and bad leverage points and vertical outliers, by decreasing both masking and swamping effects for both the untransformed variables and transformed variables. The main idea is to detect of outliers before transformation (original data) and after transformation. The results of generation study and numerical indicate that modified generalized DIFFITS (different of fit) against the Diagnostic Robust Generalized Potential (MGDFF-DRGP) successfully detect outliers in the data
2023-04
General Letters in Mathematics
(Issue : 2)
(Volume : 12)
Estimating Regression Coefficients using Robust Bootstrap with application to Covid-19 Data
The linear regression model is often used by researchers and data analysts for predictive, descriptive, and inferential purposes. When working with empirical data, this model is based on a set of assumptions that are not always satisfied. In this situation, using more complicated regression algorithms that do not strictly rely on the same assumptions might be one answer. Nevertheless, transformations provide a simpler technique for improving the validity of model assumptions and allow the user to continue using the well-known model of linear regression. The main objective of this project is to provide a transformation for the linear model’s response and predictor variables, as well as parameter estimation methods before the transformation and after the transformation. The bootstrap approach has been effectively used for many statistical estimates and inference issues, according to the paper.
2022-08
2022 International Conference on Computer Science and Software Engineering (CSASE)
(Issue : 6)
A Nonlinear Transformation Methods Using Covid-19 Data in the Kurdistan Region
Ordinary Least squares (OLS) are the most widely
used due to tradition and their optimal properties to estimate
the parameters of linear and nonlinear regression models.
Nevertheless, in the presence of outliers in the data, estimates
of OLS become inefficient, and even a single unusual point can
have a significant impact on the estimation of parameters. In the
presence of outliers is the use of robust estimators rather than
the method of OLS. They are finding a suitable nonlinear
transformation to reduce anomalies, including non-additivity,
heteroscedasticity, and non-normality in multiple nonlinear
regression. It might be beneficial to transform the response
variable or predictor variable, or both together to present the
equation in a simple, functional form that is linear in the
transformed variables. To illustrate the superior transformation
function, we compare the squared correlation coefficient
(coefficient of determination), Breusch-Pagan test, and Shapiro
Wilk test between the transformation functions.
2022-04
Thesis
2023
Using Nonlinear Transformations with Robust and Bootstrap Regression Analysis
Transformations of the dependent or independent variables or both together can improve the fit and correct violations of model assumptions: constant error variance or normality or linear relation between dependent and independent variables.
Furthermore, Ordinary Least Squares (OLS) is the most widely used approach to estimate the parameters of linear regression models. However, in the presence of outliers, robust estimators are used rather than the OLS method. In this study, we used different nonlinear transformation functions with OLS and robust regression models. To illustrate the superior transformation function, we compared the coefficient of determination, Breusch-Pagan test and Shapiro-Wilk test between the transformations function before and after the transformation. The bootstrap approach was also used, which has been effectively used for many statistical inference problems.
In this thesis, we used an R package called “trafo”, which makes it simple for the user to decide which transformation function is suitable for fulfilling the assumptions. In practice, it is often the case that the assumption of linear regression is violated, such as when highly influential outliers exist in the dataset, which will adversely impact the validity of the statistical analysis. Finding outliers is important because they are responsible for invalid inferences and inaccurate predictions as they have a greater influence on the calculated values of different estimations. The outliers are divided into Vertical Outliers (VOs), Good Leverage points (GLPs), and Bad Leverage Points (BLPs) but only the VOs and BLPs have an undue effect on parameter estimations. We compared several outlier detection techniques using a robust diagnostic plot to classify between VOs and BLPs, by decreasing both swamping and masking effects for both the untransformed and transformed variables.
The results indicated that finding the transformation function of independent and dependent variable will be suitable and beneficial in obtaining a more correct regression model in data. When data contains outliers, the Robust MM (Modified Maximum Likelihood) and the proposed bootstrap robust MM-estimator (Boot-MM) is thus recommended as the best estimate for fitting regression. The thesis indicated that modified generalized DIFFITS (different of fit) against the Diagnostic Robust Generalized Potential (MGDFF-DRGP) successfully detects outliers in the data. All the results and figures were obtained by using the R program.
2025
Conference
2022 International Conference on Computer Science and Software Engineering (CSASE)
2022-03
A Nonlinear Transformation Methods Using Covid-19 Data in the Kurdistan Region
Ordinary Least squares (OLS) are the most widely used due to tradition and their optimal properties to estimate the parameters of linear and nonlinear regression models. Nevertheless, in the presence of outliers in the data, estimates of OLS become inefficient, and even a single unusual point can have a significant impact on the estimation of parameters. In the presence of outliers is the use of robust estimators rather than the method of OLS. They are finding a suitable nonlinear transformation to reduce anomalies, including non-additivity, heteroscedasticity, and non-normality in multiple nonlinear regression. It might be beneficial to transform the response variable or predictor variable, or both together to present the equation in a simple, functional form that is linear in the transformed variables. To illustrate the superior transformation function, we compare the squared correlation coefficient (coefficient of determination), Breusch-Pagan test, and Shapiro-Wilk test between the transformation functions.
Training Course
2023-07-06,2023-12-21
Course on Pedagogical Training for Teacher
Pedagogical Training and Academic Development Center-University of Zakho, Iraq
2023
2023-04-08,2023-05-23
English Language Proficiency Course with Pre-Intermediate level
Training Course on English
2023
2023-01-17,2023-03-02
English Language Proficiency Course with Elementary level