You seem to be mixing realtionships, error models, and linearity in the relationship and in the predictors...
OLS = ordinary least squares. This is a criterion for the stochastic part of the model (error mpdel) to identify a "good fit": the values for the coefficients are chosen in a way that will minimize the "sum of squared errors". This relates only to the properties of the stochastic part, not to the functional part of the model. This means that the relationship between your variables can have any form (linear, quadratic, exponential, sinuosid, logarithmic, logistic, etc. etc.) - that does not matter. What matters is that the probability distribution for the response, conditional on the values of the predictor[s], is approximately normal: y ~ N( E(Y|X), s² ). Then the OLS solution happens to give those values for the coefficients for which the observed data will be most likely (given your model) (this is the "maximum likelihood solution" of which the OLS is just a sepcial case).
The solution ("fit") can be found by some relatively simple matrix algebra - if the coefficients are not transformed or inside some non-linear function. Models with coefficients only in their "linear" form are called linear models. Linear models can model non-linear relationships between variables. Only the coefficients must not be inside some non-linear term. So
y = bx + a
is linear, because the coefficients a and b are both "linear". This model also models a straight-line (linear) relationship between x and y. The models
y = cx² + bx + a
y = b*exp(x) + axy = c*sin(x) + bx³ +a
are also linear models, but they do not model linear relationships. Finally, the models
y = sqrt(bx) + exp(x/a)
y = b*sin(ax)
are examples for non-linear models.
The solution of such non-linear models may also be the OLS only that this solution cant be found by matrix algebra and usually has to be searched (using some more or less smart algorithm).
When the conditional probability distribution of the response is not approximately normal, the OLS solution does not give the coefficient values for which the observed data would be most likely. Here the likelihood function of the data has to be determined more or less explicitly to find its maximum (= the combination of values for the coefficients for which the data is most likely). The "shortcut" via the minmization of the "sum of squared errors" is not possible anymore. Again, this is not related to the functional relationship between the variables. As an example think of the number of radioactive decay events over time: the expected number increases linearily with time, but the numbers show a Poisson distribution (which is not normal). So the functional part of the model is linear and the stochastic part is Poisson.
The maximum likelhood solution will converge against the OLS solution when n -> Inf. Therefore, if your sample size is large, OLS may give you a useful approximative solution even for cases where the conditional probability distribution of the response is not normal.
Hi thanks for your answer. I am really graeful. I am a bit confused now. Please see the attached file in which I have shown the exact issue. I hope you can help me out.
Using the normal error model for this data is a bit optimistic. It is clear from the plot that the response is not normal distributed (the distribution at each "boardsizew" seems right-skewed). I have no ide what "absoluteofdiscretionaryaccrww" is, so I can't guess what distribution would be sensible to assume.
Whatever the relationship is, the noise in your data is considerably larger, so there won't be anything much you can infer about the relationship, no matter what model you chose.
The r²=0.0000 does not indicate that the relationship is "not linear". It just sais: using a linear relationship to model the data and fitting the coefficients by minimizing the sum of squared residuals, the model is able to explain less than 0.005% of the variance of the observed values.
Using your model (the linear functional model with assumed normal errors), the slope of the regression line is estimated to be 0.003 (per "boardsizeww"), and your data is considered "not too unexpected" for similar models with slopes between -0.0090 to +0.0095 (as given by the 95% confidence interval for "boardsizew"). Given the model specification would be more or less sensible, the next step would be to think if this range covers some relevant value, or if such slopes are completely irrelevant.
OLS regression can still be used to test many kinds of curvilinear relationships, if you add polynomial terms. However, as Jochen mentioned, your data might be problematic for other reasons; the outcome variable looks skewed, and the predictor variable has very few levels. You might want to treat the predictor as ordinal rather than continuous, and you also might want to do bootstrap regression rather than OLS.
thanks everyone for your valued out put. Sergiy please find attached my data. x14 x15 and x17 are dummy variables coded either 1 or 0. I hope you will be able to help. Looking forward.
Well, I made scatterplots of your data, but beyond that I don't know what more I can do to help. You haven't specified which variable you are interested in, and you have 18 variables of different types, which have different issues (e.g., some are categorical and some are not; some have normal distributions and some have serious abnormalities) and thus will need to be treated in different ways. If you don't state which specific variable you are interested in, you probably won't get any specific suggestions.
But in any case, inferential statistics should not be done without specific predictions anyway. If you just wanted to explore a lot of variables and visualize their effects, you have already done that just by looking at the scatterplots. If you didn't have clear predictions about the pattern of effects, and then you see from the scatterplot that a pair of variables shows a "rather random" relationship, then the proper conclusion is that you didn't find a significant relationship; it wouldn't be honest to try hacking the statistics to try to make a relationship appear.
Well it wouldd be difficult to explain the hypothesis here but I have just figured out that I made a mistake in quantifying a few variables. and some of them are normally distributed.