Approach for comparing linear, non-linear and different parameterization non-linear modelsLeave one out cross validation with lm function in RSave multiple linear regression modelsIs there a way to mimic the nls nls.control(warnOnly=TRUE) when using nlsLM?Using generalized linear models to compare group means in RCross-validation for non-linear regression using nls in RHow to calculate mean values from a linear model in R?Modeling a repeated measures logistic growth curveNon-linear regression vs log modelUse VAE approach in a Linear Regression model to compute output uncertaintyExtract and add to the data values of the probability density function based on a stan linear model

What's the purpose of autocorrelation?

Paradox regarding phase transitions in relativistic systems

What can I actually do with a high credit score?

What was the deeper meaning of Hermione wanting the cloak?

EU compensation - fire alarm at the Flight Crew's hotel

We suspect colleague is stealing company code - what do we do?

Lumens specs when buying a flashlight/torch. Why maglite has only 680lm but cheap ones have 900,000?

Why are there two bearded faces wearing red hats on my stealth bomber icon?

Carroll's interpretation of 1-forms

Is it safe to put a microwave in a walk-in closet?

Madrid to London w/ Expired 90/180 days stay as US citizen

How can I create folders in folders in terminal

Is there any actual security benefit to restricting foreign IPs?

Why do things cool off?

What is the word for a person who destroys monuments?

Persuading players to be less attached to a pre-session 0 character concept

Can I separate garlic into cloves for storage?

Where did Otto von Bismarck say "lying awake all night, hating"?

How could artificial intelligence harm us?

What happens when I use Drawmij's Instant Summons on Dimensional Shackles?

Can Brexit be undone in an emergency?

Cemented carbide swords - worth it?

Floating Point XOR

Applications of mathematics in clinical setting



Approach for comparing linear, non-linear and different parameterization non-linear models


Leave one out cross validation with lm function in RSave multiple linear regression modelsIs there a way to mimic the nls nls.control(warnOnly=TRUE) when using nlsLM?Using generalized linear models to compare group means in RCross-validation for non-linear regression using nls in RHow to calculate mean values from a linear model in R?Modeling a repeated measures logistic growth curveNon-linear regression vs log modelUse VAE approach in a Linear Regression model to compute output uncertaintyExtract and add to the data values of the probability density function based on a stan linear model






.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;








0















I search for one approach for comparing linear, non-linear and different parameterization non-linear models. For this:



#Packages
library(nls2)
library(minpack.lm)

# Data set - Diameter in function of Feature and Age
Feature<-sort(rep(c("A","B"),22))
Age<-c(60,72,88,96,27,
36,48,60,72,88,96,27,36,48,60,72,
88,96,27,36,48,60,27,27,36,48,60,
72,88,96,27,36,48,60,72,88,96,27,
36,48,60,72,88,96)
Diameter<-c(13.9,16.2,
19.1,19.3,4.7,6.7,9.6,11.2,13.1,15.3,
15.4,5.4,7,9.9,11.7,13.4,16.1,16.2,
5.9,8.3,12.3,14.5,2.3,5.2,6.2,8.6,9.3,
11.3,15.1,15.5,5,7,7.9,8.4,10.5,14,14,
4.1,4.9,6,6.7,7.7,8,8.2)
d<-dados <- data.frame(Feature,Age,Diameter)
str(d)


I will create three different models, two non-linear models with specific parametization and one linear model. In my example
a suppose that all the coefficients of each mode were significant (and not considering real results).



# Model 1 non-linear
e1<- Diameter ~ a1 * Age^a2
#Algoritm Levenberg-Marquardt
m1 <- nlsLM(e1, data = d,
start = list(a1 = 0.1, a2 = 10),
control = nls.control(maxiter = 1000))

# Model 2 linear
m2<-lm(Diameter ~ Age, data=d)

# Model 3 another non-linear
e2<- Diameter ~ a1^(-Age/a2)
m3 <- nls2(e2, data = d, alg = "brute-force",
start = data.frame(a1 = c(-1, 1), a2 = c(-1, 1)),
control = nls.control(maxiter = 1000))


Now, my idea is comparing the "better" model despite the different nature of each model, than I try a proportional measure
and for this I use each mean square error of each model comparing of total square error in data set, when a make this I have if
a comparing model 1 and 2:



## MSE approach (like pseudo R2 approach)

#Model 1
SQEm1<-summary(m1)$sigma^2*summary(m1)$df[2]# mean square error of model
SQTm1<-var(d$Diameter)*(length(d$Diameter)-1)#total square error in data se
R1<-1-SQEm1/SQTm1
R1

#Model 2
SQEm2<-summary(m2)$sigma^2*summary(m2)$df[2]# mean square error of model
R2<-1-SQEm2/SQTm1
R2


In my weak opinion model 1 is "better" that model 2. My question is, does this approach sounds correct? Is there any way to compare these models types?



Thanks in advance!










share|improve this question





















  • 2





    this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

    – gfgm
    Mar 28 at 14:01






  • 1





    IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

    – G. Grothendieck
    Mar 28 at 17:29


















0















I search for one approach for comparing linear, non-linear and different parameterization non-linear models. For this:



#Packages
library(nls2)
library(minpack.lm)

# Data set - Diameter in function of Feature and Age
Feature<-sort(rep(c("A","B"),22))
Age<-c(60,72,88,96,27,
36,48,60,72,88,96,27,36,48,60,72,
88,96,27,36,48,60,27,27,36,48,60,
72,88,96,27,36,48,60,72,88,96,27,
36,48,60,72,88,96)
Diameter<-c(13.9,16.2,
19.1,19.3,4.7,6.7,9.6,11.2,13.1,15.3,
15.4,5.4,7,9.9,11.7,13.4,16.1,16.2,
5.9,8.3,12.3,14.5,2.3,5.2,6.2,8.6,9.3,
11.3,15.1,15.5,5,7,7.9,8.4,10.5,14,14,
4.1,4.9,6,6.7,7.7,8,8.2)
d<-dados <- data.frame(Feature,Age,Diameter)
str(d)


I will create three different models, two non-linear models with specific parametization and one linear model. In my example
a suppose that all the coefficients of each mode were significant (and not considering real results).



# Model 1 non-linear
e1<- Diameter ~ a1 * Age^a2
#Algoritm Levenberg-Marquardt
m1 <- nlsLM(e1, data = d,
start = list(a1 = 0.1, a2 = 10),
control = nls.control(maxiter = 1000))

# Model 2 linear
m2<-lm(Diameter ~ Age, data=d)

# Model 3 another non-linear
e2<- Diameter ~ a1^(-Age/a2)
m3 <- nls2(e2, data = d, alg = "brute-force",
start = data.frame(a1 = c(-1, 1), a2 = c(-1, 1)),
control = nls.control(maxiter = 1000))


Now, my idea is comparing the "better" model despite the different nature of each model, than I try a proportional measure
and for this I use each mean square error of each model comparing of total square error in data set, when a make this I have if
a comparing model 1 and 2:



## MSE approach (like pseudo R2 approach)

#Model 1
SQEm1<-summary(m1)$sigma^2*summary(m1)$df[2]# mean square error of model
SQTm1<-var(d$Diameter)*(length(d$Diameter)-1)#total square error in data se
R1<-1-SQEm1/SQTm1
R1

#Model 2
SQEm2<-summary(m2)$sigma^2*summary(m2)$df[2]# mean square error of model
R2<-1-SQEm2/SQTm1
R2


In my weak opinion model 1 is "better" that model 2. My question is, does this approach sounds correct? Is there any way to compare these models types?



Thanks in advance!










share|improve this question





















  • 2





    this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

    – gfgm
    Mar 28 at 14:01






  • 1





    IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

    – G. Grothendieck
    Mar 28 at 17:29














0












0








0


1






I search for one approach for comparing linear, non-linear and different parameterization non-linear models. For this:



#Packages
library(nls2)
library(minpack.lm)

# Data set - Diameter in function of Feature and Age
Feature<-sort(rep(c("A","B"),22))
Age<-c(60,72,88,96,27,
36,48,60,72,88,96,27,36,48,60,72,
88,96,27,36,48,60,27,27,36,48,60,
72,88,96,27,36,48,60,72,88,96,27,
36,48,60,72,88,96)
Diameter<-c(13.9,16.2,
19.1,19.3,4.7,6.7,9.6,11.2,13.1,15.3,
15.4,5.4,7,9.9,11.7,13.4,16.1,16.2,
5.9,8.3,12.3,14.5,2.3,5.2,6.2,8.6,9.3,
11.3,15.1,15.5,5,7,7.9,8.4,10.5,14,14,
4.1,4.9,6,6.7,7.7,8,8.2)
d<-dados <- data.frame(Feature,Age,Diameter)
str(d)


I will create three different models, two non-linear models with specific parametization and one linear model. In my example
a suppose that all the coefficients of each mode were significant (and not considering real results).



# Model 1 non-linear
e1<- Diameter ~ a1 * Age^a2
#Algoritm Levenberg-Marquardt
m1 <- nlsLM(e1, data = d,
start = list(a1 = 0.1, a2 = 10),
control = nls.control(maxiter = 1000))

# Model 2 linear
m2<-lm(Diameter ~ Age, data=d)

# Model 3 another non-linear
e2<- Diameter ~ a1^(-Age/a2)
m3 <- nls2(e2, data = d, alg = "brute-force",
start = data.frame(a1 = c(-1, 1), a2 = c(-1, 1)),
control = nls.control(maxiter = 1000))


Now, my idea is comparing the "better" model despite the different nature of each model, than I try a proportional measure
and for this I use each mean square error of each model comparing of total square error in data set, when a make this I have if
a comparing model 1 and 2:



## MSE approach (like pseudo R2 approach)

#Model 1
SQEm1<-summary(m1)$sigma^2*summary(m1)$df[2]# mean square error of model
SQTm1<-var(d$Diameter)*(length(d$Diameter)-1)#total square error in data se
R1<-1-SQEm1/SQTm1
R1

#Model 2
SQEm2<-summary(m2)$sigma^2*summary(m2)$df[2]# mean square error of model
R2<-1-SQEm2/SQTm1
R2


In my weak opinion model 1 is "better" that model 2. My question is, does this approach sounds correct? Is there any way to compare these models types?



Thanks in advance!










share|improve this question
















I search for one approach for comparing linear, non-linear and different parameterization non-linear models. For this:



#Packages
library(nls2)
library(minpack.lm)

# Data set - Diameter in function of Feature and Age
Feature<-sort(rep(c("A","B"),22))
Age<-c(60,72,88,96,27,
36,48,60,72,88,96,27,36,48,60,72,
88,96,27,36,48,60,27,27,36,48,60,
72,88,96,27,36,48,60,72,88,96,27,
36,48,60,72,88,96)
Diameter<-c(13.9,16.2,
19.1,19.3,4.7,6.7,9.6,11.2,13.1,15.3,
15.4,5.4,7,9.9,11.7,13.4,16.1,16.2,
5.9,8.3,12.3,14.5,2.3,5.2,6.2,8.6,9.3,
11.3,15.1,15.5,5,7,7.9,8.4,10.5,14,14,
4.1,4.9,6,6.7,7.7,8,8.2)
d<-dados <- data.frame(Feature,Age,Diameter)
str(d)


I will create three different models, two non-linear models with specific parametization and one linear model. In my example
a suppose that all the coefficients of each mode were significant (and not considering real results).



# Model 1 non-linear
e1<- Diameter ~ a1 * Age^a2
#Algoritm Levenberg-Marquardt
m1 <- nlsLM(e1, data = d,
start = list(a1 = 0.1, a2 = 10),
control = nls.control(maxiter = 1000))

# Model 2 linear
m2<-lm(Diameter ~ Age, data=d)

# Model 3 another non-linear
e2<- Diameter ~ a1^(-Age/a2)
m3 <- nls2(e2, data = d, alg = "brute-force",
start = data.frame(a1 = c(-1, 1), a2 = c(-1, 1)),
control = nls.control(maxiter = 1000))


Now, my idea is comparing the "better" model despite the different nature of each model, than I try a proportional measure
and for this I use each mean square error of each model comparing of total square error in data set, when a make this I have if
a comparing model 1 and 2:



## MSE approach (like pseudo R2 approach)

#Model 1
SQEm1<-summary(m1)$sigma^2*summary(m1)$df[2]# mean square error of model
SQTm1<-var(d$Diameter)*(length(d$Diameter)-1)#total square error in data se
R1<-1-SQEm1/SQTm1
R1

#Model 2
SQEm2<-summary(m2)$sigma^2*summary(m2)$df[2]# mean square error of model
R2<-1-SQEm2/SQTm1
R2


In my weak opinion model 1 is "better" that model 2. My question is, does this approach sounds correct? Is there any way to compare these models types?



Thanks in advance!







r linear-regression lm nls non-linear-regression






share|improve this question















share|improve this question













share|improve this question




share|improve this question








edited Mar 28 at 23:01









Steve

4,8101 gold badge8 silver badges31 bronze badges




4,8101 gold badge8 silver badges31 bronze badges










asked Mar 28 at 13:44









LeprechaultLeprechault

2552 silver badges11 bronze badges




2552 silver badges11 bronze badges










  • 2





    this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

    – gfgm
    Mar 28 at 14:01






  • 1





    IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

    – G. Grothendieck
    Mar 28 at 17:29













  • 2





    this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

    – gfgm
    Mar 28 at 14:01






  • 1





    IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

    – G. Grothendieck
    Mar 28 at 17:29








2




2





this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

– gfgm
Mar 28 at 14:01





this way of comparing models doesn't penalize models for complexity and risks overfitting the data. you would be safer comparing your models via cross-validation

– gfgm
Mar 28 at 14:01




1




1





IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

– G. Grothendieck
Mar 28 at 17:29






IF the models have the same number of parameters as is the case here then you can just use the sum of squares of residuals: deviance(m1); deviance(m2) where lower is better. Also graph the fit superimposed on the data and that may make it obvious which model fits best. Be sure to sort the data on Age so that the plots come out right.

– G. Grothendieck
Mar 28 at 17:29













1 Answer
1






active

oldest

votes


















1
















#First cross-validation approach ------------------------------------------

#Cross-validation model 1
set.seed(123) # for reproducibility

n <- nrow(d)
frac <- 0.8
ix <- sample(n, frac * n) # indexes of in sample rows

e1<- Diameter ~ a1 * Age^a2
#Algoritm Levenberg-Marquardt
m1 <- nlsLM(e1, data = d,
start = list(a1 = 0.1, a2 = 10),
control = nls.control(maxiter = 1000), subset = ix)# in sample model

BOD.out <- d[-ix, ] # out of sample data
pred <- predict(m1, new = BOD.out)
act <- BOD.out$Diameter
RSS1 <- sum( (pred - act)^2 )
RSS1
#[1] 56435894734

#Cross-validation model 2
m2<-lm(Diameter ~ Age, data=d,, subset = ix)# in sample model
BOD.out2 <- d[-ix, ] # out of sample data
pred <- predict(m2, new = BOD.out2)
act <- BOD.out2$Diameter
RSS2 <- sum( (pred - act)^2 )
RSS2
#[1] 19.11031

# Sum of squares approach -----------------------------------------------
deviance(m1)
#[1] 238314429037

deviance(m2)
#[1] 257.8223


Based in gfgm and G. Grothendieck comments, RSS2 has lower error that RSS1 and comparing deviance(m2) and deviance(m2) too, than model 2 is better than model 1.






share|improve this answer


























    Your Answer






    StackExchange.ifUsing("editor", function ()
    StackExchange.using("externalEditor", function ()
    StackExchange.using("snippets", function ()
    StackExchange.snippets.init();
    );
    );
    , "code-snippets");

    StackExchange.ready(function()
    var channelOptions =
    tags: "".split(" "),
    id: "1"
    ;
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function()
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled)
    StackExchange.using("snippets", function()
    createEditor();
    );

    else
    createEditor();

    );

    function createEditor()
    StackExchange.prepareEditor(
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader:
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/4.0/"u003ecc by-sa 4.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    ,
    onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    );



    );














    draft saved

    draft discarded
















    StackExchange.ready(
    function ()
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55399164%2fapproach-for-comparing-linear-non-linear-and-different-parameterization-non-lin%23new-answer', 'question_page');

    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    1
















    #First cross-validation approach ------------------------------------------

    #Cross-validation model 1
    set.seed(123) # for reproducibility

    n <- nrow(d)
    frac <- 0.8
    ix <- sample(n, frac * n) # indexes of in sample rows

    e1<- Diameter ~ a1 * Age^a2
    #Algoritm Levenberg-Marquardt
    m1 <- nlsLM(e1, data = d,
    start = list(a1 = 0.1, a2 = 10),
    control = nls.control(maxiter = 1000), subset = ix)# in sample model

    BOD.out <- d[-ix, ] # out of sample data
    pred <- predict(m1, new = BOD.out)
    act <- BOD.out$Diameter
    RSS1 <- sum( (pred - act)^2 )
    RSS1
    #[1] 56435894734

    #Cross-validation model 2
    m2<-lm(Diameter ~ Age, data=d,, subset = ix)# in sample model
    BOD.out2 <- d[-ix, ] # out of sample data
    pred <- predict(m2, new = BOD.out2)
    act <- BOD.out2$Diameter
    RSS2 <- sum( (pred - act)^2 )
    RSS2
    #[1] 19.11031

    # Sum of squares approach -----------------------------------------------
    deviance(m1)
    #[1] 238314429037

    deviance(m2)
    #[1] 257.8223


    Based in gfgm and G. Grothendieck comments, RSS2 has lower error that RSS1 and comparing deviance(m2) and deviance(m2) too, than model 2 is better than model 1.






    share|improve this answer































      1
















      #First cross-validation approach ------------------------------------------

      #Cross-validation model 1
      set.seed(123) # for reproducibility

      n <- nrow(d)
      frac <- 0.8
      ix <- sample(n, frac * n) # indexes of in sample rows

      e1<- Diameter ~ a1 * Age^a2
      #Algoritm Levenberg-Marquardt
      m1 <- nlsLM(e1, data = d,
      start = list(a1 = 0.1, a2 = 10),
      control = nls.control(maxiter = 1000), subset = ix)# in sample model

      BOD.out <- d[-ix, ] # out of sample data
      pred <- predict(m1, new = BOD.out)
      act <- BOD.out$Diameter
      RSS1 <- sum( (pred - act)^2 )
      RSS1
      #[1] 56435894734

      #Cross-validation model 2
      m2<-lm(Diameter ~ Age, data=d,, subset = ix)# in sample model
      BOD.out2 <- d[-ix, ] # out of sample data
      pred <- predict(m2, new = BOD.out2)
      act <- BOD.out2$Diameter
      RSS2 <- sum( (pred - act)^2 )
      RSS2
      #[1] 19.11031

      # Sum of squares approach -----------------------------------------------
      deviance(m1)
      #[1] 238314429037

      deviance(m2)
      #[1] 257.8223


      Based in gfgm and G. Grothendieck comments, RSS2 has lower error that RSS1 and comparing deviance(m2) and deviance(m2) too, than model 2 is better than model 1.






      share|improve this answer





























        1














        1










        1









        #First cross-validation approach ------------------------------------------

        #Cross-validation model 1
        set.seed(123) # for reproducibility

        n <- nrow(d)
        frac <- 0.8
        ix <- sample(n, frac * n) # indexes of in sample rows

        e1<- Diameter ~ a1 * Age^a2
        #Algoritm Levenberg-Marquardt
        m1 <- nlsLM(e1, data = d,
        start = list(a1 = 0.1, a2 = 10),
        control = nls.control(maxiter = 1000), subset = ix)# in sample model

        BOD.out <- d[-ix, ] # out of sample data
        pred <- predict(m1, new = BOD.out)
        act <- BOD.out$Diameter
        RSS1 <- sum( (pred - act)^2 )
        RSS1
        #[1] 56435894734

        #Cross-validation model 2
        m2<-lm(Diameter ~ Age, data=d,, subset = ix)# in sample model
        BOD.out2 <- d[-ix, ] # out of sample data
        pred <- predict(m2, new = BOD.out2)
        act <- BOD.out2$Diameter
        RSS2 <- sum( (pred - act)^2 )
        RSS2
        #[1] 19.11031

        # Sum of squares approach -----------------------------------------------
        deviance(m1)
        #[1] 238314429037

        deviance(m2)
        #[1] 257.8223


        Based in gfgm and G. Grothendieck comments, RSS2 has lower error that RSS1 and comparing deviance(m2) and deviance(m2) too, than model 2 is better than model 1.






        share|improve this answer















        #First cross-validation approach ------------------------------------------

        #Cross-validation model 1
        set.seed(123) # for reproducibility

        n <- nrow(d)
        frac <- 0.8
        ix <- sample(n, frac * n) # indexes of in sample rows

        e1<- Diameter ~ a1 * Age^a2
        #Algoritm Levenberg-Marquardt
        m1 <- nlsLM(e1, data = d,
        start = list(a1 = 0.1, a2 = 10),
        control = nls.control(maxiter = 1000), subset = ix)# in sample model

        BOD.out <- d[-ix, ] # out of sample data
        pred <- predict(m1, new = BOD.out)
        act <- BOD.out$Diameter
        RSS1 <- sum( (pred - act)^2 )
        RSS1
        #[1] 56435894734

        #Cross-validation model 2
        m2<-lm(Diameter ~ Age, data=d,, subset = ix)# in sample model
        BOD.out2 <- d[-ix, ] # out of sample data
        pred <- predict(m2, new = BOD.out2)
        act <- BOD.out2$Diameter
        RSS2 <- sum( (pred - act)^2 )
        RSS2
        #[1] 19.11031

        # Sum of squares approach -----------------------------------------------
        deviance(m1)
        #[1] 238314429037

        deviance(m2)
        #[1] 257.8223


        Based in gfgm and G. Grothendieck comments, RSS2 has lower error that RSS1 and comparing deviance(m2) and deviance(m2) too, than model 2 is better than model 1.







        share|improve this answer














        share|improve this answer



        share|improve this answer








        edited Mar 28 at 19:07

























        answered Mar 28 at 19:00









        LeprechaultLeprechault

        2552 silver badges11 bronze badges




        2552 silver badges11 bronze badges





















            Got a question that you can’t ask on public Stack Overflow? Learn more about sharing private information with Stack Overflow for Teams.







            Got a question that you can’t ask on public Stack Overflow? Learn more about sharing private information with Stack Overflow for Teams.




















            draft saved

            draft discarded















































            Thanks for contributing an answer to Stack Overflow!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid


            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.

            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function ()
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f55399164%2fapproach-for-comparing-linear-non-linear-and-different-parameterization-non-lin%23new-answer', 'question_page');

            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            Kamusi Yaliyomo Aina za kamusi | Muundo wa kamusi | Faida za kamusi | Dhima ya picha katika kamusi | Marejeo | Tazama pia | Viungo vya nje | UrambazajiKuhusu kamusiGo-SwahiliWiki-KamusiKamusi ya Kiswahili na Kiingerezakuihariri na kuongeza habari

            Swift 4 - func physicsWorld not invoked on collision? The Next CEO of Stack OverflowHow to call Objective-C code from Swift#ifdef replacement in the Swift language@selector() in Swift?#pragma mark in Swift?Swift for loop: for index, element in array?dispatch_after - GCD in Swift?Swift Beta performance: sorting arraysSplit a String into an array in Swift?The use of Swift 3 @objc inference in Swift 4 mode is deprecated?How to optimize UITableViewCell, because my UITableView lags

            Access current req object everywhere in Node.js ExpressWhy are global variables considered bad practice? (node.js)Using req & res across functionsHow do I get the path to the current script with Node.js?What is Node.js' Connect, Express and “middleware”?Node.js w/ express error handling in callbackHow to access the GET parameters after “?” in Express?Modify Node.js req object parametersAccess “app” variable inside of ExpressJS/ConnectJS middleware?Node.js Express app - request objectAngular Http Module considered middleware?Session variables in ExpressJSAdd properties to the req object in expressjs with Typescript