Model run times

What settings have the most impact on run times? I have already seen that if I cap the number of predictors from the default of 50 to 12 this has a big impact. I would assume looking for higher order terms is another one.
1

Question on Iterations

User question received today. "When I’m running various iterations of a model (let’s saying Categorical Binning Yes/No and Numeric Binning Yes/No) there are 4 iterations. However, I can’t analyze each iteration separately. For example there are 4 iterations, with different numeric lifts. However, if I want to explore the model coefficients for the 4 iterations, I can’t. I am only left to look at iteration 2. I also couldn’t (for example) choose to put version 1.2.4 into production. I could only
1

Prunning Trees

Does PredictIQ have the ability to prune trees within models to avoid over fitting. I found this article which talks about this. http://www.statmethods.net/advstats/cart.html
1

I turned off numeric and categorical binning, but I'm still getting bins?

LityxIQ may still do binning in some situations depending on other settings like choosing to have numeric missing replaced with categories or having too many unique values in a field. It will bin for efficiency sake if the field has too many categories. Specifically, some/most algorithms do not scale well if the number is more than 25 to 30. For numerics, if "categorize" is checked in the missing variable area, then it will be binned if there are missing values in the field, even if the algorith
not commented yet

XGBoost Overfitting?

XGBoost is great, but there may be concern that it results in models that are too good and overfitted. Are there ways to reduce the "tightness" of the model? In the estimation of the predictor coefficients? Related, is there any generally accepted measure of overfitting?
1

If I use an oversampled file for model building can I show the results as if I had not used an oversample file?

Within the PredictIQ model build settings the last tab called output allows you to ‘set population average for target’. This results in the predicted scores being scaled down to reflect the population average. It does not however make any adjustments to the actual results or the associated lifts from the model. This can be accomplished by appending model scores to the full modeling file and using the distribution of the targets and non-targets by decile proportionally increase the non-targets ba
not commented yet

Pruning Trees

Does PREDICT have the ability to prune trees within models to avoid over fitting. I found this article which talks about this. http://www.statmethods.net/advstats/cart.html
1

Lift vs. Numeric Lift what is the difference?

Building a revenue model and I see that Lift 1 over random =3.2 so decile 1 is 3.2 higher than the random. However, Numeric 1 vs 10 is only 2.83 -- what is numeric 1 vs 10 -- would expect it to be decile 1 vs decile 10, but of course it would be higher than 2.83 if that was the case.
1

Model Run Time

Is there a way that I can tell how long a model run took?
1

Question on model weights

If you took all responders and 1/3 of non-responders, I would historically build models using the Weight statement since not including it would affect the model calibration. In addition, the performance statistics such as lift need to be off the weighted data. How would you recommend this be handled in LityxIQ?
1

Scoring Code Output

Can I output the scoring code for a model I create in LityxIQ?
1