This may be a stupid question, but I was just wondering why you guys use a convex combination of the two regularisers? This isn't necessary to maintain convexity. Am I correct in saying that any linear combination of the regularizers will still form a convex problem?

I can see that the alpha effectively weighs in group sparsity or more general sparsity, but what if we want to look at a case where we have high group sparsity and high general sparsity? Did you consider treating the regularizers with different lambdas?

I have seen the convex combination (1-alpha) R1 +alpha*R2 used for regularization in other papers also, so perhaps there is something obvious I am missing.

Nice paper though!

Article A Sparse-Group Lasso

Similar questions and discussions