This may be a stupid question, but I was just wondering why you guys use a convex combination of the two regularisers? This isn't necessary to maintain convexity. Am I correct in saying that any linear combination of the regularizers will still form a convex problem?
I can see that the alpha effectively weighs in group sparsity or more general sparsity, but what if we want to look at a case where we have high group sparsity and high general sparsity? Did you consider treating the regularizers with different lambdas?
I have seen the convex combination (1-alpha) R1 +alpha*R2 used for regularization in other papers also, so perhaps there is something obvious I am missing.
Nice paper though!
Article A Sparse-Group Lasso