Given group lasso penalty that is defined as $$g(W)= \lambda \sum_j \|W_j\|_2$$ is it possible to drive an smooth approximation this group penalty. I looked at following paper by Nestrove, but it was not easy to understand it.
Smooth minimization of non-smooth functions
I would like to solve this optimization problem very efficently. \begin{equation} \hat{W}={arg\,min}_W \frac{1}{2}||Y-XW||^2 + \lambda ||W||_1 + \gamma \sum_{i=1}^{k} ||W_i||_2 \end{equation} Whre $W \in R{M \times T}$ , $W_i$ $ith$ row of $W$ and $Y \in R^{N \times T}$. I would like to write original problem as : \begin{equation} \hat{W}={arg\,min}_W g(W) + \lambda ||W||_1 + h(W) \end{equation} Where $g(W) =\frac{1}{2}||Y-XW||^2$, $h(W)$: Smooth approximation of $\gamma \sum_{i=1}^{k} ||W_i||_2$, then $F(W)=g(W)+h(W)$ will be differential function and I can easily compute it's gradient.This can help me to solve original problem : $$F(W) + \lambda ||W||_1$$ using proximal gradient decent, since proximal of $||W||_1$ is known and can be computed very fast.