A simple and implementable two-loop smoothing method for semi-infinite minimax problem is given with the discretization parameter and the smoothing parameter being updated adaptively. We prove the global convergence of the algorithm when the steepest descent method or a BFGS type quasi-Newton method is applied to the smooth subproblems. The strategy for updating the smoothing parameter can not only guarantee the convergence of the algorithm but also considerably reduce the ill-conditioning caused by increasing the value of the smoothing parameter. Numerical tests show that the algorithm is robust and effective.