We can add a smoothness objective by adding
-norm of the
derivative to the objective function.
![]() |
(4.79) |
The
-norm only cares about the maximum derivative.
Large
means we put more weight on the smoothness than the
side-lobe level.
This can be formulated as an LP by adding one optimization parameter
which bounds all derivatives.
![]() |
(4.80) |
![]() |
![]() |
![]() |
![]() |
(4.81) |
The result of adding the Chebyshev norm of diff(h) to the
objective function to be minimized (
) is shown in
Fig.3.39. The result of increasing
to 20 is
shown in Fig.3.40.