Is health insurance mandatory in Hawaii?

Is health insurance mandatory in Hawaii?

On January 1, 19’75, Hawaii became the first State to have a mandated health insurance pro- gram in effect. The law, enacted June 12, 19’74, requires employers to provide protection against the costs of hospital and medical care for their employees.

See also  When did Humana sell its hospitals?