Defines the length semantics to be used in programs.
Define the FGL_LENGTH_SEMANTICS environment variable to specify byte or character length semantics, by setting the value to BYTE or CHAR, respectively.
If the variable is not set, byte length semantics will be used by default.
When using a single-byte character set such as ISO-8859-1, use byte length semantics (the default). If the application character set is UTF-8, you should use char length semantics.