> Am Fri, 20 Aug 2010 13:06:27 +0900 schrieb Andrew A. Adams: > > > I recently upgraded my Fedora Core 10 to Fedora Core 13. I'm getting a very > > strange behaviour from processing latex files including Japanese text and > > processed using xelatex. I've created a minimal input source file which > > demonstrates the problem, which is that the unicode characters in the input > > file are not the ones that appear in the output. It's possible that somehow > > I'm getting Chinese characters instead of the Japanese ones in my original > > file. I create my files in xemacs, and set the buffer encoding to UTF-8. I > > use a script to process the file using xelatex with my default options: > > > > xelatex -interaction=nonstopmode -output-driver="xdvipdfmx -p a4 -V5 " > > $1.tex > > && acroread -tempFile $1.pdf > > > > Attached are the sample tex file, the resulting output file, the log file > > from manual xelatex processing and the output from manual xdvipdfmx > > processing. > > Don't use inputenc with xelatex. Never! inputenc is meant for > 8-bit-machines. It breaks with xelatex. If your file is utf8 or > utf16 there is no nead to declare the encoding.
inputenc is not the problem. It persists even when I take out that line, and when I have the file encoded in UTF-8 (sorry for the UTF-16 version I posted earlier). -- Professor Andrew A Adams a...@meiji.ac.jp Professor at Graduate School of Business Administration, and Deputy Director of the Centre for Business Information Ethics Meiji University, Tokyo, Japan http://www.a-cubed.info/ -------------------------------------------------- Subscriptions, Archive, and List information, etc.: http://tug.org/mailman/listinfo/xetex