diff etc/unicode/README @ 779:a29c4eef8f00

[xemacs-hg @ 2002-03-18 09:40:27 by ben] add more translation tables [from mule-ucs], fixup README's
author ben
date Mon, 18 Mar 2002 09:40:41 +0000
parents 943eaba38521
children
line wrap: on
line diff
--- a/etc/unicode/README	Sat Mar 16 10:39:19 2002 +0000
+++ b/etc/unicode/README	Mon Mar 18 09:40:41 2002 +0000
@@ -9,34 +9,5 @@
 
 http://oss.software.ibm.com/icu/charset/
 
-Someone needs to write a simple program to parse these tables.  You
-should use the tables in unicode-consortium/; the ones in ibm/ can be
-used to supplement or check the accuracy of the others.
-
-Perhaps the best way is to put some C code in XEmacs, probably in the
-form of a Lisp primitive, to parse a table in a specified file and add
-the appropriate Unicode mappings using set_unicode_conversion.  Then
-it will be easy to read the tables at dump time.  Doing it this way
-avoids the need to create large Elisp files solely to initialize the
-tables, or embed a bunch of initializing data in the C code.
-
-I'd suggest this:
-
-DEFUN ("parse-unicode-translation-table", ..., 2, 5, 0 /*
-Parse Unicode translation data in FILENAME for CHARSET.
-Data is text, in the form of one translation per line -- charset codepoint
-followed by Unicode codepoint.  Numbers are decimal or hex (preceded by 0x).
-Comments are marked with a #.
-
-If START and END are given, only charset codepoints within the given range
-will be processed.  If OFFSET is given, that value will be added to all
-charset codepoints in the file to obtain the internal charset codepoint.
-
-(#### This still doesn't handle Big5 tables.  Either we need to special-case
-this or allow a CCL program or Lisp routine to do the conversion.)
-*/
-     (filename, charset, start, end, offset))
-{
-
-}
-
+The tables in unicode-consortium/ should be used as source data; the ones
+in ibm/ can be used to supplement or check the accuracy of the others.