|Babel user information|
|Users by language|
I am Artoria2e5 on The Test Wiki. My global user page contains a few more userboxes, so check them out.
⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜ ⬜⬛⬛⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬛⬜⬛⬛⬛⬛⬜⬜ ⬜⬜⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬛⬜ ⬜⬜⬛⬜⬜⬛⬛⬛⬜⬛⬜⬛⬜⬛⬜⬛⬛⬛⬛⬜⬜ <-- Emoj-ixel layout experiment; See zh:表情包 ⬜⬛⬜⬜⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬜⬜ ⬜⬛⬛⬛⬜⬛⬜⬛⬜⬜⬛⬜⬛⬜⬜⬛⬜⬜⬜⬜⬜ ⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜ ＼／＼／＼／＼／＼／＼／＼／＼／ ｃｏｒｎｅｒ ｒｅｆｌｅｃｔｏｒ
Document PUA code points used by old decoders for old CJK(V?) encodings, as normalizing PUA artifacts is as important as Unicode normalization. Most of the problem should arise in Chinese and Han Nom characters, as there are more characters to screw up.
- Microsoft mappings (reference implementation)
- commit logs from ICU and glibc (same)
- "CJKV Information Processing" by Ken Lunde
- Add the charts (or Python scripts with
str.translate) to stanfordnlp/CoreNLP wiki, then open an issue to suggest inclusion in .
- Wiki ready w/ Python script and chart links: 
What if we write something that automatically generates bad jokes by substituting random words in a Wikipedia article for some
boring:INPUT → funny:OUTPUT analogies? Word vectors can do that pretty well.
There are some good stuff I can backport to here from the zh adaptions. Some messy "refactor" diffs coming up someday.
- insert breaks to loops
- replace ad-hoc string ops with not-very-ad-hoc ones (aliases, etc.)
- probably go for a string table like CS1 is doing?
- indents. let's face it we don't care about dirty diffs if it's fixed once and for all.
- some styles like actual table literals.
- and yeah we don't need to write chrTextTable out like that.
- do early returns. why nest it when you can jump out of the wrong ones
- use long string literals.
- some hrt translation to zh.wp i guess
- gcj-02, damn it: User:Artoria2e5/coord-notice, phab:T162331
- idk, maybe some madgenderscience-level fuckup