From Wikipedia, the free encyclopedia
I am Artoria2e5 on The Test Wiki. My global user page contains a few more userboxes, so check them out if you are looking for social pages.
⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜
⬜⬛⬛⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬛⬜⬛⬛⬛⬛⬜⬜
⬜⬜⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬛⬜
⬜⬜⬛⬜⬜⬛⬛⬛⬜⬛⬜⬛⬜⬛⬜⬛⬛⬛⬛⬜⬜ <-- Emoj-ixel layout experiment; See zh:表情包
⬜⬛⬜⬜⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬛⬜⬜⬜⬜⬜
⬜⬛⬛⬛⬜⬛⬜⬛⬜⬜⬛⬜⬛⬜⬜⬛⬜⬜⬜⬜⬜
⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜⬜
\/\/\/\/\/\/\/\/
corner reflector
I forget about things. Why not run me on XTools?
Document PUA code points used by old decoders for old CJK(V?) encodings, as normalizing PUA artifacts is as important as Unicode normalization. Most of the problem should arise in Chinese and Han Nom characters, as there are more characters to screw up.
Useful references:
- Microsoft mappings (reference implementation)
- commit logs from ICU and glibc (same)
- WHATWG
- "CJKV Information Processing" by Ken Lunde
Other plans:
- Add the charts (or Python scripts with
str.translate
) to stanfordnlp/CoreNLP wiki, then open an issue to suggest inclusion in [1].
- Wiki ready w/ Python script and chart links: [2]
What if we write something that automatically generates bad jokes by substituting random words in a Wikipedia article for some boring:INPUT → funny:OUTPUT
analogies? Word vectors can do that pretty well.
There are some good stuff I can backport to here from the zh adaptions. Some messy "refactor" diffs coming up someday.
- insert breaks to loops
- replace ad-hoc string ops with not-very-ad-hoc ones (aliases, etc.)
- probably go for a string table like CS1 is doing?
- indents. let's face it we don't care about dirty diffs if it's fixed once and for all.
- some styles like actual table literals.
- and yeah we don't need to write chrTextTable out like that.
- do early returns. why nest it when you can jump out of the wrong ones
- use long string literals.
- Hugin, stitching, surprisingly cool shite
- GIMP
Todo bucket: