In the acknowledgements of my paper I have to thank a German professor and his last name contains an ö. Every time I compile the file I end up getting an A with a tilde and an ü instead of an ö (see picture).
Does anyone know how to circumvent this?
Thanks for the support.
The characters ä,ö,ü are called "Umlaute" in german.
Maybe that does help you in some way.
I found this TeX answer.
According to it, the encoding of ö should be {\"o} in LaTeX.
You can use \ddot{o}
If you need an ä \ddot{o}
Related
I have two versions for my store - English and French. And I am doing the translation from English to French in app/locale/fr_FR/Mage_Page.csv
I notice that I have to use some codes for certain French characters, such as En-tête de page for tête de page.
So if I have French words like 100% Magasinage sécurisé, how can use convert it into codes like En-tête de page?
I think it can come from the encoding of your file Mage_Page.php. If you use linux, I think that you have it in your file's properties and if you use Windows, you can check with notepad++. Richard B. gave a good link that in my opinion might solve the problem.
I'm using french on all the magento website that we produce and I do not have any problem like that so it must come from the encoding of your file.
One last thing : if your file is utf8 encoded, it means that your text was imported from a different one and editor did what it could but not successfully to change the characters (can happen with text from excel).
The inspiration for this question comes from this SO answer: https://stackoverflow.com/a/1732454/11125
I want to write a script to convert standard looking characters (e.g. letters in the English alphabet), to UTF-8 characters that are unusual and makes me appear crazy.
Did kiamlaluno sift through the thousands of UTF-8 characters to find similar looking characters to English letters, or is there possibly a more automated way to do this?
I would like the answer in Python, but any programming language will probably give me a suitable solution.
For example, convert:
PONY
to
P̯͍̭O̚N̐Y̡
A decent excerpt from the SO answer:
.
co*̶͑̾̾̅ͫ͏̙̤g͇̫͛͆̾ͫ̑͆l͖͉̗̩̳̟̍ͫͥͨe̠̅sZA̡͊͠͝LGΌ ISͮ̂҉̯͈͕̹̘̱ TO͇̹̺ͅƝ̴ȳ̳ TH̘Ë͖́̉ ͠P̯͍̭O̚N̐Y̡ H̸̡̪̯ͨ͊̽̅̾̎Ȩ̬̩̾͛ͪ̈́̀́͘ ̶̧̨̱̹̭̯ͧ̾ͬC̷̙̲̝͖ͭ̏ͥͮ͟Oͮ͏̮̪̝͍M̲̖͊̒ͪͩͬ̚̚͜Ȇ̴̟̟͙̞ͩ͌͝S̨̥̫͎̭ͯ̿̔̀ͅ?
There's a python module called pyzalgo.
>>> import zalgo
>>> print zalgo.zalgo('hello')
Ḩ̝̳͓̪̲͌̌͗̔́͌̄͢͢͢͢͢͢E̶̢̯̭̟̥̖̯̙̬ͪ̄̉ͭ͗̏́̈́L̢̢͓̺͇̠̠͉͙̪͍̽ͣͩ͂̾͒̊ͦL̵̶̶̢̬͈̬̦̜̥̲̜͓͎̝̣̙̞̝Ơ̷̸̴̷̘͙͕͈̞̜̠͙̰̬̰̣̟̭
It appears to be pretty straightforward and customisable after a quick look at the code.
Thanks for the info #hobbs.
Perhaps not "crazy looking", but something in the same vein:
https://github.com/reinderien/mimic
I have a AjaXplorer installation on a Linux Webserver.
One of the plugins of AjaXplorer is Codemirror - to view and edit text files.
Now I have the following situation: If I create a txt-File on Windows (ANSI) and upload it into Ajaxplorer (UTF-8), Codemirror shows every special character as a question mark. Consequently the whole file will be saved with question marks instead of the special characters.
But if a file once is saved in UTF-8, the special characters will be saved correctly.
So the problem exists in opening the ANSI-File. This is the point where I have to implement the solution, for example convert ANSI to UTF-8.
The 'funny' thing is, that if I open a fresh uploaded ANSI-File and a saved UTF-8 file with VIM on the Linux console, they look exactly the same, but the output in codemirror is different.
This is a uploaded ANSI-File with special characters like ä and ö and ü
Output in Codemirror: 'like ? and ? and ?'
This is a saved UTF8-File with special characters like ä and ö and ü
Output in Codemirror 'like ä and ö and ü'
This is the CodeMirror-Class of AjaXplorer and I think here must be the point where I could intervene:
https://github.com/mattleff/AjaXplorer/blob/master/plugins/editor.codemirror/class.CodeMirrorEditor.js
As you may see, I'm not a pro and I already tried some code pieces - otherwise I already had the solution ;-) I would be happy if someone gives me a hint! Thank you!!!
I'm using Par (in linux) to get nice comments formatting quickly. The problem is that now I want to introduce comments that include some international characters, like áéíóú or äëïöü...
The program Berkeley Par considers these international characters as 2 ASCII characters (I believe) and it outputs the comments somehow broken because it doesn't count characters properly.
Did you face this problem before? Do you have any solution? Ideas?
You mean the code from Add multibyte characters support in "par" (or just the patches applied to the original source) don't work for you?
Then maybe it is a problem with your shell or the font it uses. Are you sure the shell and font you use is able to reproduce unicode characters
Par, as distributed in Ubuntu from Hardy on, is supposed to handle multi-byte encodings.
http://packages.ubuntu.com/hardy/par
I've never even heard of this tool, but check out par 1.52.
The latest version of Par, released on 2001-Apr-29, tar'd and gzip'd. The only real change is better support for 8-bit character sets (as opposed to just 7-bit ASCII), but see also the release notes.
Edit: On the page, see par_1.52-i18n.3.diff.gz:
A patch by Jérôme Pouiller that adds
support for multibyte charsets (like
UTF-8), plus Debian packaging. Copied
from http://sysmic.org/par/debian/.
See also his original announcement.
I am working on an application that allows users to input Japanese language characters. I am trying to come up with a way to determine whether the user's input is a Japanese kana (hiragana, katakana, or kanji).
There are certain fields in the application where entering Latin text would be inappropriate and I need a way to limit certain fields to kanji-only, or katakana-only, etc.
The project uses UTF-8 encoding. I don't expect to accept JIS or Shift-JIS input.
Ideas?
It sounds like you basically need to just check whether each Unicode character is within a particular range. The Unicode code charts should be a good starting point.
If you're using .NET, my MiscUtil library has some Unicode range support - it's primitive, but it should do the job. I don't have the source to hand right now, but will update this post with an example later if it would be helpful.
Not sure of a perfect answer, but there is a Unicode range for katakana and hiragana listed on Wikipedia. (Which I would expect are also available from unicode.org as well.)
Hiragana: Unicode: 3040-309F
Katakana: Unicode: 30A0–30FF
Checking those ranges against the input should work as a validation for hiragana or katakana for Unicode in a language-agnostic manner.
For kanji, I would expect it to be a little more complicated, as I
expect that the Chinese characters used in Chinese and Japanese are both included in the same range, but then again, I may be wrong here. (I can't expect that Simplified Chinese and Traditional Chinese to be included in the same range...)
oh oh! I had this one once... I had a regex with the hiragana, then katakana and then the kanji. I forget the exact codes, I'll go have a look.
regex is great because you double the problems. And I did it in PHP, my choice for extra strong auto problem generation
--edit--
$pattern = '/[^\wぁ-ゔァ-ヺー\x{4E00}-\x{9FAF}_\-]+/u';
I found this here, but it's not great... I'll keep looking
--edit--
I looked through my portable hard drive.... I thought I had kept that particular snippet from the last company... sorry.