I have input and output text files which can contain non-ascii characters. Sometimes I need to escape them and sometimes I need to write the non-ascii characters. Basically if I get “Bürgerhaus” I need to output “Bu00FCrgerhaus”. If I get “Bu00FCrgerhaus” I need to output “Bürgerhaus”.
One direction goes fine:
>>> s1 = "Bu00FCrgerhaus"
>>> print(s1)
Bürgerhaus
however in the other direction I do not get the expected result (‘Bu00FCrgerhaus’):
>>> s2 = "Bürgerhaus"
>>> s2_trans = s2.encode('utf8').decode('unicode_escape')
>>> print(s2_trans)
Bürgerhaus
I read that unicode-escape needs latin-1, I tried to encode it to it, but this did not product a result either. What am I doing wrong?
(PS: Thank you Matthias for reminding me that the conversion in the first example was not necessary.)
Advertisement
Answer
You could do something like this:
charList=[]
s1 = "Bürgerhaus"
for i in [ord(x) for x in s1]:
# Keep ascii characters, unicode characters 'encoded' as their ordinal in hex
if i < 128: # not sure if that is right or can be made easier!
charList.append(chr(i))
else:
charList.append('\u%04x' % i )
res = ''.join(charList)
print(f"Mixed up sting: {res}")
for myStr in (res, s1):
if '\u' in myStr:
print(myStr.encode().decode('unicode-escape'))
else:
print(myStr)
Out:
Mixed up sting: Bu00fcrgerhaus
Bürgerhaus
Bürgerhaus
Explanation:
We are going to covert each character to it’s corresponding Unicode code point.
print([(c, ord(c)) for c in s1])
[('B', 66), ('ü', 252), ('r', 114), ('g', 103), ('e', 101), ('r', 114), ('h', 104), ('a', 97), ('u', 117), ('s', 115)]
Regular ASCII characters decimal values are < 128, bigger values, like Eur-Sign, german Umlauts … got values >= 128 (detailed table here).
Now, we are going to ‘encoded’ all characters >= 128 with their corresponding unicode representation.