I felt that when he said *83h400+93)*38hpfhi0
The thing that I find the most funny about this post, is the fact that you call this Italian
Blud could’ve chosen Runic, Egyptian, Ancient Romanian used by Vlad the Impaler, Mesapotamian or even Harappan Indic. But Italian is it.
Blud I’m gonna be fr no cap rn but wtf does blud mean I’ve been meaning to ask for months and I still don’t get it
It’s a Jamaican slang for ‘friend’ or ‘brother’.
Thanks blud.
how am i supposed to know how italians speak. i’ve never seen one
It’s a me, Mario!
They’re not real, but they can hurt you.
like reverse vampires ?
Ne sei sicuro?
That’s right! None of us knows how Italians can speak in the dark 🤌
From my experience, they speak mostly with their hands
🫰🤙🫵👌✊🫳🫸🤲🤌
Prego
Ditto
Typical 'muricans being unable to comprehend anything besides English.
/s i don't mean to be racist
yes i was a r/2we4u user, how’d you know?
Well, it certainly doesn’t overflow on 32 bit systems
Never go full APL
Ah, I see you’re using FartGPT instead of ChatGPT
French pronunciation intensifies
Cat, I farted.
is that the new model ?
Wow, an alien ion drive formula! Try to get warp drive out of it too!
Which language uses these signs? It truly looks like some kind of alien language
Unown
Glagolitic script. Oldest known Slavic alphabet according to Wikipedia.
They should revive this script. I like it more than Cyrillic.
I found it! its the Glagolitic script used in the 9th century before Cyrillic took over:
ⰀⰁⰂⰃⰄⰅⰆⰇⰈⰉⰊⰋⰌⰍⰎⰏⰐⰑⰒⰓⰔⰕⰖⰗⰘⰙⰚⰛⰜⰝⰞⰟⰠⰡⰢⰣⰤⰥⰦⰧⰨⰩⰪⰫⰬⰭⰮⰰⰱⰲⰳⰴⰵⰶⰷⰸⰹⰺⰻⰼⰽⰾⰿⱀⱁⱂⱃⱄⱅⱆⱇⱈⱉⱊⱋⱌⱍⱎⱏⱐⱑⱒⱓⱔⱕⱖⱗⱘⱙⱚⱛⱜⱝⱞ
I think it’s the Ge’ez script used in Ethiopian.
Doesn’t look like it to me:
ልዩ ጊዜ ነበር። አሁን የሚሆነውን ለማስተዋል የኢንተርኔት አውራጃ ማረጋገጥ ነበር።
Yeah, you are right.
I would like to know too! Never saw that writing system before.
APL?
No that looks like
⌶⌷⌸⌹⌺⌻⌼⌽⌾⌿⍀⍁⍂⍃⍄⍅⍆⍇⍈⍉⍊⍋⍌⍍⍎⍏⍐⍑⍒⍓⍔⍕⍖⍗⍘⍙⍚⍛⍜⍝⍞⍟⍠⍡⍢⍣⍤⍥⍦⍧⍨⍩⍪⍫⍬⍭⍮⍯⍰⍱⍲⍳⍴⍵⍶⍷⍸⍹⍺
Damn, wild Glagolitic script found. I didn’t even realise it was in the Unicode standard.
That’s not italian that’s obviously Unown
It looks so badass, I could have used that script now because im Ukrainian but instead I have cyrillic script which is so boring
rebel against Russian imperialism, return to glagolitic
It’s not russian, If my bulgarian friend is right then it was created by a bulgarian guy
There is no single person responsible for Cyrillic script. It is mostly believed to be created by mixing and changing Greek and Glagolic scripts by the scholars of Preslav Literary School, which was indeed in Bulgaria. After a while, Peter the Great changed it a lot. And then Stalin stomped out almost all the deviations in the usage of the script.
The last part is mostly why it is considered Russian. A lot of languages suffered because of Moscow just forcing them to use the version of Cyrillic that Russians were using.
Cyrillic is literally greek+glagolitic and it was partly a diplomatic creation of the Eastern Roman Empire(aka Byzantine Empire), in order to bring the slavs culturally closer to them.
Russians have nothing to do with it, other than them claiming they are the continuation of Eastern Roman Empire, something which is kinda laughable but whatever dont let your dreams be dreams.
We are so cooked
Looks like UiUa: uiua.org
This might be happening because of the ‘elegant’ (incredibly hacky) way openai encodes multiple languages into their models. Instead of using all character sets, they use a modulo operator on each character, to make all Unicode characters represented by a small range of values. On the back end, it somehow detects which language is being spoken, and uses that character set for the response. Seeing as the last line seems to be the same mathematical expression as what you asked, my guess is that your equation just happened to perfectly match some sentence that would make sense in the weird language.
I suppose it’s conceivable that there’s a bug in converting between different representations of Unicode, but I’m not buying and of this “detected which language is being spoken” nonsense or the use of character sets. It would just use Unicode.
The modulo idea makes absolutely no sense, as LLMs use tokens, not characters, and there’s soooooo many tokens. It would make no sense to make those tokens ambiguous.
I completely agree that it’s a stupid way of doing things, but it is how openai reduced the vocab size of gpt-2 & gpt-3. As far as I know–I have only read the comments in the source code– the conversion is done as a preprocessing step. Here’s the code to gpt-2: https://github.com/openai/gpt-2/blob/master/src/encoder.py I did apparently make a mistake, as the vocab reduction is done through a lut instead of a simple mod.
Do you have a source for that? Seems like an internal detail a corpo wouldn’t publish
Can’t find the exact source–I’m on mobile right now–but the code for the gpt-2 encoder uses a utf-8 to unicode look up table to shrink the vocab size. https://github.com/openai/gpt-2/blob/master/src/encoder.py
Seriously? Python for massive amounts of data? It’s a nice scripting language, but it’s excruciatingly slow
There are bindings in java and c++, but python is the industry standard for AI. The libraries for machine learning are actually written in c++, but use python language bindings. Python doesn’t tend to slow things down since machine learning is gpu-bound anyway. There are also library specific programming languages which urges the user to make pythonic code that can be compiled into c++.