Hacker Newsnew | past | comments | ask | show | jobs | submit | numpad0's commentslogin

Nothing will happen to it. Someone will eventually complete the loop between slop generators to human reward systems to turn the first order derivative of content supply back up to pre-COVID levels.

If you don't speak the local language anyway, you can't decode pronounced spoken local language names anyway. Your speech sub-systems can't lock and sync to the audio track containing languages you don't speak. Let alone transliterate or pronounce.

Multilingual doesn't mean language agnostic. We humans are always monolingual, just multi-language hot-swappable if trained. It's more like you can make;make install docker, after which you can attach/detach into/out of alternate environments while on terminal to do things or take in/out notes.

People sometimes picture multilingualism as owning a single joined-together super-language in the brain. That usually doesn't happen. Attempting this especially at young age could lead a person into a "semi-lingual" or "double-limited" state where they are not so fluent or intelligent in any particular languages.

And so, trying to make an omnilingual TTS for criticizing someone not devoting significant resources at it, don't make much sense.


> If you don't speak the local language anyway, you can't decode pronounced spoken local language names anyway

This is plainly not true.

> Multilingual doesn't mean language agnostic. We humans are always monolingual, just multi-language hot-swappable if trained

This and the analogy make no sense to me. Mind you I am trilingual.

I also did not imply that the model itself needs to be multilingual. I implied that the software that uses the model to generate speech must be multilingual and support language change detection and switching mid-sentence.


> it must be multilingual and dynamically switch between languages pretty much per word

Not abundantly obviously a satire and so interjecting: humans, including professional "simultaneous" interpreters, can't do this. This is not how languages work.


You can speak one language, switch to another language for one word, and continue speaking in the previous language.

But that's my point. You'll stop, switch, speak, stop, switch, resume. You're not going to be "I was in 東京 yesterday" as a single continuous sentence. It'll have to be broken up to three separate sentences spoken back to back, even for humans.

>"I was in 東京 yesterday"

I think it's the wrong example, because this is actually very common if you're a Chinese speaker.

Actually, people tend to say the name of the cities in their own countries in their native language.

> I went to Nantes [0], to eat some kouign-amann [1].

As a French, both [0] and [1] will be spoken the French way on the fly in the sentence, while the other words are in English. Switching happens without any pause whatsoever (because there is really only one single way to pronounce those names in my mind, no thinking required).

Note that with Speech Recognition, it is fairly common to have models understanding language switches within a sentence like with Parakeet.


Okay, it's getting clear that I'm in the wrong here with my insistence that languages don't mix and foreign words can't be inserted mid-sentence, yet that is my experience as well as behaviors of people sharing the language, incidentally including GP who suggested that I can always do the switching dance - people can if wanted, but normally don't. It's considered a show-off if the inserted word could be understood at all.

Perhaps I have to admit that my particular primary language is officially a human equivalent of an esoteric language; the myth that it's a complex language is increasingly becoming obsolete(for good!), but maybe it still qualify as being esoteric one that are not insignificantly more incompatible with others.


I think this is totally wrong. When you have both parties speaking multiple languages this happens all the time. You see this more with English being the loaner more often than it is the borrower, due to the reach that the language has. Listen to an Indian or Filipino speak for a while, it's interspersed with English words ALL the time. It happens less in English as there is not the universal knowledge base of one specific other language, but it does happen sometimes when searching for a certain, je ne sais pas.

Not really, most multilinguals switch between languages so seamlessly that you wouldn't even notice it! It even has given birth to new "languages", take for example Hinglish!!

Surströmming, the Swedish can of fermented fish, is strongly recommended to be punctured while submerged in tap water. It is not pasteurized and is actively fermenting in storage, and the content will spray around if opened under atmospheric conditions.

When transported on cargo flights, they are double packed as cans in a barrel in a crate, and considered UN classified "miscellaneous dangerous goods" with identification number UN3334 "Aviation regulated liquid, n.o.s." with accompanying scary(albeit monochromatic) warning stickers, if at all accepted. When transported on ocean going vessels, they are often required to be in its own shipping container, again double packaged and correctly labeled.


reminds me ... I opened a can of Illy coffee today. It has a big red warning saying similar under pressure warning. I went...."pfhhh, how much pressure could be in this little can?" and it went ba-boom, coffee grounds all over me and my kitchen. I will heed the Sustromming warning if I ever come across a tin

You forgot to mention that it is actually quite delicous.

I hope you didn't find that first paragraph out the hard way today!

Haha, no, I just came across these trivia... it scares me more than it intrigues me :p

It is also mostly eaten outdoors :)

Is it even good?

its an acquired taste.

good when you pair it with the right stuff - dill, capers, etc and on kinda neutral crackers


If I were to guess, probably because Musk achieved self-fulfilling prophecy of hate and discriminatory handling against him, and now any obviously him related content gets massive, organic, figurative, score penalties.

Tesla and SpaceX posts used to routinely hit the top spots and accumulate thousands of comments here, now they hardly stay an hour on the first page. Someone on the Internet's first headphone amp is now considered more important to people here than the world's largest rocket flying, if that comes with Musk attached.

Obviously as anybody knows, that's how `hate` actually works: silent exclusion, not posturing. But that was what they advocated for years, so, here's my slow claps...


Isn't age guesstimation by appearance, even with advanced machine learning techniques, even if attempted by real person with honest effort, just total snake oil? This ongoing age verification push with weird emphasis on generating name-face pairs is beyond fishy.

> ... while they’re 100% against using AI to generate art or game design, when you ask them about using AI tools to build software or websites ...

And this is not complicated at all. It's the quality of output.

Users appreciate vibecoded apps but developers are universally unfazed about vibecoded pull requests. Lots of same devs use AI for "menial" tasks and business emails. And this is NOT a double standard: people are clearly ok when generative AI outputs may exist but aren't exposed to unsuspecting human eyes, and it's not ok if they are exposed to human eyes, because the data AIs generate haven't exceeded the low-side threshold of some sort. Maybe SAN values.

(also: IIUC, cults and ponzi scheme recruitment are endemic in tabletop game communities. so board game producers distancing from anything hot in those circles, even if it were slightly irrational to do so, also makes sense.)


SiLabs RS9116, Nordic nRF7002, TI CC3235, Dialog DA16200, ...

They exist, they just don't ship to individual nobodies without NDAs and months of talks and supply commitments, therefore we might as well consider them nonexistent. And that's sinking the whole world into oblivion, and no one(sane) is doing anything about it.

Chinese equivalents of Western products ship same day to mail drops at two digits below USD denominated global market manufacturing costs. That's their secret sauce. Or tried and true East Asian miracles strategy, of exploiting material independence to vacuum in foreign currency that are short of cost but are just trustless bundles of paper anyway. Economic competition is not possible when that's possible for them and nobody else - NOBODY else, not like "only for Asian backwater whatever failed state", but China, for now, specifically.

I'm actually from Japan and, with goggles of a maker on, DJI behavior feels reminiscent of Sony until it sank. The tech is top notch, and prices make progressively less sense towards higher ends. That kills competitions by denying sales of high end products(IMO. I'm not a sane person).


I've found another[1] on a blog post[2], captioned as follows:

  Frontispiece 1. Art drawn by chimpanzee Ai using sharpies(Saito, 2008)[p.19]
  Frontispiece 2. Art styles of 4 adult chimpanzees(Saito, 2008). Guess which one was by Ai[p.20]
Not sure what the background of the author is, but this essay/lecture note discusses ego or literal self-awareness of apes contrasted against human children, using quotes from books. Apparently apes don't exhibit explosive growth of vocabulary, show use of syntax etc etc, and are therefore not able to acquire language. The post later also argues their ego may be on the edge of formulating but must be weak/incomplete.

There's also magazine excerpt[3] on a page on relevant Kyoto University research center comparing an inpainting task done by a chimpanzee and a human child of 3 years old, showing that chimpanzees can only recognize and trace existing patterns, whereas kids go and complete the face with eyes, nose and mouth.

  1: https://kyoikugenri2019.up.seesaa.net/image/2017-10-132018.11.52.jpg
  2: https://kyoikugenri2019.seesaa.net/article/471281414.html
  3: https://www.wrc.kyoto-u.ac.jp/ja/publications/AyaSaito/kagaku084.html

I don't get the text obsession beyond LLMs being immensely useful that you might as well use LLM for <insert tasks here>. I believe that some things live in text, some in variable size n-dimensional array, or in fixed set of parameters, and so on - I mean, our brains don't run on text alone.

But our brains do map high-dimensionality input to dimensions low enough to be describable with text.

You can represent a dog as a specific multi-dimensional array (raster image), but the word dog represents many kinds of images.


Yeah, so, that's a lossy/ambiguous process. That represent_in_text(raster_image) -> "dog" don't contain a meaningful amount of the original data. The idea of LLM aided CAD sounds to me like, a sufficiently long hash should contain data it represents. That doesn't make a lot of sense to me.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: