Found it
Nope. We have no Japanese sentences on the website created by LLMâs. We do however have some (temporary) English translations of those sentences that were translated by GPT. However even those we have less and less of recently as we have been focusing on eliminating them. They were just placeholders until we were able to manually check the English.
No native content that we use will ever be created by ChatGPT.
Itâs come a very long way. A year or two ago I tried using ChatGPT to correct my journal entries in Japanese and the suggestions it gave me were always awful, but these days it seems to do exactly what I ask. I guess I canât really judge it since Iâm not a native speaker but it seems pretty natural to me. Of course it does still make mistakes, but I feel like so long as youâre not a beginner itâs within a range that you will probably notice.
Not all textbooks are created equal. Some textbooks are written by native speakers who are licensed teachers and have passed the same test that youâre studying for (as part of getting their license) and were created in order to be used for curriculum at the school where they teach, and then on the other hand you have stuff like Georgeâs books. And books with varying levels of quality in-between.
Thereâs some amazing textbooks out there, but you need to be careful and research into books before you buy them.
@marziotta - I use it all of the time, and have found it really helpful.
I read NHK Easy news each morning, and will toss the articles into Chatgpt so I can see what it thinks the translation is. It has taken a while to train it to do as literal translations as possible in a format I like, but I think it has been very worth it. After doing this for several months, I am getting good enough at reading the NHK Easy news that I am not sure how much longer I will need to have Chatgpt translate them unless they are strange or on a new topic.
I also love tossing out specific sentences and having chatgpt break it down, especially the grammar principles. It has seemed to always get things right with me, when I cross reference it with my other learning materials.
There are also times when I am reading things and I donât recognize the kanji, but it has furigana. I can ask chatgp what it is. It also works vice versa, when sometimes there is something in a dialect or is super colloquial chatgp can tell me what it means, and several times none of my other tools could.
Mostly AI is helpful, especially when translating phrases that are hard to prase. But when I do a deep dive analysis on something outside the training guardrails, hallucinations start to reign. These images arenât LLM, but itâs the same sort of idea.
That sounds like you actually messed up on the input of that question. words of kanji+katakana? that makes it sound like you donât know what kanji means in this context, so it just assumed you wanted words and you thought that Romaji is âkatakanaâ itâs a case of âgarbage in, garbage outâ. Though it also appears to be an older model, or worse, a google model.
In my experience, there are very very very very few kanji+katakana okurigana, if any. so it had no choice but to try to compromise because these systems are too dumb to be disagreeable or tell you something doesnât exist.
If you ask questions with clear boundries that arenât confusing, it does pretty well.
But as said previously, use it as a suppliment, not a replacement.
Itâs been a while since Iâve had somebody phrase things in such an insecure way to me. Donât be so on edge, weâre all here to learn Japanese and have a good time doing it.
Indeed, there arenât many Kanji+katakana compound words, thatâs why I was hoping the AI could list some. I donât see how my question didnât have clear boundaries though, and itâs certainly not true that it should have told me kanji+katakana words donât exist. ăăŒçăă·ăŁăăłçăăĄă€ăć«è¶ăă”ă©éă ă”ă©éć°çăèźă«ă©ăăłăăŒæ©, to give a few examples. All these words are on Bunpro too, have a look.
thanks for the clarification, I really appreciate the amount of work you are putting into this app.
Insecure? Could you please investigate what that word means, instead of trying to insult someone? Those words there that you are using are paired words or slang. Itâs like saying âice creamâ is a single word. Japanese also love to combine words to make them shorter, making it appear to be a single word when itâs a mashup of two words. While yeah they often are together, itâs not a true word by itself.
As for the AI prompt âgarbage in, garbage outâ is a real thing. Iâm not sure of your age, but remember old Google? Knowing How to look something up was more important than the question itself. AI is just an advanced internet search at the end of the day. Iâm sorry you misinterpreted me as trying to insult you, but at least try to learn what youâre talking about before saying such things.
this is a fundamental misinterpretation of what an LLM is, itâs a text generator. The answer is simply a generated sequence of words based on what the most probable word is after the previous one. They are trained on a very simple task: Predict the next token (word or subword) given a squence of previous tokens.
Given: The cat sat on the ___
Learn to output: mat
this is called âcasual language modelingâ.
Doing this on a massive corpora of text (books, websites, code, more) exposes the LLM to a lot of linguistic variety. So much that it might give us the illusion that it is reasoning. But in the end itâs just a text generator. It will therefore hallucinate on the simplest topics with no guarantee whatsoever that what it is outputting makes the slightest sense. And trusting what it says comes with other dangers as per this study made by Microsoft.
Further reads:
Hereâs a very simple example on how the chosen words (the token sequence) determine the response without any regard to what it actually outputted.
@BreadmanNin @Rukifellth Letâs both not get off track here please. I doubt anyone meant anything negatively .
About the con cafe. The llm have a tendency to never say they do not know something, so they make up stuff. From an European point of view, I see a bit of the extreme confidence I often see in some Americans, that are always right. I do not think all Americans do that, but this is a trait that is surely promoted in many cases.
They should learn the quote âAll I know is I know nothing.â. I donât know who said that.
this is because âI donât knowâ is rarely present in the learning set. Wikipedia, online discussion platforms, scientific papers and so on, they have an explanatory tone. People usually do not reply âSorry I have no clueâ to a question on Reddit or StackExchange for example (which is part of the training set for pretty much any LLM). If they donât know, they simply donât reply.
Hence the overwhelming majority of the training data of an LLM is that to a question, a confident answer follows. Thatâs what usually happens on the internet. Because the LLM does not truly think but only concatenates tokens by probability to follow the previous tokens (starting with your prompt), you will almost always get an answer that is explanatory and confident in tone. However, the confidence in the reply is not representative of the usefulness of the given information.
Exactly this! Iâm a 3D artist and the first thing I had to learn was, ALWAYS use reference images. That is literally an industry standard workflow. The reference either comes from photos other people have taken, from paintings other people have painted, or from 3D art other people have sculpted/modeled.
People that create moves, music and other forms of media didnât create those things out of thin air, they were all influenced by things that theyâve heard and watched in their lives.
At least you are able to tell an explanation was bad, or youâve read reviews saying that textbook X provides better explanations than textbook Y. With AI, you are alone talking to a bot. If you want to confirm something, you fall back to a textbook or google.
I find Claude pretty invaluable for asking questions like âCan you unpack the grammar in this Japanese sentence? [sentence]. I donât understand the role that [word] is playing.â It usually gives an excellent explanation in response to that kind of thing.
Donât get into a long conversation with it though, just start a new chat, ask the question, get the answer, and move on. The longer the conversation gets, the more likely it is to start making up random stuff, in my experience.
Also, for complicated or obscure questions (especially about Japanese culture), it helps if you say âanswer first in Japanese and then in English.â The Japanese explanation will usually be better than the one it would give in English because itâs predicting what a Japanese speaker would say instead of predicting what an English speaker would say. So if it writes it in Japanese first and then translates it can give higher quality answers.
Iâm very torn on AI use. I do sometimes use it both for language learning and my programming job, but in both cases I find that you should absolutely never blindly trust it. So for instance I find that generative AI can be pretty ok to generate code boilerplate if I already know what I expect and I can easily check the output for correctness.
For learning a language however thatâs trickier, because by definition you probably donât have the skill level required to vet the answers and identify incorrect or inaccurate statements. I still use it from time to time when I really struggle to parse a sentence in something I read, it usually does a good job at pure translation and a somewhat worse job at explaining the grammar.
But I would absolutely never trust it with a prompt like âcan you explain Japanese ă-adjective conjugations to meâ or something similarly generic. I know some people use AI like they would a textbook and that seems like a terrible idea to me. Even generating text in Japanese for reading practice for instance seems so odd to me, you have access to a virtually infinite amount of Japanese for basically any level online, why risk having AI generate something broken?