What he's reading there is just the following sentence in the book, separated with a comma...
> TextBlob does a pretty good job at the translation: "C'est une vérité universellement reconnue, qu'un homme célibataire en possession d'une bonne fortune doit avoir besoin d'une femme!".
It can be argued that TextBlob's translation is far more exact, in fact, than the 1932 French translation of the book by V. Leconte and Ch. Pressoir:
"C'est une vérité universelle qu'un célibataire pourvu d'une belle fortune doit avoir envie de se marier, et, si peu que l'on sache de son sentiment à cet egard, lorsqu'il arrive dans une nouvelle résidence, cette idée est si bien fixée dans l'esprit de ses voisins qu'ils le considèrent sur-le-champ comme la propriété légitime de l'une ou l'autre de leurs filles."
In this case, the translation informed by ML does a better job than the human translator who is unnecessarily putting words in the original author's mouth for 'clarity'.
https://github.com/microsoft/ML-For-Beginners/blob/main/6-NL...
> [x] Try some more sentences. Which is better, ML or human translation? In which cases?
Any English second language speaker just knows this: There are basically zero cases where ML translation is in any way better or more accurate. That includes DeepL and ChatGPT w/GPT-4 proper. It's just tap water at restaurants. Which is a fantastic choice to pair with actual drinks. It's weird this sentence appears here at all.
> [...]and why is TextBlob so good at translation? Well, behind the scenes, it's using Google translate, a sophisticated AI able to parse millions of[...]
And this part. Maybe it's just me, but I think it might be showing that the author first tried to hand-roll translation, as seemingly needlessly lengthily elaborated up to this part. It could be that they then either faced technical challenges or failed validation by certified Frenchmen, and had to rewrite the section as a guide to use Google Translate API.
There seems to be an endemic misconception, mostly seen among but not limited to American people, that American English is a perfect language that is also completely disconnected from internal thoughts and intents and monologues, that simplifies language translation problem to a simple matter of "convert[ing] the formal grammar rules for one language, such as English, into a non-language dependent structure, and then translate it by converting back to another language" as the author claims, while in reality even the possibility of "non-language dependent structure" is still under debate. This kinds of attitude always existed, but now it's borderline beyond annoying.
Too big, too slow, too much resources etc. And it’s not even clear to me (mind, who is ignorant) that the LLM is some generic model suitable for all AI like tasks.
Making a big splash right now to be sure, but seems to me there’s still room for the core concepts folks have been working on for a long time.
Edit: At least that’s what I’m doing. I could be wrong though.
1. there is no reason to pay the api costs for an LLM to ingest data for you and do something with it when basically all it will be doing is writing the python codes for you that you will eventually be using
2. the LLM doesn't represent some sort of conceptual understanding of whatever you are trying to do to solve your ML problem, so you can't rely on it to be clever and answer questions or brain storm new ideas
3. even if you have a reason to use an LLM in some data processing pipeline it will only be one stop on the information super highway you are trying to create. you probably are going to use it to do something, but you probably also are going to be doing other things (e.g., image segmentation, time series analysis, etc.).
LLMs are great. but they are really just like, one more tool to have, they aren't the only tool.
Depends on your goals though, using llm just seems like using any other api to me.
Two questions related to learning in general:
1) I feel I have a good undergrad level grasp of ML but not at a grad level grasp. The math is a bit overwhelming. I am not a fan of conferences like Neurips. Any one else try to conquer this challenge and have a story to share?
2) A bit off topic but the XR course they have linked on the page is also cute. It lists both Unity and MRTK. I have a Hololens2 and am curious about spatial mapping and awareness (just to learn at the moment). Any suggestions on what is a good stack for this area? I have very little 3D graphics background .. unity seems a bit too high for serious work and an engine seems too level :(
Currently working through some chapters of “AI. A modern approach “ book.