June 20th, 2024

How babies and young children learn to understand language

Babies and young children learn language from birth, showing preference for caregivers' speech rhythm. By age one, they start speaking, forming sentences by age four. Infants use statistical learning to identify word boundaries in speech, sparking ongoing linguistic research.

Read original articleLink Icon
How babies and young children learn to understand language

Babies and young children learn to understand language through a complex process that starts even before birth. Infants as young as three days old show a preference for the rhythm of their caregivers' language, indicating some learning occurs in the womb. By around one year of age, children start saying their first words, eventually stringing them together to form simple sentences by age four. Researchers have discovered that infants use statistical learning to identify word boundaries within a continuous stream of speech. This was demonstrated in a study where eight-month-old infants could differentiate between words and non-words based on transitional probabilities between syllables. The ability of babies to acquire language without formal teaching has been a subject of interest for linguists, leading to ongoing research on how infants find words in speech and learn their meanings. Steven Mithen, a professor of early prehistory, explores these concepts in his book "The Language Puzzle: Piecing Together the Six-Million-Year Story of How Words Evolved."

Related

Smalltalk syntax in 7 minutes [video]

Smalltalk syntax in 7 minutes [video]

The YouTube video explains Smalltalk syntax, emphasizing readability and message-based object interaction. It covers keywords, arrays, closures, and method execution in Pharo Smalltalk, providing a practical example and additional learning resources.

Optimizing AI Inference at Character.ai

Optimizing AI Inference at Character.ai

Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.

Generating audio for video

Generating audio for video

Google DeepMind introduces V2A technology for video soundtracks, enhancing silent videos with synchronized audio. The system allows users to guide sound creation, aligning audio closely with visuals for realistic outputs. Ongoing research addresses challenges like maintaining audio quality and improving lip synchronization. DeepMind prioritizes responsible AI development, incorporating diverse perspectives and planning safety assessments before wider public access.

GitHub – Karpathy/LLM101n: LLM101n: Let's Build a Storyteller

GitHub – Karpathy/LLM101n: LLM101n: Let's Build a Storyteller

The GitHub repository "LLM101n: Let's build a Storyteller" offers a course on creating a Storyteller AI Large Language Model using Python, C, and CUDA. It caters to beginners, covering language modeling, deployment, programming, data types, deep learning, and neural nets. Additional chapters and appendices are available for further exploration.

Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]

Francois Chollet – LLMs won't lead to AGI – $1M Prize to find solution [video]

The video discusses limitations of large language models in AI, emphasizing genuine understanding and problem-solving skills. A prize incentivizes AI systems showcasing these abilities. Adaptability and knowledge acquisition are highlighted as crucial for true intelligence.

Link Icon 16 comments
By @TacticalCoder - 4 months
Words beginning and endings are learned too because people raising young kids are spending a great deal of time decomposing and telling them simple words. Repeating "cat", "cat", "cat" with pauses.

Statistical learning --and there are studies about this-- is also obvious when multi-lingual kids make up words that do not exists.

They'll use words from one of the language they know to come up with words (or words beginning / ending) in another language. These words, statistically, could make sense. And they'll pronounce them "properly". Yet they don't exist.

So it's not just the words: it's the pronunciation too.

As the father of a fully bilingual kid (french / english) that was fascinating to watch.

By @alexey-salmin - 4 months
> Think about listening to a language unknown to you, one with different words, grammar and prosody. You will be at an utter loss to identify its words, let alone their meaning.

My experience learning French basically. I'd say understanding where one word ends and another starts was much easier for English and German. On paper I was able to grasp the rough meaning very quickly thanks to vocabulary shared with English and Latin, but listening took a year: I was facing a solid wall of sound, no cracks.

By @nickpsecurity - 4 months
It's good that they mentioned babies hearing in the womb. I've known many mothers that read to or play music for their babies. They say they feel them respond to some things, too, where they seem to sense their surroundings. I'm just reporting what they told me since I haven't studied the literature on this stuff.

One thing I didn't like was the paragraph on how they differentiate words with no formal training. I feel that gives a false impression. Parents usually teach their children a lot about language. They give them visual cues, speak at varying rates, change their own tone in some situations, and so on.

The babies are soaking up the world on their own using one set of mechanisms. They also often receive highly-supervised training from a trusted source. Later, they get formal training on top of that. Even when not training, much of the content they see and hear is presented in a structured way that helps connect ideas. For instance, listening to the radio or TV with their parents would let them hear a lot of structured speech.

Babies are highly trained. They might also do the statistical learning. They're a mix of the two.

By @IG_Semmelweiss - 4 months
The article starts getting good in the last 2 paragraphs, explaining the actual science of observed transitional probabilities, then it suddenly stops.

Its almost as if the writer ran out of coffee, or his scientific mind went on strike. What was that ?

By @HumblyTossed - 4 months
Tip for parents - teach your babies Sign Language. They can communicate with you much earlier than they can learn to talk clearly. So much easier when a baby can tell you they're all done or want more than to have to figure that out without communication.
By @keiferski - 4 months
Does anyone here have direct experience raising multilingual kids? Specifically the situation where one partner speaks English natively and the other speaks two languages natively? For sake of example, let's say they're German and Spanish. (And you want the child to learn all three natively.)

My plan is to divide the languages by person and place:

- always talk to Parent 1 in English, no matter the location

- talk to Parent 2 in Spanish at home and German when outside the home, adhering strictly to this location-based method. The extended family mostly speaks Spanish, which makes the "home" association stronger.

This seems easier to me than dividing the languages by time (only speak Spanish on M/W/F, German on Tu/Th/Sat) or other divisions, but I'm open to any suggestions.

By @isolli - 4 months
I learnt German at the age of 25.

I distinctly remember the first time I was exposed to it (before learning it), it sounded like water gently flowing down a creek. Then I learnt the basics, and my brain started to catch on to patterns.

However, I had to go through the written form to learn properly. I found it hard to parse and remember words when I was only hearing them. Unlike young children, obviously.

By @foobarqux - 4 months
Chomsky describes why the word identification method described in the article can’t work in English (though a more elaborate method is one of the few successes he knows in applying statistical methods to language): https://youtu.be/92GOS1VIGxY

The arguments against universal grammar are no good either: For example even though children may hear lots of examples it isn’t nearly enough to derive a hierarchical grammar. It also doesn’t explain why language is hierarchical (just like it doesn’t explain why we can’t speak and hear like a modem)

By @fhe - 4 months
> When we write, we leave spaces between words. Readingwordswithoutsuchspacesisdifficult.

immediately losing credibility because quite a number of languages, such as Chinese and Japanese (and I think Korean too), are written without space between words or characters. In fact, until quite recently (last 100 years or so), written Chinese had no punctuations.

By @Anotheroneagain - 4 months
I believe it doesn't start with an explicit attempt to identify words, but from the attempt to compress the sound into fewest meaningful dimensions, and the words naturally result from it. It's a low level process that creates these words, as well as the apparent boundaries.
By @ErigmolCt - 4 months
I've always been curious whether learning a second language from childhood has a positive or negative impact on a child's development.
By @nop_slide - 4 months
Related, but I'm currently reading "Thirty Million Words: Building a Child's Brain", it's fascinating!
By @moi2388 - 4 months
Makes sense, perhaps this is also why nursery rhymes chop up words into syllables following the melody.
By @smeej - 4 months
> Just think how you pronounce the syllable ham when referring to a piece of meat and when talking about a fury animal—a long ham and a short ham-ster.

Hang on. Y'all pronounce these differently? I've lived in four U.S. regions and have a pretty generic middle-American accent and I'm having trouble even thinking what the distinction might be.

By @pcloadletter_ - 4 months
I gave up trying to read this after closing the 3rd popup
By @YeGoblynQueenne - 4 months
What a silly question to ask. They learn by training on the entire web, hiding one word at a time.

/s