Imagine if your GPS could not only give you directions but also predict where you’d want to go next. Sounds a bit like science fiction, right? Well, the world of artificial intelligence is full of surprises! Just like how we sometimes finish each other’s sentences, AI is getting pretty good at understanding and predicting our language. And it’s not just about talking to your virtual assistant; it’s about making machines think a bit more like us. Let’s dive into how this all works and why it’s exciting for both science nerds and everyday folks alike.
With generative artificial intelligence (GenAI) revolutionising our social interactions, large language models (LLMs) have taken centre stage. These models, trained using deep-learning algorithms, have shown remarkable abilities to process and generate human language. A recent study by the Hong Kong Polytechnic University (PolyU) discovered that LLMs behave more like the human brain when trained in ways that mimic human language processing, offering valuable insights into brain function and AI development.
Typically, LLMs rely on a single method of pre-training – predicting the next word in a sentence. This straightforward technique, when paired with vast amounts of data, has led to the success of models like ChatGPT. While effective, this approach doesn’t fully capture how humans understand language, which involves more than just word prediction. We also consider the overall context and meaning of sentences.
Professor Li Ping and his team at PolyU explored the next sentence prediction (NSP) task to improve LLMs. NSP helps models understand if pairs of sentences are coherent, mimicking how our brains process connected ideas. The team trained two models: one with NSP and one without, both capable of word prediction. Using fMRI data, they compared how each model’s performance aligned with human brain activity during reading.
Their findings showed that the NSP-enhanced model better matched brain activity patterns, particularly in areas involved in understanding connected sentences. This model also provided insights into how different parts of the brain contribute to comprehending longer texts. Interestingly, it even predicted reading speed, suggesting that simulating discourse comprehension helps AI models better understand human language.
Prof Li Ping emphasised that simply scaling up data and model size has its limits. He advocated for more efficient training methods, like incorporating diverse tasks such as NSP, to make LLMs more human-like and intelligent. This approach not only enhances AI performance but also bridges the gap between AI and brain research, fostering collaboration and new discoveries in both fields.
So next time you’re chatting with your virtual assistant, remember, it’s not just learning your words; it’s trying to get inside your head – quite literally! Who knew our brain’s secrets could help make AI so much smarter? Maybe one day, they’ll even start finishing our jokes too!
ammar.shakr@gmail.com