Vox - Large language models are astonishingly good at understanding and producing language. But there’s an often overlooked bias toward languages that are already well-represented on the internet. That means some languages might lose out in AI’s big technical advances. Some researchers are looking into how that works — and how to possibly shift the balance from these “high resource” languages to ones that haven’t yet had a huge online footprint. These approaches range from original dataset creation, to studying the outputs of large language models, to training open source alternatives.
Related article - Gresham College - Diseases of the Heart Structure, Muscle and Valves
Vox