<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="https://seacrowd.org/feed.xml" rel="self" type="application/atom+xml" /><link href="https://seacrowd.org/" rel="alternate" type="text/html" /><updated>2026-02-01T23:55:33+00:00</updated><id>https://seacrowd.org/feed.xml</id><title type="html">SEACrowd: AI for SEA</title><subtitle>AI for Southeast Asia, by Southeast Asians</subtitle><entry><title type="html">From Tagalog to Filipino: Deciphering Language and Identity</title><link href="https://seacrowd.org/posts/from-tagalog-to-filipino.html" rel="alternate" type="text/html" title="From Tagalog to Filipino: Deciphering Language and Identity" /><published>2026-01-27T00:00:00+00:00</published><updated>2026-01-27T00:00:00+00:00</updated><id>https://seacrowd.org/posts/from-tagalog-to-filipino</id><content type="html" xml:base="https://seacrowd.org/posts/from-tagalog-to-filipino.html"><![CDATA[<p>If you’ve ever spent time with Filipinos, you’ve probably heard the words “Tagalog” and “Filipino” used almost interchangeably. For many, the difference feels minor, but behind those two labels is a long history shaped by migration, colonization, nationalism, and modern technology. Understanding the distinction isn’t just about semantics, it’s about how the Philippines defines its national identity, and increasingly, how that identity is represented in digital spaces and an increasingly AI-centric world.</p>

<h2 id="from-austronesian-roots-to-a-colonial-transformation">From Austronesian roots to a colonial transformation</h2>

<p>To understand Filipino, we start with its base: Tagalog. As part of the <a href="https:/www.britannica.com/topic/Philippine-languages">Austronesian language family</a>, Tagalog is related to Indonesian, Malay, and many languages across Southeast Asia and the Pacific. Its story traces back more than 4,000 years to Proto-Austronesian communities in Taiwan.</p>

<p>Long before the Spaniards arrived in the 16th century, Tagalog was already the main language of Central and Southern Luzon. Continuous trade with Chinese, Indian, and Southeast Asian cultures enriched the vocabulary and ideas embedded in the language. At the time, the Tagalogs used the indigenous Baybayin script for <a href="https:/narrastudio.com/blogs/journal/baybayin-the-ancient-filipino-script-lives-on">personal communication, poetry, and short messages</a>, while longer histories and epics were <a href="http:/paulmorrow.ca/bayeng1.htm">preserved through a rich oral tradition</a>.</p>

<p><img src="/assets/images/posts/filipino-tagalog/2_baybayin.png" alt="The Baybayin Script" width="500" /></p>

<p><em><a href="https:/64.media.tumblr.com/tumblr_lutfskLZxZ1qkow7co1_r1_1280.png">Fig 1</a>. An example of the pre-colonial Baybayin script.</em></p>

<p>Spanish colonization (1565-1898) systematically sidelined Baybayin, but distinctively, it colonized the script without replacing the tongue. As the friars prioritized evangelization over assimilation, they used the Latin alphabet to <a href="https:/hiphilangsci.net/2013/07/10/early-writing-and-printing-in-the-philippines/">translate religious texts into local dialects</a> rather than teaching Spanish to the masses. With official documents and church texts shifting to Latin-based Tagalog—and the Latin script proving more practical for new Spanish-influenced sounds—Baybayin quickly lost relevance and faded from use.</p>

<p><img src="/assets/images/posts/filipino-tagalog/3_thomasite.jpg" alt="The Thomasites" width="500" /></p>

<p><em><a href="https:/blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEidMLgbsFg5RsnKUiorIfVb8K4s0p-v8Z-xGuN_OwbTz-z6hZJ_6BsOWBv0kMrhUZDqG8h9BH1dLwrr-aZi8qozAj_EXTpa-dI3oKlm5xrCqoY9N9N9ZFaSV-nkGsN-K_DDIzFpQxvZHEB5/s1600/thomasite.jpg">Fig 2</a>. A female Thomasite teacher instructing Filipino students</em></p>

<p>Following the Spanish-American War in 1898 up until the end of the American occupation in 1946, the United States utilized a radically different strategy: <a href="https:/en.wikipedia.org/wiki/Thomasites#:~:text=Our,schools">assimilation through mass education</a>. Unlike the exclusive nature of Spanish rule, the Americans established a public schooling system through <a href="https:/elibrary.judiciary.gov.ph/thebookshelf/showdocs/28/15605#:~:text=SEC%2E%2014,teachers">Act No. 704 of 1901</a>, that enforced English as the sole medium of instruction. By making English the new entry point for commerce and bureaucracy, the Americans effectively institutionalized a foreign language in mere decades, reshaping the linguistic landscape far more aggressively than Spain and pushing Tagalog into a new role during <a href="https:/ncca.gov.ph/about-culture-and-arts/in-focus/andres-bonifacio-and-the-1896-revolution">the rise of nationalism</a>.</p>

<h2 id="constructing-a-national-language">Constructing a national language</h2>

<p>The dream of a unified Filipino nation has always been deeply tied to the dream of a single national language. In 1937, President Manuel L. Quezon declared Tagalog as the <a href="https:/elibrary.judiciary.gov.ph/thebookshelf/showdocs/5/77655#:~:text=the%20selection%20of%20Tagalog%20as%20the%20basis%20of%20the%20national%20language%20as%20it%20has%20been%20found%20to%20be%20used%20and%20accepted%20by%20the%20greatest%20number%20of%20Filipinos">basis for the national language</a>. This decision was driven by a mix of commercial practicality and political dominance. As the language of the capital, Tagalog had long served as the default medium for trade. However, it also reflected the influence of the region’s elite. With the seat of government and the center of the revolution historically situated in the Tagalog regions, the dialect naturally became the language of power (refer to the <a href="https:/en.wikisource.org/wiki/Provisional_Constitution_of_the_Philippines_(1897)#:~:text=Tagalog%20shall%20be%20the%20official%20language%20of%20the%20Republic%2E">revolutionary constitution</a>).</p>

<p>But in a country home to over 175 distinct languages, this decision was met with controversy. Speakers of Cebuano, Ilocano, and other major languages, felt excluded, sparking a debate about “<a href="https:/www.studocu.com/ph/document/cagayan-national-high-school/filipino/jmmd0190487-possible-referencesource/32948530">Tagalog imperialism</a>“—a dynamic which can be likened to the elevation of the Beijing dialect to Standard Mandarin in <a href="https:/www.youtube.com/watch?v=qzr09A6Zb48">unifying 20th century China</a>.</p>

<p>To create a more inclusive national language, the term “Pilipino” was introduced in 1959. But it wasn’t until the 1987 Constitution that “Filipino” became the official national language, mandating that it be <a href="https:/lawphil.net/consti/cons1987.html#:~:text=developed%20and%20enriched"><em>developed and enriched</em></a> by all Philippine languages. This mandate represented a big conceptual leap. The 1987 Constitution imagined Filipino as an inclusive national language drawing from all Philippine languages, but the reality is that <a href="https:/asianstudies.cornell.edu/research/tagalog-filipino#:~:text=Filipino%20and%20Tagalog%20are%20varieties%20or%20dialects%20of%20the%20same%20language">Filipino remains fundamentally Tagalog</a>; its grammar, syntax, and core vocabulary are still Tagalog at their base, even as it absorbs words from other languages.</p>

<h2 id="does-filipino-represent-all-filipinos">Does ‘Filipino’ represent all Filipinos?</h2>

<p>Today, the linguistic environment is complex. While Filipino is the national language, is it spoken everywhere? Yes and no. Thanks to mass media and the education system, almost all Filipinos understand it. However, in regions like Visayas and Mindanao (⅔ of the major islands of the Philippines), it often serves as a <em>lingua franca</em> for speaking with outsiders (i.e. business), whereas the local vernacular dominates daily life.</p>

<p><img src="/assets/images/posts/filipino-tagalog/4_langmap.png" alt="Linguistic Map of the Philippines" width="500" /></p>

<p><em><a href="https:/www.geocurrents.info/wp-content/uploads/2016/01/Maritime-Linkages-Philippine-Languages-Map.png">Fig 3</a>. A map showing the linguistic complexity of the Philippine archipelago.</em></p>

<!-- https:/www.geocurrents.info/wp-content/uploads/2016/01/Maritime-Linkages-Philippine-Languages-Map.png -->

<p>In fact, resentment toward the centralization of Tagalog persists. In areas where Cebuano (part of the Bisaya languages) is dominant, pushback against “Imperial Manila” remains tangible, with some <a href="https:/e-journal.usd.ac.id/index.php/LLT/article/view/2581#:~:text=favored%20English%20to%20be%20the%20language%20used%20for%20formal%20communication">locals preferring to use English</a> rather than Filipino in formal settings. This friction has even migrated to digital spaces, where the viral “Tagalog vs. Bisaya” meme trend frequently resurfaces these long-standing regional rivalries.</p>

<p>This linguistic tension has found new life online, where language itself is evolving faster than ever. Social media has become a laboratory for linguistic evolution: Gen Z slang like <em>charot</em> (just kidding) and <em>sana all</em> (wish it was like that for everyone) among others spread virally across platforms. Tiktok, Twitter/X, and Facebook comments sections show Filipino absorbing not just English loan words, but internet-native expressions that would make purists wince. These digital spaces are where the “development and enrichment” mandated by the constitution actually happens—not in government committees, but through millions of Filipinos creating language in real-time.</p>

<p>Furthermore, the Philippine linguistic ecosystem outside the capital is fragile. While major regional languages (like Ilocano, Hiligaynon, Bikolano, etc.) remain robust, the pressure of learning Filipino and English is impacting smaller languages. Dozens of indigenous languages, such as those spoken by the Itneg and Batak communities are currently <a href="https:/www.facebook.com/photo/?fbid=134899795998833&amp;set=a.116046591217487">classified as dying or endangered</a> as younger generations shift away from them due to the <a href="https:/www.sil.org/system/files/reapdata/79/35/83/79358370809751253082780911549103284404/Pascasio.pdf">cultural and economic hegemony of the two national languages.</a></p>

<h2 id="tagalog-vs-filipino-whats-the-actual-difference">Tagalog vs Filipino: What’s the actual difference?</h2>

<p>In practice, linguists generally agree on this:</p>

<blockquote>
  <p><strong>Filipino</strong> = the standardized, modern, national-language variety of Tagalog.</p>
</blockquote>

<p>Structurally, they share the same grammar. The difference lies in:</p>

<h3 id="1-vocabulary-breadth">1. Vocabulary breadth</h3>

<p>Filipino absorbs words not only from Tagalog, but also Cebuano, Ilocano, Spanish, English, and more.</p>

<ul>
  <li>
    <p><strong>Katarungan</strong> (Justice) - adopted from the Bisaya root <em>tarong</em> (straight/right), used interchangeably with the Spanish-derived <em>hustisya</em></p>
  </li>
  <li>
    <p><strong>Trabaho</strong> (Work) - a loan word from the Spanish word: <em>trabajo</em></p>
  </li>
  <li>
    <p><strong>Kompyuter</strong> (Computer) - a loan word ‘Filipinized’ from the English word: computer</p>
  </li>
</ul>

<h3 id="2-alphabet-expansion">2. Alphabet expansion</h3>

<ul>
  <li>
    <p><strong>Baybayin:</strong> 17 characters</p>
  </li>
  <li>
    <p><strong>Abakada (Tagalog):</strong> 20 letters</p>
  </li>
  <li>
    <p><strong>Modern Filipino:</strong> 28 letters; adding F, J, Ñ, Z, among others, to accommodate a wider range of words</p>
  </li>
</ul>

<h3 id="3-everyday-usage">3. Everyday usage</h3>

<p>Modern Filipino reflects how people actually speak today, full of code-switching between Filipino, English, and regional languages.</p>

<p><strong>Examples:</strong></p>

<ul>
  <li>
    <p><em>Mag-picture tayo.</em> (Let’s take a picture together.) [Filipino+English]</p>
  </li>
  <li>
    <p><em>Na-late ako sa work dahil sa traffic.</em> (I was late for work because of traffic.) [Filipino+English]</p>
  </li>
  <li>
    <p><em>I am going to the market, bibili lang ako ng pagkain, para naay makaon sa balay.</em> (I am going to the market, I will just buy some food, so there is something to eat at home.) [Filipino+English+Cebuano]</p>
  </li>
</ul>

<p>For many Filipinos, especially in urban areas, this trilingual juggling act is the norm. A Cebuano speaker in Manila might switch between Cebuano with family, Filipino with colleagues, English in formal emails, and a fluid mix of all three when texting friends. For NLP, this means our “Filipino” data isn’t just Tagalog with borrowed words, it’s a dynamic ecosystem where boundaries between languages blur constantly.</p>

<h2 id="why-this-matters-for-nlp">Why this matters for NLP</h2>

<p>For NLP researchers, the distinction between Tagalog and Filipino isn’t about grammar as structurally, they follow the same system. What matters is knowing what our data actually contains.</p>

<p>Most online Philippine text, whether scraped in <a href="https:/github.com/commoncrawl/web-languages/blob/main/living/tagalog.md">Common Crawl</a> or a dataset on HuggingFace—uses the ISO code <code class="language-plaintext highlighter-rouge">tl</code> (Tagalog). But the actual content inside is overwhelmingly Filipino, the modern, evolving variety that naturally encompasses code-switching and usage of loan words.</p>

<p><img src="/assets/images/posts/filipino-tagalog/5_datacard.png" alt="Dataset Card Screenshot" width="500" /></p>

<p><em><a href="https:/huggingface.co/datasets/uonlp/CulturaX#:~:text=0%2E00-,69,0%2E00">Fig 4</a>. A standard dataset card showing the language code ‘tl’ (Tagalog) but labeled Filipino</em></p>

<!-- https:/imgur.com/a/KQcl6RY -->

<p>In practice, this means we’re already working with Filipino, even though the metadata in <a href="https:/huggingface.co/datasets/uonlp/CulturaX#:~:text=0%2E00-,69,0%2E00">data cards</a>, language tags in model configurations, and standard ISO codes are universally labeled “Tagalog” (<code class="language-plaintext highlighter-rouge">tl</code>). We don’t need to change the ISO standard or relabel existing corpora. What’s important is recognizing the distinction so that we understand the linguistic reality behind the data we use.</p>

<h2 id="filipino-as-a-linguistic-bridge">Filipino as a linguistic bridge</h2>

<p>Just as Indonesian serves as an anchor for regional languages like Javanese or Sundanese, Filipino can serve as the pivot language for Philippine NLP. Because it shares core grammar and cognates with <a href="https:/www.researchgate.net/publication/396574502_Morphological_Analysis_of_Tagalog_Hiligaynon_and_Cebuano_Languages">Cebuano, Hiligaynon</a>, Waray, Kapampangan, Bikol, and many others, a strong Filipino model creates a foundation that accelerates progress for every other language in the country.</p>

<table>
  <thead>
    <tr>
      <th style="text-align: left">English</th>
      <th style="text-align: left">Tagalog (Pivot)</th>
      <th style="text-align: left">Cebuano</th>
      <th style="text-align: left">Ilocano</th>
      <th style="text-align: left">Kapampangan</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td style="text-align: left"><strong>Eye</strong></td>
      <td style="text-align: left">Mata</td>
      <td style="text-align: left">Mata</td>
      <td style="text-align: left">Mata</td>
      <td style="text-align: left">Mata</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>Two</strong></td>
      <td style="text-align: left">Dalawa</td>
      <td style="text-align: left">Duha</td>
      <td style="text-align: left">Dua</td>
      <td style="text-align: left">Adwa</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>House</strong></td>
      <td style="text-align: left">Bahay</td>
      <td style="text-align: left">Balay</td>
      <td style="text-align: left">Balay</td>
      <td style="text-align: left">Bale</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>New</strong></td>
      <td style="text-align: left">Bago</td>
      <td style="text-align: left">Bag-o</td>
      <td style="text-align: left">Baro</td>
      <td style="text-align: left">Bayu</td>
    </tr>
  </tbody>
</table>

<p><em>Table 1. A lexical comparison showing high similarity (cognates) across major Philippine languages.</em></p>

<p>This allows researchers to utilize <a href="https:/www.geeksforgeeks.org/nlp/transfer-learning-in-nlp/">transfer learning</a>, by effectively “recycling” what an AI learns from the data rich Filipino language and applying it to regional languages that have less data. Instead of building models from scratch for every dialect and language, a good Filipino system can accelerate the creation of translation tools and datasets for the rest of the archipelago.</p>

<p>Understanding the Tagalog-Filipino distinction helps us build models that reflect how people actually communicate today, without needing to replace existing standards.</p>

<h2 id="the-double-edged-sword-progress-and-preservation">The double-edged sword: Progress and preservation</h2>

<p>Here’s the cold truth: as we build better Filipino NLP models, we might be accelerating the very <a href="https:/www.nature.com/articles/s41599-023-01911-8#:~:text=caused%20a%20language%20shift%20and%20language%20extinction">problem we hoped to solve.</a> With every successful Filipino chatbot, translation tool, or voice assistant, we reinforce Filipino’s dominance, creating economic incentives/dependencies which push communities away from their local languages.</p>

<p>But the same technology could be their lifeline. If Filipino serves as a true pivot language, transfer learning makes it economically feasible to build tools for languages with tiny datasets. Imagine a Filipino to Itneg translation model helping grandparents connect with their <em>apo’s</em> who never learned the language, or documentation systems preserving Batak oral traditions before the last fluent speakers pass away.</p>

<p><img src="/assets/images/posts/filipino-tagalog/6_bridge.jpg" alt="Dataset Card Screenshot" width="500" /></p>

<p><em><a href="https:/gttp./assets/images/posts/filipino-tagalog/.tshiftcdn.com/270079/x/0/san-juanico-bridge-in-leyte.jpg">Fig 5</a>. The San Juanico Bridge spans the strait between Samar and Leyte—a symbol of connection in an archipelago of over 7,600 islands.</em></p>

<!-- https:/gttp./assets/images/posts/filipino-tagalog/.tshiftcdn.com/270079/x/0/san-juanico-bridge-in-leyte.jpg -->

<p>The choice isn’t inevitable. It depends on whether researchers treat Filipino as the <em>destination</em> or as the <em>bridge</em>. If we stop at building excellent Filipino models, we’ve simply digitized linguistic hegemony. But if we use Filipino’s advantage to pull smaller languages into the digital age, we might preserve what otherwise would be lost.</p>

<h2 id="preserving-identity-in-the-algorithm">Preserving identity in the algorithm</h2>

<p>The difference between Tagalog and Filipino isn’t just academic trivia—it reflects how the Philippines defines itself as a nation. Acknowledging Filipino as the national, standardized, evolving language means acknowledging our linguistic diversity while recognizing that its structure remains inherently rooted in Tagalog.</p>

<p>For NLP researchers, this awareness shapes everything we build. It helps us interpret our data honestly, design models that reflect how Filipinos actually communicate, and most critically, ensures that as we bring Philippine languages into the age of AI, we’re creating tools that connect rather than divide.</p>

<p>The algorithms we create today will shape which languages survive tomorrow. That’s not just a technical challenge, it’s a responsibility to the 175+ languages that make up who we are.</p>]]></content><author><name>Isaiah Flores</name></author><category term="Society &amp; AI" /><category term="linguistics" /><category term="nlp" /><category term="filipino" /><category term="tagalog" /><category term="history" /><category term="data-culture" /><summary type="html"><![CDATA[Is it Tagalog or Filipino? We explore how a regional language became a national standard, and why this evolution matters for training robust, inclusive NLP models.]]></summary></entry><entry><title type="html">Growing SEA-Native AI Research Talent: Retrospective on the SEACrowd Apprentice Program 2024–25</title><link href="https://seacrowd.org/posts/apprentice-program-retrospection.html" rel="alternate" type="text/html" title="Growing SEA-Native AI Research Talent: Retrospective on the SEACrowd Apprentice Program 2024–25" /><published>2025-12-03T00:00:00+00:00</published><updated>2025-12-03T00:00:00+00:00</updated><id>https://seacrowd.org/posts/apprentice-program-retrospection</id><content type="html" xml:base="https://seacrowd.org/posts/apprentice-program-retrospection.html"><![CDATA[<p>I’m Samuel Cahyawijaya, a Member of Technical Staff at <a href="https://cohere.com/">Cohere</a> and one of the initiators of <a href="/about">SEACrowd</a>.</p>

<p>Over 2024-2025, I served as a lead mentor for the <strong><a href="/apprenticeship">SEACrowd Apprentice Program</a></strong>. I watched early-career researchers across SEA go from “I’m curious about AI research” to co-authoring papers and releasing a multilingual, culturally grounded dialogue dataset.</p>

<p>This post shares what we built, what was hard, what we’re changing, and how you can <a href="/apprenticeship">join Batch 2026</a>.</p>

<p>TLDR;</p>

<ul>
  <li>We ran our first SEACrowd Apprentice Program cohort focused on real research.</li>
  <li>Along the way, we hit hard problems: uneven commitment, skill gaps inside teams, mentor bandwidth issues, and losing momentum over a long timeline.</li>
  <li>We’re changing the structure for <strong><a href="/apprenticeship">Batch 2026</a></strong> with clearer selection, midterm and end-term milestones, external evaluations, and better-matched teams.</li>
  <li><a href="/apprenticeship">Application is open</a> until Dec 17, 2025 11:59PM UTC-12. Go to our page for more details.</li>
</ul>

<h2 id="about-seacrowd">About SEACrowd</h2>

<p>SEACrowd is a community-run initiative focused on AI research and development for Southeast Asia. Rather than just debating SEA’s lack of representation, we <a href="/resources">build models and datasets</a>, and <a href="/publications">do research</a> that center SEA languages, cultures, and realities.</p>

<p>For this, we bring together researchers, students, and practitioners who care about SEA languages and communities. Read more about us <a href="/about">here</a>.</p>

<h2 id="why-we-started-the-seacrowd-apprentice-program">Why We Started the SEACrowd Apprentice Program</h2>

<p>As SEACrowd grew, we kept bumping into the same problem:</p>

<blockquote>
  <p>Many talented people in Southeast Asia <em>want</em> to do AI research but don’t see a clear path from “interested” to “actually doing research that gets published.”</p>
</blockquote>

<p>The gaps are recurring:</p>

<ol>
  <li><strong>Lack of mentorship.</strong> It’s hard to get regular feedback from experienced researchers who understand both the technical side <em>and</em> the regional context.</li>
  <li><strong>No structured, low-barrier entry point.</strong> Many people in the region see research and postgraduate study as distant or already out of reach, especially when pitted against peers from better-resourced countries and institutions.</li>
  <li><strong>Limited exposure to full projects.</strong> Many only see pieces of the workflow (e.g., coding, reading papers), not the entire journey from idea → experiments → writing → submission.</li>
</ol>

<p>These issues motivated us to create a program that would:</p>

<ol>
  <li>Design around real, scoped, publishable projects</li>
  <li>Pair experienced researchers with early-career talent across Southeast Asia for hands-on learning</li>
</ol>

<p>We wanted this to go far beyond a reading group or workshop: to meaningfully upgrade mentees’ research skills and strengthen their chances for PhD applications and AI research careers. And all of this concretized in our first <strong>SEACrowd Apprentice Program</strong> cohort in 2024-25.</p>

<h2 id="what-we-accomplished-in-batch-202425">What We Accomplished in Batch 2024–25</h2>

<p>Looking back at the first cohort, I’m genuinely proud of what the teams pulled off. In one cycle, mentees contributed to:</p>

<ol>
  <li>A new technique for controlling language use in multilingual LLMs
→ <a href="https://aclanthology.org/2025.mrl-main.30/"><em>Language Surgery in Multilingual Large Language Models</em></a>, published at <a href="https://sigtyp.github.io/ws2025-mrl.html">Multilingual Representation Learning (MRL) Workshop @ EMNLP 2025</a></li>
  <li>A way to recover linguistic typology from language model behavior
→ <a href="https://aclanthology.org/2025.mrl-main.29/"><em>Entropy2Vec: Crosslingual Language Modeling Entropy as End-to-End Learnable Language Representations</em></a> at <a href="https://sigtyp.github.io/ws2025-mrl.html">MRL 2025</a></li>
  <li>A culturally grounded dialogue dataset &amp; benchmark for SEA languages
→ <a href="https://arxiv.org/abs/2508.07069"><em>SEADialogues: A Multilingual Culturally Grounded Multi-turn Dialogue Dataset on Southeast Asian Languages</em></a></li>
</ol>

<p>Click on each project below to have a closer look!</p>

<div class="accordion" id="projects-accordion">
    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-projects-accordion-1">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#projects-accordion-1" aria-expanded="false" aria-controls="projects-accordion-1" data-bs-parent="#projects-accordion">
      Project 1: Language Surgery in Multilingual Large Language Models
    </button>
  </h3>
  <div id="projects-accordion-1" class="accordion-collapse collapse " aria-labelledby="heading-projects-accordion-1" data-bs-parent="#projects-accordion">
    <div class="accordion-body">
      <p>👥 <strong>Mentees:</strong> Joanito Agili Lopo, Muhammad Ravi Shulthan Habibi, Tack Hwa Wong</p>

<p>→ <a href="https://aclanthology.org/2025.mrl-main.30/">Read paper, published at MRL 2025</a></p>

<h4 id="the-problem">The problem</h4>

<p>Multilingual LLMs support many languages, but anyone who’s tried prompting them in under-resourced languages has probably seen this:</p>

<p>You prompt in one language. The model partly responds in another (often English), drifts into mixed-language output, or drops in quality compared to high-resource languages like English or Chinese.</p>

<p>So, we wanted to understand:</p>

<ul>
  <li>How do multilingual LLMs organize languages internally?</li>
  <li>Can we influence which language they “think in” or generate at inference time without retraining the model?</li>
</ul>

<h4 id="the-approach">The approach</h4>

<p>The team studied how multilingual LLMs organize languages in their latent space (the internal continuous representation). Specifically, they looked at how representations shift when the model handles different languages, and how those representations relate across languages.</p>

<p>They developed <strong>Inference-Time Language Control (ITLC)</strong>, a method for nudging models to produce more consistent language outputs from cross-lingual prompts, <em>without</em> retraining.</p>

<h4 id="why-this-matters">Why this matters</h4>

<p>Low-resource languages often get 1) less stable outputs, 2) higher rates of language switching, and 3) lower average quality.</p>

<p>ITLC offers a way to get more stable behavior in underrepresented languages using existing models where retraining is expensive or impractical. It also helps us understand how languages are arranged inside these models.</p>

    </div>
  </div>
</div>

    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-projects-accordion-2">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#projects-accordion-2" aria-expanded="false" aria-controls="projects-accordion-2" data-bs-parent="#projects-accordion">
      Project 2: Entropy2Vec — Crosslingual Language Modeling Entropy
    </button>
  </h3>
  <div id="projects-accordion-2" class="accordion-collapse collapse " aria-labelledby="heading-projects-accordion-2" data-bs-parent="#projects-accordion">
    <div class="accordion-body">
      <p>👥 <strong>Mentees:</strong> Patrick Amadeus Irawan, Ryandito Diandaru, Belati Jagad Bintang Syuhada, Randy Zakya Suchrady</p>

<p>→ <a href="https://aclanthology.org/2025.mrl-main.29/">Read paper, published at MRL 2025</a></p>

<h4 id="the-problem">The problem</h4>

<p>Linguistic typology—mapping how languages relate to each other—has traditionally come from decades of manual documentation, linguistic expertise, or resources like <a href="https://www.ethnologue.com/">Ethnologue</a> and <a href="https://glottolog.org/">Glottolog</a>).</p>

<p>Modern LLMs, meanwhile, have learned behavior across many languages through large-scale training with rich, internal knowledge about which tokens are likely in which contexts.</p>

<p>Given the ability of LLMs to learn multiple languages through the large-scale training, we believe that this knowledge could provide a fine-grained representation that captures the dynamic across different languages as we know in linguistic typology. But <strong>can the internal language knowledge of LLMs can actually reflect the linguistic typological relationships similar to those in Ethnologue and Glottolog?</strong></p>

<h4 id="the-approach">The approach</h4>

<p>The team extracted language representations from cross-lingual language modeling <strong>entropy</strong> collected from a diverse set of monolingual language models (LMs). Entropy is a measure of how unexpected the next token is, according to the model; if a model shows similar prediction patterns in two languages, those languages might be structurally similar.</p>

<p>They built <strong>Entropy2Vec</strong>, a language embedding where each language is a continuous vector in a latent space, and distances between vectors match linguistic typology relationships. Instead of understanding low-level handcrafted features like word order, syntax, or morphological complexity, they used the model’s own internal knowledge (i.e., how it predicts next tokens) to recover typological structure.</p>

<h4 id="why-this-matters">Why this matters</h4>

<p>The learned language vectors mimic typological structure (i.e., languages that are close in the typology tree end up close in the embedding space) and can regularize fine-tuning to improve adaptability to unseen languages.</p>

<p>For multilingual research in SEA, where many languages are under-documented, this line of work opens the door to:</p>

<ul>
  <li>reusing existing models to infer structural relationships</li>
  <li>inform better transfer learning strategies</li>
  <li>helping models generalize better to languages with very little data</li>
</ul>

    </div>
  </div>
</div>

    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-projects-accordion-3">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#projects-accordion-3" aria-expanded="false" aria-controls="projects-accordion-3" data-bs-parent="#projects-accordion">
      Project 3: SEADialogues — Culturally Grounded Dialogue Dataset
    </button>
  </h3>
  <div id="projects-accordion-3" class="accordion-collapse collapse " aria-labelledby="heading-projects-accordion-3" data-bs-parent="#projects-accordion">
    <div class="accordion-body">
      <p>👥 <strong>Mentees:</strong> Muhammad Dehan Al Kautsar, Aswin Candra, Muhammad Alif Al Hakim, Maxalmina Satria Kahfi</p>

<p>→ <a href="https://huggingface.co/datasets/SEACrowd/SEADialogues">View dataset on HuggingFace</a> &amp;&amp; <a href="https://arxiv.org/abs/2508.07069">Read paper</a></p>

<h4 id="the-problem">The problem</h4>

<p>Even as LLMs improve, underrepresented languages perform worse and frequently fail to align with local norms, values, and expectations in multi-turn conversations.</p>

<h4 id="the-approach">The approach</h4>

<p>The team built a culturally grounded, multi-turn dialogue dataset for SEA languages. They collected, curated, and annotated conversations that reflect real cultural practices—local forms of politeness, family structures, community dynamics. The resulting dataset is suitable for both training and evaluating value-aware conversational models.</p>

<h4 id="why-this-matters">Why this matters</h4>

<p>SEADialogues provides a targeted resource for improving conversational AI in SEA languages, a benchmark for evaluating value alignment in culturally rich settings, and a starting point for future work on dialogue safety and social norms in SEA contexts.</p>

    </div>
  </div>
</div>

    
  </div>

<h2 id="what-we-learned-and-what-were-changing">What We Learned (and What We’re Changing)</h2>

<p>Running the first cohort was encouraging and humbling. Other than the successes, we hit challenges around time, skill differences, mentor bandwidth, and sustaining momentum. Here’s what observed and how we’re changing:</p>

<h3 id="1-commitment-and-time-constraints">1. Commitment and Time Constraints</h3>

<p>Mentees juggled full-time obligations, different experience levels, and life events. Even when motivation was high, sustained commitment across several months on a full research project proved challenging.</p>

<p><strong>So we’re changing with</strong>:</p>

<ul>
  <li>Refined selection criteria assessing prior exposure to coding and machine learning (ML), time availability, and clarity of motivation.</li>
  <li>Clearer upfront communication about timeline and commitment expectations, especially where the heavy lifts are, and what level of commitment is realistic.</li>
</ul>

<h3 id="2-skill-gaps-within-teams">2. Skill Gaps Within Teams</h3>

<p>Mentees had varying experience: some with prior NLP research, others new but determined. Diversity is a strength, but when one or two people carry most of the technical load, they burn out. When others feel they’re slowing the team, they disengage.</p>

<p><strong>So we’re changing with:</strong></p>

<ul>
  <li>More thoughtful team formation with shared baselines and lead mentees for task distribution.</li>
  <li>Projects structured so that different contributions matter—design, implementation, analysis, writing. The goal is to make it easier for mentees with different strengths to stay engaged and grow while still involved in most of the steps.</li>
</ul>

<h3 id="3-mentor-availability-and-consistency">3. Mentor Availability and Consistency</h3>

<p>Mentors are busy professionals or researchers volunteering their time for SEA talent training.</p>

<p>We also faced challenges with scheduling across time zones, supporting mentees when mentors got busy, and ensuring consistent feedback and guidance when a team had more than one mentor.</p>

<p><strong>So we’re changing with</strong>:</p>

<ul>
  <li>More explicit mentor roles within each team</li>
  <li>Coordinated mentor expectations to avoid conflicting directions</li>
  <li>Built-in redundancy so teams aren’t stranded if one mentor gets overloaded</li>
  <li>Shared guidelines for mentors about program goals and balancing ambition with feasibility</li>
</ul>

<p>We hope to make the mentee experience more coherent, regardless of which team they’re on.</p>

<h3 id="4-long-project-timeline-and-momentum">4. Long Project Timeline and Momentum</h3>

<p>Initially, we planned for the program to end whenever the write-up was ready for submission, but that open-ended timeline ended up hurting motivation and discipline.</p>

<p>On top of that, progress naturally came in waves—exciting starts, dragging middles, intense finals. Keeping everyone motivated through the uncertain middle was one of the toughest parts, especially when early experiments failed or teams weren’t sure their direction was still viable.</p>

<p><strong>So we’re changing with</strong>:</p>

<ul>
  <li>Fixed four-month window with clear checkpoints.</li>
  <li>Midterm presentations where teams present their progress to a broader group, get feedback from other mentors and can adjust scope and direction if needed.</li>
  <li>Clear end-of-term assessment with cross-team evaluation considering both research quality and team process.</li>
</ul>

<h2 id="so-was-it-worth-it">So Was It Worth It?</h2>

<p>The first SEACrowd Apprentice Program started as an experiment:</p>

<blockquote>
  <p>Could we meaningfully support early-career SEA researchers to produce real, impactful work in a relatively short time, across countries and time zones, online?</p>
</blockquote>

<p>The answer is: <strong>Yes, absolutely!</strong></p>

<p>Batch 2024-25 delivered two accepted workshop papers at MRL 2025, one new multilingual dialogue dataset, and a group of mentees who now have hands-on research experience, co-authorship in respected venues, and a stronger network across SEA and beyond.</p>

<p>We’re making structural changes for Batch 2026 because we now have a clearer picture of where early-career SEA researchers struggle, what support is most valuable, and how to design a program that respects constraints while demanding real work.</p>

<h2 id="how-to-join-batch-2026">How to Join Batch 2026</h2>

<p>If you’ve read this far, there’s a good chance you want to try this. Consider applying to the <strong><a href="/apprenticeship">SEACrowd 2026 Apprentice Program</a></strong> if you:</p>

<ul>
  <li>Are based in or connected to Southeast Asia</li>
  <li>Have basic Python coding and AI/ML/NLP familiarity</li>
  <li>Can commit consistent time over several months</li>
  <li>Are curious about research beyond using APIs or off-the-shelf models</li>
  <li>Want a path toward PhD programs or AI research roles</li>
</ul>

<p>We can’t guarantee a paper for every team, but we offer a structured environment, mentors who care about your growth, and an honest path from “I’m curious about AI research” to “I’ve actually done it.”</p>

<ul>
  <li>🗓️ Mentee application deadline: Dec 17, 2025 (UTC-12).</li>
  <li>🌐 Full details, requirements, and the application form, see our <a href="/apprenticeship">Apprentice Program page</a>.</li>
</ul>

<p>If you’d like to help empower the next generation of SEA researchers with SEACrowd, you can support us at <a href="/donate">this link</a>.</p>]]></content><author><name>Samuel Cahyawijaya</name></author><category term="Apprenticeship" /><category term="apprenticeship" /><category term="restropection" /><summary type="html"><![CDATA[What we learned from running the first SEACrowd Apprentice Program cohort, and how we're improving for Batch 2026.]]></summary></entry><entry><title type="html">Mind the Language Gap: Building an Inclusive AI Future for Southeast Asia</title><link href="https://seacrowd.org/posts/tfgi-mind-language-gap.html" rel="alternate" type="text/html" title="Mind the Language Gap: Building an Inclusive AI Future for Southeast Asia" /><published>2025-09-22T00:00:00+00:00</published><updated>2025-09-22T00:00:00+00:00</updated><id>https://seacrowd.org/posts/tfgi-mind-language-gap</id><content type="html" xml:base="https://seacrowd.org/posts/tfgi-mind-language-gap.html"><![CDATA[<p><em>Originally posted as a perspectives piece on <a href="https://techforgoodinstitute.org/blog/perspectives/mind-the-language-gap-building-an-inclusive-ai-future-for-southeast-asia/">Tech for Good Institute</a></em></p>

<p>Southeast Asia’s (SEA) digital ambitions are accelerating. Governments across the region are launching national AI strategies, digitizing public services, and investing in infrastructure to drive economic growth and social development. Initiatives like the <a href="https://academic.oup.com/oodh/article/doi/10.1093/oodh/oqae011/7611743">ATIPAN project</a> in the Philippines and <a href="https://www.equity.tech/work/mit-solve">MediBot</a> in Timor-Leste—bringing AI-powered healthcare to remote communities—demonstrate how transformative these technologies can be. Yet amid this momentum lies a quiet but urgent gap: the AI systems shaping SEA’s digital future often fail to represent its languages, cultures, or lived realities.</p>

<p>With over 100 ethnic groups who speak over a thousand living languages and dialects, SEA is one of the most linguistically diverse regions on Earth. Yet <a href="https://aclanthology.org/2024.findings-emnlp.766.pdf">modern AI systems are largely trained on English and a few other global languages</a>, leaving most SEA communities (speakers of Javanese, Tagalog, Burmese, and many others) underrepresented or <a href="https://www.techpolicy.press/how-multilingual-ai-can-protect-language-and-improve-global-technology/">invisible</a> in the AI development.</p>

<p>Why does this matter? This language gap isn’t just technical; it’s a barrier to equitable digital inclusion,
as <a href="https://techforgoodinstitute.org/blog/perspectives/the-power-of-language-diversity-in-the-ai-era/">language is deeply tied to identity, trust, and nuance</a>. When AI systems misinterpret what users say—or fail to speak in ways that feel natural or respectful—they risk delivering harmful advice, misclassifying inputs, or simply being ignored. A major <a href="https://aclanthology.org/2024.emnlp-main.296/">2024 study published at EMNLP</a> confirmed this: <a href="https://techforgoodinstitute.org/research/research-commentary/the-rise-of-regional-language-models-in-southeast-asia/">large language models struggle with low-resource SEA languages</a>, leading to mistranslations, cultural misinterpretations, and even harmful outputs. For example, AI might misread a patient’s expression of fatigue as benign when, in cultural context, it hints at something more serious like vertigo, and then make non-localized recommendations, suggesting an urgent clinic visit which may be helpful in Singapore but unrealistic or distressing in rural Indonesia, where the nearest doctor may be hours away.</p>

<p>For example, when <a href="https://www.simpol.ph/living-ai-bias-in-southeast-asia/">asked</a> to define the Filipino word <em>“sapot”</em>, ChatGPT interpreted it as “mental confusion” or “being tangled.” In context, it actually refers to a state of psychological unease—like a web of emotions—best understood through lived experience. Such misinterpretations aren’t trivial: they erode trust, miss nuance, and limit the emotional accuracy essential in mental health tools. While AI is scaling and improving mental health care across the English-speaking world, local language speakers cannot benefit.</p>

<p>As AI becomes more embedded in commerce, education, and governance, how can the region ensure that the future of AI truly reflects its people, languages, and lived realities?</p>

<h2 id="challenges--barriers">Challenges &amp; Barriers</h2>

<p>Southeast Asia’s push for inclusive AI faces four interconnected challenges: data scarcity, fragmented development, limited market incentives, and gaps in trust.</p>

<div class="accordion accordion-flush" id="challenges-accordion">
    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-challenges-accordion-1">
    <button class="accordion-button " type="button" data-bs-toggle="collapse" data-bs-target="#challenges-accordion-1" aria-expanded="true" aria-controls="challenges-accordion-1" data-bs-parent="#challenges-accordion">
      Data Scarcity
    </button>
  </h3>
  <div id="challenges-accordion-1" class="accordion-collapse collapse show" aria-labelledby="heading-challenges-accordion-1" data-bs-parent="#challenges-accordion">
    <div class="accordion-body">
      <p>Most SEA languages lack the large, high-quality datasets needed to train robust models. Where data exists, it is often scattered across informal sources and hard to standardise. The problem is worse for languages with strong oral traditions, which may have little or no digital footprint. Building quality datasets requires more than literal translation, which risks producing awkward “translationese”; it demands deep cultural grounding. A 2024 study by SEACrowd showed that popular global models underperform on SEA language tasks., particularly in generating natural-sounding text. Even when technically included, model performance for languages with limited digital presence fall behind, mirroring the hierarchy of data availability. Small language groups, already excluded from services, risk further marginalisation when AI tools bypass them.</p>

<p><img src="/assets/images/posts/tfgi-language-gap/culturally-unaware-ai.png" alt="Culturally-unaware image generation of 'people eating chicken pastel', a popular dish in the Philippines, using Stable Diffusion 3.5 Large" /></p>

<p><em>Culturally-unaware image generation of “people eating chicken pastel”, a popular dish in the Philippines, using stability.ai’s AI model <a href="https://huggingface.co/stabilityai/stable-diffusion-3.5-large">Stable Diffusion 3.5 Large</a>.</em></p>

    </div>
  </div>
</div>

    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-challenges-accordion-2">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#challenges-accordion-2" aria-expanded="false" aria-controls="challenges-accordion-2" data-bs-parent="#challenges-accordion">
      Fragmented Development
    </button>
  </h3>
  <div id="challenges-accordion-2" class="accordion-collapse collapse " aria-labelledby="heading-challenges-accordion-2" data-bs-parent="#challenges-accordion">
    <div class="accordion-body">
      <p>National AI strategies often prioritise infrastructure, data governance, and economic competitiveness, sidelining linguistic inclusion. Policy approaches vary widely between countries, and without regional coordination or data-sharing frameworks (e.g., common formats, ethical standards, pooled compute resources), efforts remain siloed. Some promising local initiatives are beginning to emerge. Thailand’s <a href="https://opentyphoon.ai">Typhoon</a> model, an accessible Thai-centric Large Language Model (LLM), was also trained on informal language to capture stylistic nuances that global models often overlook. Indonesia’s <a href="https://aclanthology.org/2023.findings-acl.868/">NusaCrowd</a> curated high-quality open datasets for low-resource languages, including widely spoken Javanese and Sundanese, as well as endangered tongues like Lampung and Buginese, capturing the breadth of linguistic diversity and cultural contexts such as code-switching and shifting levels of formality. Yet, without sustained investment and alignment with broader ASEAN strategies, their long-term support and interoperability remain limited. Regional collaboration is especially crucial in Southeast Asia, where many languages—like Malay, Khmer, and Hmong—cross national borders, and individual countries may lack the capacity to build full-stack AI pipelines independently.</p>

    </div>
  </div>
</div>

    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-challenges-accordion-3">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#challenges-accordion-3" aria-expanded="false" aria-controls="challenges-accordion-3" data-bs-parent="#challenges-accordion">
      Limited Market Incentives
    </button>
  </h3>
  <div id="challenges-accordion-3" class="accordion-collapse collapse " aria-labelledby="heading-challenges-accordion-3" data-bs-parent="#challenges-accordion">
    <div class="accordion-body">
      <p>Because big tech companies prioritise mainstream languages with existing commercial value, indigenous and low-resource languages are rarely incorporated into their models or business strategies. Meanwhile, local startups, academic labs, and grassroots groups often lack the computing power and funding needed to build language-specific tools. The region also faces a shortage of skilled NLP researchers and data engineers experienced in low-resource AI development, leaving the ecosystem under-resourced.</p>

<p>For perspective, <a href="https://arxiv.org/abs/2504.05747">SEA-LION, Southeast Asia’s flagship open-source LLM project</a>, was built by <strong>31 authors</strong>—compared to <strong>199</strong> for <a href="https://arxiv.org/abs/2501.12948">China’s DeepSeek-RI</a>, and a staggering <strong>3,295</strong> contributors behind <a href="https://arxiv.org/abs/2507.06261">Google’s latest Gemini model</a>.</p>

    </div>
  </div>
</div>

    
      
      
      







<div class="accordion-item">
  <h3 class="accordion-header" id="heading-challenges-accordion-4">
    <button class="accordion-button collapsed" type="button" data-bs-toggle="collapse" data-bs-target="#challenges-accordion-4" aria-expanded="false" aria-controls="challenges-accordion-4" data-bs-parent="#challenges-accordion">
      Gaps in Accessibility and Trust
    </button>
  </h3>
  <div id="challenges-accordion-4" class="accordion-collapse collapse " aria-labelledby="heading-challenges-accordion-4" data-bs-parent="#challenges-accordion">
    <div class="accordion-body">
      <p>In much of Southeast Asia, AI adoption is constrained by foundational infrastructure challenges: limited connectivity, unreliable power, costly or low-spec devices, and insufficient digital literacy. Even when tools are available, widespread usage is not guaranteed. Poor localisation—beyond mere translation—can result in awkward tone, cultural mismatches, or unfamiliar interfaces. In the region, this may manifest as overly formal language, failure to interpret code-switching (the blending of languages), or disregard for indirect communication norms. When tools feel extractive or culturally alien, they risk eroding user trust.</p>

    </div>
  </div>
</div>

    
  </div>

<h2 id="opportunities-and-solutions-building-inclusive-ai-from-the-ground-up">Opportunities and Solutions: Building Inclusive AI from the Ground Up</h2>

<p>Despite the barriers, SEA has a unique opportunity to lead in creating AI
that is truly inclusive and culturally grounded.
The region can chart its own path—treating linguistic and cultural diversity as assets, not obstacles.
With its deep traditions of multilingualism, code-switching, oral storytelling, and cultural hybridity,
SEA is well-placed to pioneer flexible, context-aware AI systems that handle code-switching,
shifting levels of formality, and socially complex communication.<br />
AI attuned to SEA’s complexity could enable trust-sensitive applications,
from health promotion in conservative areas to crisis communication across multiple languages and dialects.</p>

<h3 id="1-local-innovation-and-homegrown-solutions">1. Local Innovation and Homegrown Solutions</h3>

<p>A growing ecosystem of regional initiatives is tackling SEA’s unique linguistic challenges, blending grassroots energy with institutional support. Community-led efforts like <a href="https://seacrowd.github.io">SEACrowd</a> are making a significant impact—<a href="https://seacrowd.github.io/seacrowd-catalogue/">curating hundreds of corpora</a> covering nearly 1,000 languages, building performance benchmarks in 38 Southeast Asian languages (for comparison, OpenAI’s latest model only benchmarked performance on five SEA languages), while <a href="/apprenticeship">nurturing local AI talent</a>.
SEACrowd also collaborates with global open source initiatives such as
<a href="https://mlcommons.org">ML Commons</a>, <a href="https://commoncrawl.org">Common Crawl</a>, and <a href="https://www.masakhane.io">Masakhane</a>
to share lessons and enable the global shift toward community-led, <a href="https://www.taipeitimes.com/News/editorials/archives/2025/06/29/2003839418">inclusive</a> AI development. The Singapore-based <a href="https://techforgoodinstitute.org/blog/perspectives/increasing-southeast-asian-representation-in-ai/">SEA-LION initiative is creating open-source LLMs trained on 11 Southeast Asian languages to capture cultural nuances</a>, while Thailand’s Typhoon model and Indonesia’s <a href="https://arxiv.org/pdf/2309.10661">NusaWrites</a> are building datasets and models rooted in local context. Together, these efforts offer a powerful alternative to global models that often overlook the region’s linguistic diversity.</p>

<p><img src="/assets/images/posts/tfgi-language-gap/emnlp-2023-launch.png" alt="Launch of SEACrowd at EMNLP 2023" /></p>

<p><em>Launch of <a href="/">SEACrowd</a> at <a href="https://2023.emnlp.org/">EMNLP 2023</a>.</em></p>

<p>Beyond technology, these initiatives play a vital preservation role. <a href="https://www.unesco.org/en/days/mother-language">UNESCO</a> warns that nearly 40% of the world’s languages—many in SEA—are <a href="https://seatongue.com/blog/the-asian-languages-most-at-risk-of-disappearing-in-the-ai-era">endangered</a>. By creating a <a href="https://arxiv.org/abs/2502.18148">digital footprint</a>, these initiatives help safeguard not only languages but also the cultural knowledge embedded within them.</p>

<h3 id="2-regional-coordination-and-shared-infrastructure">2. Regional Coordination and Shared Infrastructure</h3>

<p>To break silos, ASEAN—working alongside universities, community groups, international organisations such as UNESCO, and global open-source initiatives–should support interoperable data frameworks and shared standards. Projects like <a href="https://aclanthology.org/2025.acl-long.916/">SEA-VL</a>—pairing over a million culturally relevant images with local-language captions—show both the value and complexity of cross-border collaboration. A Southeast Asian NLP Commons could standardise benchmarks, ethics, and governance, especially for indigenous and low-resource languages. India’s <a href="https://ai4bharat.iitm.ac.in">AI4Bharat</a> offers a model, funding open datasets in over 20 Indian languages with government, academic, and civil society support.</p>

<h3 id="3-enabling-ecosystems-through-policy-and-incentives">3. Enabling Ecosystems through Policy and Incentives</h3>

<p>Governments can treat linguistic datasets as public digital goods and fund open-source AI for regional languages. Procurement policies, tax incentives, and grants can spur business investment in inclusion. Policymakers are starting to take notice—<a href="https://asean.org/wp-content/uploads/2024/02/ASEAN-Guide-on-AI-Governance-and-Ethics_beautified_201223_v2.pdf">ASEAN’s Guide on AI</a> <a href="https://asean.org/wp-content/uploads/2024/02/ASEAN-Guide-on-AI-Governance-and-Ethics_beautified_201223_v2.pdf">Governance</a> <a href="https://asean.org/wp-content/uploads/2024/02/ASEAN-Guide-on-AI-Governance-and-Ethics_beautified_201223_v2.pdf">and Ethics</a> and <a href="https://aiverifyfoundation.sg/resources/mgf-gen-ai/">Singapore’s IMDA</a> emphasise inclusive data practices. However, unless language equity becomes a core pillar of digital transformation, SEA risks developing AI that speaks over its people.</p>

<h3 id="4-trust-transparency-inclusion">4. Trust, Transparency, Inclusion</h3>

<p>Language inclusion must be <a href="https://arxiv.org/pdf/2506.07506">participatory</a>. Co-governance models—where contributors shape data practices and evaluation—build awareness, trust, and ownership. Investing in mentorship, transparency, and shared control ensures SEA’s digital future reflects its full diversity.</p>

<h2 id="conclusion">Conclusion</h2>

<p>Southeast Asia’s digital future depends on closing the language gap in AI. The region’s linguistic diversity is a strategic asset, yet current systems exclude many communities. This article outlined the barriers—like missing datasets and weak incentives—and showcased efforts like SEACrowd, Typhoon, and Nusa Collection that chart a new path. Policymakers must treat local language data as essential infrastructure, while industry and communities build AI that genuinely reflects SEA voices. Inclusive AI is not optional—it is a strategic imperative to ensure digital transformation benefits everyone. By investing in linguistic inclusion, SEA doesn’t just close a gap-it leads the world in building AI that truly belongs to everyone.</p>

<h2 id="about-the-authors">About the Authors</h2>

<p>Onno Kampman is an AI Scientist at Singapore’s MOH Office for Healthcare Transformation (MOHT) and a Visiting Scientist at the University of Cambridge. He leads pioneering projects that apply AI to mental health care transformation, and contributes to SEACrowd’s mission to boost Southeast Asian AI capabilities.</p>

<p>Holy Lovenia is the Lead of SEACrowd. Based in London and affiliated with AI Singapore, she drives SEACrowd’s strategy to unify and scale AI resources across Southeast Asia—most recently through initiatives like SEACrowd’s multilingual benchmarks and the SEA‑VL vision‑language dataset.</p>

<h2 id="about-the-organisation">About the Organisation</h2>

<p><a href="/about">SEACrowd</a> is a research community advancing Southeast Asia-focused AI and empowering the next generation of AI researchers in the region. The organisation envisions a future where Southeast Asia’s AI ecosystem is mature, globally competitive, and grounded in the region’s diverse linguistic and cultural contexts.</p>

<p>SEACrowd’s initiatives include leading data collection and model development efforts tailored to Southeast Asia, building and connecting a regional research network, and supporting early-career talent through mentorship and hands-on experience via the <a href="/apprenticeship">SEACrowd Apprentice Program</a>.</p>]]></content><author><name>Onno P. Kampman</name></author><category term="perspectives" /><category term="cross-posting" /><summary type="html"><![CDATA[Southeast Asia's AI future depends on closing the language gap, ensuring that the region's rich linguistic and cultural diversity is reflected in the data and models driving technological progress. Grassroots initiatives like SEACrowd show that inclusive, community-driven AI is not only possible but essential for equitable digital development.]]></summary></entry><entry><title type="html">SEACrowd forms legal entity in Indonesia</title><link href="https://seacrowd.org/posts/incorporation.html" rel="alternate" type="text/html" title="SEACrowd forms legal entity in Indonesia" /><published>2024-08-22T00:00:00+00:00</published><updated>2024-08-22T00:00:00+00:00</updated><id>https://seacrowd.org/posts/incorporation</id><content type="html" xml:base="https://seacrowd.org/posts/incorporation.html"><![CDATA[<p>Today, SEACrowd has formed a legal entity in Indonesia. 🇮🇩</p>

<p>This marks a significant milestone in our growth and commitment to
fostering local collaborations and innovation across Southeast Asia.
We look forward to strengthening our presence and
contributing to AI in SEA even more.</p>

<div class="row g-3 my-4">
  <div class="col-sm-4">
    <img src="/assets/images/posts/sign-holy.jpg" alt="Holy signing documents" class="img-fluid" />
  </div>
  <div class="col-sm-4">
    <img src="/assets/images/posts/sign-aji.jpg" alt="Aji signing documents" class="img-fluid" />
  </div>
  <div class="col-sm-4">
    <img src="/assets/images/posts/sign-sabil.jpg" alt="Sabil signing documents" class="img-fluid" />
  </div>
</div>

<p>The structure of our legal entity is as follows:</p>

<ul>
  <li>Board of Trustees
    <ul>
      <li>Alham Fikri Aji (<em>Chief</em>)</li>
      <li>Samuel Cahyawijaya</li>
    </ul>
  </li>
  <li>Executive Board
    <ul>
      <li>Holy Lovenia (<em>Chief</em>)</li>
      <li>Salsabil Maulana Akbar (<em>Secretary</em>)</li>
      <li>Frederikus Hudi (<em>Treasury</em>)</li>
    </ul>
  </li>
  <li>Supervisory Board
    <ul>
      <li>Genta Indra Winata (<em>Chief</em>)</li>
      <li>Fajri Koto</li>
      <li>Jennifer Santoso</li>
    </ul>
  </li>
</ul>]]></content><author><name></name></author><category term="info" /><summary type="html"><![CDATA[Today, SEACrowd has formed a legal entity in Indonesia. 🇮🇩]]></summary></entry></feed>