NAACL 2024: International Conference Spotlights AI and NLP Progress In Mexico City

Yesterday, the Mexican NLP Summer School ushered in the second day of enlightening presentations, tutorials and knowledge sharing from experts globally in the context of NAACL 2024 (North American Association of Computational Linguistics). The event preempts the commencement of the much-anticipated NAACL 2024 conference set to take place in Mexico City.

The two-day summer school was sponsored in part by private enterprise, like Grammarly and Bloomberg LP, while the organizing aspect was managed by researchers Thamar Solorio from Mohamed bin Zayed University of Artificial Intelligence, Ximena Gutierrez-Vasques and other members of the UNAM and Mexico City’s vast network of research faculty as well as student volunteers from abroad and within Mexico. Notably, Helena Gomez, a researcher and professor from UNAM, coordinated vast amounts of the events as both attested by the website for NAACL and on the ground participants.

Points Of Contact Between Global Industry, Research Outside of Global North & LATAM

Some of the interest from private industry and international universities reflects both a belief in the vast network of talent in Mexico City as well as shared problems. While large language models are both a puzzling and effective means towards greater applications of NLP and conversational AI, they are also difficult to explain and too crude to be applied with elegance towards more sophisticated settings. The probabilistic nature of output makes for the possibility of erroneous output given a conversational context, known as ‘hallucinations’.

In the case of Bloomberg, they have internally tested their own GPT model for finance applications with a paper published on the matter earlier this year. The model at the time of launch was the largest domain specific model in the world, concentrating in the finance domain, with refinements taking place internally.

Meanwhile, MBZAI’s university has built their own Large Language Models to execute applications in the Conversational AI space. The university has claimed the model outperforms Llama-2.

The spirit of both projects resonates well with presenters who seek to not necessarily discredit the hype around LLM’s, but to make strides in more specific applications with their considerations taking center stage. Particularly, the call to make strides outside of the corporate-academic nexus between American and European Industry and their regional powerhouse universities.

Highlights

A wide variety of research topics have been set to unravel in this conference. Ranging from zero-shot hierarchical table analysis using Augmented Language Models, presentations on fairness, safety, and multilinguality to discussions on biomedical entities in low-resource settings, the conference portfolio promises to offer an inspiring agglomeration of the latest AI and NLP research.

Forums to explain the need for grounded research in LLM applications was highlighted in a panel discussion conducted by both academic and industry experts based in Mexico City or somehow connected to regional network. Specifically, the need for regional experts to define goals beyond the usual metrics linked to performance and define projects that meet human needs.

One of the most comprehensive tutorials was given by Danae Sánchez who reviewed the first Bag of Word methods all the way through to LLM’s most modern incarnation, as well as the grounbreaking research conducted by her team at the University of Copenhagen. Using the PIXELS model, the team has managed to skip over the traditional NLP pipeline and train language models as if they were vision models. Surely, this aforementioned description does not do the team justice. Nonetheless, for now and to capture the moment, we leave her tweet below:

From the Hilton in Mexico City, the NAACL conference presents an incredible opportunity for scholars to showcase their original research. Among these distinguished contributors, attendees were eager to present papers that tackled various subjects. One of the highlights, “E5: Zero-shot Hierarchical Table Analysis using Augmented LLMs via Explain, Extract, Execute, Exhibit, and Extrapolate”, while the full program is available online.