Code and data for the EMNLP 2020 paper: "Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank"
-
Updated
Feb 10, 2023 - Python
Code and data for the EMNLP 2020 paper: "Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank"
Match celebrity users with their respective tweets by making use of Semantic Textual Similarity on over 900+ celebrity users' 2.5 million+ scraped tweets utilizing SBERT, streamlit, tweepy and FastAPI
In this project, we compared Spanish BERT and Multilingual BERT in the Sentiment Analysis task.
Cross-lingual misinformation detection
Language Model Decomposition: Quantifying the Dependency and Correlation of Language Models
Align Parallel Sentence of 104 Languages with the help of mBERT and LaBSE
Codes for the short essay project for Mphil TAL by Yingjia Wan in 2023.
Repository for the paper titled: "When is BERT Multilingual? Isolating Crucial Ingredients for Cross-lingual Transfer"
submission from team import-winning-model
This work was done as part of Kaggle InClass competition hosted by Sharechat. We had to develop AI solutions for predicting abusive comments posted on the Moj app in 10+ languages given natural language data and user context data.
This repository contains a number of experiments with Multi Lingual Transformer models (Multi-Lingual BERT, DistilBERT, XLM-RoBERTa, mT5 and ByT5) focussed on the Dutch language.
Zero-shot and Translation Experiments on XQuAD, MLQA and TyDiQA
By using the hypothesis of historical linguistics, we found a way to improve the performance of multilingual transformers with limited amount of data
CS224n Homeworks and final project
Add a description, image, and links to the multilingual-bert topic page so that developers can more easily learn about it.
To associate your repository with the multilingual-bert topic, visit your repo's landing page and select "manage topics."