Informal to formal dataset mask MLM
-
Updated
Oct 11, 2024 - Python
Informal to formal dataset mask MLM
Calculate perplexity on a text with pre-trained language models. Support MLM (eg. DeBERTa), recurrent LM (eg. GPT3), and encoder-decoder LM (eg. Flan-T5).
TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis
translatorlab: a machine translation tool that uses artificial intelligence models to provide accurate and fast translations between different languages
CodeGen is a family of open-source model for program synthesis. Trained on TPU-v4. Competitive with OpenAI Codex.
[NeurIPS'22 Spotlight] A Contrastive Framework for Neural Text Generation
Embeddings: State-of-the-art Text Representations for Natural Language Processing tasks, an initial version of library focus on the Polish Language
This is the official code for the paper CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning (NeurIPS22).
Simple next word prediction model from scratch, implemented using only numpy.
Code for "Semi-supervised Formality Style Transfer using Language Model Discriminator and Mutual Information Maximization"
A 78.5% word sense disambiguator based on Transformers and RoBERTa (PyTorch)
高性能小模型测评 Shared Tasks in NLPCC 2020. Task 1 - Light Pre-Training Chinese Language Model for NLP Task
Add a description, image, and links to the languagemodel topic page so that developers can more easily learn about it.
To associate your repository with the languagemodel topic, visit your repo's landing page and select "manage topics."