中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
-
Updated
May 23, 2024 - Python
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
A project demonstrating the use of Large Language Models (LLMs) for text classification using the RoBERTa model.
BERT classification model for processing texts longer than 512 tokens. Text is first divided into smaller chunks and after feeding them to BERT, intermediate results are pooled. The implementation allows fine-tuning.
Official implementation of the papers "GECToR – Grammatical Error Correction: Tag, Not Rewrite" (BEA-20) and "Text Simplification by Tagging" (BEA-21)
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
Classification of medical texts to differentiate between human medical and veterinary subjects.
The programming environment »Open Roberta Lab« by Fraunhofer IAIS enables children and adolescents to program robots. A variety of different programming blocks are provided to program motors and sensors of the robot. Open Roberta Lab uses an approach of graphical programming so that beginners can seamlessly start coding. As a cloud-based applica…
Stylometry approach detecting writing patterns and changings using NLTK, XML-roBERTa, Gensim topic modelling and unsupervised-PCA learning
news-please - an integrated web crawler and information extractor for news that just works
The Role of Model Architecture and Scale in Predicting Molecular Properties: Insights from Fine-Tuning RoBERTa, BART, and LLaMA
Happy Transformer makes it easy to fine-tune and perform inference with NLP Transformer models.
Plugin repository to perform sentiment analysis on comments in Kotlin files
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
ACM RecSys Challenge 2018 - Track2Vec and Artist2Vec for Spotify playlist continuation
This study aims to investigate the effectiveness of three Transformers (BERT, RoBERTa, XLNet) in handling data sparsity and cold start problems in the recommender system. We present a Transformer-based hybrid recommender system that predicts missing ratings and ex- tracts semantic embeddings from user reviews to mitigate the issues.
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
Simple State-of-the-Art BERT-Based Sentence Classification with Keras / TensorFlow 2. Built with HuggingFace's Transformers.
Add a description, image, and links to the roberta topic page so that developers can more easily learn about it.
To associate your repository with the roberta topic, visit your repo's landing page and select "manage topics."