⛔ [DEPRECATED] Adapt Transformer-based language models to new text domains
-
Updated
Feb 21, 2024 - Jupyter Notebook
⛔ [DEPRECATED] Adapt Transformer-based language models to new text domains
We use phonetics as a feature to create a joint semantic-phonetic embedding and improve the neural machine translation between Chinese and Japanese. 🥳
This project leverages deep learning transformers to classify YouTube comments into six distinct emotions.
Hugging Face Transformers offer a powerful framework for state-of-the-art NLP, with the Pipeline API for easy inference, Tokenization for efficient preprocessing, and Quantization for optimized deployment.
Sesta attività di Big Data Analytics
Add a description, image, and links to the huggingface-tokenizers topic page so that developers can more easily learn about it.
To associate your repository with the huggingface-tokenizers topic, visit your repo's landing page and select "manage topics."