Difference between revisions of "NLP"
Moskalenko (talk | contribs) |
|||
Line 25: | Line 25: | ||
*'''parlai:''' Conversational AI framework by Facebook, includes a wide variety of models from 110M to 9B parameters. | *'''parlai:''' Conversational AI framework by Facebook, includes a wide variety of models from 110M to 9B parameters. | ||
+ | ==Large Language Models== | ||
+ | Few large models are available for open source download, and the data quality of those models may be low or unknown. Starter LLMs trained using Megatron-LM are available in the examples and reference data folder. These models can be applied as is, trained more, or fine-tuned. Starter models include a 20B and 40B parameter GPTs and a 9B parameter BERT. Please contact ericeric@ufl.edu for more information. | ||
==Examples and Reference Data== | ==Examples and Reference Data== | ||
Please see <code>/data/ai/</code> folder, [[AI_Examples]], and [[AI_Reference_Datasets]] for helpful resources. Notebooks and batch scripts cover everything from pretraining and inferencing to summarization, information extraction, and topic modeling. Addition reference data, including benchmarks such as the popular [https://super.gluebenchmark.com/tasks superglue], are already available in <code>/data/ai/benchmarks/nlp</code>. | Please see <code>/data/ai/</code> folder, [[AI_Examples]], and [[AI_Reference_Datasets]] for helpful resources. Notebooks and batch scripts cover everything from pretraining and inferencing to summarization, information extraction, and topic modeling. Addition reference data, including benchmarks such as the popular [https://super.gluebenchmark.com/tasks superglue], are already available in <code>/data/ai/benchmarks/nlp</code>. |
Revision as of 14:31, 9 June 2023
Description
This page describes natural language processing software and resources on HiperGator. NLP is involved in many other fields of AI, such as image recognition. Research Computing can help with language modeling for knowledge exploration, measurement, classification, summarization, conversational AI, or other uses via support requests or consulting.
Environment Modules for NLP
- nlp:
module load nlp
provides a Python environment with pytorch, torchtext, nltk, Spacy, transformers, sentence-transformers, Flair, BERTopic for topic modeling, sentencepiece, RAPIDSai for data processing and machine learning algorithms, gensim, scikit-learn, and more.
- ngc-pytorch:
module load ngc-pytorch
will provides a singularity container Python environment with pytorch including the Nvidia Apex optimizers required for Megatron-LM. Research computing has pretrained, large parameter Megatron language models available to HiperGator users. See /data/ai/examples/nlp or AI_Examples for more information.
- Flair NLP: See FlairNLP for more information.
- nemo:
module load nemo
will provide a singularity container environment with Python and Nvidia NeMo. NeMo has NLP task training, plus speech-to-text and text-to-speech models, and the option to apply your own pretrained Megatron language models.
- pytorch or tensorflow: Note, use
module spider pytorch
ortensorflow
to list the version we have available. If the nlp environments or these environments do not have libraries you require, you made need to create a Conda environment. See Conda and Managing_Python_environments_and_Jupyter_kernels for more details.
- spark-nlp: See our Spark help doc to start a Spark cluster. Spark-nlp Python module is available in tensorflow/2.4.1.
- parlai: Conversational AI framework by Facebook, includes a wide variety of models from 110M to 9B parameters.
Large Language Models
Few large models are available for open source download, and the data quality of those models may be low or unknown. Starter LLMs trained using Megatron-LM are available in the examples and reference data folder. These models can be applied as is, trained more, or fine-tuned. Starter models include a 20B and 40B parameter GPTs and a 9B parameter BERT. Please contact ericeric@ufl.edu for more information.
Examples and Reference Data
Please see /data/ai/
folder, AI_Examples, and AI_Reference_Datasets for helpful resources. Notebooks and batch scripts cover everything from pretraining and inferencing to summarization, information extraction, and topic modeling. Addition reference data, including benchmarks such as the popular superglue, are already available in /data/ai/benchmarks/nlp
.