Natural language processing software and resources on HiperGator include several different software environments and examples with Nvidia Megatron and other software. NLP is involved in many other fields of AI, such as image recognition. Research Computing can help with language modeling for knowledge exploration, measurement, classification, summarization, conversational AI, or other uses via support requests or consulting.
Environment Modules for NLP
module load nlpprovides a Python environment with pytorch, torchtext, nltk, Spacy, transformers, sentence-transformers, Flair, BERTopic for topic modeling, sentencepiece, RAPIDSai for data processing and machine learning algorithms, gensim, scikit-learn, and more.
module load ngc-pytorchwill provides a singularity container Python environment with pytorch including the Nvidia Apex optimizers required for Megatron-LM. Research computing has pretrained, large parameter Megatron language models available to HiperGator users. See /data/ai/examples/nlp or AI_Examples for more information.
- Flair NLP: See FlairNLP for more information.
module load nemowill provide a singularity container environment with Python and Nvidia NeMo. NeMo has NLP task training, plus speech-to-text and text-to-speech models, and the option to apply your own pretrained Megatron language models.
- pytorch or tensorflow: Note, use
module spider pytorchor
tensorflowto list the version we have available. If the nlp environments or these environments do not have libraries you require, you made need to create a Conda environment. See Conda and Managing_Python_environments_and_Jupyter_kernels for more details.
- spark-nlp: See our Spark help doc to start a Spark cluster. Spark-nlp Python module is available in tensorflow/2.4.1.
- parlai: Conversational AI framework by Facebook, includes a wide variety of models from 110M to 9B parameters.
Large Language Models
Many large models are available for open source download, although may require different software frameworks or end user license agreements. Starter LLMs trained using Megatron-LM are available in the examples and reference data folder. These models can be applied as is, trained more, or fine-tuned. Starter models include a 20B GPT and a 9B parameter BERT. Please create a help ticket for more information.
Examples and Reference Data
/data/ai/ folder, AI_Examples, and AI_Reference_Datasets for helpful resources. Notebooks and batch scripts cover everything from pretraining and inferencing to summarization, information extraction, and topic modeling. Addition reference data, including benchmarks such as the popular superglue, are already available in