NLP

From UFRC
Revision as of 19:57, 8 May 2024 by Zhao.qian (talk | contribs)
Jump to navigation Jump to search

Natural language processing software and resources on HiperGator include several different software environments and examples with Nvidia Megatron and other software. NLP is involved in many other fields of AI, such as image recognition. Research Computing can help with language modeling for knowledge exploration, measurement, classification, summarization, conversational AI, or other uses via support requests or consulting.

Environment Modules for NLP

  • nlp: module load nlp provides a Python environment with pytorch, torchtext, nltk, Spacy, transformers, sentence-transformers, Flair, BERTopic for topic modeling, sentencepiece, RAPIDSai for data processing and machine learning algorithms, gensim, scikit-learn, and more.


  • ngc-pytorch: module load ngc-pytorch will provides a singularity container Python environment with pytorch including the Nvidia Apex optimizers required for Megatron-LM. Research computing has pretrained, large parameter Megatron language models available to HiperGator users. See /data/ai/examples/nlp or AI_Examples for more information.


  • Flair NLP: See FlairNLP for more information.


  • nemo: module load nemo will provide a singularity container environment with Python and Nvidia NeMo. NeMo has NLP task training, plus speech-to-text and text-to-speech models, and the option to apply your own pretrained Megatron language models.


  • pytorch or tensorflow: Note, use module spider pytorch or tensorflow to list the version we have available. If the nlp environments or these environments do not have libraries you require, you made need to create a Conda environment. See Conda and Managing_Python_environments_and_Jupyter_kernels for more details.


  • spark-nlp: See our Spark help doc to start a Spark cluster. Spark-nlp Python module is available in tensorflow/2.4.1.


  • parlai: Conversational AI framework by Facebook, includes a wide variety of models from 110M to 9B parameters.

Large Language Models

Many large models are available for open source download, although may require different software frameworks or end user license agreements. Starter LLMs trained using Megatron-LM are available in the examples and reference data folder. These models can be applied as is, trained more, or fine-tuned. Starter models include a 20B GPT and a 9B parameter BERT. Please create a help ticket for more information.


Examples and Reference Data

Please see /data/ai/ folder, AI_Examples, and AI_Reference_Datasets for helpful resources. Notebooks and batch scripts cover everything from pretraining and inferencing to summarization, information extraction, and topic modeling. Addition reference data, including benchmarks such as the popular superglue, are already available in /data/ai/benchmarks/nlp.