Installation |
|
---|---|
Install text required python packages in conda or virtualenv environment |
|
Uninstall textrpp conda environment |
|
Initialize text required python packages |
|
Transform text to word embeddings |
|
Extract layers and aggregate them to word embeddings, for all character variables in a given dataframe. |
|
Change the names of the dimensions in the word embeddings. |
|
Extract layers of hidden states (word embeddings) for all character variables in a given dataframe. |
|
Select and aggregate layers of hidden states to form a word embeddings. |
|
Pre-trained dimension reduction (experimental) |
|
Applies word embeddings from a given decontextualized static space (such as from Latent Semantic Analyses) to all character variables |
|
Fine-tuning |
|
Task Adapted Pre-Training (experimental) |
|
Language Analysis Tasks |
|
Predict label and probability of a text using a pretrained classifier language model. (experimental) |
|
Predicts the words that will follow a specified text prompt. (experimental) |
|
Named Entity Recognition. (experimental) |
|
Summarize texts. (experimental) |
|
Question Answering. (experimental) |
|
Translation. (experimental) |
|
Zero Shot Classification (Experimental) |
|
Train word embeddings |
|
Train word embeddings to a numeric (ridge regression) or categorical (random forest) variable. |
|
Individually trains word embeddings from several text variables to several numeric or categorical variables. It is possible to have word embeddings from one text variable and several numeric/categprical variables; or vice verse, word embeddings from several text variables to one numeric/categorical variable. It is not possible to mix numeric and categorical variables. |
|
Train word embeddings to a numeric variable. |
|
Train word embeddings to a categorical variable using random forrest. |
|
Predict from word embeddings |
|
Predict scores or classification from, e.g., textTrain. |
|
Significance testing correlations If only y1 is provided a t-test is computed, between the absolute error from yhat1-y1 and yhat2-y1. |
|
Predict from several models, selecting the correct input |
|
Semantic similarities and distances |
|
Compute the semantic similarity between two text variables. |
|
Compute the semantic distance between two text variables. |
|
Compute semantic similarity scores between all combinations in a word embedding |
|
Compute semantic distance scores between all combinations in a word embedding |
|
Compute the semantic similarity between a text variable and a word norm (i.e., a text represented by one word embedding that represent a construct). |
|
Compute the semantic distance between a text variable and a word norm (i.e., a text represented by one word embedding that represent a construct/concept). |
|
Plot words in the word embedding space |
|
Compute Supervised Dimension Projection and related variables for plotting words. |
|
Plot words from textProjection() or textWordPrediction(). |
|
Plot words according to Supervised Dimension Projection. |
|
Compute predictions based on single words for plotting words. The word embeddings of single words are trained to predict the mean value associated with that word. P-values does NOT work yet. |
|
Compute semantic similarity score between single words' word embeddings and the aggregated word embedding of all words. |
|
Plot words according to semantic similarity to the aggregated word embedding. |
|
Compute 2 PCA dimensions of the word embeddings for individual words. |
|
Plot words according to 2-D plot from 2 PCA components. |
|
View or delete downloaded HuggingFace models in R |
|
Check downloaded, available models. |
|
Get the number of layers in a given model. |
|
Delete a specified model and model associated files. |
|
Miscellaneous |
|
Compute descriptive statistics of character variables. |
|
Tokenize according to different huggingface transformers |
|
Example Data |
|
Text and numeric data for 10 participants. |
|
Word embeddings for 4 text variables for 40 participants |
|
Word embeddings from textEmbedRawLayers function |
|
Example text and numeric data. |
|
Data for plotting a Dot Product Projection Plot. |
|
Example data for plotting a Semantic Centrality Plot. |
|
Example data for plotting a Principle Component Projection Plot. |