Evaluation

The module includes performance metrics for neural graph embeddings models, along with model selection routines, negatives generation, and an implementation of the learning-to-rank-based evaluation protocol used in literature.

Metrics

Learning-to-rank metrics to evaluate the performance of neural graph embedding models.

rank_score(y_true, y_pred[, pos_lab]) Rank of a triple
mrr_score(ranks) Mean Reciprocal Rank (MRR)
mr_score(ranks) Mean Rank (MR)
hits_at_n_score(ranks, n) Hits@N

Negatives Generation

Negatives generation routines. These are corruption strategies based on the Local Closed-World Assumption (LCWA).

generate_corruptions_for_eval(X, …[, …]) Generate corruptions for evaluation.
generate_corruptions_for_fit(X[, …]) Generate corruptions for training.

Evaluation & Model Selection

Functions to evaluate the predictive power of knowledge graph embedding models, and routines for model selection.

evaluate_performance(X, model[, …]) Evaluate the performance of an embedding model.
select_best_model_ranking(model_class, X, …) Model selection routine for embedding models.

Helper Functions

Utilities and support functions for evaluation procedures.

train_test_split_no_unseen(X[, test_size, seed]) Split into train and test sets.
create_mappings(X) Create string-IDs mappings for entities and relations.
to_idx(X, ent_to_idx, rel_to_idx) Convert statements (triples) into integer IDs.