You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This release includes general improvements to the library and new metrics within the NLP domain.
Natural language processing is arguably one of the most exciting areas of machine learning, with models such as BERT, ROBERTA, GPT-3 etc., really pushing what automated text translation, recognition, and generation systems are capable of.
With the introduction of these models, many metrics have been proposed that measure how well these models perform. TorchMetrics v0.5 includes 4 such metrics: BERT score, BLEU, ROUGE and WER.
num_thresholds argument in BinnedPrecisionRecallCurve
Fixed
Fixed bug where classification metrics with average='macro' would lead to wrong result if a class was missing (Fix metrics in macro average #303)
Fixed weighted, multi-class AUROC computation to allow for 0 observations of some class, as contribution to final AUROC is 0 (Weighted AUROC to omit empty classes #376)
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
[0.5.0] - 2021-08-09
This release includes general improvements to the library and new metrics within the NLP domain.
Natural language processing is arguably one of the most exciting areas of machine learning, with models such as BERT, ROBERTA, GPT-3 etc., really pushing what automated text translation, recognition, and generation systems are capable of.
With the introduction of these models, many metrics have been proposed that measure how well these models perform. TorchMetrics v0.5 includes 4 such metrics: BERT score, BLEU, ROUGE and WER.
Detail changes
Added
MetricTrackerwrapper metric for keeping track of the same metric over multiple epochs (metric tracker #238)nDCGmetric for target with values larger than 1 (Allow target nDCG metric to be integer larger than 1 #349)nDCGmetric (fix nDCG can not be called with negative relevance targets #378)Noneas reduction option inCosineSimilaritymetric (Add None as reduction option in CosineSimilarity #400)AveragePrecision(multilabel for AveragePrecision #386)Changed
psnrandssimfromfunctional.regression.*tofunctional.image.*(move functional psnr & ssim to image #382)image_gradientfromfunctional.image_gradientstofunctional.image.gradients(Move image gradient #381)R2Scorefromregression.r2scoretoregression.r2(cleaning & prune re-definine #371)torch.argmaxinstead oftorch.topkwhenk=1for better performance (Use argmax when topk=1 #419)Deprecated
r2score>>r2_scoreandkldivergence>>kl_divergenceinfunctional(cleaning & prune re-definine #371)bleu_scorefromfunctional.nlptofunctional.text.bleu(Added Blue Score the respective folders #360)Removed
thresholdhas to be in (0,1) range to support logit input (Allow threshold to be outside (0,1) domain #351, Remove remaining threshold checks #401)predscould not be bigger thannum_classesto support logit input (Remove check that preds value need to be smaller than num_classes #357)regression.psnrandregression.ssim(move functional psnr & ssim to image #382):functional.mean_relative_errornum_thresholdsargument inBinnedPrecisionRecallCurveFixed
average='macro'would lead to wrong result if a class was missing (Fix metrics in macro average #303)weighted,multi-classAUROC computation to allow for 0 observations of some class, as contribution to final AUROC is 0 (Weighted AUROC to omit empty classes #376)_forward_cacheand_computedattributes are also moved to the correct device if metric is moved (Move forward cache and computed to device #413)IoUmetric when usingignore_indexargument (fix ignore_index in the computation of IoU #328)Contributors
@BeyondTheProof, @Borda, @CSautier, @discort, @edwardclem, @gagan3012, @hugoperrin, @karthikrangasai, @paul-grundmann, @quancs, @rajs96, @SkafteNicki, @vatch123
If we forgot someone due to not matching commit email with GitHub account, let us know :]
This discussion was created from the release Text-related (NLP) metrics.
Beta Was this translation helpful? Give feedback.
All reactions