Introducing LibAUC 1.4.0
We are releasing a new version of LibAUC at long last!
New Features
- Added support for optimizing mAP at top-k positions by setting
mAPLoss(top_k=k). - Integrated 8 commonly used graph neural networks into the library, i.e.,
GCN,DeeperGCN,GIN,GINE,GAT,MPNN,GraphSAGEandPNA. - Introduced cosine gamma schedule in self-supervised contrastive learning, which has been demonstrated in our recent work. To use cosine gamma schedule, we can set
GCLoss('unimodal', gamma_schedule = 'cosine')andGCLoss('bimodal', gamma_schedule = 'cosine'). - Added
datasets.webdataset, which supports loading image-text dataset stored in WebDataset format.
Tutorial Update
- Provided a tutorial to show how to train a GNN model by optimizing AUPRC with our novel APLoss and SOAP optimizer on a binary molecule classification task from ogbg-molpcba dataset.
- Provided a tutorial on optimizing global contrastive loss with SogCLR and cosine gamma schedule.
- Updated previous tutorials by adding a recommended pretraining part to significantly boost models’ performance.
What's Changed
- A new page was added on LibAUC website to show our active users.
- Added
return_indexargument to the classdatasets.ImageFolderand added the explanation ofreturn_indexto our documentation. - Changed the default
etavalue to0.1inpAUC_CVaR_Loss. Previously, the value is inconsistent with the documentation. - Fixed the learning rate setting in SogCLR tutorial. Previously, the learning rate is incorrectly passed into the function.
- Fixed the circular import error when importing from
metrics#57.
Acknowledgment
Team: Gang Li, Xiyuan Wei, Siqi Guo, Zihao Qiu, Tianbao Yang (Advisor)