Skip to content

This repository provides the code for our paper "Adaptive Regularized Warped Gradient Descent Enhances Model Generalization and Meta-learning for Few-shot learning" submitted to Neurocomputing.

License

Notifications You must be signed in to change notification settings

susan-Rao/ARWarpGrad

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Adaptive Regularized Warped Gradient Descent Enhances Model Generalization and Meta-learning for Few-shot learning

This repository provides the code for our paper "Adaptive Regularized Warped Gradient Descent Enhances Model Generalization and Meta-learning for Few-shot learning" submitted to Neurocomputing.

Acknowledgement

The main structure of this code is based on [WarpGrad] (https://github.com/flennerhag/warpgrad). We sincerely thank the authors for sharing the codes for their great works.

Dependencies

Datasets:

Refers to WarpGard (https://openreview.net/forum?id=rkeiQlBFPB).

Train ARWarpGrad

cd src/omniglot
python main.py

The settings of hyperparameters can be changed in options.py.

Citation

Please cite our paper if you find the code or dataset useful for your research.

License

This library is licensed under the Apache 2.0 License.

About

This repository provides the code for our paper "Adaptive Regularized Warped Gradient Descent Enhances Model Generalization and Meta-learning for Few-shot learning" submitted to Neurocomputing.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages