__init__.py 1.4 KB

1234567891011121314151617181920212223242526272829303132333435363738
  1. """
  2. :mod:`torch.optim` is a package implementing various optimization algorithms.
  3. Most commonly used methods are already supported, and the interface is general
  4. enough, so that more sophisticated ones can also be easily integrated in the
  5. future.
  6. """
  7. from . import lr_scheduler, swa_utils
  8. from .adadelta import Adadelta
  9. from .adagrad import Adagrad
  10. from .adam import Adam
  11. from .adamax import Adamax
  12. from .adamw import AdamW
  13. from .asgd import ASGD
  14. from .lbfgs import LBFGS
  15. from .nadam import NAdam
  16. from .optimizer import Optimizer
  17. from .radam import RAdam
  18. from .rmsprop import RMSprop
  19. from .rprop import Rprop
  20. from .sgd import SGD
  21. from .sparse_adam import SparseAdam
  22. del adadelta # type: ignore[name-defined] # noqa: F821
  23. del adagrad # type: ignore[name-defined] # noqa: F821
  24. del adam # type: ignore[name-defined] # noqa: F821
  25. del adamw # type: ignore[name-defined] # noqa: F821
  26. del sparse_adam # type: ignore[name-defined] # noqa: F821
  27. del adamax # type: ignore[name-defined] # noqa: F821
  28. del asgd # type: ignore[name-defined] # noqa: F821
  29. del sgd # type: ignore[name-defined] # noqa: F821
  30. del radam # type: ignore[name-defined] # noqa: F821
  31. del rprop # type: ignore[name-defined] # noqa: F821
  32. del rmsprop # type: ignore[name-defined] # noqa: F821
  33. del optimizer # type: ignore[name-defined] # noqa: F821
  34. del nadam # type: ignore[name-defined] # noqa: F821
  35. del lbfgs # type: ignore[name-defined] # noqa: F821