File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change 1515
1616import torch
1717import torch .distributed as dist
18- from torch ._six import inf
1918import torch .distributed as dist
2019
20+ try :
21+ from torch ._six import inf as inf
22+ except ModuleNotFoundError :
23+ from torch import inf as inf
24+
2125from deepspeed .utils import logger
2226from numpy import prod
2327
Original file line number Diff line number Diff line change 66from torch .distributed .distributed_c10d import _get_global_rank
77import torch .distributed as dist
88import math
9- from torch ._six import inf
109from torch .autograd import Variable
1110
1211import collections
1312
1413from deepspeed .runtime .fp16 .loss_scaler import LossScaler , DynamicLossScaler
15- from deepspeed .runtime .utils import see_memory_usage , is_model_parallel_parameter
14+ from deepspeed .runtime .utils import inf , see_memory_usage , is_model_parallel_parameter
1615from deepspeed .runtime .zero .config import ZERO_OPTIMIZATION_GRADIENTS
1716from deepspeed .ops .adam import DeepSpeedCPUAdam
1817from deepspeed .ops .op_builder import UtilsBuilder
Original file line number Diff line number Diff line change 1111from torch .distributed .distributed_c10d import _get_global_rank
1212import torch .distributed as dist
1313import math
14- from torch ._six import inf
1514from torch .autograd import Variable
1615
1716from deepspeed .utils .logging import logger
1817from deepspeed .runtime .fp16 .loss_scaler import LossScaler , DynamicLossScaler
19- from deepspeed .runtime .utils import see_memory_usage , is_model_parallel_parameter
18+ from deepspeed .runtime .utils import inf , see_memory_usage , is_model_parallel_parameter
2019from deepspeed .runtime .zero .partition_parameters import *
2120from deepspeed .runtime .zero .partition_parameters import _init_external_params
2221from deepspeed .runtime .zero .constants import ZERO_OPTIMIZATION_WEIGHTS
You can’t perform that action at this time.
0 commit comments