torch.Storage¶
A torch.Storage is a contiguous, one-dimensional array of a single
data type.
Every torch.Tensor has a corresponding storage of the same data type.
-
class
torch.DoubleStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.FloatStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.HalfStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.LongStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.IntStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.ShortStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.CharStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.ByteStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.BoolStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.BFloat16Storage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.ComplexDoubleStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.ComplexFloatStorage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.QUInt8Storage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.QInt8Storage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.QInt32Storage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-
-
class
torch.QUInt4x2Storage(*args, **kwargs)[source]¶ -
bfloat16()¶ Casts this storage to bfloat16 type
-
bool()¶ Casts this storage to bool type
-
byte()¶ Casts this storage to byte type
-
char()¶ Casts this storage to char type
-
clone()¶ Returns a copy of this storage
-
complex_double()¶ Casts this storage to complex double type
-
complex_float()¶ Casts this storage to complex float type
-
copy_(source, non_blocking=None)¶
-
cpu()¶ Returns a CPU copy of this storage if it’s not already on the CPU
-
cuda(device=None, non_blocking=False, **kwargs)¶
-
data_ptr()¶
-
property
device¶
-
double()¶ Casts this storage to double type
-
element_size()¶
-
fill_(value)¶
-
float()¶ Casts this storage to float type
-
classmethod
from_buffer(*args, **kwargs)¶
-
classmethod
from_file(filename, shared, size)¶
-
get_device()¶
-
half()¶ Casts this storage to half type
-
int()¶ Casts this storage to int type
-
property
is_cuda¶
-
is_pinned()¶
-
is_sparse= False¶
-
long()¶ Casts this storage to long type
-
nbytes()¶
-
pickle_storage_type()¶
-
pin_memory()¶ Coppies the storage to pinned memory, if it’s not already pinned.
-
resize_(size)¶
Moves the storage to shared memory.
This is a no-op for storages already in shared memory and for CUDA storages, which do not need to be moved for sharing across processes. Storages in shared memory cannot be resized.
Returns: self
-
short()¶ Casts this storage to short type
-
size()¶
-
tolist()¶ Returns a list containing the elements of this storage
-
type(dtype=None, non_blocking=False)¶
-