Functions
FileCatalogFactory (name, filetype[, examples]) |
Factory method to create a CatalogSource that uses a subclass of nbodykit.io.base.FileType to read data from disk. |
Classes
BigFileCatalog (*args, **kwargs) |
A CatalogSource that uses BigFile to read data from disk. |
BinaryCatalog (*args, **kwargs) |
A CatalogSource that uses BinaryFile to read data from disk. |
CSVCatalog (*args, **kwargs) |
A CatalogSource that uses CSVFile to read data from disk. |
FITSCatalog (*args, **kwargs) |
A CatalogSource that uses FITSFile to read data from disk. |
FileCatalogBase (filetype[, args, kwargs, …]) |
Base class to create a source of particles from a single file, or multiple files, on disk. |
Gadget1Catalog (*args, **kwargs) |
A CatalogSource that uses Gadget1File to read data from disk. |
HDFCatalog (*args, **kwargs) |
A CatalogSource that uses HDFFile to read data from disk. |
TPMBinaryCatalog (*args, **kwargs) |
A CatalogSource that uses TPMBinaryFile to read data from disk. |
nbodykit.source.catalog.file.
FileCatalogFactory
(name, filetype, examples=None)[source]¶Factory method to create a CatalogSource
that uses a subclass of nbodykit.io.base.FileType
to read
data from disk.
Parameters: |
|
---|---|
Returns: | the |
Return type: | subclass of |
nbodykit.source.catalog.file.
FileCatalogBase
(filetype, args=(), kwargs={}, comm=None, use_cache=False)[source]¶Base class to create a source of particles from a single file, or multiple files, on disk.
Files of a specific type should be subclasses of this class.
Parameters: |
|
---|
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
get_hardcolumn
(col)[source]¶Return a column from the underlying file source.
Columns are returned as dask arrays.
hardcolumns
¶The union of the columns in the file and any transformed columns.
nbodykit.source.catalog.file.
CSVCatalog
(*args, **kwargs)¶A CatalogSource that uses CSVFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Examples
Please see the documentation for examples.
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
BinaryCatalog
(*args, **kwargs)¶A CatalogSource that uses BinaryFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Examples
Please see the documentation for examples.
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
BigFileCatalog
(*args, **kwargs)¶A CatalogSource that uses BigFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Examples
Please see the documentation for examples.
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
HDFCatalog
(*args, **kwargs)¶A CatalogSource that uses HDFFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Examples
Please see the documentation for examples.
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
TPMBinaryCatalog
(*args, **kwargs)¶A CatalogSource that uses TPMBinaryFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
Gadget1Catalog
(*args, **kwargs)¶A CatalogSource that uses Gadget1File
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |
nbodykit.source.catalog.file.
FITSCatalog
(*args, **kwargs)¶A CatalogSource that uses FITSFile
to read data from disk.
Multiple files can be read at once by supplying a list of file
names or a glob asterisk pattern as the path
argument. See
Reading Multiple Data Files at Once for examples.
Parameters: |
|
---|
Examples
Please see the documentation for examples.
Attributes
Index |
The attribute giving the global index rank of each particle in the list. |
attrs |
A dictionary storing relevant meta-data about the CatalogSource. |
columns |
All columns in the CatalogSource, including those hard-coded into the class’s defintion and override columns provided by the user. |
csize |
The total, collective size of the CatalogSource, i.e., summed across all ranks. |
hardcolumns |
The union of the columns in the file and any transformed columns. |
size |
The number of objects in the CatalogSource on the local rank. |
use_cache |
If set to True , use the built-in caching features of dask to cache data in memory. |
Methods
Selection () |
A boolean column that selects a subset slice of the CatalogSource. |
Value () |
When interpolating a CatalogSource on to a mesh, the value of this array is used as the Value that each particle contributes to a given mesh cell. |
Weight () |
The column giving the weight to use for each particle on the mesh. |
compute (*args, **kwargs) |
Our version of dask.compute() that computes multiple delayed dask collections at once. |
copy () |
Return a shallow copy of the object, where each column is a reference of the corresponding column in self . |
get_hardcolumn (col) |
Return a column from the underlying file source. |
gslice (start, stop[, end, redistribute]) |
Execute a global slice of a CatalogSource. |
make_column (array) |
Utility function to convert an array-like object to a dask.array.Array . |
read (columns) |
Return the requested columns as dask arrays. |
save (output, columns[, datasets, header]) |
Save the CatalogSource to a bigfile.BigFile . |
sort (keys[, reverse, usecols]) |
Return a CatalogSource, sorted globally across all MPI ranks in ascending order by the input keys. |
to_mesh ([Nmesh, BoxSize, dtype, interlaced, …]) |
Convert the CatalogSource to a MeshSource, using the specified parameters. |
view ([type]) |
Return a “view” of the CatalogSource object, with the returned type set by type . |