Skip to content
This repository was archived by the owner on Feb 26, 2025. It is now read-only.

Refactor edge_index::resolve.#314

Merged
1uc merged 2 commits intomasterfrom
1uc/refactor-edge_index-resolve
Nov 29, 2023
Merged

Refactor edge_index::resolve.#314
1uc merged 2 commits intomasterfrom
1uc/refactor-edge_index-resolve

Conversation

@1uc
Copy link
Collaborator

@1uc 1uc commented Nov 14, 2023

The point is to split reading of the dataset into a separate function, and then
make resolve safe for collective IO (assuming the newly introduced function
is).

The overload for reading a single nodeID is removed as it's unused now.

@1uc
Copy link
Collaborator Author

1uc commented Nov 14, 2023

This is a preparatory PR for #307

@1uc 1uc marked this pull request as ready for review November 14, 2023 10:24
@1uc 1uc force-pushed the 1uc/refactor-edge_index-resolve branch from be7ba48 to 09bb030 Compare November 22, 2023 14:50
Copy link
Contributor

@mgeplf mgeplf left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

No complaints on my side, looks good.

@1uc 1uc force-pushed the 1uc/refactor-edge_index-resolve branch from 72d102e to c05f24f Compare November 29, 2023 07:23
1uc added 2 commits November 29, 2023 08:31
The point is to split reading of the dataset into a separate function, and then
make `resolve` safe for collective IO (assuming the newly introduced function
is).

The overload for reading a single `nodeID` is removed as it's unused now.
@1uc 1uc force-pushed the 1uc/refactor-edge_index-resolve branch from c05f24f to 7634c20 Compare November 29, 2023 07:32
@1uc 1uc merged commit a7d1453 into master Nov 29, 2023
@1uc 1uc deleted the 1uc/refactor-edge_index-resolve branch November 29, 2023 08:02
1uc added a commit that referenced this pull request Nov 30, 2023
The point is to split reading of the dataset into a separate function, and then
make `resolve` safe for collective IO (assuming the newly introduced function
is).

The overload for reading a single `nodeID` is removed as it's unused now.
WeinaJi pushed a commit to BlueBrain/neurodamus that referenced this pull request Jan 29, 2024
## Context
When using `WholeCell` load-balancing, the access pattern when reading
parameters during synapse creation is extremely poor and is the main
reason why we see long (10+ minutes) periods of severe performance
degradation of our parallel filesystem when running slightly larger
simulations on BB5.

Using Darshan and several PoCs we established that the time required to
read these parameters can be reduced by more than 8x and IOps can be
reduced by over 1000x when using collective MPI-IO. Moreover, the
"waiters" where reduced substantially as well. See BBPBGLIB-1070.

Following those finding we concluded that neurodamus would need to use
collective MPI-IO in the future.

We've implemented most of the required changes directly in libsonata
allowing others to benefit from the same optimizations should the need
arise. See,
BlueBrain/libsonata#309
BlueBrain/libsonata#307

and preparatory work:
BlueBrain/libsonata#315
BlueBrain/libsonata#314
BlueBrain/libsonata#298 

By instrumenting two simulations (SSCX and reduced MMB) we concluded
that neurodamus was almost collective. However, certain attributes where
read in different order on different MPI ranks. Maybe due to salting
hashes differently on different MPI ranks.

## Scope
This PR enables neurodamus to use collective IO for the simulation
described above.

## Testing
<!--
Please add a new test under `tests`. Consider the following cases:

1. If the change is in an independent component (e.g, a new container
type, a parser, etc) a bare unit test should be sufficient. See e.g.
`tests/test_coords.py`
2. If you are fixing or adding components supporting a scientific use
case, affecting node or synapse creation, etc..., which typically rely
on Neuron, tests should set up a simulation using that feature,
instantiate neurodamus, **assess the state**, run the simulation and
check the results are as expected.
    See an example at `tests/test_simulation.py#L66`
-->
We successfully ran the reduced MMB simulation, but since SSCX hasn't
been converted to SONATA, we can't run that simulation.

## Review
* [x] PR description is complete
* [x] Coding style (imports, function length, New functions, classes or
files) are good
* [ ] Unit/Scientific test added
* [ ] Updated Readme, in-code, developer documentation

---------

Co-authored-by: Luc Grosheintz <luc.grosheintz@gmail.ch>
WeinaJi pushed a commit to BlueBrain/neurodamus that referenced this pull request Oct 14, 2024
## Context
When using `WholeCell` load-balancing, the access pattern when reading
parameters during synapse creation is extremely poor and is the main
reason why we see long (10+ minutes) periods of severe performance
degradation of our parallel filesystem when running slightly larger
simulations on BB5.

Using Darshan and several PoCs we established that the time required to
read these parameters can be reduced by more than 8x and IOps can be
reduced by over 1000x when using collective MPI-IO. Moreover, the
"waiters" where reduced substantially as well. See BBPBGLIB-1070.

Following those finding we concluded that neurodamus would need to use
collective MPI-IO in the future.

We've implemented most of the required changes directly in libsonata
allowing others to benefit from the same optimizations should the need
arise. See,
BlueBrain/libsonata#309
BlueBrain/libsonata#307

and preparatory work:
BlueBrain/libsonata#315
BlueBrain/libsonata#314
BlueBrain/libsonata#298 

By instrumenting two simulations (SSCX and reduced MMB) we concluded
that neurodamus was almost collective. However, certain attributes where
read in different order on different MPI ranks. Maybe due to salting
hashes differently on different MPI ranks.

## Scope
This PR enables neurodamus to use collective IO for the simulation
described above.

## Testing
<!--
Please add a new test under `tests`. Consider the following cases:

1. If the change is in an independent component (e.g, a new container
type, a parser, etc) a bare unit test should be sufficient. See e.g.
`tests/test_coords.py`
2. If you are fixing or adding components supporting a scientific use
case, affecting node or synapse creation, etc..., which typically rely
on Neuron, tests should set up a simulation using that feature,
instantiate neurodamus, **assess the state**, run the simulation and
check the results are as expected.
    See an example at `tests/test_simulation.py#L66`
-->
We successfully ran the reduced MMB simulation, but since SSCX hasn't
been converted to SONATA, we can't run that simulation.

## Review
* [x] PR description is complete
* [x] Coding style (imports, function length, New functions, classes or
files) are good
* [ ] Unit/Scientific test added
* [ ] Updated Readme, in-code, developer documentation

---------

Co-authored-by: Luc Grosheintz <luc.grosheintz@gmail.ch>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants