A core design goal is to facilitate the rendering of any data registered to a reference atlas. To this end,
brainrender facilitates the creation of 3D objects from many different types of data (e.g. cell locations, brain regions) within minimal need for the development of dedicated code. In addition,
brainrender is fully integrated with BrainGlobe's atlasAPI ensuring that you can use
brainrender with any atlas supported by the API with no need for any changes in your code.
The general workflow for any
brainrender visualization consists of just a few steps:
Load your data and generate a
Actor. This can be done using custom code, or with the dedicated
Actor classes provided by
brainrender which can be used to render most types of data.
Add your data to a
Render your scene, or use it to create animated videos.
To learn more in detail how to use
brainrender, keep reading this documentation and when you're ready check out the examples at the GitHub repository.
brainrender can be used with Jupyter notebooks, but some care must be used when doing that. Find more details here.