A brief overview of brainrender`s functionality
This documentation is for the new brainrender 2.0!
If are getting started with brainrender, carry on. But if you have been using previous versions of the software, you might want to check the guide we've put together to transition to the new brainrender.
This section explains the logic behind how
brainrenderworks. If you want more details about which classes/methods are part of each script, and what they do, check the automatically generated docs here.
A core design goal is to facilitate the rendering of any data registered to a reference atlas. To this end,
brainrenderfacilitates the creation of 3D objects from many different types of data (e.g. cell locations, brain regions) within minimal need for the development of dedicated code. In addition,
brainrenderis fully integrated with BrainGlobe's atlasAPI ensuring that you can use
brainrenderwith any atlas supported by the API with no need for any changes in your code.
Overview of brainrender's workflow
The general workflow for any
brainrendervisualization consists of just a few steps:
- 1.Load your data and generate a
Actor. This can be done using custom code, or with the dedicated
Actorclasses provided by
brainrenderwhich can be used to render most types of data.
To learn more in detail how to use
brainrender, keep reading this documentation and when you're ready check out the examples at the GitHub repository.
brainrendercan be used with Jupyter notebooks, but some care must be used when doing that. Find more details here.
If you've found
brainrenderuseful in your work, please cite