You can use boolean logic (e.g. AND/OR/NOT) for complex search queries. For more help and examples, see the search documentation.
Search by package name:
my-package (implicit)
name:my-package (explicit)
Search by package filename:
filename:my-package.ext
Search by package tag:
tag:latest
Search by package version:
version:1.0.0
prerelease:true (prereleases)
prerelease:false (no prereleases)
Search by package architecture:
architecture:x86_64
Search by package distribution:
distribution:el
Search by package license:
license:MIT
Search by package format:
format:deb
Search by package status:
status:in_progress
Search by package file checksum:
checksum:5afba
Search by package security status:
severity:critical
Search by package vulnerabilities:
vulnerabilities:>1
vulnerabilities:<1000
Search by # of package downloads:
downloads:>8
downloads:<100
Search by package type:
type:binary
type:source
Search by package size (bytes):
size:>50000
size:<10000
Search by dependency name/version:
dependency:log4j
dependency:log4j=1.0.0
dependency:log4j>1.0.0
Search by uploaded date:
uploaded:>"1 day ago"
uploaded:<"August 14, 2022 EST"
Search by entitlement token (identifier):
entitlement:3lKPVJPosCsY
Search by policy violation:
policy_violated:true
deny_policy_violated:true
license_policy_violated:true
vulnerability_policy_violated:true
Search by repository:
repository:repo-name
Search by last download date:
last_downloaded:<"30 days ago"
last_downloaded:>"August 14, 2022 EST"
Search queries for all Debian-specific (and related) package types
Search by component:
deb_component:unstable
Search queries for all Maven-specific (and related) package types
Search by group ID:
maven_group_id:org.apache
Search queries for all Docker-specific (and related) package types
Search by image digest:
docker_image_digest:sha256:7c5..6d4
(full hashref only)
Search by layer digest:
docker_layer_digest:sha256:4c4..ae4
(full hashref only)
Search queries for all Generic-specific package types
Search by file path:
generic_filepath:path/to/file.txt
Search by directory:
generic_directory:path/to
Field type modifiers (depending on the type, you can influence behaviour)
For all queries, you can use:
~foo for negation
For string queries, you can use:
^foo to anchor to start of term
foo$ to anchor to end of term
foo*bar for fuzzy matching
For number/date or version queries, you can use:
>foo for values greater than
>=foo for values greater / equal
<foo for values less than
<=foo for values less / equal
Need a secure and centralised artifact repository to deliver Alpine,
Cargo,
CocoaPods,
Composer,
Conan,
Conda,
CRAN,
Dart,
Debian,
Docker,
Generic,
Go,
Helm,
Hex,
HuggingFace,
LuaRocks,
Maven,
npm,
NuGet,
P2,
Python,
RedHat,
Ruby,
Swift,
Terraform,
Vagrant,
VSX,
Raw & More packages?
Cloudsmith is the new standard in Package / Artifact Management and Software Distribution.
With support for all major package formats, you can trust us to manage your software supply chain.
brainlit
0.2.0
One-liner (summary)
Description
# Brainlit []() [](https://travis-ci.com/neurodata/brainlit) [](https://badge.fury.io/py/brainlit) [](https://github.com/psf/black) [](https://codecov.io/gh/neurodata/brainlit) [](https://opensource.org/licenses/Apache-2.0) This repository is a container of methods that Neurodata usees to expose their open-source code while it is in the process of being merged with larger scientific libraries such as scipy, scikit-image, or scikit-learn. Additioanlly, methods for computational neuroscience on brains too specific for a general scientific library can be found here, such as image registration software tuned specifically for large brain volumes.

- [Motivation](#motivation)
- [Installation](#installation) * [Environment](#environment) * [Install from pypi](#install-from-pypi) * [Install from source](#install-from-source)
- [How to Use Brainlit](#how-to-use-brainlit) * [Data Setup](#data-setup) * [Create a Session](#create-a-session)
- [Features](#features) * [Registration](#registration)
- [Core](#core) * [Push/Pull Data](#push-and-pull-data) * [Visualize](#visualize) * [Manually Segment](#manually-segment) * [Automatically Segment](#automatically-and-semi-automatically-segment)
- [API reference](#api-reference)
- [Tests](#tests)
- [Contributing](#contributing)
- [Credits](#credits)
## Motivation The repository originated as the project of a team in Joshua Vogelstein's class Neurodata at Johns Hopkins University. This project was focused on data science towards the [mouselight data](https://www.hhmi.org/news/mouselight-project-maps-1000-neurons-and-counting-in-the-mouse-brain). It becme apparent that the tools developed for the class would be useful for other groups doing data science on large data volumes. The repository can now be considered a "holding bay" for code developed by Neurodata for collaborators and researchers to use.
## Installation ### Environment
- [get conda](https://docs.conda.io/projects/conda/en/latest/user-guide/getting-started.html)
- create a virtual environment with python>=3.7 via conda create --name brainlit python=3.7
- activate the environment via conda activate brainlit
- ### Install from pypi
- install brainlit via pip install brainlit
- ### Install from source
- clone the repo via git clone https://github.com/neurodata/brainlit.git
- cd into the repo via cd brainlit
- install brainlit via pip install -e .
## How to use Brainlit ### Data setup The source data directory should look something like an octree data structure with optional swc folder
- data/
- default.0.tif
- 1/
- default.0.tif
- 1/ ... 8/
- 2/ ... 8/
- transform.txt
- consensus-swcs (optional, for .swc files)
First, decide for your team where you'd like to store the data - whether it will be on a local machine or on the cloud. If on the cloud, each collaborator will need to create a file at ~/.cloudvolume/secrets/x-secret.json, where x is one of [aws, gc, azure] which contains your id and secret key for your cloud platform.
### Create a session Each user will start their scripts with approximately the same lines: ``` from brainlit.utils.ngl import NeuroglancerSession
session = NeuroglancerSession(url='file:///abc123xyz') ``` From here, any number of tools can be run such as the visualization or annotation tools. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/visualization/visualization.ipynb).
## Features
### Registration The registration subpackage is a facsimile of ARDENT, a pip-installable (pip install ardent) package for nonlinear image registration wrapped in an object-oriented framework for ease of use. This is an implementation of the LDDMM algorithm with modifications, written by Devin Crowley and based on "Diffeomorphic registration with intensity transformation and missing data: Application to 3D digital pathology of Alzheimer's disease." This paper extends on an older LDDMM paper, "Computing large deformation metric mappings via geodesic flows of diffeomorphisms."
This is the more recent paper:
Tward, Daniel, et al. "Diffeomorphic registration with intensity transformation and missing data: Application to 3D digital pathology of Alzheimer's disease." Frontiers in neuroscience 14 (2020).
https://doi.org/10.3389/fnins.2020.00052
This is the original LDDMM paper:
Beg, M. Faisal, et al. "Computing large deformation metric mappings via geodesic flows of diffeomorphisms." International journal of computer vision 61.2 (2005): 139-157.
https://doi.org/10.1023/B:VISI.0000043755.93987.aa
A tutorial is available in docs/notebooks/registration_demo.ipynb.
## Core The core brain-lit package can be described by the diagram at the top of the readme:
### (Push and Pull Data) Brainlit uses the Seung Lab's [Cloudvolume](https://github.com/seung-lab/cloud-volume) package to push and pull data through the cloud or a local machine in an efficient and parallelized fashion. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/utils/uploading_brains.ipynb). The only requirement is to have an account on a cloud service on s3, azure, or google cloud.
Loading data via local filepath of an octree structure is also supported. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/utils/upload_brains.ipynb).
### Visualize Brainlit supports many methods to visualize large data. Visualizing the entire data can be done via Google's [Neuroglancer](https://github.com/google/neuroglancer), which provides a web link as shown below.
screenshot
Brainlit also has tools to visualize chunks of data as 2d slices or as a 3d model. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/visualization/visualization.ipynb).
screenshot
### Manually Segment Brainlit includes a lightweight manual segmentation pipeline. This allows collaborators of a projec to pull data from the cloud, create annotations, and push their annotations back up as a separate channel. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/pipelines/manual_segementation.ipynb).
### Automatically and Semi-automatically Segment Similar to the above pipeline, segmentations can be automatically or semi-automatically generated and pushed to a separate channel for viewing. [Interactive demo](https://github.com/neurodata/brainlit/blob/master/docs/notebooks/pipelines/seg_pipeline_demo.ipynb).
## API Reference [](https://brainlight.readthedocs.io/en/latest/?badge=latest) The documentation can be found at [https://brainlight.readthedocs.io/en/latest/](https://brainlight.readthedocs.io/en/latest/).
## Tests Running tests can easily be done by moving to the root directory of the brainlit package ant typing pytest tests or python -m pytest tests. Running a specific test, such as test_upload.py can be done simply by ptest tests/test_upload.py.
## Contributing Contribution guidelines can be found via [CONTRIBUTING.md](https://github.com/neurodata/brainlit/blob/master/CONTRIBUTING.md)
## Credits Thanks to the neurodata team and the group in the neurodata class which started the project. This project is currently managed by Tommy Athey and Bijan Varjavand.
| Status | Completed |
|---|---|
| Checksum (MD5) | df462cfd7bb1e364cdf0384deaa02b4c |
| Checksum (SHA-1) | 8480882171b9a00d4eb710d1887944e8f0dae091 |
| Checksum (SHA-256) | 1bfce914c509959d3036be43bd3c95add0b9ac4d3073ee15eef51e392f4d15aa |
| Checksum (SHA-512) | 2efdd70e72f2f99c927e78860a1704bd2939f1ccc0ea5ec9d501d0fe70a4539913… |
| GPG Signature | |
| GPG Fingerprint | 6811684bac0b8895434e97bdd4391b8fb999e537 |
| Storage Region | Dublin, Ireland |
| Type | Binary (contains binaries and binary artifacts) |
| Uploaded At | 4 months, 3 weeks ago |
| Uploaded By |
|
| Slug Id | brainlit-020-py3-none-anywhl-zlzu |
| Unique Id | IV8DuOTAavfae2Y8 |
| Version (Raw) | 0.2.0 |
| Version (Parsed) |
|
| extended metadata | |
| Author | ('Bijan Varjivand, Ryan Lu, Matt Figdore, Alex Fiallos, Stanley Wang, Victor Wang',) <bvarjav1@jhu.edu> |
| Classifiers | Development Status :: 3 - Alpha | Intended Audience :: Science/Research | License :: OSI Approved :: Apache Software License | Programming Language :: Python :: 3 | Programming Language :: Python :: 3.5 | Programming Language :: Python :: 3.6 | Programming Language :: Python :: 3.7 | Topic :: Scientific/Engineering :: Mathematics |
| Homepage URL | https://github.com/neurodata/brainlit |
| Metadata Version | 2.1 |
| Py Filetype | bdist_wheel |
| Py Version | py3 |
| Requires Dist | cloud-volume (>=1.6.0) | napari (>=0.2.11) | networkx (>=2.1) | nibabel (>=2.4.1) | nilearn (>=0.5.2) | numpy (>=1.8.1) | opencv-python (==4.2.0.32) | scikit-learn (>=0.19.1) | scipy (>=1.1.0) | seaborn (>=0.9.0) | simpleitk (>=1.2.4) | statsmodels (>=0.11.1) | tifffile (>=2020.7.17) |
| pkg | brainlit-0.2.0-py3-none-any.whl |
4
111.2 KB |
md5 | sha1 | sha256 | sha512 |
This package has 66 files/directories.
Last scanned
4 months, 3 weeks ago
Scan result
Clean
Vulnerability count
0
Max. severity
UnknownPackage statistics are no longer available on cloudsmith.io. Please visit our new web app to access this feature.
You can embed a badge in another website that shows this or the latest version of this package.
To embed the badge for this specific package version, use the following:
[](https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/)
|This version of 'brainlit' @ Cloudsmith|
.. |This version of 'brainlit' @ Cloudsmith| image:: https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true
:target: https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/
image::https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true[link="https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/",title="This version of 'brainlit' @ Cloudsmith"]
<a href="https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/"><img src="https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/0.2.0/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true" alt="This version of 'brainlit' @ Cloudsmith" /></a>
rendered as:
To embed the badge for the latest package version, use the following:
[](https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/)
|Latest version of 'brainlit' @ Cloudsmith|
.. |Latest version of 'brainlit' @ Cloudsmith| image:: https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true&show_latest=true
:target: https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/
image::https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true&show_latest=true[link="https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/",title="Latest version of 'brainlit' @ Cloudsmith"]
<a href="https://cloudsmith.io/~demo-docs/repos/awesome-repo/packages/detail/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/"><img src="https://api.cloudsmith.com/v1/badges/version/demo-docs/awesome-repo/python/brainlit/latest/a=noarch;xf=bdist_wheel;xn=brainlit;xv=py3/?render=true&show_latest=true" alt="Latest version of 'brainlit' @ Cloudsmith" /></a>
rendered as:
These instructions assume you have setup the repository first (or read it).
To install/use brainlit @ version 0.2.0 ...
pip install 'brainlit==0.2.0'
You can also install the latest version of this package:
pip install --upgrade 'brainlit'
If necessary, you can specify the repository directly:
pip install \
--index-url=https://dl.cloudsmith.io/public/demo-docs/awesome-repo/python/simple/ \
brainlit==0.2.0
If you've got a project requirements.txt file, you can specify this as a dependency:
--index-url=https://dl.cloudsmith.io/public/demo-docs/awesome-repo/python/simple/
brainlit==0.2.0
In addition, you can use this repository as an extra index url. However, please read our documentation on this parameter before using it. For example in a requirements.txt file:
--extra-index-url=https://dl.cloudsmith.io/public/demo-docs/awesome-repo/python/simple/
brainlit==0.2.0