Cortex Operations

Docker Deployment

A Cortex may be quickly deployed using pre-built docker containers which are built with each Synapse release. The Cortex docker container is configured to allocate all local storage within the directory /vertex/storage which may be mapped in as a volume.


It is strongly recommended that production deployments specify a specific container version such as vertexproject/synapse-cortex:v2.4.0 to make version updates an opt-in behavior.

The following example docker-compose.yml file will execute a cortex with an external storage volume mapped and ports mapped to expose the Cortex:

    image: vertexproject/synapse-cortex:master
    - 443:4443
    - 27492:27492
    - ./core_000:/vertex/storage
version: '3'


The SYN_CORTEX_AUTH_PASSWD environment variable or auth:passwd key in cell.yaml will initialize the password for the root user. Additional custom Cortex configuration may be done via additional environment variables or options within the cell.yaml.

You can now connect to the Cortex using the cmdr tool:

python -m tcp://root:secret@localhost:27492

Configuring A Mirror

A Cortex mirror replicates an entire Cortex to another host allowing it to act as a parallel query execution engine or hot-spare. Mirrors may be created in “tree” configurations where one mirror is used to synchronize several downstream mirrors to offload replication load from the main Cortex. Additionally, Cortex mirrors support write-back, allowing the mirror to appear to be read-write by passing edits to the upstream Cortex and awaiting synchronization to apply the change locally.

To deploy a Cortex mirror, you must first ensure that the nexuslog:en setting on your Cortex is enabled in cell.yaml. This is enabled by default, but may be disabled in some high-performance custom configurations. The initial state of a mirror is created by creating a backup of the Cortex using the standard synapse service backup tool:

python -m /path/to/cortex /path/to/mirror/cortex

A single backup may be copied multiple times to deploy several mirrors. Once the backup is created, edit the cell.yaml of the new mirror to include a telepath URL to the upstream Cortex:

mirror: "tcp://user:passwd@cortexhost/"


The telepath URL must contain a valid user/password with global admin priviledges on the upstream Cortex.

Once the mirror option is configured, the new mirror Cortex may be deployed normally and will stay synchronized in near real-time provided it can maintain the network link.

Configuring A Remote Axon

By default a Cortex will initialize a local Axon for general object / blob storage. This allows certain Cortex functionality to work without additional configuration. The local Axon is not exposed in a remote fashion.

An Axon can instead be deployed as a remote server (see Axon Operations) and the Cortex can be configured to be aware of it, by specifying the Axon’s Telepath URL in the axon configuration parameter (see Cortex Configuration Options).

For example, if the remote Axon is listening on port 27592, and has a service user core00, then the Cortex cell.yaml file could have the following configuration:

axon: tcp://core00:secret@<axon_host_ip>:27592

For interacting with byte storage inside of Storm, see $lib.bytes for APIs related to interacting with the Axon.

0.1.x to 2.x.x Migration


The 0.1.x to 2.x.x migration tools have been removed in Synapse v2.9.0.

For information about migrating older Cortexes to v2.8.0, please refer to the v2.8.0 documentation here.