Abstract
Single-molecule localization microscopy resolves objects below the diffraction limit of light via sparse, stochastic detection of target molecules. Single molecules appear as clustered detection events after image reconstruction. However, identification of clusters of localizations is often complicated by the spatial proximity of target molecules and by background noise. Clustering results of existing algorithms often depend on user-generated training data or user-selected parameters, which can lead to unintentional clustering errors. Here we suggest an unbiased algorithm (FINDER) based on adaptive global parameter selection and demonstrate that the algorithm is robust to noise inclusion and target molecule density. We benchmark the most common density based clustering algorithms in test scenarios which we designed based on experimental datasets. We show that FINDER can keep the number of false positive inclusions low while also maintaining a low number of false negative detections in densely populated regions.
Introduction
Super-resolution microscopy has opened up new frontiers in biological and biomedical research by providing unprecedented molecular insights into the inner workings of cells [1, 2]. Classical light microscopy can only resolve structural features that are larger than the diffraction limit of light of a few hundred nanometers [3]. By overcoming the diffraction limit, super resolution microscopy has revealed long hidden mechanisms underlying intracellular transport processes [4] and the spatial organization of mRNA translation [5, 6]. The major feature of single-molecule localization microscopy (SMLM) is its ability to exploit the stochastic and sparse switching of fluorescence emission of specific labels binding to target molecules [7]. For example, in DNA-based point accumulation for imaging in nanoscale topography (DNA PAINT), each labelled target molecule is stochastically and repeatedly detected by transient binding of imager and docking strands [8]. Over time, each target molecule generates fluorescent detection events that cluster in space. Target molecules can be actin cytoskeleton [9], AMPA receptors [10] – or a well-defined DNA origami [11]. However, SMLM data sets often contain localizations events that represent ambiguous information. For instance, a super resolved image may contain false positive localizations (i.e. background noise from nonspecific fluorescent signal that does not originate from a target molecule). Furthermore, in high density regions of target molecules, localizations from multiple target molecules can overlap or form complex structures [12].
Current state-of-the-art cluster detection algorithms rely on some form of prior, user-provided information. This information can be the type of localization patterns to be detected or prior experience. One of the most widely used [13, 14, 15, 16, 17] algorithms is the ‘density-based spatial clustering algorithm’ (DBSCAN) [18, 19]. It is intuitive, simple, fast for 2D datasets [20] – but the parameter choice suffers from human bias. To identify the ‘most probable’ clustering parameters based on prior knowledge, empirical methods for parameter identification have been proposed [19] (see Table S1). OPTICS [21] circumvent the use of a global clustering parameter by defining borders between clusters of localizations through changes in local point density. Fundamentally different approaches to clustering have also gained traction to circumvent the shortcomings of DBSCAN [22, 23, 24, 25]. In 2015, Rubin-Delanchy et al. have introduced a Bayesian parameter finding approach [26, 27] which starts from a user generated prior parameter set for a cluster-proposing algorithm, and subsequently computes a posterior probability for each parameter set. Also in 2015, Levet et al. introduced SR Tesselation, an algorithm which segments large-scale images into polygonal regions [28, 29] and is specialized to reveal spatial structures at multiple scales [30]. Most recently in 2020, Williamson et al. developed a machine-learning approach [31] that classifies localizations depending on their local neighborhood. This approach does not require the users to provide parameters. It also beats most classical algorithms in selected clustering challenges [31], but depends on training data sets.
Despite their successes, none of the current approaches have fully removed the dependency on prior knowledge – either a statistical model is needed, a reference density needs to be set, or a machine-learning model needs to be trained on a pre-selected data set. We address the problem of parameter sensitivity and user-generated bias by building on the widely applied DBSCAN [13, 32, 33, 34, 35] algorithm and propose an unbiased parameter selection which we call FINDER. FINDER minimizes dependencies on prior knowledge by leveraging what is usually seen as a distractor: false positive localizations, or ‘noise’. The core principle of FINDER is to use information about the clustering variability with respect to the variation of the radial parameter, which is also described in [36], to then select the most robust clustering. Global noise levels therefore act as a lower boundary for the sensitivity of the algorithm, preventing over-segmentation and minimizing false positive cluster inclusions. To validate this approach, we use clusters which we identified in super-resolution images [8] to produce synthetic test sets, and compare the performance of FINDER with one of the currently best performing clustering algorithms, the adaptive machine-learning algorithm CAML [31]. We also compare the performance to classical DBSCAN and density-based OPTICS clustering approaches [21]. We show that our FINDER algorithm is independent of training data, it is computationally tractable while also showing a similar or better performance as measured by true positive detections, and reducing false-positive cluster detections and allowing for better interpretability.
Results
The assignment of clusters of localizations in super resolution microscopy is not trivial. Consider, for example, DNA PAINT data sets of super resolved neuronal AMPA receptor localizations. In Fig. 1 (left) one can see how difficult it is to identify clusters of localizations due to variable cluster sizes, high-noise and overlap. Similarly, data sets of DNA origami trimers (Fig. 1 right) indicate that identifying localizations that belong to single molecules of interest and separating these locations from the background noise is also a challenge. Algorithms can propose candidate clusters that correspond to single molecules of interest. However, verifying the reliability of the result is difficult because the ground truth is not always known and different parameter settings within the algorithm can lead to different outcomes. This means that the presence or absence of molecules of interest and their inferred location will vary depending on the parameter settings of the algorithm. In general, parameter settings, algorithm training and selection can become sources for clustering bias or errors.
To solve these challenges we propose the clustering algorithm, which we call FINDER and which we outline in Fig. 2 and detail in the method section “FINDER algorithm”. To benchmark FINDER, we used super resolution images of structurally well-defined DNA origami trimers and tetramers. DNA origami are folded DNA nanostructures with well- defined binding sites for fluorophores [11]. As such, the ground truth of the geometry of binding sites is known: The algorithms should identify not more than 3 and 4 subclusters for each DNA origami trimer and tetramer, respectively. Note that the subclusters of the identified oligomers exhibit a considerable heterogeneity in their number of localizations per subcluster.
We start by comparing FINDER to the recently proposed machine-learning based clustering algorithms (CAML), which outperformed most classical clustering algorithms in selected clustering test cases [31]. CAML feeds density variations of the neighborhood of each point into a trained classifier to identify clusters. Here, we use the pre-trained models ‘CAML 07VEJJ’ and ‘CAML 87B144’ from Ref. [31], which consider the first 100 and 1000 neighboring points, respectively. In contrast, FINDER identifies an optimal radial parameter based on a similarity measure that is computed across an interval of probable parameters (see Methods for details). FINDER uses a density-based approach which is based on the core points defined in DBSCAN and then optimizes this parameter based on the parameter phase space in the particular data set. Core points are points that have at least minPts neighbors within radius ε. In DBSCAN, these core points are used to initiate clusters. Instead, in FINDER, all non-core points are identified as noise localizations and iteratively removed. We refer to this more conservative definition of core points as ‘noise-free DBSCAN’. Our tests suggest that this self-contained definition of core points is more robust to noise and leads to a lower number of false positive cluster detections (see Figs. S6–S11). In the supplemental information we also compare the performance of FINDER with a version of FINDER using the classical density-based algorithm DBSCAN and with OPTICS, showing a higher false positive detection rate and a lower robustness, respectively.
In Fig. 3, we show that the FINDER algorithm accurately predicts the number of binding sites of the DNA origami oligomer, even though the density of localizations is highly heterogeneous and the final result applies one radial parameter to the full image. Notably, the adaptive CAML algorithms lead to a wide variety of detected subclusters of localizations. CAML 07VEJJ detects 3-mers most often, but fails in detecting 4-mers. In constrast, CAML 87B144 is more accurate in detecting 4-mers than 3-mers. One explanation for this discrepancy is that, on average, the detected 3-mers have 156 localizations per subcluster, but 4-mers have 285 localizations per subcluster. This could explain why CAML 07VEJJ, which considers only the first 100 neighboring points, performs poorly for the tetramer dataset, but it does not explain the performance of CAML 87B144. It also does not explain the segmentation failure of CAML 87B144 if no random noise is added (see Fig. S5). This results suggest that a version of CAML which considers the first 1000 points would need to be retrained for such 3-mer and 4-mer datasets. Retraining of the model is one possibility to include global information into local clustering decisions. However, the ground truth statistics regarding inter-and intra-cluster distance in experimentally recorded data sets is not a priori known. It is therefore hard to define a good training data set without introducing user-generated training biases. This highlights a challenge that adaptive algorithms share: information about the local neighborhood is used for clustering decisions – but often, these decisions need information about the global properties of the dataset – such as noise intensity or cluster separation. Often, these global properties are assumed a priori through user-defined parameters or training data. This potential source of bias is avoided by FINDER, which systematically probes the full dataset to identify one set of global parameters for an easily interpretable density-based clustering algorithm. As a second benchmark, we employed two libraries of unit clusters of localizations: (1) manually identified clusters in a SMLM dataset of a synapse [10], which contain on average only 17 localizations per cluster (see Fig. S1 a) and (2) manually identified sub-clusters of DNA-origami trimers, with on average 113 localizations per cluster (see Fig. S1 b). We re-arranged these unit clusters in three different configurations and added random noise points. These surrogate test data sets provide a ground-truth, while also retaining the biological variability of the cluster geometry. Our clustering outcomes are summarized in Fig 4. As expected, the CAML 07VEJJ model fails for the set of unit clusters where the number of points per cluster can be larger than the number of points considered (100, see Fig. S1 b). Interestingly, in several test cases CAML 87B144, which considers the first 1000 neighboring points, also fails for the smaller set of unit clusters (see Fig. S1 a). If the ‘correct’ CAML model is chosen, performance is good, with a high number of true positive and low number of false negative detections. For all test cases we considered here, FINDER leads to a similar number of true positive cluster detections as the CAML model that performs better for the given configuration. FINDER consistently results in the lowest number of false positive cluster detections. See Figs. S6-S11 for more expansive tests. This suggests that FINDER is able to identify global parameters which give robust results with low false positive detection rates.
Finally, we applied FINDER and other clustering algorithms on SMLM datasets for which the ground truth is not known. In Fig. 5, we show the clustering results for single-molecule localization DNA-PAINT data of newly synthesized proteins after global homeostatic scaling in neuronal dendrites [6] (see Fig. S13 for an analogous analysis for a dataset of neuronal AMPA-receptors). Most clusters of localizations detected by FINDER and the two CAML algorithms have 50 or fewer localizations. CAML (07VEJJ) leads to an abrupt cutoff of cluster sizes at 100, suggesting that typical clusters can exceed that size, and therefore more than 100 neighbors need to be included in the analysis. The cluster size distribution obtained using CAML (87B144) leads to a tail in which clusters with more than 150 localizations are found, and 40% of all localizations are identified as noise. In comparison, FINDER identifies 27% of all localizations as noise localizations. FINDER therefore includes more large clusters, which usually represent molecular aggregations, than CAML (87B144). The distribution of cluster sizes for CAML (87B144) (see second bin) and visual inspection of the clusters suggests that there is an over-segmentation, which may require additional filtering. We conclude that the cluster-size distributions are overall robust with respect to the choice of the algorithm. We also note that clustering-results provided by FINDER have the advantage of being easily interpretable and reproducible: For example here, a cluster is defined if 10 core points are found within 80.4nm, and a noise localization is defined if it does not have 10 localizations within that radius.
Discussion
The identification of clustered localizations in single-molecule localization microscopy data is a crucial step for quantifying proteins within nano-clusters [23, 37]. A bottleneck for automated image analysis is the identification of appropriate clustering parameters, particularly if target molecules themselves are clustered or if noise levels are high. Clustering algorithms used in scientific studies of SMLM datasets therefore need to fulfill many – sometimes contradicting – requirements. They have to be fast, robust with respect to parameter choice, and offer results that are easily interpretable. Manual parameter tuning needs to be avoided, as detailed knowledge about the system may not be readily available and can introduce human bias. Because control experiments with known ground truth clusterings are not always available, false positive cluster inclusions need to be minimized in order to avoid over- or misinterpretations of the data.
DBSCAN is one of the most popular clustering algorithms for SMLM data [14, 15, 16, 17, 13, 32, 33, 34, 35]. Part of its appeal is the fact that its density-based clustering rule is intuitive and it can offer close-to-optimal results if parameters are chosen carefully [22]. It has two input parameters: a radial parameter ∊ and the minimal number of points minPts within this radial parameter needed to initiate a cluster. However it is not always clear how to set the radial parameter manually for heterogeneous cluster sizes or datasets with changing density. Heuristic rules for setting these parameters exist [19, 22], but in many examples lead to ambiguity. For instance, the optimal value for the radial parameter ε is commonly estimated from the k-th-neighbor curves, which may be not well-defined and vary with k [22]. Several adaptations and improvements of DBSCAN have been proposed to deal with density-variations, and the parameter estimation problem [38, 39]. Reverse-nearest-neighbor approaches such as RECORD [40], IS-DBSCAN [41], ISB-DBSCAN [42] and RNN-DBSCAN [43] only require setting one parameter k, and can deal with density-variations within the dataset. In many applications, however, parameters are manually chosen (see Table S1) which may not lead to optimal results when analyzing many datasets with differing localization densities.
Here, we pursue an approach that does not locally adapt to varying densities, but which sets a global threshold for the required cluster density for scenarios in which the radial parameter cannot unambiguously be extracted from k-th-neighbor curves. The idea is to leverage the information given by overall noise levels to set a global threshold for cluster inclusions and therefore minimize the number of false positive cluster inclusions. To tackle the issue of parameter choice in this scenario, while retaining the advantages of density-based clustering methods such as DBSCAN, we presented an unbiased, automated parameter identification algorithm FINDER. FINDER combines the following benefits: First, it does not require prior knowledge about the structure of clustered localizations and by that it side-steps the need for a statistical model, the need to perform supervised training on annotated data sets, or the need of manual parameter choices that might encompass user-generated biases. Second, FINDER leads to easily interpretable results and automatically defines parameters which are then applied across the full dataset. Finally, FINDER can be applied to large, heterogeneous data sets with clustering results which are robust to noise and signal overlap (Figs. 2, 3, 4).
In the absence of a known ground truth- as is the case in most scientific exploratory data analysis-clustering algorithms need to be transparent, easily interpretable, and minimize false positive detections to avoid misinterpretation of data. In local adaptive algorithms, local density changes are employed to detect clusters of localizations. But it is often unclear on what basis the threshold for the local density change ist set. In 2020, the local machine-learning based clustering method (CAML [31]) has been proposed. If the correctly trained CAML algorithm is chosen, clustering results are generally very good, with high true positive and low false positive detections (see Fig. 4). However in some cases, these methods can lead to severe over-segmentation (Figs. 3, 4, 5) and in other cases to under-counting of clusters (Fig. 3). For exploratory data analysis with an unknown ground truth, this sensitivity is critical, because one does not know which datasets are outside of the validity regime of a given trained model.
This oversegmentation can be avoided if the global parameter choice is used to transfer information from the global to the local scale. For instance, information at the global scale is the overall amount of noise, or the general heterogeneity of the clusters. The local scale is the neighborhood of each localization. By avoiding pre-filtering of noise, and by performing both noise-filtering and clustering in one step, FINDER uses the global information to set the local radial parameter ε, which then defines a threshold for clustering. In brief, FINDER selects the radial parameter which leads to the most robust clustering. Let us mention that for our test scenarios we set the minPts parameter to 10 localizations. This is in contrast to minPts = 4 recommended for all 2D datasets by Ester et al. [18]. In particular, systematic phase space explorations revealed that minPts = 4 lead to high sensitivity of the results to noise, over-segmentation, and as a result a large number of false positive cluster detections (see Fig. S17 and Fig. S18). We also note that a more stringent cluster point definition based on iteratively removing non-core points leads to more robust clustering results that the classical DBSCAN algorithm.
We systematically benchmarked FINDER against existing algorithms using two sets of experimentally recorded clusters of localizations. We found that – despite using the same parameters across the full dataset – the cluster inclusion and exclusion criteria used in FINDER perform robustly compared to trained machine-learning models. Our tests on synthetic data sets in the high noise and the high overlap regimes showed that density variations due to noise can lead to over-segmentation in adaptive clustering algorithms if an incorrect algorithm is used (see Fig. 4). For example, for synthetic reconstructions, FINDER performed similar to the better of two pre-trained CAML models. FINDER was able to minimize the number of false positive clusters while maintaining a high ratio of true positive clusters. We also tested FINDER on a dataset of DNA-origami trimers and tetramers and showed that FINDER reduces the number of false positive cluster detections and at the same time retains many true positive detections – leading to an accurate prediction of the underlying molecular structure.
In conclusion, we showed that performing noise identification, parameter-choice and clustering in one single post-processing step, such as proposed in FINDER, provides a reliable and unbiased method for a spatial analysis of SMLM data sets. In most experimental settings, the ground truth is not known, and therefore minimizing the number of false positive cluster detections is important to avoid erroneous interpretations of experimental results. We showed here that an all-in-one cluster identification can help limit the effect of human biases, and can speed up the interpretation of single molecule microscopy datasets.
Methods
FINDER algorithm
The FINDER algorithm identifies the hyperparameters for a cluster-proposing algorithm. Here, we employ two cluster proposing algorithms: DBSCAN, as well as a version of DBSCAN based on iterative removal of non-core points (see Methods-section on “Noise-free DBSCAN” for details). Both algorithms take two parameters, which are a fixed minimum number of neighboring points (minPts), and a typical distance (ε). FINDER determines the radial parameter through the following steps:
Compute the distribution of the distance to the kth-nearest neighbors. Here, we set k = 10 (see discussion, and Fig. S17 and Fig. S18).
Define the interval in which the algorithm will search for the second parameter ε as the interval between the 10th to the 90th percentile of the distribution of kth-nearest neighbors.
For the input dataset, compute clustering results at n logarithmically distributed points across the interval (here, n = 15, see Fig. S19 for robustness), using the cluster-proposing algorithm.
Compute the similarity score for each clustering result, see Methods section “Similarity of clusterings and similarity score”.
The parameters that correspond to the clustering with the largest similarity score is selected.
Noise-free DBSCAN cluster definition
DBSCAN initiates clusters using core points. Core points are points which have a at least minPts neigh-boring points within a distance ε. Here, we modified this classic DBSCAN cluster definition to make it more robust to noise. First, we iteratively remove all non-core points from the dataset of localizations X such that only core points remain (see Fig. S2). Next, FINDER partitions the remaining core points into clusters. The algorithm is illustrated by the following pseudo-code:
In all figures of the main manuscript, we used the noise-free DBSCAN cluster definition inside FINDER. For comparison with the classic DBSCAN cluster definition see Figs. S6-S11 in the supplementary material.
Similarity of clusterings and the definition of the similarity score
Let be a clustering of a set of points . We define the similarity score of two clusterings and as the sum of similar subsets within the partitions:
Two subsets Xi and Yj are said to be similar if the number of overlapping points is larger than the number of non-overlapping points for each of the subsets:
The similarity score of a clustering within an assembly of clusterings is defined as the sum of similarity scores:
Generation of surrogate test data from DNA origami images
To benchmark the performance of FINDER on experimental data sets and compare its performance to alternative existing clustering methods we used a DNA origami data set with a known cluster structure. We considered images of DNA origami containing three or four binding sites that we measured using DNA-PAINT [8]. Even though three or four localization clusters are expected for this DNA origami data set and this knowledge can provide a ground truth for clustering outcomes, we found that all clustering methods consistently detected a varying fraction of dimers. Upon visual inspection, we found that some expected trimers or tetramers appeared incomplete because some fluorophore binding sites were absent from some origami. We thus divided the trimer data into two groups: visibly resolved and visibly un-resolved trimers. We also introduced the group of resolved tetramers, in order to test the performance of all algorithms on a different geometric configuration See Fig. 3 for analysis of both groups.
In a second test, we accessed the robustness of the algorithms in the limit of overlapping clusters, in the high-noise limit, and with geometric constraints. We manually selected clusters representing single binding sites from resolved trimers (see Fig. 4 b) and, in a second set, we manually selected from AMPA receptor images (see Fig. S1). We then re-assembled those monomers in pre-defined grid and path geometries, and with varying levels of added random noise localizations.
Definition of true and false positive cluster detections
In Fig. 4, a cluster X from the ground truth clustering is counted as being correctly detected by cluster Y from clustering if the overlap between both clusters covers at least 30% of the points of cluster X, i.e. if |X ∩ Y | > 0.3|X| and if cluster Y does not detect any other clusters of clustering , i.e. one cluster Y cannot detect two clusters from . All clusters from that can be attributed to a cluster of in such a way are counted as true positives, and the remaining clusters of count as false positives. In Fig. S16, we vary the overlap threshold between 0% and 100%, and show robustness of the results with respect to this variation.
Note that other metrics for the similarity of two clusterings [43] such as the Adjusted Rand Index (ARI) [44] or the Normalized Mutual Information (NMI) [45], mix the similarity and the number of correctly identified clusters. In contrast, for the second benchmark in Fig. 4, we focus on how many clusters have been correctly and incorrectly identified. We therefore use a metric that uses a hard, binary, threshold for individual clusters.
Experimentally recorded super-resolution microscopy data
DNA Origami data
DNA origami containing 3 (’trimers’, 3-fold symmetry, 55 nm interspacing) and 4 (’tetramers’, 4 fold symmetry, 40 nm interspacing) binding sites (containing P1 docking oligos) were imaged on the same N-STORM system (Nikon, Japan) as the above reported AMPA-receptor data: an Eclipse Ti-E inverted microscope, equipped with a Perfect Focus System (Ti-PSF) and a motorized x-y stage. Total internal reflection fluorescence (TIRF) was adjusted using a motorized TIRF illuminator in combination with a 100 x oil-immersion objective (CFL Apo TIRF, NA 1.49) with a final pixel size of 158 nm. For imaging, 647 nm excitation wavelength was used, housed in a MLC400B (Agilent) laser combiner. An optical fiber guided the laser beam to the microscope body and via a dichroic mirror (T660LPXR, Chroma) to the sample plane. Fluorescence emission was separated from excitation light via a bandpass filter (ET705/72m, Chroma) and detected by an iXon Ultra EMCCD camera (DU - 897U-CS0-23 #BV, Andor). The software NIS-Elements Ar/C (Nikon) and μManager were used to control the setup and the camera. TIRF illumination was used for super-resolution acquisitions of DNA origami data with a power of 30-40 mW, which was determined directly after the objective and under wide-field configuration. Time-lapse datasets with 24000 frames for DNA origami trimers and 10773 frames for DNA origami tetramers and 16 bit depth were acquired at 3.3 Hz frame rate and 5 MHz camera read-out bandwidth; pre-amplification: 3; electron multiplying gain: 50. For DNA-PAINT imaging, the imaging buffer contains P1-Atto655 (CTAGATGTAT-Atto655, Eurofins Ge-nomics) in 500 mM NaCl, pH 7.3. The P1-Atto655 concentration was 10 nM for origami trimer and tetramer data, and 0.5 nM for AMPA receptor PAINT [10].
DNA-PAINT acquisitions were reconstructed using Picasso:Localize, a module of the Picasso software, by applying a minimal net gradient of 1500. With Picasso:Render, drift corrections were applied based on the redundant cross-correlation (RCC), with a segmentation of 1000 was applied. Drift-corrected data was filtered using Picasso:Filter. To generate the DNA origami trimer and tetramer cluster datasets for cluster-identification validation, trimer and tetramer clusters were identified by eye using Picasso:Render and manually selected with a picking diameter of 2 camera pixels.
Newly synthesized protein data
Newly synthesized protein data was previously reported (see Ref. [6]). In brief, cultured neuron was incubated in a growth medium containing (Tetrodotoxin) TTX for 1 hr. 15 mins before the treatment ended, the neuron was metabolically labelled with AHA. The immuno-stained neuron samples were then imaged using DNA-PAINT [8].
AMPA-receptor data
AMPA-receptor validation data was previously reported (see Ref. [10]). In brief, cultured neurons were stained by primary antibody against AMPA receptor GluA2 subunit before fixation and secondary antibody staining, in which the secondary antibody was modified to carry a P1 docking oligo. The immuno-stained neuron samples were then imaged using DNA-PAINT [8].
Supplemental Information
Acknowledgements
C.S. and A.N. acknowledge the support by the Add-on Fellowship for Interdisciplinary Life Science (Project number: 850027) of the Joachim-Herz Foundation. C.S. is supported by an EMBO long-term postdoctoral fellowship (EMBO ALTF 860-2018), HFSP Cross-Disciplinary Fellowship (LT000737/2019-C). M.H. is funded by the German Science Foundation (DFG, grant SFB 902) and the Bundesministerium für Bildung und Forschung (BMBF:eBio). EMS is funded by the Max Planck Society, DFG CRC 1080: Molecular and Cellular Mechanisms of Neural Homeostasis and DFG CRC 902: Molecular Principles of RNA-based Regulation and the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 743216). T.T. is funded by the Max Planck Society and DFG CRC1080: Molecular and Cellular Mechanisms of Neural Homeostasis. We thank Ann-Christin Andres and Nina Deussner for contributing the tetramer and trimer origami data. A.N. thanks Carlos Wert-Carvajal for proof-reading and valuable feedback.