49 results on '"Nayebi, Aran"'
Search Results
2. A Goal-Driven Approach to Systems Neuroscience
- Author
-
Nayebi, Aran
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Machine Learning - Abstract
Humans and animals exhibit a range of interesting behaviors in dynamic environments, and it is unclear how our brains actively reformat this dense sensory information to enable these behaviors. Experimental neuroscience is undergoing a revolution in its ability to record and manipulate hundreds to thousands of neurons while an animal is performing a complex behavior. As these paradigms enable unprecedented access to the brain, a natural question that arises is how to distill these data into interpretable insights about how neural circuits give rise to intelligent behaviors. The classical approach in systems neuroscience has been to ascribe well-defined operations to individual neurons and provide a description of how these operations combine to produce a circuit-level theory of neural computations. While this approach has had some success for small-scale recordings with simple stimuli, designed to probe a particular circuit computation, often times these ultimately lead to disparate descriptions of the same system across stimuli. Perhaps more strikingly, many response profiles of neurons are difficult to succinctly describe in words, suggesting that new approaches are needed in light of these experimental observations. In this thesis, we offer a different definition of interpretability that we show has promise in yielding unified structural and functional models of neural circuits, and describes the evolutionary constraints that give rise to the response properties of the neural population, including those that have previously been difficult to describe individually. We demonstrate the utility of this framework across multiple brain areas and species to study the roles of recurrent processing in the primate ventral visual pathway; mouse visual processing; heterogeneity in rodent medial entorhinal cortex; and facilitating biological learning., Comment: 230 pages, Stanford University PhD Thesis, March 2022: https://purl.stanford.edu/qk457cr2641
- Published
- 2023
3. Neural Foundations of Mental Simulation: Future Prediction of Latent Representations on Dynamic Scenes
- Author
-
Nayebi, Aran, Rajalingham, Rishi, Jazayeri, Mehrdad, and Yang, Guangyu Robert
- Subjects
Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Robotics ,Quantitative Biology - Neurons and Cognition - Abstract
Humans and animals have a rich and flexible understanding of the physical world, which enables them to infer the underlying dynamical trajectories of objects and events, plausible future states, and use that to plan and anticipate the consequences of actions. However, the neural mechanisms underlying these computations are unclear. We combine a goal-driven modeling approach with dense neurophysiological data and high-throughput human behavioral readouts to directly impinge on this question. Specifically, we construct and evaluate several classes of sensory-cognitive networks to predict the future state of rich, ethologically-relevant environments, ranging from self-supervised end-to-end models with pixel-wise or object-centric objectives, to models that future predict in the latent space of purely static image-based or dynamic video-based pretrained foundation models. We find strong differentiation across these model classes in their ability to predict neural and behavioral data both within and across diverse environments. In particular, we find that neural responses are currently best predicted by models trained to predict the future state of their environment in the latent space of pretrained foundation models optimized for dynamic scenes in a self-supervised manner. Notably, models that future predict in the latent space of video foundation models that are optimized to support a diverse range of sensorimotor tasks, reasonably match both human behavioral error patterns and neural dynamics across all environmental scenarios that we were able to test. Overall, these findings suggest that the neural mechanisms and behaviors of primate mental simulation are thus far most consistent with being optimized to future predict on dynamic, reusable visual representations that are useful for Embodied AI more generally., Comment: 20 pages, 10 figures, NeurIPS 2023 Camera Ready Version (spotlight)
- Published
- 2023
4. Identifying Learning Rules From Neural Network Observables
- Author
-
Nayebi, Aran, Srivastava, Sanjana, Ganguli, Surya, and Yamins, Daniel L. K.
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Machine Learning ,Statistics - Machine Learning - Abstract
The brain modifies its synaptic strengths during learning in order to better adapt to its environment. However, the underlying plasticity rules that govern learning are unknown. Many proposals have been suggested, including Hebbian mechanisms, explicit error backpropagation, and a variety of alternatives. It is an open question as to what specific experimental measurements would need to be made to determine whether any given learning rule is operative in a real biological system. In this work, we take a "virtual experimental" approach to this problem. Simulating idealized neuroscience experiments with artificial neural networks, we generate a large-scale dataset of learning trajectories of aggregate statistics measured in a variety of neural network architectures, loss functions, learning rule hyperparameters, and parameter initializations. We then take a discriminative approach, training linear and simple non-linear classifiers to identify learning rules from features based on these observables. We show that different classes of learning rules can be separated solely on the basis of aggregate statistics of the weights, activations, or instantaneous layer-wise activity changes, and that these results generalize to limited access to the trajectory and held-out architectures and learning curricula. We identify the statistics of each observable that are most relevant for rule identification, finding that statistics from network activities across training are more robust to unit undersampling and measurement noise than those obtained from the synaptic strengths. Our results suggest that activation patterns, available from electrophysiological recordings of post-synaptic activities on the order of several hundred units, frequently measured at wider intervals over the course of learning, may provide a good basis on which to identify learning rules., Comment: NeurIPS 2020 Camera Ready Version, 21 pages including supplementary information, 13 figures
- Published
- 2020
5. Learning Physical Graph Representations from Visual Scenes
- Author
-
Bear, Daniel M., Fan, Chaofei, Mrowca, Damian, Li, Yunzhu, Alter, Seth, Nayebi, Aran, Schwartz, Jeremy, Fei-Fei, Li, Wu, Jiajun, Tenenbaum, Joshua B., and Yamins, Daniel L. K.
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning ,I.4.8 ,I.2.6 - Abstract
Convolutional Neural Networks (CNNs) have proved exceptional at learning representations for visual object categorization. However, CNNs do not explicitly encode objects, parts, and their physical properties, which has limited CNNs' success on tasks that require structured understanding of visual scenes. To overcome these limitations, we introduce the idea of Physical Scene Graphs (PSGs), which represent scenes as hierarchical graphs, with nodes in the hierarchy corresponding intuitively to object parts at different scales, and edges to physical connections between parts. Bound to each node is a vector of latent attributes that intuitively represent object properties such as surface shape and texture. We also describe PSGNet, a network architecture that learns to extract PSGs by reconstructing scenes through a PSG-structured bottleneck. PSGNet augments standard CNNs by including: recurrent feedback connections to combine low and high-level image information; graph pooling and vectorization operations that convert spatially-uniform feature maps into object-centric graph structures; and perceptual grouping principles to encourage the identification of meaningful scene elements. We show that PSGNet outperforms alternative self-supervised scene representation algorithms at scene segmentation tasks, especially on complex real-world images, and generalizes well to unseen object types and scene arrangements. PSGNet is also able learn from physical motion, enhancing scene estimates even for static images. We present a series of ablation studies illustrating the importance of each component of the PSGNet architecture, analyses showing that learned latent attributes capture intuitive scene properties, and illustrate the use of PSGs for compositional scene inference., Comment: 23 pages; corrected affiliations and acknowledgments
- Published
- 2020
6. Two Routes to Scalable Credit Assignment without Weight Symmetry
- Author
-
Kunin, Daniel, Nayebi, Aran, Sagastuy-Brena, Javier, Ganguli, Surya, Bloom, Jonathan M., and Yamins, Daniel L. K.
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Machine Learning ,Computer Science - Neural and Evolutionary Computing ,Statistics - Machine Learning - Abstract
The neural plausibility of backpropagation has long been disputed, primarily for its use of non-local weight transport $-$ the biologically dubious requirement that one neuron instantaneously measure the synaptic weights of another. Until recently, attempts to create local learning rules that avoid weight transport have typically failed in the large-scale learning scenarios where backpropagation shines, e.g. ImageNet categorization with deep convolutional networks. Here, we investigate a recently proposed local learning rule that yields competitive performance with backpropagation and find that it is highly sensitive to metaparameter choices, requiring laborious tuning that does not transfer across network architecture. Our analysis indicates the underlying mathematical reason for this instability, allowing us to identify a more robust local learning rule that better transfers without metaparameter tuning. Nonetheless, we find a performance and stability gap between this local rule and backpropagation that widens with increasing model depth. We then investigate several non-local learning rules that relax the need for instantaneous weight transport into a more biologically-plausible "weight estimation" process, showing that these rules match state-of-the-art performance on deep networks and operate effectively in the presence of noisy updates. Taken together, our results suggest two routes towards the discovery of neural implementations for credit assignment without weight symmetry: further improvement of local rules so that they perform consistently across architectures and the identification of biological implementations for non-local learning mechanisms., Comment: ICML 2020 Camera Ready Version, 19 pages including supplementary information, 10 figures
- Published
- 2020
7. From deep learning to mechanistic understanding in neuroscience: the structure of retinal prediction
- Author
-
Tanaka, Hidenori, Nayebi, Aran, Maheswaranathan, Niru, McIntosh, Lane, Baccus, Stephen A., and Ganguli, Surya
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Machine Learning ,Physics - Biological Physics - Abstract
Recently, deep feedforward neural networks have achieved considerable success in modeling biological sensory processing, in terms of reproducing the input-output map of sensory neurons. However, such models raise profound questions about the very nature of explanation in neuroscience. Are we simply replacing one complex system (a biological circuit) with another (a deep network), without understanding either? Moreover, beyond neural representations, are the deep network's computational mechanisms for generating neural responses the same as those in the brain? Without a systematic approach to extracting and understanding computational mechanisms from deep neural network models, it can be difficult both to assess the degree of utility of deep learning approaches in neuroscience, and to extract experimentally testable hypotheses from deep networks. We develop such a systematic approach by combining dimensionality reduction and modern attribution methods for determining the relative importance of interneurons for specific visual computations. We apply this approach to deep network models of the retina, revealing a conceptual understanding of how the retina acts as a predictive feature extractor that signals deviations from expectations for diverse spatiotemporal stimuli. For each stimulus, our extracted computational mechanisms are consistent with prior scientific literature, and in one case yields a new mechanistic hypothesis. Thus overall, this work not only yields insights into the computational mechanisms underlying the striking predictive capabilities of the retina, but also places the framework of deep networks as neuroscientific models on firmer theoretical foundations, by providing a new roadmap to go beyond comparing neural representations to extracting and understand computational mechanisms.
- Published
- 2019
8. Interpreting the retinal neural code for natural scenes: From computations to neurons
- Author
-
Maheswaranathan, Niru, McIntosh, Lane T., Tanaka, Hidenori, Grant, Satchel, Kastner, David B., Melander, Joshua B., Nayebi, Aran, Brezovec, Luke E., Wang, Julia H., Ganguli, Surya, and Baccus, Stephen A.
- Published
- 2023
- Full Text
- View/download PDF
9. Brain-Like Object Recognition with High-Performing Shallow Recurrent ANNs
- Author
-
Kubilius, Jonas, Schrimpf, Martin, Kar, Kohitij, Hong, Ha, Majaj, Najib J., Rajalingham, Rishi, Issa, Elias B., Bashivan, Pouya, Prescott-Roy, Jonathan, Schmidt, Kailyn, Nayebi, Aran, Bear, Daniel, Yamins, Daniel L. K., and DiCarlo, James J.
- Subjects
Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning ,Computer Science - Neural and Evolutionary Computing ,Electrical Engineering and Systems Science - Image and Video Processing ,Quantitative Biology - Neurons and Cognition - Abstract
Deep convolutional artificial neural networks (ANNs) are the leading class of candidate models of the mechanisms of visual processing in the primate ventral stream. While initially inspired by brain anatomy, over the past years, these ANNs have evolved from a simple eight-layer architecture in AlexNet to extremely deep and branching architectures, demonstrating increasingly better object categorization performance, yet bringing into question how brain-like they still are. In particular, typical deep models from the machine learning community are often hard to map onto the brain's anatomy due to their vast number of layers and missing biologically-important connections, such as recurrence. Here we demonstrate that better anatomical alignment to the brain and high performance on machine learning as well as neuroscience measures do not have to be in contradiction. We developed CORnet-S, a shallow ANN with four anatomically mapped areas and recurrent connectivity, guided by Brain-Score, a new large-scale composite of neural and behavioral benchmarks for quantifying the functional fidelity of models of the primate ventral visual stream. Despite being significantly shallower than most models, CORnet-S is the top model on Brain-Score and outperforms similarly compact models on ImageNet. Moreover, our extensive analyses of CORnet-S circuitry variants reveal that recurrence is the main predictive factor of both Brain-Score and ImageNet top-1 performance. Finally, we report that the temporal evolution of the CORnet-S "IT" neural population resembles the actual monkey IT population dynamics. Taken together, these results establish CORnet-S, a compact, recurrent ANN, as the current best model of the primate ventral visual stream., Comment: NeurIPS 2019 (Oral). Code available at https://github.com/dicarlolab/neurips2019
- Published
- 2019
10. Task-Driven Convolutional Recurrent Models of the Visual System
- Author
-
Nayebi, Aran, Bear, Daniel, Kubilius, Jonas, Kar, Kohitij, Ganguli, Surya, Sussillo, David, DiCarlo, James J., and Yamins, Daniel L. K.
- Subjects
Quantitative Biology - Neurons and Cognition ,Computer Science - Artificial Intelligence ,Computer Science - Computer Vision and Pattern Recognition ,Computer Science - Machine Learning ,Computer Science - Neural and Evolutionary Computing - Abstract
Feed-forward convolutional neural networks (CNNs) are currently state-of-the-art for object classification tasks such as ImageNet. Further, they are quantitatively accurate models of temporally-averaged responses of neurons in the primate brain's visual system. However, biological visual systems have two ubiquitous architectural features not shared with typical CNNs: local recurrence within cortical areas, and long-range feedback from downstream areas to upstream areas. Here we explored the role of recurrence in improving classification performance. We found that standard forms of recurrence (vanilla RNNs and LSTMs) do not perform well within deep CNNs on the ImageNet task. In contrast, novel cells that incorporated two structural features, bypassing and gating, were able to boost task accuracy substantially. We extended these design principles in an automated search over thousands of model architectures, which identified novel local recurrent cells and long-range feedback connections useful for object recognition. Moreover, these task-optimized ConvRNNs matched the dynamics of neural activity in the primate visual system better than feedforward networks, suggesting a role for the brain's recurrent connections in performing difficult visual behaviors., Comment: NIPS 2018 Camera Ready Version, 16 pages including supplementary information, 6 figures
- Published
- 2018
11. Unsupervised neural network models of the ventral visual stream
- Author
-
Zhuang, Chengxu, Yan, Siming, Nayebi, Aran, Schrimpf, Martin, Frank, Michael C., DiCarlo, James J., and Yamins, Daniel L. K.
- Published
- 2021
12. Biologically inspired protection of deep networks from adversarial attacks
- Author
-
Nayebi, Aran and Ganguli, Surya
- Subjects
Statistics - Machine Learning ,Computer Science - Learning ,Quantitative Biology - Neurons and Cognition - Abstract
Inspired by biophysical principles underlying nonlinear dendritic computation in neural circuits, we develop a scheme to train deep neural networks to make them robust to adversarial attacks. Our scheme generates highly nonlinear, saturated neural networks that achieve state of the art performance on gradient based adversarial examples on MNIST, despite never being exposed to adversarially chosen examples during training. Moreover, these networks exhibit unprecedented robustness to targeted, iterative schemes for generating adversarial examples, including second-order methods. We further identify principles governing how these networks achieve their robustness, drawing on methods from information geometry. We find these networks progressively create highly flat and compressed internal representations that are sensitive to very few input dimensions, while still solving the task. Moreover, they employ highly kurtotic weight distributions, also found in the brain, and we demonstrate how such kurtosis can protect even linear classifiers from adversarial attack., Comment: 11 pages
- Published
- 2017
13. Deep Learning Models of the Retinal Response to Natural Scenes
- Author
-
McIntosh, Lane T., Maheswaranathan, Niru, Nayebi, Aran, Ganguli, Surya, and Baccus, Stephen A.
- Subjects
Quantitative Biology - Neurons and Cognition ,Statistics - Machine Learning - Abstract
A central challenge in neuroscience is to understand neural computations and circuit mechanisms that underlie the encoding of ethologically relevant, natural stimuli. In multilayered neural circuits, nonlinear processes such as synaptic transmission and spiking dynamics present a significant obstacle to the creation of accurate computational models of responses to natural stimuli. Here we demonstrate that deep convolutional neural networks (CNNs) capture retinal responses to natural scenes nearly to within the variability of a cell's response, and are markedly more accurate than linear-nonlinear (LN) models and Generalized Linear Models (GLMs). Moreover, we find two additional surprising properties of CNNs: they are less susceptible to overfitting than their LN counterparts when trained on small amounts of data, and generalize better when tested on stimuli drawn from a different distribution (e.g. between natural scenes and white noise). Examination of trained CNNs reveals several properties. First, a richer set of feature maps is necessary for predicting the responses to natural scenes compared to white noise. Second, temporally precise responses to slowly varying inputs originate from feedforward inhibition, similar to known retinal mechanisms. Third, the injection of latent noise sources in intermediate layers enables our model to capture the sub-Poisson spiking variability observed in retinal ganglion cells. Fourth, augmenting our CNNs with recurrent lateral connections enables them to capture contrast adaptation as an emergent property of accurately describing retinal responses to natural scenes. These methods can be readily generalized to other sensory modalities and stimulus ensembles. Overall, this work demonstrates that CNNs not only accurately capture sensory circuit responses to natural scenes, but also yield information about the circuit's internal structure and function., Comment: L.T.M. and N.M. contributed equally to this work. Presented at NIPS 2016
- Published
- 2017
14. Scaling Properties for Artificial Neural Network Models of a Small Nervous System
- Author
-
Simeon, Quilee, primary, Venâncio, Leandro, additional, Skuhersky, Michael A., additional, Nayebi, Aran, additional, Boyden, Edward S., additional, and Yang, Guangyu Robert, additional
- Published
- 2024
- Full Text
- View/download PDF
15. Quantum algorithms for shortest paths problems in structured instances
- Author
-
Nayebi, Aran and Williams, Virginia Vassilevska
- Subjects
Quantum Physics ,Computer Science - Data Structures and Algorithms - Abstract
We consider the quantum time complexity of the all pairs shortest paths (APSP) problem and some of its variants. The trivial classical algorithm for APSP and most all pairs path problems runs in $O(n^3)$ time, while the trivial algorithm in the quantum setting runs in $\tilde{O}(n^{2.5})$ time, using Grover search. A major open problem in classical algorithms is to obtain a truly subcubic time algorithm for APSP, i.e. an algorithm running in $O(n^{3-\varepsilon})$ time for constant $\varepsilon>0$. To approach this problem, many truly subcubic time classical algorithms have been devised for APSP and its variants for structured inputs. Some examples of such problems are APSP in geometrically weighted graphs, graphs with small integer edge weights or a small number of weights incident to each vertex, and the all pairs earliest arrivals problem. In this paper we revisit these problems in the quantum setting and obtain the first nontrivial (i.e. $O(n^{2.5-\varepsilon})$ time) quantum algorithms for the problems.
- Published
- 2014
16. Mouse visual cortex as a limited resource system that self-learns an ecologically-general representation
- Author
-
Nayebi, Aran, primary, Kong, Nathan C. L., additional, Zhuang, Chengxu, additional, Gardner, Justin L., additional, Norcia, Anthony M., additional, and Yamins, Daniel L. K., additional
- Published
- 2023
- Full Text
- View/download PDF
17. Quantum lower bound for inverting a permutation with advice
- Author
-
Nayebi, Aran, Aaronson, Scott, Belovs, Aleksandrs, and Trevisan, Luca
- Subjects
Quantum Physics ,Computer Science - Computational Complexity ,Computer Science - Cryptography and Security - Abstract
Given a random permutation $f: [N] \to [N]$ as a black box and $y \in [N]$, we want to output $x = f^{-1}(y)$. Supplementary to our input, we are given classical advice in the form of a pre-computed data structure; this advice can depend on the permutation but \emph{not} on the input $y$. Classically, there is a data structure of size $\tilde{O}(S)$ and an algorithm that with the help of the data structure, given $f(x)$, can invert $f$ in time $\tilde{O}(T)$, for every choice of parameters $S$, $T$, such that $S\cdot T \ge N$. We prove a quantum lower bound of $T^2\cdot S \ge \tilde{\Omega}(\epsilon N)$ for quantum algorithms that invert a random permutation $f$ on an $\epsilon$ fraction of inputs, where $T$ is the number of queries to $f$ and $S$ is the amount of advice. This answers an open question of De et al. We also give a $\Omega(\sqrt{N/m})$ quantum lower bound for the simpler but related Yao's box problem, which is the problem of recovering a bit $x_j$, given the ability to query an $N$-bit string $x$ at any index except the $j$-th, and also given $m$ bits of advice that depend on $x$ but not on $j$., Comment: To appear in Quantum Information & Computation. Revised version based on referee comments
- Published
- 2014
18. Exponential prefixed polynomial equations
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Logic ,Mathematics - Combinatorics ,Mathematics - Number Theory ,Primary 03B70, Secondary 05C15, 05A18 - Abstract
A prefixed polynomial equation is an equation of the form $P(t_1,\ldots,t_n) = 0$, where $P$ is a polynomial whose variables $t_1,\ldots,t_n$ range over the natural numbers, preceded by quantifiers over some, or all, of its variables. Here, we consider exponential prefixed polynomial equations (EPPEs), where variables can also occur as exponents. We obtain a relatively concise EPPE equivalent to the combinatorial principle of the Paris-Harrington theorem for pairs (which is independent of primitive recursive arithmetic), as well as an EPPE equivalent to Goodstein's theorem (which is independent of Peano arithmetic). Some new devices are used in addition to known methods for the elimination of bounded universal quantifiers for Diophantine predicates.
- Published
- 2013
19. Practical intractability: a critique of the hypercomputation movement
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Logic ,Computer Science - Emerging Technologies ,Quantum Physics - Abstract
For over a decade, the hypercomputation movement has produced computational models that in theory solve the algorithmically unsolvable, but they are not physically realizable according to currently accepted physical theories. While opponents to the hypercomputation movement provide arguments against the physical realizability of specific models in order to demonstrate this, these arguments lack the generality to be a satisfactory justification against the construction of \emph{any} information-processing machine that computes beyond the universal Turing machine. To this end, I present a more mathematically concrete challenge to hypercomputability, and will show that one is immediately led into physical impossibilities, thereby demonstrating the infeasibility of hypercomputers more generally. This gives impetus to propose and justify a more plausible starting point for an extension to the classical paradigm that is physically possible, at least in principle. Instead of attempting to rely on infinities such as idealized limits of infinite time or numerical precision, or some other physically unattainable source, one should focus on extending the classical paradigm to better encapsulate modern computational problems that are not well-expressed/modeled by the closed-system paradigm of the Turing machine. I present the first steps toward this goal by considering contemporary computational problems dealing with intractability and issues surrounding cyber-physical systems, and argue that a reasonable extension to the classical paradigm should focus on these issues in order to be practically viable., Comment: To appear in Minds and Machines
- Published
- 2012
- Full Text
- View/download PDF
20. A Note on the Inverse Laplace Transformation of $f(t)$
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Classical Analysis and ODEs ,Primary 44A10, Secondary 26A33 - Abstract
Let $\mathcal{L}\{f(t)\} = \int_{0}^{\infty}e^{-st}f(t)dt$ denote the Laplace transform of $f$. It is well-known that if $f(t)$ is a piecewise continuous function on the interval $t:[0,\infty)$ and of exponential order for $t > N$; then $\lim_{s\to\infty}F(s) = 0$, where $F(s) = \mathcal{L}\{f(t)\}$. In this paper we prove that the lesser known converse does not hold true; namely, if $F(s)$ is a continuous function in terms of $s$ for which $\lim_{s\to\infty}F(s) = 0$, then it does not follow that $F(s)$ is the Laplace transform of a piecewise continuous function of exponential order., Comment: This paper has been withdrawn by the author due to an incorrect assumption based on equation (0.0.1)
- Published
- 2010
21. Upper bounds on the solutions to $n = p+m^2$
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Number Theory ,Primary 11P32, Secondary 11P55 - Abstract
Hardy and Littlewood conjectured that every large integer $n$ that is not a square is the sum of a prime and a square. They believed that the number $\mathcal{R}(n)$ of such representations for $n = p+m^2$ is asymptotically given by \mathcal{R}(n) \sim \frac{\sqrt{n}}{\log n}\prod_{p=3}^{\infty}(1-\frac{1}{p-1}(\frac{n}{p})), where $p$ is a prime, $m$ is an integer, and $(\frac{n}{p})$ denotes the Legendre symbol. Unfortunately, as we will later point out, this conjecture is difficult to prove and not \emph{all} integers that are nonsquares can be represented as the sum of a prime and a square. Instead in this paper we prove two upper bounds for $\mathcal{R}(n)$ for $n \le N$. The first upper bound applies to \emph{all} $n \le N$. The second upper bound depends on the possible existence of the Siegel zero, and assumes its existence, and applies to all $N/2 < n \le N$ but at most $\ll N^{1-\delta_1}$ of these integers, where $N$ is a sufficiently large positive integer and $0< \delta_1 \le 0.000025$.
- Published
- 2010
22. Fast matrix multiplication techniques based on the Adleman-Lipton model
- Author
-
Nayebi, Aran
- Subjects
Quantitative Biology - Quantitative Methods ,Computer Science - Data Structures and Algorithms ,Computer Science - Emerging Technologies ,65F05, 03D10 (Primary) 68Q10, 68Q05, 03D80 (Secondary) - Abstract
On distributed memory electronic computers, the implementation and association of fast parallel matrix multiplication algorithms has yielded astounding results and insights. In this discourse, we use the tools of molecular biology to demonstrate the theoretical encoding of Strassen's fast matrix multiplication algorithm with DNA based on an $n$-moduli set in the residue number system, thereby demonstrating the viability of computational mathematics with DNA. As a result, a general scalable implementation of this model in the DNA computing paradigm is presented and can be generalized to the application of \emph{all} fast matrix multiplication algorithms on a DNA computer. We also discuss the practical capabilities and issues of this scalable implementation. Fast methods of matrix computations with DNA are important because they also allow for the efficient implementation of other algorithms (i.e. inversion, computing determinants, and graph theory) with DNA., Comment: To appear in the International Journal of Computer Engineering Research. Minor changes made to make the preprint as similar as possible to the published version
- Published
- 2009
- Full Text
- View/download PDF
23. On integers as the sum of a prime and a $k$-th power
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Number Theory ,Computer Science - Data Structures and Algorithms ,11P32, 11P55, 11D85 - Abstract
Let $\mathcal{R}_k(n)$ be the number of representations of an integer $n$ as the sum of a prime and a $k$-th power. Define E_k(X) := |\{n \le X, n \in I_k, n\text{not a sum of a prime and a $k$-th power}\}|. Hardy and Littlewood conjectured that for $k = 2$ and $k=3$, E_k(X) \ll_{k} 1. In this note we present an alternative approach grounded in the theory of Diophantine equations towards a proof of the conjecture for all $k \ge 2$., Comment: This paper has been withdrawn by the author due to several errors in the manuscript, a prominent problem being that it has been known at least since Tarski that in real numbers there exists a deterministic Turing machine which determines if a variety is empty or nonempty
- Published
- 2009
24. On the distribution of Carmichael numbers
- Author
-
Nayebi, Aran
- Subjects
Mathematics - Number Theory ,Primary 11Y35, 11N60, Secondary 11N05, 11N37, 11N25, 11Y11 - Abstract
Erd\H{o}s conjectured in 1956 that there are $x^{1-o(1)}$ Carmichael numbers up to $x$. Pomerance made this conjecture more precise and proposed that there are $x^{1-{\frac{\{1+o(1)\}\log\log\log x}{\log\log x}}}$ Carmichael numbers up to $x$. At the time, his data tables up to $25 \cdot 10^{9}$ appeared to support his conjecture. However, Pinch extended this data and showed that up to $10^{21}$, Pomerance's conjecture did not appear well-supported. Thus, the purpose of this paper is two-fold. First, we build upon the work of Pomerance and others to present an alternate conjecture regarding the distribution of Carmichael numbers that fits proven bounds and is better supported by Pinch's new data. Second, we provide another conjecture concerning the distribution of Carmichael numbers that sharpens Pomerance's heuristic arguments. We also extend and update counts pertaining to pseudoprimes and Carmichael numbers, and discuss the distribution of One-Parameter Quadratic-Base Test pseudoprimes., Comment: This paper has been withdrawn by the author because I do not believe the conjectures are plausible
- Published
- 2009
25. Efficient Hybrid Algorithms for Computing Clusters Overlap
- Author
-
Javangula, Pradeep, Modarre, Kourosh, Shenoy, Paresh, Liu, Yi, and Nayebi, Aran
- Published
- 2017
- Full Text
- View/download PDF
26. For Human-Like Models, Train on Human-Like Tasks
- Author
-
Hermann, Katherine, primary, Nayebi, Aran, additional, van Steenkiste, Sjoerd, additional, and Jones, Matthew, additional
- Published
- 2023
- Full Text
- View/download PDF
27. Neural Mechanisms of Mental Simulation in Primate Frontal Cortex
- Author
-
Nayebi, Aran, primary, Rajalingham, Rishi, additional, Jazayeri, Mehrdad, additional, and Yang, Guangyu, additional
- Published
- 2023
- Full Text
- View/download PDF
28. When and why grid cells appear or not in trained path integrators
- Author
-
Sorscher, Ben, primary, Mel, Gabriel C., additional, Nayebi, Aran, additional, Giocomo, Lisa, additional, Yamins, Daniel, additional, and Ganguli, Surya, additional
- Published
- 2022
- Full Text
- View/download PDF
29. Recurrent Connections in the Primate Ventral Visual Stream Mediate a Trade-Off Between Task Performance and Network Size During Core Object Recognition
- Author
-
Nayebi, Aran, primary, Sagastuy-Brena, Javier, additional, Bear, Daniel M., additional, Kar, Kohitij, additional, Kubilius, Jonas, additional, Ganguli, Surya, additional, Sussillo, David, additional, DiCarlo, James J., additional, and Yamins, Daniel L. K., additional
- Published
- 2022
- Full Text
- View/download PDF
30. Modelling inter-animal variability
- Author
-
Sagastuy-Brena, Javier, primary, Thobani, Imran, additional, Nayebi, Aran, additional, Cao, Rosa, additional, and Yamins, Dan, additional
- Published
- 2022
- Full Text
- View/download PDF
31. Distinct in vivo dynamics of excitatory synapses onto cortical pyramidal neurons and parvalbumin-positive interneurons
- Author
-
Melander, Joshua B., Nayebi, Aran, Jongbloets, Bart C., Fortin, Dale A., Qin, Maozhen, Ganguli, Surya, Mao, Tianyi, and Zhong, Haining
- Published
- 2021
- Full Text
- View/download PDF
32. Explaining heterogeneity in medial entorhinal cortex with task-driven neural networks
- Author
-
Nayebi, Aran, primary, Attinger, Alexander, additional, Campbell, Malcolm G., additional, Hardcastle, Kiah, additional, Low, Isabel I.C., additional, Mallory, Caitlin S., additional, Mel, Gabriel C., additional, Sorscher, Ben, additional, Williams, Alex H., additional, Ganguli, Surya, additional, Giocomo, Lisa M., additional, and Yamins, Daniel L.K., additional
- Published
- 2021
- Full Text
- View/download PDF
33. Practical Intractability: A Critique of the Hypercomputation Movement
- Author
-
Nayebi, Aran
- Published
- 2014
- Full Text
- View/download PDF
34. Mouse visual cortex as a limited resource system that self-learns an ecologically-general representation
- Author
-
Nayebi, Aran, primary, Kong, Nathan C. L., additional, Zhuang, Chengxu, additional, Gardner, Justin L., additional, Norcia, Anthony M., additional, and Yamins, Daniel L. K., additional
- Published
- 2021
- Full Text
- View/download PDF
35. Distinctin vivodynamics of excitatory synapses onto cortical pyramidal neurons and inhibitory interneurons
- Author
-
Melander, Joshua B., primary, Nayebi, Aran, additional, Jongbloets, Bart C., additional, Fortin, Dale A., additional, Qin, Maozhen, additional, Ganguli, Surya, additional, Mao, Tianyi, additional, and Zhong, Haining, additional
- Published
- 2021
- Full Text
- View/download PDF
36. Recurrent Connections in the Primate Ventral Visual Stream Mediate a Tradeoff Between Task Performance and Network Size During Core Object Recognition
- Author
-
Nayebi, Aran, primary, Sagastuy-Brena, Javier, additional, Bear, Daniel M., additional, Kar, Kohitij, additional, Kubilius, Jonas, additional, Ganguli, Surya, additional, Sussillo, David, additional, DiCarlo, James J., additional, and Yamins, Daniel L. K., additional
- Published
- 2021
- Full Text
- View/download PDF
37. Distinct in vivo Dynamics of Excitatory Synapses Onto Cortical Pyramidal Neurons and Inhibitory Interneurons
- Author
-
Melander, Joshua B., primary, Nayebi, Aran, additional, Jongbloets, Bart C., additional, Fortin, Dale A., additional, Qin, Maozhen, additional, Ganguli, Surya, additional, Mao, Tianyi, additional, and Zhong, Haining, additional
- Published
- 2021
- Full Text
- View/download PDF
38. Unsupervised Neural Network Models of the Ventral Visual Stream
- Author
-
Zhuang, Chengxu, primary, Yan, Siming, additional, Nayebi, Aran, additional, Schrimpf, Martin, additional, Frank, Michael C., additional, DiCarlo, James J., additional, and Yamins, Daniel L. K., additional
- Published
- 2020
- Full Text
- View/download PDF
39. Self-supervised Neural Network Models of Higher Visual Cortex Development
- Author
-
Zhuang, Chengxu, primary, Yan, Siming, additional, Nayebi, Aran, additional, and Yamins, Daniel, additional
- Published
- 2019
- Full Text
- View/download PDF
40. CORnet: Modeling the Neural Mechanisms of Core Object Recognition
- Author
-
Kubilius, Jonas, primary, Schrimpf, Martin, additional, Nayebi, Aran, additional, Bear, Daniel, additional, Yamins, Daniel L. K., additional, and DiCarlo, James J., additional
- Published
- 2018
- Full Text
- View/download PDF
41. Convolutional recurrent neural network models of dynamics in higher visual cortex
- Author
-
Nayebi, Aran, primary, Kubilius, Jonas, additional, Bear, Daniel, additional, Ganguli, Surya, additional, DiCarlo, James, additional, and Yamins, Daniel, additional
- Published
- 2018
- Full Text
- View/download PDF
42. The dynamic neural code of the retina for natural scenes
- Author
-
Maheswaranathan, Niru, primary, McIntosh, Lane T., additional, Tanaka, Hidenori, additional, Grant, Satchel, additional, Kastner, David B., additional, Melander, Josh B., additional, Nayebi, Aran, additional, Brezovec, Luke, additional, Wang, Julia, additional, Ganguli, Surya, additional, and Baccus, Stephen A., additional
- Published
- 2018
- Full Text
- View/download PDF
43. Quantum lower bound for inverting a permutation with advice
- Author
-
Nayebi, Aran, primary, Aaronson, Scott, additional, Belovs, Aleksandrs, additional, and Trevisan, Luca, additional
- Published
- 2015
- Full Text
- View/download PDF
44. Practical Intractability: A Critique of the Hypercomputation Movement
- Author
-
Nayebi, Aran, primary
- Published
- 2013
- Full Text
- View/download PDF
45. Distinct in vivodynamics of excitatory synapses onto cortical pyramidal neurons and parvalbumin-positive interneurons
- Author
-
Melander, Joshua B., Nayebi, Aran, Jongbloets, Bart C., Fortin, Dale A., Qin, Maozhen, Ganguli, Surya, Mao, Tianyi, and Zhong, Haining
- Abstract
Cortical function relies on the balanced activation of excitatory and inhibitory neurons. However, little is known about the organization and dynamics of shaft excitatory synapses onto cortical inhibitory interneurons. Here, we use the excitatory postsynaptic marker PSD-95, fluorescently labeled at endogenous levels, as a proxy for excitatory synapses onto layer 2/3 pyramidal neurons and parvalbumin-positive (PV+) interneurons in the barrel cortex of adult mice. Longitudinal in vivoimaging under baseline conditions reveals that, although synaptic weights in both neuronal types are log-normally distributed, synapses onto PV+neurons are less heterogeneous and more stable. Markov model analyses suggest that the synaptic weight distribution is set intrinsically by ongoing cell-type-specific dynamics, and substantial changes are due to accumulated gradual changes. Synaptic weight dynamics are multiplicative, i.e., changes scale with weights, although PV+synapses also exhibit an additive component. These results reveal that cell-type-specific processes govern cortical synaptic strengths and dynamics.
- Published
- 2021
- Full Text
- View/download PDF
46. For human-like models, train on human-like tasks.
- Author
-
Hermann K, Nayebi A, van Steenkiste S, and Jones M
- Subjects
- Humans, Neural Networks, Computer, Deep Learning
- Abstract
Bowers et al. express skepticism about deep neural networks (DNNs) as models of human vision due to DNNs' failures to account for results from psychological research. We argue that to fairly assess DNNs, we must first train them on more human-like tasks which we hypothesize will induce more human-like behaviors and representations.
- Published
- 2023
- Full Text
- View/download PDF
47. Neural Foundations of Mental Simulation: Future Prediction of Latent Representations on Dynamic Scenes.
- Author
-
Nayebi A, Rajalingham R, Jazayeri M, and Yang GR
- Abstract
Humans and animals have a rich and flexible understanding of the physical world, which enables them to infer the underlying dynamical trajectories of objects and events, plausible future states, and use that to plan and anticipate the consequences of actions. However, the neural mechanisms underlying these computations are unclear. We combine a goal-driven modeling approach with dense neurophysiological data and high-throughput human behavioral readouts that contain thousands of comparisons to directly impinge on this question. Specifically, we construct and evaluate several classes of sensory-cognitive networks to predict the future state of rich, ethologically-relevant environments, ranging from self-supervised end-to-end models with pixel-wise or object-slot objectives, to models that future predict in the latent space of purely static image-pretrained or dynamic video-pretrained foundation models. We find that "scale is not all you need", and that many state-of-the-art machine learning models fail to perform well on our neural and behavioral benchmarks for future prediction. In fact, only one class of models matches these data well overall. We find that neural responses are currently best predicted by models trained to predict the future state of their environment in the latent space of pretrained foundation models optimized for dynamic scenes in a self-supervised manner. These models also approach the neurons' ability to predict the environmental state variables that are visually hidden from view, despite not being explicitly trained to do so. Finally, we find that not all foundation model latents are equal. Notably, models that future predict in the latent space of video foundation models that are optimized to support a diverse range of egocentric sensorimotor tasks, reasonably match both human behavioral error patterns and neural dynamics across all environmental scenarios that we were able to test. Overall, these findings suggest that the neural mechanisms and behaviors of primate mental simulation have strong inductive biases associated with them, and are thus far most consistent with being optimized to future predict on reusable visual representations that are useful for Embodied AI more generally.
- Published
- 2023
48. From deep learning to mechanistic understanding in neuroscience: the structure of retinal prediction.
- Author
-
Tanaka H, Nayebi A, Maheswaranathan N, McIntosh L, Baccus SA, and Ganguli S
- Abstract
Recently, deep feedforward neural networks have achieved considerable success in modeling biological sensory processing, in terms of reproducing the input-output map of sensory neurons. However, such models raise profound questions about the very nature of explanation in neuroscience. Are we simply replacing one complex system (a biological circuit) with another (a deep network), without understanding either? Moreover, beyond neural representations, are the deep network's computational mechanisms for generating neural responses the same as those in the brain? Without a systematic approach to extracting and understanding computational mechanisms from deep neural network models, it can be difficult both to assess the degree of utility of deep learning approaches in neuroscience, and to extract experimentally testable hypotheses from deep networks. We develop such a systematic approach by combining dimensionality reduction and modern attribution methods for determining the relative importance of interneurons for specific visual computations. We apply this approach to deep network models of the retina, revealing a conceptual understanding of how the retina acts as a predictive feature extractor that signals deviations from expectations for diverse spatiotemporal stimuli. For each stimulus, our extracted computational mechanisms are consistent with prior scientific literature, and in one case yields a new mechanistic hypothesis. Thus overall, this work not only yields insights into the computational mechanisms underlying the striking predictive capabilities of the retina, but also places the framework of deep networks as neuroscientific models on firmer theoretical foundations, by providing a new roadmap to go beyond comparing neural representations to extracting and understand computational mechanisms.
- Published
- 2019
49. Deep Learning Models of the Retinal Response to Natural Scenes.
- Author
-
McIntosh LT, Maheswaranathan N, Nayebi A, Ganguli S, and Baccus SA
- Abstract
A central challenge in sensory neuroscience is to understand neural computations and circuit mechanisms that underlie the encoding of ethologically relevant, natural stimuli. In multilayered neural circuits, nonlinear processes such as synaptic transmission and spiking dynamics present a significant obstacle to the creation of accurate computational models of responses to natural stimuli. Here we demonstrate that deep convolutional neural networks (CNNs) capture retinal responses to natural scenes nearly to within the variability of a cell's response, and are markedly more accurate than linear-nonlinear (LN) models and Generalized Linear Models (GLMs). Moreover, we find two additional surprising properties of CNNs: they are less susceptible to overfitting than their LN counterparts when trained on small amounts of data, and generalize better when tested on stimuli drawn from a different distribution (e.g. between natural scenes and white noise). An examination of the learned CNNs reveals several properties. First, a richer set of feature maps is necessary for predicting the responses to natural scenes compared to white noise. Second, temporally precise responses to slowly varying inputs originate from feedforward inhibition, similar to known retinal mechanisms. Third, the injection of latent noise sources in intermediate layers enables our model to capture the sub-Poisson spiking variability observed in retinal ganglion cells. Fourth, augmenting our CNNs with recurrent lateral connections enables them to capture contrast adaptation as an emergent property of accurately describing retinal responses to natural scenes. These methods can be readily generalized to other sensory modalities and stimulus ensembles. Overall, this work demonstrates that CNNs not only accurately capture sensory circuit responses to natural scenes, but also can yield information about the circuit's internal structure and function.
- Published
- 2016
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.