986 results on '"Large numbers"'
Search Results
2. Mathematical cognition related to the large numbers in early societies: A study based on 5th-century Buddhist Commentaries in Sri Lanka.
- Author
-
Jayawardana, Chandana
- Subjects
- *
MATHEMATICAL forms , *BUDDHISTS , *COGNITION , *HISTORY of mathematics , *ASIAN studies - Abstract
Universal presentations and interpretations of the history of mathematics have now been challenged, identifying the need of resurrecting hitherto marginalized contributions. In such studies, cognitive factors influencing the related development trajectories play a crucial role. Different societies would have had different mathematical cognitions contributing to the development of distinct forms of mathematics. This study attempts to surface the related cognitive contents of using large numbers. The principal source used for the study—the Buddhist Commentaries—determines the period and region covered. As such, it explores the conditions prevailed in 5th century CE (approximately) in Sri Lanka. 2000 Mathematics Subject Classification (MSC 2000) code: 01A07 [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. Numerical Implementation of Multidimensional Functions Extremum Search
- Author
-
Kovalenko, Lyudmila, Kalmykov, Oleg, Reznik, Petro, Demianenko, Ivan, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Arsenyeva, Olga, editor, Romanova, Tetyana, editor, Sukhonos, Maria, editor, Biletskyi, Ihor, editor, and Tsegelnyk, Yevgen, editor
- Published
- 2023
- Full Text
- View/download PDF
4. A million is more than a thousand: Children's acquisition of very large number words.
- Author
-
Cheung, Pierina and Ansari, Daniel
- Subjects
- *
SPEECH , *VOCABULARY , *SEMANTICS , *SPEECH processing systems - Abstract
Very large numbers words such as "hundred," "thousand," "million," "billion," and "trillion" pose a learning problem for children because they are sparse in everyday speech and children's experience with extremely large quantities is scarce. In this study, we examine when children acquire the relative ordering of very large number words as a first step toward understanding their acquisition. In Study 1, a hundred and twenty‐five 5–8‐year‐olds participated in a verbal number comparison task involving very large number words. We found that children can judge which of two very large numbers is more as early as age 6, prior to entering first grade. In Study 2, we provided a descriptive analysis on the usage of very large number words using the CHILDES database. We found that the relative frequency of large number words does not change across the years, with "hundred" uttered more frequently than others by an order of magnitude. We also found that adults were more likely to use large number words to reference units of quantification for money, weight, and time, than for discrete, physical entities. Together, these results show that children construct a numerical scale for large number words prior to learning their precise cardinal meanings, and highlight how frequency and context may support their acquisition. Our results have pedagogical implications and highlight a need to investigate how children acquire meanings for number words that reference quantities beyond our everyday experience. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
5. How to compare power towers?
- Author
-
Gryszka, Karol
- Subjects
- *
FINITE, The , *EXPONENTIATION - Abstract
We provide various tools that give an answer to the following question: given two distinct finite tower powers with entries at least 1 and (possibly) different heights, determine which power tower is greater than the other one. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
6. Number and Counting in Context, Classifications and Large Numbers
- Author
-
Owens, Kay, Lean, Glen, Ellerton, Nerida F., Series Editor, Clements, M. A. Ken, Series Editor, Owens, Kay, Lean, Glen, Paraide, Patricia, and Muke, Charly
- Published
- 2018
- Full Text
- View/download PDF
7. Rewriting the History of Number from Papua New Guinea and Oceania Evidence
- Author
-
Owens, Kay, Lean, Glen, Ellerton, Nerida F., Series Editor, Clements, M. A. Ken, Series Editor, Owens, Kay, Lean, Glen, Paraide, Patricia, and Muke, Charly
- Published
- 2018
- Full Text
- View/download PDF
8. On Practical Nearest Sub-Trajectory Queries under the Fréchet Distance
- Author
-
Joachim Gudmundsson, John Pfeifer, and Martin P. Seybold
- Subjects
Computer science ,Heuristic ,Fréchet distance ,Large numbers ,Synthetic data ,k-nearest neighbors algorithm ,Computer Science Applications ,Modeling and Simulation ,Signal Processing ,Trajectory ,Discrete Mathematics and Combinatorics ,Geometry and Topology ,Focus (optics) ,Cluster analysis ,Algorithm ,Information Systems - Abstract
We study the problem of sub-trajectory nearest-neighbor queries on polygonal curves under the continuous Fréchet distance. Given an n vertex trajectory P and an m vertex query trajectory Q , we seek to report a vertex-aligned sub-trajectory P ′ of P that is closest to Q , i.e., P′ must start and end on contiguous vertices of P . Since in real data P typically contains a very large number of vertices, we focus on answering queries, without restrictions on P or Q , using only precomputed structures of 𝒪(n) size. We use three baseline algorithms from straightforward extensions of known work; however, they have impractical performance on realistic inputs. Therefore, we propose a new Hierarchical Simplification Tree (HST) data structure and an adaptive clustering-based query algorithm that efficiently explores relevant parts of P . The core of our query methods is a novel greedy-backtracking algorithm that solves the Fréchet decision problem using 𝒪(n+m) space and 𝒪O(nm) time in the worst case. Experiments on real and synthetic data show that our heuristic effectively prunes the search space and greatly reduces computations compared to baseline approaches.
- Published
- 2023
9. Edge-sorter: A hardware sorting engine for area & power constrained edge computing devices.
- Author
-
Beitollahi, Hakem, Pandi, Marziye, and Moghaddas, Mostafa
- Subjects
- *
EDGE computing , *HARDWARE , *RESEARCH personnel , *ENGINES , *SMART devices - Abstract
In recent years, hardware sorters have been an attracted topic for researchers. Since hardware sorters play a crucial role in embedded systems, several attempts have been made to efficiently design and implement these sorters. Previous state-of-the-art hardware sorters are not suitable for embedded edge computing devices because they (1) consume high power, (2) occupy high area, (3) work for limited data-width numbers, (4) require many memory resources, and (5) finally, their architecture is not scalable with the number of input records. This paper proposes a hardware sorter for edge devices with limited hardware resources. The proposed hardware sorter, called Edge-Sorter, processes 4 bits of input records at each clock cycle. Edge-Sorter utilizes the unary processing in its main processing core. Edge-Sorter has valuable attributes compared to previous state-of-the-art techniques, including low power consumption, low area occupation, sorting numbers without storing their indices, sorting numbers with arbitrary data-width, and scalable with the number of input records. The proposed approach is evaluated and compared with previous state-of-the-art techniques with two different implementation and synthesis environments: Xilinx Vivado FPGA-based and Synopsys Design Compiler 45-nm ASIC-based. The Synthesis results of both environments indicate that both Edge-Sorter techniques reduces area and power consumption on average by 80% and 90%, respectively compared to previous techniques. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Contiguous Line Segments in the Ulam Spiral: Experiments with Larger Numbers
- Author
-
Chmielewski, Leszek J., Janowicz, Maciej, Gawdzik, Grzegorz, Orłowski, Arkadiusz, Hutchison, David, Series editor, Kanade, Takeo, Series editor, Kittler, Josef, Series editor, Kleinberg, Jon M., Series editor, Mattern, Friedemann, Series editor, Mitchell, John C., Series editor, Naor, Moni, Series editor, Pandu Rangan, C., Series editor, Steffen, Bernhard, Series editor, Terzopoulos, Demetri, Series editor, Tygar, Doug, Series editor, Weikum, Gerhard, Series editor, Rutkowski, Leszek, editor, Korytkowski, Marcin, editor, Scherer, Rafał, editor, Tadeusiewicz, Ryszard, editor, Zadeh, Lotfi A., editor, and Zurada, Jacek M., editor
- Published
- 2017
- Full Text
- View/download PDF
11. Underdetermined Blind Source Separation for Sparse Signals Based on the Law of Large Numbers and Minimum Intersection Angle Rule.
- Author
-
Xu, Pengfei, Jia, Yinjie, Wang, Zhijian, and Jiang, Mingxin
- Subjects
- *
BLIND source separation , *SIGNAL separation , *INTERSECTION numbers , *DIGITAL signal processing , *LAW of large numbers - Abstract
Underdetermined blind source separation (UBSS) is an important issue for sparse signals, and a novel two-step approach for UBSS based on the law of large numbers and minimum intersection angle rule (LM method) is presented. In the first step, an estimation of the mixed matrix is obtained by using the law of large numbers, and the number of source signals is displayed graphically. In the second step, a method of estimating the source signals by the minimum intersection angle rule is proposed. The significance of this step is that the minimum intersection rule is better than the shortest path method, and the decomposition components can be found optimally by the former. The simulation results illustrate the effectiveness of the LM method. It has a simple principle, has good transplantation capability and may be widely applied in various fields of digital signal processing. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
12. Our Intuitive Grasp of the Repugnant Conclusion
- Author
-
Gustafsson, Johan E., Arrhenius, Gustaf, book editor, Bykvist, Krister, book editor, Campbell, Tim, book editor, and Finneron-Burns, Elizabeth, book editor
- Published
- 2022
- Full Text
- View/download PDF
13. Dirac's large number hypothesis: A journey from concept to implication.
- Author
-
Ray, Saibal, Mukhopadhyay, Utpal, Ray, Soham, and Bhattacharjee, Arjak
- Subjects
- *
GRAVITATIONAL constant , *PHYSICAL constants , *DIMENSIONLESS numbers , *LIFE sciences , *HYPOTHESIS - Abstract
Large dimensionless numbers, arising out of ratios of various physical constants, intrigued many scientists, especially Dirac. Relying on the coincidence of large numbers, Dirac arrived at the revolutionary hypothesis that the gravitational constant G should vary inversely at the cosmic time t. This hypothesis of Dirac, known as the Large Number Hypothesis (LNH), sparked off many speculations, arguments and new ideas in terms of applications. Works done by several authors with LNH as their basic platform are extensively reviewed in this work. Relationship between some of those works are pointed out here elaborately. Possibility of time variations of physical constants other than G as well as large numbers in various realm of physical and biological sciences are also discussed. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
14. Coping with the Tiny: Measures, Computations, and Reasoning with Small Amounts in Jean des Murs's Quadrivial Works.
- Author
-
Husson, Matthieu
- Subjects
- *
MUSIC theory , *ASTRONOMY , *ARITHMETIC mean , *FRACTIONS , *MATHEMATICS - Abstract
This paper aims at complementing the cross-disciplinary intellectual portrait of Jean des Murs presented in this special issue from the particular perspective of his arithmetical and geometrical works. If looked at from this angle, the general question immediately implies new ones: what are the connections between Jean's arithmetical or geometrical works and those in the other fields of the quadrivium ? What is their importance in the global intellectual picture of Jean des Murs as a quadrivial author of the fourteenth century? In order to explore these different issues, I have chosen to focus on a specific question: to what extent do Jean des Murs's mathematical works reflect a persistent effort throughout his career to address the various issues related to small amounts? Focusing on this particularly significant problem will give me the opportunity to cover a representative sample of Jean des Murs's quadrivial works from a specific perspective. Music and astronomy address small amounts in different contexts and with different mathematical tools. However, some deeper similarities appear. They concern mainly the question of measure and incommensurability as well as the relation between mathematical and physical reasoning. Overall, they reflect the intellectual coherence between quadrivial sciences as attested by Jean des Murs's works. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
15. Belief functions clustering for epipole localization
- Author
-
Sylvie Le Hégarat-Mascle, Huiqin Chen, and Emanuel Aldea
- Subjects
business.industry ,Computer science ,Applied Mathematics ,Large numbers ,Pattern recognition ,0102 computer and information sciences ,01 natural sciences ,Theoretical Computer Science ,03 medical and health sciences ,0302 clinical medicine ,010201 computation theory & mathematics ,Artificial Intelligence ,Robustness (computer science) ,Outlier ,Information source ,030212 general & internal medicine ,Artificial intelligence ,Cluster analysis ,business ,Software - Abstract
This work deals with the clustering of information sources for epipole estimation in a multi-camera system. For this problem, each pair of matched visual features in the images can be considered as an elementary information source. The epipole is then estimated by combining these elementary sources taking into account their inadequacy, in particular large imprecision and presence of outliers, as well as the very large number of sources. We address the challenges introduced by a large number of sources with a strategy based on clustering and intra-cluster fusion using the Belief Functions framework. When evaluated on real data, the proposed algorithm exhibits more robustness in terms of accuracy and precision than the standard approaches which provide singular solutions.
- Published
- 2021
16. Pre-service teachers’ conceptions of the magnitude of large numbers.
- Author
-
Brass, Amber and Harkness, Shelly Sheats
- Subjects
LAW of large numbers ,MAGNITUDE estimation ,MATHEMATICS teachers - Abstract
This investigation examined the thinking of 128 pre-service teachers about the magnitude of large numbers and the relationships between large numbers. These pre-service teachers responded to a written task asking them to place one billion on a number line that included zero and one trillion and to provide an explanation for their placement. One billion was placed proportionately to zero and one trillion on the number line in only 20% of responses. The data revealed nine distinct strategies that emerged and different conceptions the pre-service teachers used for making the placements. The authors discuss possible reasons for the different conceptions. [ABSTRACT FROM PUBLISHER]
- Published
- 2017
- Full Text
- View/download PDF
17. Data Detection in Large MIMO Satellite Communication Systems
- Author
-
M K Arti
- Subjects
Beamforming ,Scheme (programming language) ,Computer science ,Data detection ,Real-time computing ,MIMO ,Large numbers ,020206 networking & telecommunications ,020302 automobile design & engineering ,02 engineering and technology ,Moment-generating function ,0203 mechanical engineering ,Control and Systems Engineering ,0202 electrical engineering, electronic engineering, information engineering ,Communications satellite ,Satellite ,Electrical and Electronic Engineering ,computer ,Computer Science::Information Theory ,computer.programming_language - Abstract
A large multiple-input multiple-output satellite communication system is proposed, where satellite and Earth stations (ESs) contain a very large number of antennas. ESs want to exchange their data via a satellite. Data of transmitting ES is transmitted by using a beamforming vector. At the receiving ES, the data of unwanted ESs is canceled by utilizing null space based approach. Then a combining vector is used at the receiving ES in order to retrieve the data. It is observed that sufficiently large number of antennas at the ESs are needed by the proposed scheme. By following the standard moment generating function based approach, analytical results of symbol error rate of the proposed system are formulated.
- Published
- 2021
18. Analysis Of Heart Disease Using Machine Learning Techniques
- Author
-
Samson Cherlapally et.al
- Subjects
Heart disease ,business.industry ,Computer science ,General Mathematics ,Large numbers ,Python (programming language) ,medicine.disease ,Machine learning ,computer.software_genre ,Education ,Computational Mathematics ,Tree (data structure) ,Computational Theory and Mathematics ,Data extraction ,Data exchange ,medicine ,Artificial intelligence ,business ,Health sector ,computer ,computer.programming_language - Abstract
There is a very large number in the health sector and special methods are also used systematically. Data exchange is one of the most commonly used methods. Heart disease is one of the leading causes of death in the world. This system predicts the possibility of heart disease. The results of this system provide a 100% risk of heart disease. The data used are categorized according to medical criteria. The system evaluates these parameters using data extraction methods used in Python using two basic machine learning algorithms, the Solution Tree Algorithm, and the algorithm that demonstrates the best accuracy in heart disease.
- Published
- 2021
19. ContractWard: Automated Vulnerability Detection Models for Ethereum Smart Contracts
- Author
-
Chunhua Su, Hao Wang, Wei Wang, Yidong Li, Jingjing Song, and Guangquan Xu
- Subjects
Dependency (UML) ,Smart contract ,Computer Networks and Communications ,Computer science ,Bigram ,Distributed computing ,Feature extraction ,Large numbers ,020206 networking & telecommunications ,02 engineering and technology ,computer.file_format ,021001 nanoscience & nanotechnology ,Symbolic execution ,Computer Science Applications ,Control and Systems Engineering ,0202 electrical engineering, electronic engineering, information engineering ,Executable ,Layer (object-oriented design) ,0210 nano-technology ,computer - Abstract
Smart contracts are decentralized applications running on Blockchain. A very large number of smart contracts has been deployed on Ethereum. Meanwhile, security flaws of contracts have led to huge pecuniary losses and destroyed the ecological stability of contract layer on Blockchain. It is thus an emerging yet crucial issue to effectively and efficiently detect vulnerabilities in contracts. Existing detection methods like Oyente and Securify are mainly based on symbolic execution or analysis. These methods are very time-consuming, as the symbolic execution requires the exploration of all executable paths or the analysis of dependency graphs in a contract. In this work, we propose ContractWard to detect vulnerabilities in smart contracts with machine learning techniques. First, we extract bigram features from simplified operation codes of smart contracts. Second, we employ five machine learning algorithms and two sampling algorithms to build the models. ContractWard is evaluated with 49502 real-world smart contracts running on Ethereum. The experimental results demonstrate the effectiveness and efficiency of ContractWard. The predictive Micro-F1 and Macro-F1 of ContractWard are over 96% and the average detection time is 4 seconds on each smart contract when we use XGBoost for training the models and SMOTETomek for balancing the training sets. © 2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
- Published
- 2021
20. A Termination Criterion for Probabilistic Point Clouds Registration
- Author
-
Domenico G. Sorrenti and Simone Fontana
- Subjects
0209 industrial biotechnology ,Computer science ,lcsh:T57-57.97 ,Work (physics) ,Point set ,Probabilistic logic ,Point cloud ,Large numbers ,ICP ,alignment ,02 engineering and technology ,020901 industrial engineering & automation ,point set ,lcsh:Applied mathematics. Quantitative methods ,point cloud registration ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Algorithm - Abstract
Probabilistic Point Clouds Registration (PPCR) is an algorithm that, in its multi-iteration version, outperformed state-of-the-art algorithms for local point clouds registration. However, its performances have been tested using a fixed high number of iterations. To be of practical usefulness, we think that the algorithm should decide by itself when to stop, on one hand to avoid an excessive number of iterations and waste computational time, on the other to avoid getting a sub-optimal registration. With this work, we compare different termination criteria on several datasets, and prove that the chosen one produces very good results that are comparable to those obtained using a very large number of iterations, while saving computational time.
- Published
- 2021
21. Noise-adaptive synthetic oversampling technique
- Author
-
Trang Nguyen, Minh Thanh Vo, H. Anh Vo, and Tuong Le
- Subjects
Computer science ,business.industry ,Supervised learning ,Large numbers ,Sample (statistics) ,02 engineering and technology ,Minority class ,Machine learning ,computer.software_genre ,Field (computer science) ,Reduction (complexity) ,ComputingMethodologies_PATTERNRECOGNITION ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Oversampling ,020201 artificial intelligence & image processing ,Noise (video) ,Artificial intelligence ,business ,computer - Abstract
In the field of supervised learning, the problem of class imbalance is one of the most difficult problems, and has attracted a great deal of research attention in recent years. In an imbalanced dataset, minority classes are those that contain very small numbers of data samples, while the remaining classes have a very large number of data samples. This type of imbalance reduces the predictive performance of machine learning models. There are currently three approaches for dealing with the class imbalance problem: algorithm-level, data-level, and ensemble-based approaches. Of these, data-level approaches are the most widely used, and consist of three sub-categories: under-sampling, oversampling, and hybrid techniques. Oversampling techniques generate synthetic samples for the minority class to balance an imbalanced dataset. However, existing oversampling approaches do not have a strategy for handling noise samples in imbalanced and noisy datasets, which leads to a reduction in the predictive performance of machine learning models. This study therefore proposes a noise-adaptive synthetic oversampling technique (NASOTECH) to deal with the class imbalance problem in imbalanced and noisy datasets. The noise-adaptive synthetic oversampling (NASO) strategy is first introduced, which is used to identify the number of samples generated for each sample in the minority class, based on the concept of the noise ratio. Next, the NASOTECH algorithm is proposed, based on the NASO strategy, to handle the class imbalance problem in imbalanced and noisy datasets. Finally, empirical experiments are conducted on several synthetic and real datasets to verify the effectiveness of the proposed approach. The experimental results confirm that NASOTECH outperforms three state-of-the-art oversampling techniques in terms of accuracy and geometric mean (G-mean) on imbalanced and noisy datasets.
- Published
- 2021
22. Face Recognition Based on Statistical Texture Features
- Author
-
Lama Akram Ibrahim, Nasser Nasser, and Majd Ali
- Subjects
business.industry ,Computer science ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Large numbers ,Image processing ,Pattern recognition ,Facial recognition system ,Image (mathematics) ,High memory ,Robustness (computer science) ,Pattern recognition (psychology) ,Artificial intelligence ,Focus (optics) ,business - Abstract
Facial recognition has attracted the attention of researchers and has been one of the most prominent topics in the fields of image processing and pattern recognition since 1990. This resulted in a very large number of recognition methods and techniques with the aim of increasing the accuracy and robustness of existing systems. Many techniques have been developed to address the challenges and reliable recognition systems have been reached but require considerable processing time, suffer from high memory consumption and are relatively complex. The focus of this paper is on extracting subset of descriptors (less correlated and less calculations) from the co-occurrence matrix with the goal of enhancing the performance of Haralick’s descriptors. Improvements are achieved by adding the image pre-processing and selecting the proper method according to the database problem and by extracting features from image local regions.
- Published
- 2021
23. Harnessing the Power of Emotion for Social Change: Review of Numbers and Nerves: Information, Emotion, and Meaning in a World of Data by Scott Slovic and Paul Slovic (2015).
- Author
-
Kelly, Anne M. W.
- Subjects
SOCIAL change ,SOCIAL psychology ,SOCIAL surveys ,EMOTIONS - Abstract
Literature and environment professor Scott Slovic, and his father, psychologist Paul Slovic, editors of this collection of essays and interviews, describe and demonstrate the psychological effects which hamper our ability to comprehend and respond appropriately to large numerical data. The collection then offers a brief survey of art works which, by first appealing to viewers' emotions, can potentially move the viewer to a better understanding of numbers. [ABSTRACT FROM AUTHOR]
- Published
- 2017
- Full Text
- View/download PDF
24. Cluster Analysis Tailored to Structure Change of Tropical Cyclones Using a Very Large Number of Trajectories
- Author
-
Elmar Schömer, Michael Riemer, Tobias Kremer, and Christian Euler
- Subjects
Atmospheric Science ,010504 meteorology & atmospheric sciences ,Climatology ,Cluster (physics) ,Structure (category theory) ,Large numbers ,010501 environmental sciences ,Tropical cyclone ,01 natural sciences ,Geology ,0105 earth and related environmental sciences - Abstract
Major airstreams in tropical cyclones (TCs) are rarely described from a Lagrangian perspective. Such a perspective, however, is required to account for asymmetries and time dependence of the TC circulation. We present a procedure that identifies main airstreams in TCs based on trajectory clustering. The procedure takes into account the TC’s large degree of inherent symmetry and is suitable for a very large number of trajectories . A large number of trajectories may be needed to resolve both the TC’s inner-core convection as well as the larger-scale environment. We define similarity of trajectories based on their shape in a storm-relative reference frame, rather than on proximity in physical space, and use Fréchet distance, which emphasizes differences in trajectory shape, as a similarity metric. To make feasible the use of this elaborate metric, data compression is introduced that approximates the shape of trajectories in an optimal sense. To make clustering of large numbers of trajectories computationally feasible, we reduce dimensionality in distance space by so-called landmark multidimensional scaling. Finally, k-means clustering is performed in this low-dimensional space. We investigate the extratropical transition of Tropical Storm Karl (2016) to demonstrate the applicability of our clustering procedure. All identified clusters prove to be physically meaningful and describe distinct flavors of inflow, ascent, outflow, and quasi-horizontal motion in Karl’s vicinity. Importantly, the clusters exhibit gradual temporal evolution, which is most notable because the clustering procedure itself does not impose temporal consistency on the clusters. Finally, TC problems are discussed for which the application of the clustering procedures seems to be most fruitful.
- Published
- 2020
25. Cryptanalysis and improvement of mutual authentication protocol for real-time data access in industrial wireless sensor networks
- Author
-
Devender Kumar, Karan Rawat, Shubham Singh Manhas, and Sai Kishore Pachigolla
- Subjects
business.industry ,Computer science ,ComputerSystemsOrganization_COMPUTER-COMMUNICATIONNETWORKS ,Physical unclonable function ,Large numbers ,020206 networking & telecommunications ,02 engineering and technology ,Computer Graphics and Computer-Aided Design ,Computer Science Applications ,law.invention ,Hardware and Architecture ,law ,Sensor node ,Mutual authentication protocol ,0202 electrical engineering, electronic engineering, information engineering ,Wireless ,ComputerSystemsOrganization_SPECIAL-PURPOSEANDAPPLICATION-BASEDSYSTEMS ,020201 artificial intelligence & image processing ,Real-time data ,Cryptanalysis ,business ,Wireless sensor network ,Software ,Computer network - Abstract
Wireless Sensor Networks (WSNs) have a very large number of applications in different domains of the industry. One of those applications is the industrial use of WSN. Industrial Wireless Sensor Net...
- Published
- 2020
26. A framework for adaptive open-pit mining planning under geological uncertainty
- Author
-
Tito Homem-de-Mello, Tomas Lagos, Denis Sauré, Margaret Armstrong, and Guido Lagos
- Subjects
Mathematical optimization ,021103 operations research ,Control and Optimization ,Adaptive optimization ,business.industry ,Computer science ,Mechanical Engineering ,0211 other engineering and technologies ,Aerospace Engineering ,Open-pit mining ,Large numbers ,02 engineering and technology ,Geostatistics ,Financial engineering ,Knapsack problem ,Production schedule ,Stochastic optimization ,021108 energy ,Electrical and Electronic Engineering ,business ,Software ,Civil and Structural Engineering - Abstract
Mine planning optimization aims at maximizing the profit obtained from extracting valuable ore. Beyond its theoretical complexity—the open-pit mining problem with capacity constraints reduces to a knapsack problem with precedence constraints, which is NP-hard—practical instances of the problem usually involve a large to very large number of decision variables, typically of the order of millions for large mines. Additionally, any comprehensive approach to mine planning ought to consider the underlying geostatistical uncertainty as only limited information obtained from drill hole samples of the mineral is initially available. In this regard, as blocks are extracted sequentially, information about the ore grades of blocks yet to be extracted changes based on the blocks that have already been mined. Thus, the problem lies in the class of multi-period large scale stochastic optimization problems with decision-dependent information uncertainty. Such problems are exceedingly hard to solve, so approximations are required. This paper presents an adaptive optimization scheme for multi-period production scheduling in open-pit mining under geological uncertainty that allows us to solve practical instances of the problem. Our approach is based on a rolling-horizon adaptive optimization framework that learns from new information that becomes available as blocks are mined. By considering the evolution of geostatistical uncertainty, the proposed optimization framework produces an operational policy that reduces the risk of the production schedule. Our numerical tests with mines of moderate sizes show that our rolling horizon adaptive policy gives consistently better results than a non-adaptive stochastic optimization formulation, for a range of realistic problem instances.
- Published
- 2020
27. Aggregation, Balancing, and Respect for the Claims of Individuals
- Author
-
Bastian Steuwer
- Subjects
Sociology and Political Science ,media_common.quotation_subject ,Small number ,05 social sciences ,Large numbers ,06 humanities and the arts ,0603 philosophy, ethics and religion ,Philosophy ,Harm ,060302 philosophy ,0502 economics and business ,Economics ,050207 economics ,Law and economics ,Skepticism ,media_common - Abstract
Most non-consequentialists “let the numbers count” when one can save either a lesser or greater number from equal or similar harm. But they are wary of doing so when one can save either a small number from grave harm or instead a very large number from minor harm. Limited aggregation is an approach that reconciles these two commitments. It is motivated by a powerful idea: our decision whom to save should respect each person who has a claim to our help, including those whom we fail to save. However, it has recently been argued that it is open to decisive objections. I develop a new limitedly aggregative view:Hybrid Balance Relevant Claims. This view is well grounded in the reasons we have to be skeptical of aggregation and resolves all recent challenges by paying careful attention to the rationale for limited aggregation.
- Published
- 2020
28. Derivation and analysis of parallel-in-time neural ordinary differential equations
- Author
-
Emmanuel Lorin
- Subjects
Artificial neural network ,Computer science ,Applied Mathematics ,Linear system ,Ode ,Large numbers ,02 engineering and technology ,Residual ,Nonlinear system ,Artificial Intelligence ,Ordinary differential equation ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Node (circuits) ,Algorithm - Abstract
The introduction in 2015 of Residual Neural Networks (RNN) and ResNET allowed for outstanding improvements of the performance of learning algorithms for evolution problems containing a “large” number of layers. Continuous-depth RNN-like models called Neural Ordinary Differential Equations (NODE) were then introduced in 2019. The latter have a constant memory cost, and avoid the a priori specification of the number of hidden layers. In this paper, we derive and analyze a parallel (-in-parameter and time) version of the NODE, which potentially allows for a more efficient implementation than a standard/naive parallelization of NODEs with respect to the parameters only. We expect this approach to be relevant whenever we have access to a very large number of processors, or when we are dealing with high dimensional ODE systems. Moreover, when using implicit ODE solvers, solutions to linear systems with up to cubic complexity are then required for solving nonlinear systems using for instance Newton’s algorithm; as the proposed approach allows to reduce the overall number of time-steps thanks to an iterative increase of the accuracy order of the ODE system solvers, it then reduces the number of linear systems to solve, hence benefiting from a scaling effect.
- Published
- 2020
29. Collusion-resistant protocols for private processing of aggregated queries in distributed databases
- Author
-
Leanne Rylands, Joe Ryan, Andrei V. Kelarev, Jennifer Seberry, Xun Yi, and Yuqing Lin
- Subjects
Protocol (science) ,Information Systems and Management ,Information retrieval ,Distributed database ,Computer science ,Large numbers ,02 engineering and technology ,16. Peace & justice ,Data structure ,Information sensitivity ,Hardware and Architecture ,020204 information systems ,Collusion ,0202 electrical engineering, electronic engineering, information engineering ,Confidentiality ,Communication complexity ,Software ,Information Systems - Abstract
Private processing of database queries protects the confidentiality of sensitive data when queries are answered. It is important to design collusion-resistant protocols ensuring that privacy remains protected even when a certain number of honest-but-curious participants collude to share their knowledge in order to gain unauthorised access to sensitive information. A novel setting arises when aggregated queries need to be answered for a large distributed database, but legal requirements or commercial interests forbid making access to records in each subdatabase available to other counterparts. For example, a very large number of medical records may be stored in a distributed database, which is a union of several separate databases from different hospitals, or even from different countries. The present article introduces and investigates two protocols for collusion-resistant private processing of aggregated queries in this novel setting: Accelerated Multi-round Iterative Protocol (AMIP) and Restricted Multi-round Iterative Protocol (RMIP). We define a large collection of query functions and show that AMIP and RMIP protocols can answer all queries in this collection. Our experiments demonstrate that the AMIP protocol outperforms all other applicable algorithms, and this achievement is especially significant in terms of the communication complexity.
- Published
- 2020
30. High-dimensional penalized arch processes
- Author
-
Jean-David Fermanian and Benjamin Poignard
- Subjects
Statistics::Theory ,Economics and Econometrics ,Positive definiteness ,Group (mathematics) ,Applied mathematics ,Large numbers ,Arch models ,High dimensional ,Arch ,Mathematics - Abstract
We introduce a general methodology to consistently estimate multidimensional ARCH models equation-by-equation, possibly with a very large number of parameters through penalization (Sparse Group Las...
- Published
- 2020
31. A Cautionary Tale for Machine Learning Design: why we Still Need Human-Assisted Big Data Analysis
- Author
-
Luca Casini, Giovanni Delnevo, Marco Roccetti, Paola Salomoni, and Roccetti, M., Delnevo, G., Casini, L., Salomoni, P.
- Subjects
Machine learning design ,Computer Networks and Communications ,Computer science ,Big data ,Extrapolation ,02 engineering and technology ,Machine learning ,computer.software_genre ,Water consumption ,Data semantics ,0202 electrical engineering, electronic engineering, information engineering ,sort ,Human-in-the-loop method ,Human-machine-Bigdata interaction loop ,Artificial neural network ,business.industry ,Large numbers ,020206 networking & telecommunications ,Hardware and Architecture ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Classifier (UML) ,Software ,Water metering and consumption ,Information Systems - Abstract
Supervised Machine Learning (ML) requires that smart algorithms scrutinize a very large number of labeled samples before they can make right predictions. And this is not always true either. In our experience, in fact, a neural network trained with a huge database comprised of over fifteen million water meter readings had essentially failed to predict when a meter would malfunction/need disassembly based on a history of water consumption measurements. With a second step, we developed a methodology, based on the enforcement of a specialized data semantics, that allowed us to extract only those samples for training that were not noised by data impurities. With this methodology, we re-trained the neural network up to a prediction accuracy of over 80%. Yet, we simultaneously realized that the new training dataset was significantly different from the initial one in statistical terms, and much smaller, as well. We had reached a sort of paradox: We had alleviated the initial problem with a better interpretable model, but we had changed the replicated form of the initial data. To reconcile that paradox, we further enhanced our data semantics with the contribution of field experts. This has finally led to the extrapolation of a training dataset truly representative of regular/defective water meters and able to describe the underlying statistical phenomenon, while still providing an excellent prediction accuracy of the resulting classifier. At the end of this path, the lesson we have learnt is that a human-in-the-loop approach may significantly help to clean and re-organize noised datasets for an empowered ML design experience.
- Published
- 2020
32. ProDCoNN: Protein design using a convolutional neural network
- Author
-
Chun-Chao Lo, Chenran Wang, Xiuwen Liu, Jinfeng Zhang, Yang Chen, Wu Wei, and Yuan Zhang
- Subjects
Computer science ,Protein design ,Biophysics ,Datasets as Topic ,Protein Engineering ,Biochemistry ,Convolutional neural network ,Protein Structure, Secondary ,Article ,03 medical and health sciences ,Protein structure ,Structural Biology ,Computational structural biology ,Amino Acid Sequence ,Databases, Protein ,Molecular Biology ,030304 developmental biology ,0303 health sciences ,business.industry ,030302 biochemistry & molecular biology ,Proteins ,Large numbers ,Pattern recognition ,Protein engineering ,Bond length ,Benchmarking ,Molecular geometry ,Artificial intelligence ,Neural Networks, Computer ,business ,Sequence Alignment ,Algorithm ,Software - Abstract
Designing protein sequences that fold to a given three-dimensional (3D) structure has long been a challenging problem in computational structural biology with significant theoretical and practical implications. In this study, we first formulated this problem as predicting the residue type given the 3D structural environment around the C(α) atom of a residue, which is repeated for each residue of a protein. We designed a nine-layer 3D deep convolutional neural network (CNN) that takes as input a gridded box with the atomic coordinates and types around a residue. Several CNN layers were designed to capture structure information at different scales, such as bond lengths, bond angles, torsion angles, and secondary structures. Trained on a very large number of protein structures, the method, called ProDCoNN (protein design with CNN), achieved state-of-the-art performance when tested on large numbers of test proteins and benchmark datasets.
- Published
- 2020
33. Front tracking transition system model with controlled moving bottlenecks and probabilistic traffic breakdowns
- Author
-
Mladen Cicic, Karl Henrik Johansson, and Igor Mikolášek
- Subjects
0209 industrial biotechnology ,Computer science ,020208 electrical & electronic engineering ,Probabilistic logic ,Large numbers ,02 engineering and technology ,Function (mathematics) ,Traffic flow ,Tracking (particle physics) ,Bottleneck ,020901 industrial engineering & automation ,Control and Systems Engineering ,Control theory ,Transition system ,0202 electrical engineering, electronic engineering, information engineering ,State (computer science) - Abstract
Cell-based approximations of PDE traffic models are widely used for traffic prediction and control. However, in order to represent the traffic state with good resolution, cell-based models often require a short cell length, which results in a very large number of states. We propose a new transition system traffic model, based on the front tracking method for solving the LWR PDE model. Assuming piecewise-linear flux function and piecewise-constant initial conditions, this model gives an exact solution. Furthermore, it is easier to extend, has fewer states and, although its dynamics are intrinsically hybrid, is faster to simulate than an equivalent cell-based approximation. The model is extended to enable handling moving bottlenecks as well as probabilistic traffic breakdowns and capacity drops at static bottlenecks. A control strategy that utilizes controlled moving bottlenecks for bottleneck decongestion is described and tested in simulation. It is shown that we are able to keep the static bottleneck in free flow by creating controlled moving bottlenecks at specific instances along on the road, and using them to regulate the incoming traffic flow.
- Published
- 2020
34. A million is more than a thousand: Children's acquisition of very large number words
- Author
-
Pierina Cheung and Daniel Ansari
- Subjects
Cognitive Neuroscience ,number words ,Developmental and Educational Psychology ,CHILDES ,large numbers ,math curriculum ,Pediatrics - Abstract
Very large numbers words such as “hundred,” “thousand,” “million,” “billion,” and “trillion” pose a learning problem for children because they are sparse in everyday speech and children's experience with extremely large quantities is scarce. In this study, we examine when children acquire the relative ordering of very large number words as a first step toward understanding their acquisition. In Study 1, a hundred and twenty-five 5–8-year-olds participated in a verbal number comparison task involving very large number words. We found that children can judge which of two very large numbers is more as early as age 6, prior to entering first grade. In Study 2, we provided a descriptive analysis on the usage of very large number words using the CHILDES database. We found that the relative frequency of large number words does not change across the years, with “hundred” uttered more frequently than others by an order of magnitude. We also found that adults were more likely to use large number words to reference units of quantification for money, weight, and time, than for discrete, physical entities. Together, these results show that children construct a numerical scale for large number words prior to learning their precise cardinal meanings, and highlight how frequency and context may support their acquisition. Our results have pedagogical implications and highlight a need to investigate how children acquire meanings for number words that reference quantities beyond our everyday experience.
- Published
- 2022
35. On some graph-based two-sample tests for high dimension, low sample size data
- Author
-
Anil K. Ghosh, Soham Sarkar, and Rahul Biswas
- Subjects
nearest neighbor ,distributions ,high-dimensional consistency ,02 engineering and technology ,minimum spanning tree ,shortest hamiltonian path ,multivariate ,Artificial Intelligence ,020204 information systems ,Euclidean geometry ,0202 electrical engineering, electronic engineering, information engineering ,Two sample ,equality ,large numbers ,laws ,Statistical hypothesis testing ,Mathematics ,non-bipartite matching ,Nominal level ,Euclidean distance ,Sample size determination ,Graph (abstract data type) ,020201 artificial intelligence & image processing ,Pairwise comparison ,Algorithm ,distance concentration ,Software ,permutation test - Abstract
Testing for equality of two high-dimensional distributions is a challenging problem, and this becomes even more challenging when the sample size is small. Over the last few decades, several graph-based two-sample tests have been proposed in the literature, which can be used for data of arbitrary dimensions. Most of these test statistics are computed using pairwise Euclidean distances among the observations. But, due to concentration of pairwise Euclidean distances, these tests have poor performance in many high-dimensional problems. Some of them can have powers even below the nominal level when the scale-difference between two distributions dominates the location-difference. To overcome these limitations, we introduce some new dissimilarity indices and use them to modify some popular graph-based tests. These modified tests use the distance concentration phenomenon to their advantage, and as a result, they outperform the corresponding tests based on the Euclidean distance in a wide variety of examples. We establish the high-dimensional consistency of these modified tests under fairly general conditions. Analyzing several simulated as well as real data sets, we demonstrate their usefulness in high dimension, low sample size situations.
- Published
- 2019
36. Efficient, robust and effective rank aggregation for massive biological datasets
- Author
-
Alain Denise, Laurent Bulteau, Bryan Brancotte, Pierre Andrieu, Adeline Pierrot, Stéphane Vialette, Sarah Cohen-Boulakia, Laboratoire Interdisciplinaire des Sciences du Numérique (LISN), Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS), Hub Bioinformatique et Biostatistique - Bioinformatics and Biostatistics HUB, Institut Pasteur [Paris] (IP)-Université Paris Cité (UPCité), Laboratoire d'Informatique Gaspard-Monge (LIGM), École des Ponts ParisTech (ENPC)-Centre National de la Recherche Scientifique (CNRS)-Université Gustave Eiffel, Institut de Biologie Intégrative de la Cellule (I2BC), Commissariat à l'énergie atomique et aux énergies alternatives (CEA)-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS), This work has been partly supported by CNRS Défi Mastodons Grant., CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS), and Institut Pasteur [Paris]-Centre National de la Recherche Scientifique (CNRS)
- Subjects
[INFO.INFO-CC]Computer Science [cs]/Computational Complexity [cs.CC] ,Computer Networks and Communications ,Computer science ,[INFO.INFO-DS]Computer Science [cs]/Data Structures and Algorithms [cs.DS] ,Context (language use) ,02 engineering and technology ,computer.software_genre ,Set (abstract data type) ,Consensus ranking ,0202 electrical engineering, electronic engineering, information engineering ,Rank (computer programming) ,Large numbers ,Approximation algorithm ,020206 networking & telecommunications ,Exact algorithm ,Ranking ,Hardware and Architecture ,Rank aggregation ,Massive biological datasets ,020201 artificial intelligence & image processing ,Data mining ,[INFO.INFO-BI]Computer Science [cs]/Bioinformatics [q-bio.QM] ,Heuristics ,Kemeny rule ,computer ,Software - Abstract
International audience; Massive biological datasets are available in various sources. To answer a biological question (e.g., ''which are the genes involved in a given disease?''), life scientists query and mine such datasets using various techniques. Each technique provides a list of results usually ranked by importance (e.g., a list of ranked genes). Combining the results obtained by various techniques, that is, combining ranked lists of elements into one list of elements is of paramount importance to help life scientists make the most of various results and prioritize further investigations. Rank aggregation techniques are particularly well-fitted with this context as they take in a set of rankings and provide a consensus, that is, a single ranking which is the ''closest'' to the input rankings. However, (i) the problem of rank aggregation is NP-hard in most cases (using an exact algorithm is currently not possible for more than a few dozens of elements) and (ii) several (possibly very different) exact solutions can be obtained. As answer to (i), many heuristics and approximation algorithms have been proposed. However, heuristics cannot guarantee how far from an exact solution the consensus ranking will be, and the approximation ratio of approximation algorithms dedicated to the problem is fairly high (not less than 3/2). No solution has yet been proposed to help true-users dealing with the problem encountered in point (ii). In this paper we present a complete system able to perform rank aggregation of massive biological datasets. Our solution is efficient as it is based on an original partitioning method making it possible to compute a high-quality consensus using an exact algorithm in a large number of cases. Our solution is robust as it clearly identifies for the user groups of elements whose relative order is the same in any optimal solution. These features provide answers to points (i) and (ii) and lie in mathematical bases offering guarantees on the computed result. Also, our solution is effective as it has been implemented into a real tool, ConquR-BioV2 is used for the life science community, and evaluated at large-scale using a very large number of datasets.
- Published
- 2021
37. Certain Finite Integrals Related to the Products of Special Functions
- Author
-
Dinesh Kumar, Frédéric Ayant, Suphawat Asawasamrit, and Jessada Tariboon
- Subjects
Class (set theory) ,Physics and Astronomy (miscellaneous) ,General Mathematics ,Multivariable calculus ,Aleph-function ,general class of multivariable polynomials ,Large numbers ,hypergeometric function ,multivariable Aleph-function ,Algebra ,Chemistry (miscellaneous) ,Special functions ,Product (mathematics) ,Computer Science (miscellaneous) ,QA1-939 ,Hypergeometric function ,Mathematics - Abstract
The aim of this paper is to establish a theorem associated with the product of the Aleph-function, the multivariable Aleph-function, and the general class of polynomials. The results of this theorem are unified in nature and provide a very large number of analogous results (new or known) involving simpler special functions and polynomials (of one or several variables) as special cases. The derived results lead to significant applications in physics and engineering sciences.
- Published
- 2021
38. Continuous Random Variables
- Author
-
Muammer Catak, Witold Pedrycz, and Tofigh Allahviranloo
- Subjects
Discrete mathematics ,Set (abstract data type) ,Transformation (function) ,Probability theory ,Large numbers ,Probability density function ,Function (mathematics) ,Set theory ,Random variable ,Mathematics - Abstract
The basic theory of the probability has been discussed in Chap. 1. That knowledge based on the set theory is essential to understand the notion of probability. However, it lacks applicability in many cases. For instance, let consider that very large number of events is under investigation. Calculations of probabilities of any arbitrary outcomes would require a great deal of effort. For this reason, the general concept of the probability theory is developed by means of random variables and corresponding probability functions, namely the probability distribution function and the probability density function. In general, a function maps elements of a well-defined set to another well-defined set. Let \(x \in A \) and \(y \in B\), a function f(x) defines the transformation from set A to set B, and it is denoted as; $$\begin{aligned} f : x \rightarrow y \end{aligned}$$ A random variable, X, is a function related with the outcomes of a certain random experiment.
- Published
- 2021
39. SiGMoiD: A super-statistical generative model for binary data
- Author
-
Purushottam D. Dixit, Germán Plata, and Xiaochuan Zhao
- Subjects
Computer science ,Entropy ,Binary number ,Animal Cells ,Biology (General) ,Neurons ,Ecology ,Principle of maximum entropy ,Physics ,Genomics ,Generative model ,Data point ,Computational Theory and Mathematics ,Community Ecology ,Medical Microbiology ,Modeling and Simulation ,Binary data ,Physical Sciences ,Probability distribution ,Thermodynamics ,Cellular Types ,Algorithm ,Network Analysis ,Algorithms ,Research Article ,Computer and Information Sciences ,QH301-705.5 ,Sample (statistics) ,Microbial Genomics ,Microbiology ,Cellular and Molecular Neuroscience ,Metabolic Networks ,Genetics ,Molecular Biology ,Ecology, Evolution, Behavior and Systematics ,Models, Statistical ,Bacteria ,Gut Bacteria ,Ecology and Environmental Sciences ,Probabilistic logic ,Organisms ,Large numbers ,Biology and Life Sciences ,Computational Biology ,Sigmoid function ,Cell Biology ,Probability Theory ,Probability Distribution ,Cellular Neuroscience ,Microbiome ,Mathematics ,Neuroscience - Abstract
In modern computational biology, there is great interest in building probabilistic models to describe collections of a large number of co-varying binary variables. However, current approaches to build generative models rely on modelers’ identification of constraints and are computationally expensive to infer when the number of variables is large (N~100). Here, we address both these issues with Super-statistical Generative Model for binary Data (SiGMoiD). SiGMoiD is a maximum entropy-based framework where we imagine the data as arising from super-statistical system; individual binary variables in a given sample are coupled to the same ‘bath’ whose intensive variables vary from sample to sample. Importantly, unlike standard maximum entropy approaches where modeler specifies the constraints, the SiGMoiD algorithm infers them directly from the data. Due to this optimal choice of constraints, SiGMoiD allows us to model collections of a very large number (N>1000) of binary variables. Finally, SiGMoiD offers a reduced dimensional description of the data, allowing us to identify clusters of similar data points as well as binary variables. We illustrate the versatility of SiGMoiD using multiple datasets spanning several time- and length-scales., Author summary Collectively varying binary variables are ubiquitous in modern biology. Given that the number of possible configurations of these systems typically far exceeds the number of available samples, generative models have become an essential tool in quantitative descriptions of binary data. The state-of-the-art approaches to build generative models have several conceptual limitations. Specifically, they rely on the modeler choosing system-appropriate constraints, which can be challenging in systems with many complex interactions. Moreover, they are computationally expensive to infer when the number of variables is large (N~100). To address this issue, we propose a theoretical generalization of the maximum entropy approach that allows us to model very high dimensional data; at least an order of magnitude higher than what is currently possible. This framework will be a significant advancement in the computational analysis of covarying binary variables.
- Published
- 2021
40. Discrimination of large quantities: Weber's law and short-term memory in angelfish, Pterophyllum scalare.
- Author
-
Gómez-Laplaza, Luis M. and Gerlai, Robert
- Subjects
- *
SCALARE , *FRESHWATER angelfishes , *BANKS (Oceanography) , *CICHLIDS , *ANIMAL ecology - Abstract
The ability to discriminate between different quantities has important ecological relevance for animals when engaging in behaviours such as forming groups, foraging or trying to avoid predators. Quantity discrimination has been shown in a diversity of human and nonhuman animal species. In angelfish this discrimination ability has been investigated using dichotomous choice tests when the numerically different stimulus groups (shoals) of conspecifics were fully visible to the test fish. Here, using a new procedure we investigated whether test fish were able to discriminate between the contrasting shoals using their memory. After a period of full visual access to the contrasted shoals on the two sides of their test tank, the test fish was required to make a choice while being able to see only a single member of the stimulus shoals on each side. With this cognitively more demanding procedure we tested discrimination between numerically large shoals (≥ four fish per stimulus shoal). As in our previous studies, we found that angelfish consistently chose the larger of the two shoals when the shoals differed by a 2:1 or higher ratio, but not those that differed by a 3:2 or 4:3 ratio. The results followed Weber's law in that performance became poorer as the ratio between the two stimulus shoals approached one. In addition, when we kept the absolute difference between the contrasted shoals constant, discrimination was less accurate as the shoal sizes increased. This pattern of results lends support for the analogue magnitude representational system in the angelfish, a nonverbal approximation system believed to be employed by a diversity of human and nonhuman animal species. Furthermore, our results also demonstrate that angelfish remember the different shoals presented to them, i.e. they make their choice based upon mental representation of the different quantities. [ABSTRACT FROM AUTHOR]
- Published
- 2016
- Full Text
- View/download PDF
41. Efficient Modeling of the Routing and Spectrum Allocation Problem for Flexgrid Optical Networks
- Author
-
Brigitte Jaumard and Quang Anh Nguyen
- Subjects
Mathematical optimization ,Exploit ,Computer science ,Testbed ,Scalability ,Large numbers ,Routing (electronic design automation) ,Heuristics ,RSA problem ,Frequency allocation - Abstract
While the problem of Routing and Spectrum Allocation (RSA) has been widely studied, very few studies attempt to solve realistic sized instances. Indeed, the state of the art is always below the standard transport capacity of a fiber link with 384 frequency slots, regardless of what the authors consider, heuristics or exact methods with a few exceptions. In this paper, we are interested in reducing the gap between realistic data sets and testbed instances that are often considered, using exact methods. Even if exact methods may fail to solve in reasonable time very large instances, they can, however, output solutions with a very good and proven accuracy. The novelty of this paper is to exploit the observations that optimal solutions contain a very large number of lightpaths associated with shortest paths or k-shortest paths with a small k. We propose an original efficient large-scale optimization model and decomposition algorithm to solve the RSA problem for flexgrid optical networks. It allows the exact or near optimal solution of much larger instances than in the literature.
- Published
- 2021
42. Extended granular micromechanics approach: a micromorphic theory of degreen
- Author
-
Nima Nejadsadeghi and Anil Misra
- Subjects
Collective behavior ,General Mathematics ,Science and engineering ,Micromechanics ,Large numbers ,02 engineering and technology ,021001 nanoscience & nanotechnology ,Granular material ,Degree (music) ,020303 mechanical engineering & transports ,0203 mechanical engineering ,Mechanics of Materials ,General Materials Science ,Statistical physics ,0210 nano-technology ,Mathematics - Abstract
For many problems in science and engineering, it is necessary to describe the collective behavior of a very large number of grains. Complexity inherent in granular materials, whether due the variability of grain interactions or grain-scale morphological factors, requires modeling approaches that are both representative and tractable. In these cases, continuum modeling remains the most feasible approach; however, for such models to be representative, they must properly account for the granular nature of the material. The granular micromechanics approach has been shown to offer a way forward for linking the grain-scale behavior to the collective behavior of millions and billions of grains while keeping within the continuum framework. In this paper, an extended granular micromechanics approach is developed that leads to a micromorphic theory of degree n. This extended form aims at capturing the detailed grain-scale kinematics in disordered (mechanically or morphologically) granular media. To this end, additional continuum kinematic measures are introduced and related to the grain-pair relative motions. The need for enriched descriptions is justified through experimental measurements as well as results from simulations using discrete models. Stresses conjugate to the kinematic measures are then defined and related, through equivalence of deformation energy density, to forces conjugate to the measures of grain-pair relative motions. The kinetic energy density description for a continuum material point is also correspondingly enriched, and a variational approach is used to derive the governing equations of motion. By specifying a particular choice for degree n, abridged models of degrees 2 and 1 are derived, which are shown to further simplify to micro-polar or Cosserat-type and second-gradient models of granular materials.
- Published
- 2019
43. One-Bit Sigma-Delta MIMO Precoding
- Author
-
A. Lee Swindlehurst, Qiang Li, Wing-Kin Ma, and Mingjie Shao
- Subjects
Signal Processing (eess.SP) ,FOS: Computer and information sciences ,Optimization problem ,Computer science ,Information Theory (cs.IT) ,Computer Science - Information Theory ,MIMO ,Sigma ,Large numbers ,020206 networking & telecommunications ,Data_CODINGANDINFORMATIONTHEORY ,02 engineering and technology ,Topology ,Delta-sigma modulation ,Precoding ,Hardware_GENERAL ,Modulation ,Signal Processing ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Electrical Engineering and Systems Science - Signal Processing ,Electrical and Electronic Engineering ,Computer Science::Information Theory ,Communication channel - Abstract
Coarsely quantized MIMO signalling methods have gained popularity in the recent developments of massive MIMO as they open up opportunities for massive MIMO implementation using cheap and power-efficient radio-frequency front-ends. This paper presents a new one-bit MIMO precoding approach using spatial Sigma-Delta ( $\Sigma \Delta$ ) modulation. In previous one-bit MIMO precoding research, one mainly focuses on using optimization to tackle the difficult binary signal optimization problem that arises from the precoding design. Our approach attempts a different route. Assuming angular MIMO channels, we apply $\Sigma \Delta$ modulation—a classical concept in analog-to-digital conversion of temporal signals—in space. The resulting $\Sigma \Delta$ precoding approach has two main advantages: First, we no longer need to deal with binary optimization in $\Sigma \Delta$ precoding design. Particularly, the binary signal restriction is replaced by peak signal amplitude constraints. Second, the impact of the quantization error can be well controlled via modulator design and under appropriate operating conditions. Through symbol error probability analysis, we reveal that the very large number of antennas in massive MIMO provides favorable operating conditions for $\Sigma \Delta$ precoding. In addition, we develop a new $\Sigma \Delta$ modulation architecture that is capable of adapting the channel to achieve nearly zero quantization error for a targeted user. Furthermore, we consider multi-user $\Sigma \Delta$ precoding using the zero-forcing and symbol-level precoding schemes. These two $\Sigma \Delta$ precoding schemes perform considerably better than their direct one-bit quantized counterparts, as simulation results show.
- Published
- 2019
44. Molecular discovery by optimal sequential search
- Author
-
Genyuan Li
- Subjects
010304 chemical physics ,Property (programming) ,business.industry ,Applied Mathematics ,010102 general mathematics ,Process (computing) ,Large numbers ,General Chemistry ,Composition (combinatorics) ,01 natural sciences ,Automation ,Task (computing) ,Large set (Ramsey theory) ,0103 physical sciences ,0101 mathematics ,business ,Algorithm ,Linear search - Abstract
In the development of a new compound in chemistry and molecular biology, especially a new medicine in pharmaceutical industry, we often need to find candidate(s), a molecule or molecules, with the best desired property (e.g., binding affinity in medicine) from a large set of molecules with the same scaffold but m distinct functional substitutes at each of its n different sites. The total number $$N_{\mathrm{lib}}$$ of molecules in this library is $$m^n$$ . In some cases, $$N_{\mathrm{lib}}$$ can be a very large number (e.g., millions). This is a challenging task because it is costly and often infeasible to synthesize and test all of these molecules. A new algorithm referred to as optimal sequential search is developed to overcome this difficulty. Especially, this algorithm is chemically intuitive which only uses the information of molecule composition, and accessible to practical chemists. The algorithm can be applied to small, medium and large size molecule libraries. With syntheses and property measurements for a limited number of molecules, the top best candidate molecules can be effectively captured from the whole library. Three examples with library size 64, 160,000 and 1,048,576, respectively, are used for illustration. For the first small library, syntheses and property measurements of 17 molecules are sufficient to capture the top 7 best candidate molecules; for the two medium and large libraries, syntheses and property measurements of about one thousand molecules can capture most or a large part of the top 500, especially the top 100 best candidate molecules. However, the algorithm needs to perform multiple (e.g., hundreds of) iterative syntheses and property measurements. The time cost may not be acceptable if the algorithm is performed manually. To make the algorithm practical, automation of the sequential searching process is the following task.
- Published
- 2019
45. Computing the closest real normal matrix and normal completion
- Author
-
Nicola Guglielmi and Carmela Scalone
- Subjects
Matrix nearness problems ,Closest normal matrix ,Closest real normal matrix ,Matrix ODEs ,Normal completion ,Normal completion of minimal norm ,Applied Mathematics ,media_common.quotation_subject ,Matrix norm ,Large numbers ,010103 numerical & computational mathematics ,01 natural sciences ,Normal matrix ,Square (algebra) ,010101 applied mathematics ,Combinatorics ,Computational Mathematics ,Matrix (mathematics) ,0101 mathematics ,Normality ,Eigenvalues and eigenvectors ,Dedekind–MacNeille completion ,Mathematics ,media_common - Abstract
In this article, we consider the problems (unsolved in the literature) of computing the nearest normal matrix X to a given non-normal matrix A, under certain constraints, that are (i) if A is real, we impose that also X is real; (ii) if A has known entries on a given sparsity pattern Ω and unknown/uncertain entries otherwise, we impose to X the constraint xij = aij for all entries (i,j) in the pattern Ω. As far as we know, there do not exist in the literature specific algorithms aiming to solve these problems. For the case in which all entries of A can be modified, there exists an algorithm by Ruhe, which is able to compute the closest normal matrix. However, if A is real, the closest computed matrix by Ruhe’s algorithm might be complex, which motivates the development of a different algorithm preserving reality. Normality is characterized in a very large number of ways; in this article, we consider the property that the square of the Frobenius norm of a normal matrix is equal to the sum of the squares of the moduli of its eigenvalues. This characterization allows us to formulate as equivalent problem the minimization of a functional of an unknown matrix, which should be normal, fulfill the required constraints, and have minimal distance from the given matrix A.
- Published
- 2019
46. Development of a Hybrid Real Estate Recommender System
- Author
-
Hilal Erdogan Sumnu, Bahadir Gokoz, Tevfik Aytekin, and Hakan Tas
- Subjects
Product (business) ,Renting ,Work (electrical) ,Computer science ,business.industry ,Final product ,Large numbers ,Real estate ,Recommender system ,business ,Data science ,Content filtering - Abstract
Today consumers are confronted with a very large number of products and services to choose from. This makes it difficult for users to find relevant products among a huge number of alternatives. Recommender systems help users to find products of interest by analyzing past user transactions such as product views and purchases. There is a similar problem in the real estate industry where hundreds of thousands of properties are available for rentals or sales. In this work we present the details of a real estate recommender system developed for Zingat.com. The system developed is a hybrid of collaborative and content filtering approaches. We will explain the challenges we face, the recommendation techniques we use to overcome these challenges, and the final product used for recommendation.
- Published
- 2019
47. Protein-Templated Dynamic Combinatorial Chemistry
- Subjects
IDENTIFICATION ,Target-directed dynamic combinatorial chemistry ,Biochemical activity ,HYDRAZONE FORMATION ,Medicinal chemistry ,NUCLEOPHILIC CATALYSIS ,LIBRARIES ,STD NMR ,Hit identification ,Protein stability ,LARGE NUMBERS ,LIGAND ,INHIBITORS ,OPTIMIZATION ,GENERATION - Abstract
Dynamic combinatorial chemistry (DCC) is a powerful tool to identify bioactive compounds. This efficient technique allows the target to select its own binders and circumvents the need for synthesis and biochemical evaluation of all individual derivatives. An ever-increasing number of publications report the use of DCC on biologically relevant target proteins. This minireview complements previous reviews by focusing on the experimental protocol and giving detailed examples of essential steps and factors that need to be considered, such as protein stability, buffer composition and cosolvents.
- Published
- 2019
48. GUIMesh: A tool to import STEP geometries into Geant4 via GDML
- Author
-
Patrícia Gonçalves and Marco Pinto
- Subjects
Engineering drawing ,Markup language ,business.industry ,Computer science ,Margin of error ,General Physics and Astronomy ,Large numbers ,CAD ,Python (programming language) ,computer.software_genre ,01 natural sciences ,010305 fluids & plasmas ,Software ,Hardware and Architecture ,0103 physical sciences ,Computer Aided Design ,010306 general physics ,business ,Physics - Computational Physics ,computer ,Graphical user interface ,computer.programming_language - Abstract
Detailed radiation analysis of instruments flown in space is critical to ensure mission safety, often requiring the use of state-of-the-art particle transport simulation tools. Geant4 is one of the most powerful toolkits to simulate the interaction and the passage of particles through matter, but it is not prepared to receive Standard for The Exchange of Product data (STEP) files, the most versatile Computer-Aided Design (CAD) format, as input, requiring previous conversion to other CAD formats. This may lead to loss of detail and under or overestimation of the quantities under study, especially when the instruments have complex shapes, and/or a large number of volumes. Though several solutions have been proposed to import complex geometries from STEP files into Geant4, so far, only commercial options are available. In this paper we present a new tool, GUIMesh, that embeds FreeCAD libraries, an open-source CAD editor, to tessellate volumes, and convert them to Geometry Description Markup Language (GDML), a Geant4 readable format , in a straightforward way. Several degrees of freedom are given to the user regarding mesh precision and choice of material. Different geometries were tested for material definition, geometry and navigation errors, and the method used was successfully validated. Program Summary Program Title: GUIMesh Program Files doi: http://dx.doi.org/10.17632/c3c9xbspzp.1 Licensing provisions : GNU General Public License 3 (GPLv3) Programming language: Python Nature of problem: No open-source software allows to import STEP geometries intro Geant4, one of the most powerful toolkits to simulate radiation interaction with matter. Since CAD software is extensively used in the design of particle and radiation detection instruments, it is highly desirable for physicists that STEP geometries can be imported to Geant4 with little effort instead of having to code structures made of a very large number of solid volumes which are difficult to accurately reproduce with Geant4 C++ classes. Solution method : STEP geometries are converted to tessellated volumes (with some margin of error) using FreeCAD libraries. A Python script then writes GDML files based on the results allowing users to import these geometries with Geant4. A graphical user interface provides several options to the user, including material assignment and mesh precision setting for each volume.
- Published
- 2019
49. Modeling of Aggregated IoT Traffic and Its Application to an IoT Cloud
- Author
-
Poul E. Heegaard, Samuel Kounev, Florian Metzger, Tobias Hobfeld, and Andre Bauer
- Subjects
business.industry ,Stochastic process ,Computer science ,Distributed computing ,media_common.quotation_subject ,Large numbers ,Fidelity ,Cloud computing ,Context (language use) ,Poisson distribution ,Telecommunications network ,symbols.namesake ,Scalability ,Computer Science::Networking and Internet Architecture ,symbols ,Electrical and Electronic Engineering ,business ,media_common - Abstract
As the Internet of Things (IoT) continues to gain traction in telecommunication networks, a very large number of devices are expected to be connected and used in the near future. In order to appropriately plan and dimension the network, as well as the back-end cloud systems and the resulting signaling load, traffic models are employed. These models are designed to accurately capture and predict the properties of IoT traffic in a concise manner. To achieve this, Poisson process approximations, based on the Palm-Khintchine theorem, have often been used in the past. Due to the scale (and the difference in scales in various IoT networks) of the modeled systems, the fidelity of this approximation is crucial, as, in practice, it is very challenging to accurately measure or simulate large-scale IoT deployments. The main goal of this paper is to understand the level of accuracy of the Poisson approximation model. To this end, we first survey both common IoT network properties and network scales as well as traffic types. Second, we explain and discuss the Palm-Khintiche theorem, how it is applied to the problem, and which inaccuracies can occur when using it. Based on this, we derive guidelines as to when a Poisson process can be assumed for aggregated periodic IoT traffic. Finally, we evaluate our approach in the context of an IoT cloud scaler use case. © 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
- Published
- 2019
50. Enhancing the order picking process through a new storage assignment strategy in forward-reserve area
- Author
-
Behnam Bahrami, El-Houssaine Aghezzaf, and Veronique Limère
- Subjects
0209 industrial biotechnology ,Order picking ,021103 operations research ,Computer science ,Strategy and Management ,0211 other engineering and technologies ,Process (computing) ,Large numbers ,02 engineering and technology ,Management Science and Operations Research ,Industrial engineering ,Industrial and Manufacturing Engineering ,Warehouse ,020901 industrial engineering & automation - Abstract
This paper reexamines the order picking process in a warehouse facing the challenges that e-commerce brings about and which are characterised by a very large number of small sized orders and return...
- Published
- 2019
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.