Irad Ben-Gal
Tel Aviv-Yafo, Tel Aviv District, Israel
8K עוקבים
מעל 500 קשרים
על אודות
Known expert in AI, machine learning and predictive analytics with more than 20 years of…
מאמרים מאת Irad
פעילות
-
הסבלנות מרה אך פירותיה מתוקים. בחמישי האחרון קיימנו אירוע סופר מרתק בשיתוף פעולה עם Drive TLV שדוגם מקרי בוחן אמיתיים על איך חברות בתחום המוביליטי…
הסבלנות מרה אך פירותיה מתוקים. בחמישי האחרון קיימנו אירוע סופר מרתק בשיתוף פעולה עם Drive TLV שדוגם מקרי בוחן אמיתיים על איך חברות בתחום המוביליטי…
נוסף לייק על ידי Irad Ben-Gal
-
I was very honored to be invited to the 75th anniversary of the department of statistics of Virginia Tech together with Bart De Ketelaere and Dewi…
I was very honored to be invited to the 75th anniversary of the department of statistics of Virginia Tech together with Bart De Ketelaere and Dewi…
נוסף לייק על ידי Irad Ben-Gal
ניסיון
חינוך
פרסומים
-
A Risk-Scoring Feedback Model for Webpages and Web Users based on Browsing Behavior. Accepted,
ACM Transactions on Intelligent Systems and Technology
It has been claimed that many security breaches are often caused by vulnerable (naïve) employees within the organization [Ponemon Institute LLC 2015a]. Thus, the weakest link in security is often not the technology itself but rather the people who use it [Schneier 2003]. In this paper, we propose a machine learning scheme for detecting risky webpages and risky browsing behavior, performed by naïve users in the organization. The scheme analyzes the interaction between two modules: one represents…
It has been claimed that many security breaches are often caused by vulnerable (naïve) employees within the organization [Ponemon Institute LLC 2015a]. Thus, the weakest link in security is often not the technology itself but rather the people who use it [Schneier 2003]. In this paper, we propose a machine learning scheme for detecting risky webpages and risky browsing behavior, performed by naïve users in the organization. The scheme analyzes the interaction between two modules: one represents naïve users, while the other represents risky webpages. It implements a feedback loop between these modules such that if a webpage is exposed to a lot of traffic from risky users, its "risk score" increases, while in a similar manner, as the user is exposed to risky webpages (with a high "risk score"), his own "risk score" increases. The proposed scheme is tested on a real-world dataset of HTTP logs provided by a large American toolbar company. The results suggest that a feedback learning process involving webpages and users can improve the scoring accuracy and lead to the detection of unknown malicious webpages.
אַחֵר הכותבים -
Shortening the Average Path Length for users in Websites
Quality and Reliability Engineering International
This work proposes a method to improve the QoS provided to internet users by website servers. In particular, the goal is to minimize the expected number of browsing steps (clicks), also known as the expected path length, which are required to reach a website page by a community of users. We use Markov chain modeling to represent the transition probabilities from one webpage to another, and the first passage from page to page that can be retrieved from web server logs. The proposed
method…This work proposes a method to improve the QoS provided to internet users by website servers. In particular, the goal is to minimize the expected number of browsing steps (clicks), also known as the expected path length, which are required to reach a website page by a community of users. We use Markov chain modeling to represent the transition probabilities from one webpage to another, and the first passage from page to page that can be retrieved from web server logs. The proposed
method deletes links among webpages to minimize the expected path length of the website. Three different methods for achieving this goal are examined: (i) a greedy deletion heuristic; (ii) an approximated branch and bound algorithm; and (iii) a cross-entropy metaheuristic. Numerical studies show that the proposed greedy heuristic results in the optimal solution in more than 60% of the tested cases, while in almost 90% of the cases, the obtained solution is within 10% of the optimal solution. Both the approximated branch and bound and the cross-entropy methods achieved optimality in more than 80% of the tested cases; however, this came with a much higher computational cost.אַחֵר הכותבים -
Book: Search and Foraging: Individual Motion and Swarm Dynamics
Taylor and Francis
-
Parallel Construction of Decision Trees with Consistently Non Increasing Expected Number of Tests
Applied Stochastic Models in Business and Industry, Vol. 31(1) 64-78
In recent years, with the emergence of big data and online Internet applications, the ability to classify huge amounts of objects in a short time has become extremely important. Such a challenge can be achieved by constructing decision trees (DTs) with a low expected number of tests (ENT). We address this challenge by proposing the ‘save favorable general optimal testing algorithm’ (SFGOTA)
that guarantees, unlike conventional look-ahead DT algorithms, the construction of DTs with monotonic…In recent years, with the emergence of big data and online Internet applications, the ability to classify huge amounts of objects in a short time has become extremely important. Such a challenge can be achieved by constructing decision trees (DTs) with a low expected number of tests (ENT). We address this challenge by proposing the ‘save favorable general optimal testing algorithm’ (SFGOTA)
that guarantees, unlike conventional look-ahead DT algorithms, the construction of DTs with monotonic non-increasing ENT. The proposed algorithm has a lower complexity in comparison to conventional look-ahead algorithms. It can utilize parallel processing to reduce the execution time when needed. Several numerical studies exemplify how the proposed SF-GOTA generates
efficient DTs faster than standard look-ahead algorithms, while converging to a DT with a minimum ENT.אַחֵר הכותבים -
A Group-Testing Algorithm with Online Informational Learning
IIE Transactions, 46:2, 164-184
An online group testing method to search for a hidden object in a discrete search space is proposed. A relevant example is a search after a nonconforming unit in a batch, while many other applications can be related. A probability mass function is defined over the search space to represent the probability of an object (e.g., a nonconforming unit) to be located at some point or subspace. The suggested
method follows a stochastic local search procedure and can be viewed as a generalization of…An online group testing method to search for a hidden object in a discrete search space is proposed. A relevant example is a search after a nonconforming unit in a batch, while many other applications can be related. A probability mass function is defined over the search space to represent the probability of an object (e.g., a nonconforming unit) to be located at some point or subspace. The suggested
method follows a stochastic local search procedure and can be viewed as a generalization of the Learning Real-Time A∗ (LRTA∗) search algorithm, while using informational distance measures over the searched space. It is proved that the proposed Informational LRTA∗ (ILRTA∗) algorithm converges and always terminates. Moreover, it is shown that under relevant assumptions, the proposed algorithm generalizes known optimal information-theoretic search procedures, such as the offline Huffman search or the generalized
optimum testing algorithm. However, the ILRTA∗ can be applied to new situations, such as a search with side information or an online search where the probability distribution changes. The obtained results can help to bridge the gap between different search procedures that are related to quality control, artificial intelligence, and information theory.אַחֵר הכותבים -
Efficient Construction of Decision Trees by the Dual Information Distance Method
Quality Technology & Quantitative Management (QTQM), 11( 1), 133-147.
The construction of efficient decision and classification trees is a fundamental task in Big Data
analytics which is known to be NP-hard. Accordingly, many greedy heuristics were suggested for the construction of decision-trees, but were found to result in local-optimum solutions. In this work we present the dual information distance (DID) method for efficient construction of decision trees that is computationally attractive, yet relatively robust to noise. The DID heuristic selects features…The construction of efficient decision and classification trees is a fundamental task in Big Data
analytics which is known to be NP-hard. Accordingly, many greedy heuristics were suggested for the construction of decision-trees, but were found to result in local-optimum solutions. In this work we present the dual information distance (DID) method for efficient construction of decision trees that is computationally attractive, yet relatively robust to noise. The DID heuristic selects features by considering both their immediate contribution to the classification, as well as their future potential effects. It represents the construction of classification trees by finding the shortest paths over a graph of partitions
that are defined by the selected features. The DID method takes into account both the orthogonality between the selected partitions, as well as the reduction of uncertainty on the class partition given the selected attributes. We show that the DID method often outperforms popular classifiers, in terms of average depth and classification accuracy.אַחֵר הכותבים -
Fuzzy Implementation of Qubits Operators
Journal of Computer Science & Systems Biology 2014
In the paper, a complete minimal system of fuzzy logic operators that implements the operations over qubits is suggested. The model is based on a parametric system of the fuzzy not and phase operators and represents operators of quantum information theory acting on real amplitudes. An example of the system application is provided by its implementation to a mobile robot control, and its correspondence with the previously suggested models is illustrated by numerical simulations.
אַחֵר הכותבים -
Condition-Based Maintenance via Simulation and a Targeted Bayesian Network Metamodel
Taylor & Francis
Condition-based maintenance (CBM) is increasingly applied to operational systems to reduce lifecycle costs. Predicting the performance of various CBM policies is a challenging task addressed in this work. We suggest a CBM framework that is based on system simulations and a targeted Bayesian network model. Simulations explore the robustness of various CBM policies under different scenarios. The Bayesian network, which is learned from the simulation data, is then used as an explanatory compact…
Condition-based maintenance (CBM) is increasingly applied to operational systems to reduce lifecycle costs. Predicting the performance of various CBM policies is a challenging task addressed in this work. We suggest a CBM framework that is based on system simulations and a targeted Bayesian network model. Simulations explore the robustness of various CBM policies under different scenarios. The Bayesian network, which is learned from the simulation data, is then used as an explanatory compact metamodel for failure prediction. The framework is demonstrated through a study of an operator of a freight rail fleet. This study demonstrates a significant profit improvement compared to other methods.
אַחֵר הכותביםראה פרסום -
Capacity Reallocation in Open Jackson Networks
IIE Transactions, 45, 259–272.
This article investigates methods for reallocation of service capacities in open Jackson networks in order to minimize either a system’smean total work-in-process or its response time. The focus is mainly on a method called node generation, by which capacity can be transferred from a node in echelon j to a newly generated node in echelon j + 1. The proposed procedure is compared with
the more conventional capacity redistribution method, by which capacity can be transferred from any node in…This article investigates methods for reallocation of service capacities in open Jackson networks in order to minimize either a system’smean total work-in-process or its response time. The focus is mainly on a method called node generation, by which capacity can be transferred from a node in echelon j to a newly generated node in echelon j + 1. The proposed procedure is compared with
the more conventional capacity redistribution method, by which capacity can be transferred from any node in echelon j to existing
successor nodes in echelon j + 1. Formulation of each method as a mathematical programming problem reveals the structure of the optimal solution for both problems. The motivation for considering these approaches stems from real-life settings, in particular, from a production line or supply chains where the two types of capacity reallocation are applied. Heuristic methods are developed to solve relatively large networks in tractable time. Numerical results and analyses are presented.אַחֵר הכותבים -
Fractal Geometry Based Statistical Process Control for Non-Linear Auto-correlated Processes
IIE Transactions, 45,373-391
This article suggests a new Statistical Process Control (SPC) approach for data-rich environments. The proposed approach is based on the theory of fractal geometry. In particular, a monitoring scheme is developed that is based on fractal representation of the monitored data at each stage to account for online changes in monitored processes. The proposed fractal-SPC enables a dynamic inspection of non-linear and state-dependent processes with a discrete and finite state space. It is aimed for…
This article suggests a new Statistical Process Control (SPC) approach for data-rich environments. The proposed approach is based on the theory of fractal geometry. In particular, a monitoring scheme is developed that is based on fractal representation of the monitored data at each stage to account for online changes in monitored processes. The proposed fractal-SPC enables a dynamic inspection of non-linear and state-dependent processes with a discrete and finite state space. It is aimed for use with both univariate and multivariate data. The SPC is accomplished by applying an iterated function system to represent a process as a fractal and exploiting the fractal dimension as an important monitoring attribute. It is shown that data patterns can be transformed into representing fractals in a manner that preserves their reference (in control) correlations and dependencies. The fractal statistics can then be used for anomaly detection, pattern analysis, and root cause analysis. Numerical examples and comparisons to conventional SPC methods are given.
אַחֵר הכותבים -
Improving Information Retrieval in Peer-to-Peer Networks using Shared Content Clustering
Knowledge and Information Systems (KAIS) journal, Vol. 39 (2), 383-408 Springer Verlag
Peer-to-peer (p2p) networks are used by millions for searching and downloading
content. Recently, clustering algorithms were shown to be useful for helping users find
content in large networks. Yet, many of these algorithms overlook the fact that p2p networks
follow graph models with a power-law node degree distribution. This paper studies
the obtained clusters when applying clustering algorithms on power-law graphs and their
applicability for finding content. Driven by the…Peer-to-peer (p2p) networks are used by millions for searching and downloading
content. Recently, clustering algorithms were shown to be useful for helping users find
content in large networks. Yet, many of these algorithms overlook the fact that p2p networks
follow graph models with a power-law node degree distribution. This paper studies
the obtained clusters when applying clustering algorithms on power-law graphs and their
applicability for finding content. Driven by the observed deficiencies, a simple yet efficient
clustering algorithm is proposed, which targets a relaxed optimization of a minimal distance
distribution of each cluster with a size balancing scheme. A comparative analysis using a
song-similarity graph collected from 1.2 million Gnutella users reveals that commonly used
efficiency measures often overlook search and recommendation applicability issues and provide
the wrong impression that the resulting clusters are well suited for these tasks. We show
that the proposed algorithm performs well on various measures that are well suited for the
domain.אַחֵר הכותבים -
Moving Target Search Algorithm with Informational Distance Measures”,
The Open Applied Informatics Journal. 6, PP.1-10
We consider an Ishida and Korf Moving Target Search (MTS) algorithm with informational distance measures. Similarly to the previously defined Informational Learning Real-Time A* algorithm, the suggested algorithm acts on the set of partitions of the sample space, on which the probability mass function is defined. The information-based Rokhlin metric and its lower bound – Ornstein metric, give the necessary distance measures. We prove that similarly to the Ishida
and Korf MTS algorithm, the…We consider an Ishida and Korf Moving Target Search (MTS) algorithm with informational distance measures. Similarly to the previously defined Informational Learning Real-Time A* algorithm, the suggested algorithm acts on the set of partitions of the sample space, on which the probability mass function is defined. The information-based Rokhlin metric and its lower bound – Ornstein metric, give the necessary distance measures. We prove that similarly to the Ishida
and Korf MTS algorithm, the proposed Informational MTS (IMTS) algorithm always terminates and finds the target. The comparison of the IMTS algorithm with known models shows that it outperforms known Markov decision process model of search with probabilistic and informational decision criteria. These findings help to construct a unified framework of search after both static and moving targets, and to bridge the gap between different search procedures that are related to
both artificial intelligence and information theory.אַחֵר הכותבים -
Predictive Maintenance Framework through Condition-based Maintenance Bayesian Network and Monte Carlo Simulation
Quality Engineering, 25(4), 370-384
Condition-based maintenance (CBM) is increasingly applied to operational systems to reduce lifecycle costs. Predicting the performance of various CBM policies is a challenging task addressed in this work.
We suggest a CBM framework that is based on system simulations and a targeted Bayesian network model. Simulations explore the robustness of various
CBM policies under different scenarios. The Bayesian network, which is learned from the simulation data, is then used as an explanatory…Condition-based maintenance (CBM) is increasingly applied to operational systems to reduce lifecycle costs. Predicting the performance of various CBM policies is a challenging task addressed in this work.
We suggest a CBM framework that is based on system simulations and a targeted Bayesian network model. Simulations explore the robustness of various
CBM policies under different scenarios. The Bayesian network, which is learned from the simulation data, is then used as an explanatory compact metamodel for failure prediction. The framework is demonstrated through a study of an operator of a freight rail fleet. This study demonstrates a significant profit improvement compared to other methodsאַחֵר הכותבים -
Efficient Bayesian Network Learning for System Optimization in Reliability Engineering
Quality Technology & Quantitative Management, Vol. 9, No. 1, pp. 97-114
We present a new Bayesian network modeling that learns the behavior of an unknown system
from real data and can be used for reliability engineering and optimization processes in industrial systems. The suggested approach relies on quantitative criteria for addressing the trade-off between the complexity of a learned model and its prediction accuracy. These criteria are based on measures from Information Theory as they predetermine both the accuracy as well as the complexity of the model. We…We present a new Bayesian network modeling that learns the behavior of an unknown system
from real data and can be used for reliability engineering and optimization processes in industrial systems. The suggested approach relies on quantitative criteria for addressing the trade-off between the complexity of a learned model and its prediction accuracy. These criteria are based on measures from Information Theory as they predetermine both the accuracy as well as the complexity of the model. We illustrate the proposed method by a classical example of system reliability engineering. Using computer experiments, we show how in a targeted Bayesian network learning, a tremendous reduction in the model complexity can be accomplished, while maintaining most of the essential information for optimizing the system.אַחֵר הכותבים -
Predicting Stock Returns Using a Variable Order Markov Tree Model
Studies in Nonlinear Dynamics & Econometric, Vol. 16, No. 5, 1-35
The weak form of the Efficient Market Hypothesis (EMH) states that the current market price
fully reflects the information of past prices and rules out predictions based on price data alone. In
an efficient market, consistent prediction of the next outcome of a financial time series is problematic
because there are no reoccurring patterns that can be used for a reliable prediction.
This research offers an alternative test of the weak form of the EMH. It uses a universal…The weak form of the Efficient Market Hypothesis (EMH) states that the current market price
fully reflects the information of past prices and rules out predictions based on price data alone. In
an efficient market, consistent prediction of the next outcome of a financial time series is problematic
because there are no reoccurring patterns that can be used for a reliable prediction.
This research offers an alternative test of the weak form of the EMH. It uses a universal prediction
algorithm based on the Variable Order Markov tree model to identify re-occurring patterns
in the data, constructs explanatory models, and predicts the next time-series outcome. Based on
these predictions, it rejects the EMH for certain stock markets while accepting it for other markets.
The weak form of the EMH is tested for four international stock exchanges: the German DAX
index; the American Dow-Jones30 index; the Austrian ATX index and the Danish KFX index.
The universal prediction algorithm is used with sliding windows of 50, 75, and 100 consecutive
daily returns for periods of up to 12 trading years. Statistically significant predictions are detected
for 17% to 81% of the ATX, KFX and DJ30 stock series for about 3% to 30% of the trading days.
A summary prediction analysis indicates that for a confidence level of 99% the more volatile German
(DAX) and American (DJ30) markets are indeed efficient. The algorithm detects periods of
potential market inefficiency in the ATX and KFX markets that may be exploited for obtaining
excess returns.
אַחֵר הכותבים -
A new standardization model for physician staffing at hospitals"
The International Journal of Production and Performance Management, Volume 59, Issue 8, 796-812
Purpose
The purpose of this study is to develop a model for physician staffing requirements
that could be generally applied to any hospital department, taking into account factors such
as occupancy level, professional absences, allowances, physician work duties and
patient service levels.
Design/methodology/approach
The ability to generalize the model was tested via its implementation to five hospital
departments considered to represent a cross-section of all hospital…Purpose
The purpose of this study is to develop a model for physician staffing requirements
that could be generally applied to any hospital department, taking into account factors such
as occupancy level, professional absences, allowances, physician work duties and
patient service levels.
Design/methodology/approach
The ability to generalize the model was tested via its implementation to five hospital
departments considered to represent a cross-section of all hospital requirements: Internal
Medicine, Surgery, Orthopedics Children's (Pediatrics), and Gynecology. The work is
based on a combination of a survey, work-sampling and direct time study, conducted
by professional observers with a high degree of frequency and over a relatively long
period of time.
Findings
The model that we developed is based on the concept of "required work capacity" of
physicians. The model makes it possible to account for such factors as yearly
capacity, level of desired service, increments for work conditions, roster duties and
necessary absences.אַחֵר הכותבים -
Optimization-Seeking Experimentations: Design of an RL Circuit via the Vs-optimality criterion
Quality and Reliability Engineering International, Volume 26, 147-155
In this paper we explore the Vs-optimality criterion that was proposed in Ginsburg and Ben-Gal (IIE Trans. 2006; 38:445–461) as a new design-of-experiment (DOE) alphabetic optimality criterion. The Vs-optimality criterion seeks to minimize the variance of the optimal solution of an empirically fitted model. We show that the Vs-optimality citerion is well related to known alphabetic DOE criteria. However, it focuses on an ‘optimization-seeking’ experimental approach rather than an…
In this paper we explore the Vs-optimality criterion that was proposed in Ginsburg and Ben-Gal (IIE Trans. 2006; 38:445–461) as a new design-of-experiment (DOE) alphabetic optimality criterion. The Vs-optimality criterion seeks to minimize the variance of the optimal solution of an empirically fitted model. We show that the Vs-optimality citerion is well related to known alphabetic DOE criteria. However, it focuses on an ‘optimization-seeking’ experimental approach rather than an ‘information-seeking’ approach, which is often adopted by traditional optimality criteria. We illustrate the differences between these two approaches by a detailed example of a robust design of an RL-circuit.
אַחֵר הכותבים -
Kahan T., Bukchin Y., Menassa R. and Ben-Gal I., "Backup Strategy for Robots' Failures in an Automotive Assembly system."
International Journal of Production Economics 120 (2009) 315–326
Automotive assembly lines are often characterized by robots’ failures that may result in
stoppages of the lines and manual backup of tasks. The phenomena tend to impair
throughput rate and products’ quality. This paper presents a backup strategy in which
working robots perform tasks of failed robots. The proposed Mixed-Integer LinearProgramming
based approach minimizes the throughput loss by utilizing the robots’
redundancy in the system. Two algorithms are developed to comply…Automotive assembly lines are often characterized by robots’ failures that may result in
stoppages of the lines and manual backup of tasks. The phenomena tend to impair
throughput rate and products’ quality. This paper presents a backup strategy in which
working robots perform tasks of failed robots. The proposed Mixed-Integer LinearProgramming
based approach minimizes the throughput loss by utilizing the robots’
redundancy in the system. Two algorithms are developed to comply with stochastic
conditions of a real-world environment. The performance of these algorithms is
compared with several heuristics, and the downstream-backup based algorithm is
found superior to all other methods. -
Measuring the Efficiency of the Intraday Forex Market with a Universal Data Compression Algorithm
Computational Economics, Vol. 33 (2) 131-154.
Universal compression algorithms can detect recurring patterns in any
type of temporal data—including financial data—for the purpose of compression.
The universal algorithms actually find a model of the data that can be used for either
compression or prediction. We present a universal Variable Order Markov (VOM)
model and use it to test the weak form of the Efficient Market Hypothesis (EMH).
The EMH is tested for 12 pairs of international intra-day currency exchange rates…Universal compression algorithms can detect recurring patterns in any
type of temporal data—including financial data—for the purpose of compression.
The universal algorithms actually find a model of the data that can be used for either
compression or prediction. We present a universal Variable Order Markov (VOM)
model and use it to test the weak form of the Efficient Market Hypothesis (EMH).
The EMH is tested for 12 pairs of international intra-day currency exchange rates for
one year series of 1, 5, 10, 15, 20, 25 and 30 min. Statistically significant compression
is detected in all the time-series and the high frequency series are also predictable
above random. However, the predictability of the model is not sufficient to generate
a profitable trading strategy, thus, Forex market turns out to be efficient, at least most
of the time.אַחֵר הכותבים -
Bayesian Networks
Encyclopedia of Statistics in Quality and Reliability, Ruggeri F., Faltin F. & Kenett R. (Eds.), Wiley & Sons
-
Evaluation of Gene-Expression Clustering by Mutual Information Distance Measures
BMC Bioinformatics, 8(1):111
Background: The definition of a distance measure plays a key role in the evaluation of different
clustering solutions of gene expression profiles. In this empirical study we compare different
clustering solutions when using the Mutual Information (MI) measure versus the use of the well
known Euclidean distance and Pearson correlation coefficient.
Results: Relying on several public gene expression datasets, we evaluate the homogeneity and
separation scores of different clustering…Background: The definition of a distance measure plays a key role in the evaluation of different
clustering solutions of gene expression profiles. In this empirical study we compare different
clustering solutions when using the Mutual Information (MI) measure versus the use of the well
known Euclidean distance and Pearson correlation coefficient.
Results: Relying on several public gene expression datasets, we evaluate the homogeneity and
separation scores of different clustering solutions. It was found that the use of the MI measure
yields a more significant differentiation among erroneous clustering solutions. The proposed
measure was also used to analyze the performance of several known clustering algorithms. A
comparative study of these algorithms reveals that their "best solutions" are ranked almost
oppositely when using different distance measures, despite the found correspondence between
these measures when analysing the averaged scores of groups of solutions.
Conclusion: In view of the results, further attention should be paid to the selection of a proper
distance measure for analyzing the clustering of gene expression data.אַחֵר הכותבים -
Recognition of CIS-Regulatory Elements with VOMBAT
The Journal of Bioinformatics and Computational Biology, vol. 5, (02B), 561-579
Variable order Markov models and variable order Bayesian trees have been proposed
for the recognition of cis-regulatory elements, and it has been demonstrated that they
outperform traditional models such as position weight matrices, Markov models, and
Bayesian trees for the recognition of binding sites in prokaryotes. Here, we study to which
degree variable order models can improve the recognition of eukaryotic cis-regulatory
elements. We find that variable order models can…Variable order Markov models and variable order Bayesian trees have been proposed
for the recognition of cis-regulatory elements, and it has been demonstrated that they
outperform traditional models such as position weight matrices, Markov models, and
Bayesian trees for the recognition of binding sites in prokaryotes. Here, we study to which
degree variable order models can improve the recognition of eukaryotic cis-regulatory
elements. We find that variable order models can improve the recognition of binding sites
of all of the studied transcription factors. To ease a systematic evaluation of different
model combinations based on problem-specific data sets and allow genomic scans of
cis-regulatory elements based on fixed and variable order Markov models and Bayesian
trees, we provide the VOMBAT server to the public community.אַחֵר הכותבים -
Remote Learning for the Manipulation and Control of Robotic Cells
The European Journal of Engineering Education, vol. 32 (4), 481-494
This work proposes an approach to remote learning of robotic cells based on internet and simulation tools. The proposed approach, which integrates remote-learning and tele-operation into a generic scheme, is designed to enable students and developers to set-up and manipulate a robotic cell remotely. Its implementation is based on a dedicated website that supports a full three-dimensional simulation and a manual online control of the cell. The website provides high flexibility for making changes…
This work proposes an approach to remote learning of robotic cells based on internet and simulation tools. The proposed approach, which integrates remote-learning and tele-operation into a generic scheme, is designed to enable students and developers to set-up and manipulate a robotic cell remotely. Its implementation is based on a dedicated website that supports a full three-dimensional simulation and a manual online control of the cell. The website provides high flexibility for making changes in the design of the work area, independent of the user’s location. The website can be used to further integrate full-scale remote learning labs in collaboration with other organizations
אַחֵר הכותבים -
The Funnel Experiment: A Markov-Based SPC approach
Quality and Reliability Engineering International, 23:899–913
The classical funnel experiment was used by Deming to promote the idea of statistical
process control (SPC). The popular example illustrates that the implementation of
simple feedback rules to stationary processes violates the independence assumption
and prevents the implementation of conventional SPC. However, Deming did
not indicate how to implement SPC in the presence of such feedback rules.
This pedagogical gap is addressed here by introducing a simple feedback rule to
the…The classical funnel experiment was used by Deming to promote the idea of statistical
process control (SPC). The popular example illustrates that the implementation of
simple feedback rules to stationary processes violates the independence assumption
and prevents the implementation of conventional SPC. However, Deming did
not indicate how to implement SPC in the presence of such feedback rules.
This pedagogical gap is addressed here by introducing a simple feedback rule to
the funnel example that results in a nonlinear process to which the traditional SPC
methods cannot be applied. The proposed method of Markov-based SPC, which is a
simplified version of the context-based SPC method, is shown to monitor the modified
process wellאַחֵר הכותבים -
Using a VOM Model for Reconstructing Potential Coding Regions in EST Sequences
Journal of Computational Statistics, vol. 22, no. 1, 49-69
This paper presents a method for annotating coding and noncoding DNA regions by using variable order Markov (VOM) models. A main advantage in using VOM models is that their order may vary for different sequences, depending on the sequences’ statistics. As a result, VOM models are more flexible with respect to model parameterization and can be trained on relatively short sequences and on low-quality datasets, such as expressed sequence tags (ESTs). The paper presents a modified VOM model for…
This paper presents a method for annotating coding and noncoding DNA regions by using variable order Markov (VOM) models. A main advantage in using VOM models is that their order may vary for different sequences, depending on the sequences’ statistics. As a result, VOM models are more flexible with respect to model parameterization and can be trained on relatively short sequences and on low-quality datasets, such as expressed sequence tags (ESTs). The paper presents a modified VOM model for detecting and correcting insertion and deletion sequencing errors that are commonly found in ESTs. In a series of experiments the proposed method is found to be robust to random errors in these sequences
אַחֵר הכותבים -
Designing Experiments for Robust Optimization Problems: The Vs-optimality criterion.
IIE Transactions on Quality and Reliability, Vol. 38, 445 – 461.
We suggest an experimentation strategy for the robust design of empirically fitted models. The suggested approach is used to design experiments that minimize the variance of the optimal robust solution. The new design-of-experiment optimality criterion, termed Vs-optimal, prioritizes the estimation of a model’s coefficients, such that the variance of the optimal solution is minimized by the
performed experiments. It is discussed how the proposed criterion is related to known optimality…We suggest an experimentation strategy for the robust design of empirically fitted models. The suggested approach is used to design experiments that minimize the variance of the optimal robust solution. The new design-of-experiment optimality criterion, termed Vs-optimal, prioritizes the estimation of a model’s coefficients, such that the variance of the optimal solution is minimized by the
performed experiments. It is discussed how the proposed criterion is related to known optimality criteria. We present an analytical formulation of the suggested approach for linear models and a numerical procedure for higher-order or nonpolynomial models. In comparison with conventional robust-design methods, our approach provides more information on the robust solution by umerically generating its multidimensional distribution. Moreover, in a case study, the proposed approach results in a better robust solution in comparison with these standard methods.אַחֵר הכותבים -
Gene-Finding with the VOM Model
Journal of Computational Methods in Sciences and Engineering, Vol.7, pp. 45-54
We present the architecture of an elementary gene-finding algorithm that is based on a Variable Order Markov model (VOM). The VOM model is a generalization of the traditional Markov model that can cope with varying memory dependencies. The VOM model is more efficient in terms of its parameterization and therefore can be trained on relatively short sequences. Experiments with the proposed gene-finder (GF) on three prokaryotic genomes indicate its potential advantage on the detection of short…
We present the architecture of an elementary gene-finding algorithm that is based on a Variable Order Markov model (VOM). The VOM model is a generalization of the traditional Markov model that can cope with varying memory dependencies. The VOM model is more efficient in terms of its parameterization and therefore can be trained on relatively short sequences. Experiments with the proposed gene-finder (GF) on three prokaryotic genomes indicate its potential advantage on the detection of short genes.
אַחֵר הכותבים -
VOMBAT: Prediction of Transcription Factor Binding Sites using Variable Order Bayesian Trees
Nucleic Acids Research, vol. 34, issue W529–W533.
Variable order Markov models and variable order
Bayesian trees have been proposed for the recognition
of transcription factor binding sites, and it
could be demonstrated that they outperform traditional
models, such as position weight matrices,
Markov models and Bayesian trees. We develop a
web server for the recognition of DNA binding sites
based on variable order Markov models and variable
order Bayesian trees offering the following functionality:
(i) given datasets…Variable order Markov models and variable order
Bayesian trees have been proposed for the recognition
of transcription factor binding sites, and it
could be demonstrated that they outperform traditional
models, such as position weight matrices,
Markov models and Bayesian trees. We develop a
web server for the recognition of DNA binding sites
based on variable order Markov models and variable
order Bayesian trees offering the following functionality:
(i) given datasets with annotated binding
sites and genomic background sequences, variable
order Markov models and variable order Bayesian
trees can be trained; (ii) given a set of trained
models, putative DNA binding sites can be predicted
in a given set of genomic sequences and (iii)
given a dataset with annotated binding sites and
a dataset with genomic background sequences,
cross-validation experiments for different model
combinations with different parameter settings can
be performed. Several of the offered services are
computationally demanding, such as genome-wide
predictions of DNA binding sites in mammalian
genomes or sets of 104
-fold cross-validation experiments
for different model combinations based on
problem-specific data sets. In order to execute these
jobs, and in order to serve multiple users at the
same time, the web server is attached to a Linux
cluster with 150 processors.אַחֵר הכותבים -
Economic optimization of off-line inspection in a process subject to failure and recovery
IIE Transactions on Quality and Reliability, 37, 995–1009.
In certain types of processes, verification of the quality of the output units is possible only after the entire batch has been processed. We develop a model that prescribes which units should be inspected and how the units that were not inspected should be disposed of, in order to minimize the expected sum of inspection costs and disposition error costs, for processes that are subject to random failure and recovery. The model is based on a dynamic programming algorithm that has a low…
In certain types of processes, verification of the quality of the output units is possible only after the entire batch has been processed. We develop a model that prescribes which units should be inspected and how the units that were not inspected should be disposed of, in order to minimize the expected sum of inspection costs and disposition error costs, for processes that are subject to random failure and recovery. The model is based on a dynamic programming algorithm that has a low computational complexity. The study also includes a sensitivity analysis under a variety of cost and probability scenarios, supplemented by an analysis of the smallest batch that requires inspection, the expected number of inspections, and the performance of an easy to implement heuristic.
אַחֵר הכותבים -
Identification of Transcription Factor Binding Sites with Variable-order Bayesian Networks,
Bioinformatics,vol. 21, no. 11, 2657-2666.
Motivation: We propose a new class of variable-order Bayesian network (VOBN) models for the identification of transcription factor binding sites (TFBSs). The proposed models generalize the widely used position weight matrix (PWM) models, Markov models and Bayesian network models. In contrast to these models, where for each position a fixed subset of the remaining positions is used to model dependencies, in VOBN models, these subsets may vary based on the specific nucleotides observed, which are…
Motivation: We propose a new class of variable-order Bayesian network (VOBN) models for the identification of transcription factor binding sites (TFBSs). The proposed models generalize the widely used position weight matrix (PWM) models, Markov models and Bayesian network models. In contrast to these models, where for each position a fixed subset of the remaining positions is used to model dependencies, in VOBN models, these subsets may vary based on the specific nucleotides observed, which are called the context. This flexibility turns out to be of advantage for the classification and analysis of TFBSs, as statistical dependencies between nucleotides in different TFBS positions (not necessarily adjacent) may be taken into account efficiently—in a position-specific and context-specific manner.
Results: We apply the VOBN model to a set of 238 experimentally verified sigma-70 binding sites in Escherichia coli. We find that the VOBN model can distinguish these 238 sites from a set of 472 intergenic ‘nonpromoter’ sequences with a higher accuracy than fixed-order Markov models or Bayesian trees. We use a replicated stratified-holdout experiment having a fixed true-negative rate of 99.9%. We find that for a foreground inhomogeneous VOBN model of order 1 and a background
homogeneous variable-order Markov (VOM) model of order 5, the obtained mean true-positive (TP) rate is 47.56%. In comparison, the best TP rate for the conventional models is 44.39%, obtained from a foreground PWM model and a background 2nd-order Markov model. As the standard deviation of the estimated TP rate is 0.01%, this improvement is highly significant.
אַחֵר הכותבים -
On the Use of Data Compression Measures to Assess Robust Designs
IEEE Trans. on Reliability, Vol. 54, no. 3, 381-388.
—In this paper, we suggest a potential use of data compression
measures, such as the Entropy, and the Huffman Coding,
to assess the effects of noise factors on the reliability of tested systems.
In particular, we extend the Taguchi method for robust design
by computing the entropy of the percent contribution values
of the noise factors. The new measures are computed already at
the parameter-design stage, and together with the traditional S/N
ratios enable the specification…—In this paper, we suggest a potential use of data compression
measures, such as the Entropy, and the Huffman Coding,
to assess the effects of noise factors on the reliability of tested systems.
In particular, we extend the Taguchi method for robust design
by computing the entropy of the percent contribution values
of the noise factors. The new measures are computed already at
the parameter-design stage, and together with the traditional S/N
ratios enable the specification of a robust design. Assuming that
(some of) the noise factors should be naturalized, the entropy of a
design reflects the potential efforts that will be required in the tolerance-design
stage to reach a more reliable system. Using a small
example, we illustrate the contribution of the new measure that
might alter the designer decision in comparison with the traditional
Taguchi method, and ultimately obtain a system with a lower
quality loss.
Assuming that the percent contribution values can reflect the
probability of a noise factor to trigger a disturbance in the system
response, a series of probabilistic algorithms can be applied to the
robust design problem. We focus on the Huffman coding algorithm,
and show how to implement this algorithm such that the designer
obtains the minimal expected number of tests in order to find the
disturbing noise factor. The entropy measure, in this case, provides
the lower bound on the algorithm’s performance. -
Outlier detection
In: Data Mining and Knowledge Discovery Handbook: A Complete Guide for Practitioners and Researchers, Maimon O. and Rockach L. (Editors.) Kluwer Academic Publishers
-
Statistical Process Control of the Stochastic Complexity of Discrete Processes
Communications on Dependability and Quality Management in Engineering, Vol. 8, no. 3, 55-61.
Changes in stochastic processes often affect their description length, and reflected by their stochastic complexity measures. Monitoring the stochastic complexity of a sequence (or, equivalently, its code length) can detect process changes that may be undetectable by traditional SPC methods. The context tree is proposed here as a universal compression algorithm for measuring the stochastic complexity of a state-dependent discrete process. The advantage of
the proposed method is in the…Changes in stochastic processes often affect their description length, and reflected by their stochastic complexity measures. Monitoring the stochastic complexity of a sequence (or, equivalently, its code length) can detect process changes that may be undetectable by traditional SPC methods. The context tree is proposed here as a universal compression algorithm for measuring the stochastic complexity of a state-dependent discrete process. The advantage of
the proposed method is in the reduced number of samples that are needed for reliable monitoring.
אַחֵר הכותבים -
Throughput of Multiple Part-Type Systems: A Subjective Linear Measure
The International Journal for Manufacturing Science and Production (IJMSP), Vol. 6(3), 143-151.
The term throughput, which is commonly used as a performance measure of various production systems, has not been uniquely defined for multiple part-type systems. In some cases, the analytical procedures that were developed to maximize throughput of multiple part-type systems are difficult to evaluate and justify. Moreover, in some cases an inaccurate definition of this term motivated incorrect operational concepts. This paper discusses some of the problems in the traditional definition of…
The term throughput, which is commonly used as a performance measure of various production systems, has not been uniquely defined for multiple part-type systems. In some cases, the analytical procedures that were developed to maximize throughput of multiple part-type systems are difficult to evaluate and justify. Moreover, in some cases an inaccurate definition of this term motivated incorrect operational concepts. This paper discusses some of the problems in the traditional definition of throughput and suggests a new, more suitable definition for a multiple-product manufacturing system.
אַחֵר הכותבים -
Using a Compressibility Measure to Distinguish Coding and Noncoding DNA
FEJournal of Theoretical Statistics (FEJT).Vol. 13(2), 215-235
-
Using a Pseudo-Stochastic Approach for Multiple-Parts Scheduling on an Unreliable Machine,
IIE Transactions on Operations Engineering, vol. 37, no. 3, pp. 189 - 199
In this paper we follow previous “pseudo-stochastic” approaches that solve stochastic control problems by using deterministic optimal control methods. In a similar manner to the certainty equivalence principle, the suggested model maximizes a given profit function of the expected system outcome. However, unlike the certainty equivalence principle, we model the expected influences of all future events (including those that are expected beyond the planning horizon), as encapsulated by their…
In this paper we follow previous “pseudo-stochastic” approaches that solve stochastic control problems by using deterministic optimal control methods. In a similar manner to the certainty equivalence principle, the suggested model maximizes a given profit function of the expected system outcome. However, unlike the certainty equivalence principle, we model the expected influences of all future events (including those that are expected beyond the planning horizon), as encapsulated by their density functions and not only by their mean values. The model is applied to the optimal scheduling of multiple part-types on a single machine that is subject to random
failures and repairs. The objective of the scheduler is to maximize the profit function of the produced multiple-part mix. A numerical study is performed to evaluate the suggested pseudo-stochastic solutions under various conditions. These solutions are compared to a profit upper bound of the stochastic optimal control solutionsאַחֵר הכותבים -
An Upper Bound of the Weight-Balanced Testing Procedure with Multiple Testers
IIE Transactions on Quality and Reliability, vol. 36, no. 5, 481-491.
This paper presents the performance of the Weight-Balanced Testing (WBT)algorithm with multiple testers. The WBT algorithm aims to minimize the expected number of (round of)tests and has been proposed for coding, memory storage, search and testing applications. It often provides reasonable results if used with a single tester. Yet, the performance of the WBT algorithm with multiple testers and particularly its upper bound have not been previously analyzed, despite the large body of literature…
This paper presents the performance of the Weight-Balanced Testing (WBT)algorithm with multiple testers. The WBT algorithm aims to minimize the expected number of (round of)tests and has been proposed for coding, memory storage, search and testing applications. It often provides reasonable results if used with a single tester. Yet, the performance of the WBT algorithm with multiple testers and particularly its upper bound have not been previously analyzed, despite the large body of literature that exists on the WBT algorithm, and the recent papers that suggest its use in various testing applications. Here we demonstrate that WBT algorithm with multiple testers is far from being the optimal search procedure. The main result of this paper is the generalization of the upper bound on the expected number of tests previously obtained for a single-tester WBT algorithm. For this purpose, we first draw an analogy between the WBT algorithm and alphabetic codes; both being represented by the same Q-ary search tree. The upper bound is then obtained on the expected path length of a Q-ary tree, which is constructed by the WBT algorithm. Applications to the field of testing and some numerical examples are presented for illustrative purposes.
-
Statistical Process Control via Context Modeling of Finite State Processes: An Application to Production Monitoring,
IIE Transactions on Quality and Reliability, vol. 36, no. 5, 401-415.
Conventional Statistical Process Control (SPC) schemes fail to monitor nonlinear and finite-state processes that often result from feedback-controlled processes. SPC methods that are designed to monitor autocorrelated processes usually assume a known model (often an ARIMA) that might poorly describe the real process. In this paper, we present a novel SPC methodology based on context
modeling of finite-state processes. The method utilizes a series of context-tree models to estimate the…Conventional Statistical Process Control (SPC) schemes fail to monitor nonlinear and finite-state processes that often result from feedback-controlled processes. SPC methods that are designed to monitor autocorrelated processes usually assume a known model (often an ARIMA) that might poorly describe the real process. In this paper, we present a novel SPC methodology based on context
modeling of finite-state processes. The method utilizes a series of context-tree models to estimate the conditional distribution of the process output given the context of previous observations. The Kullback-Leibler divergence statistic is derived to indicate significant changes in the trees along the process. The method is implemented in a simulated flexible manufacturing system in order to detect significant changes in its production mix ratio output.אַחֵר הכותבים -
CSPC: A Monitoring Procedure for State Dependent Processes,
Technometrics, Vol. 45, no. 4, pp. 293-311.
Most statistical process control (SPC) methods are not suitable for monitoring nonlinear and statedependent processes. This article introduces the context-based SPC (CSPC) methodology for statedependent data generated by a finite-memory source. The key idea of the CSPC is to monitor the statistical attributes of a process by comparing two context trees at any monitoring period of time. The first is a reference tree that represents the “in control” reference behavior of the process; the second…
Most statistical process control (SPC) methods are not suitable for monitoring nonlinear and statedependent processes. This article introduces the context-based SPC (CSPC) methodology for statedependent data generated by a finite-memory source. The key idea of the CSPC is to monitor the statistical attributes of a process by comparing two context trees at any monitoring period of time. The first is a reference tree that represents the “in control” reference behavior of the process; the second is a monitored tree, generated periodically from a sample of sequenced observations, that represents the behavior of the process at that period. The Kullback–Leibler (KL) statistic is used to measure the relative “distance” between these two trees, and an analytic distribution of this statistic is derived. Monitoring the KL statistic indicates whether there has been any significant change in the process that requires intervention. An example of buffer-level monitoring in a roduction system demonstrates the viability of the new method with respect to conventional methods
אַחֵר הכותבים -
Self-Correcting Inspection Procedure Under Inspection Errors.
IIE Transactions on Quality and Reliability, 34(6), pp. 529-540
In this paper we present a novel treatment of the inspection-system design problem when inspection is unreliable and subject to classification errors. Our approach, based on the theory of Error-Correcting Codes (ECC), leads to the development of a SelfCorrecting Inspection (SCI) decision rule that does not require complete knowledge of inspection error probabilities. We show that the proposed rule assures correct classification, if the number of inspection errors is less than a certain number…
In this paper we present a novel treatment of the inspection-system design problem when inspection is unreliable and subject to classification errors. Our approach, based on the theory of Error-Correcting Codes (ECC), leads to the development of a SelfCorrecting Inspection (SCI) decision rule that does not require complete knowledge of inspection error probabilities. We show that the proposed rule assures correct classification, if the number of inspection errors is less than a certain number. We analyze the performance of the SCI decision rule under different inspection situations, including some situations that are uncommon in the field of error-correcting codes. Then, we show how the underlying mathematical structure can be applied to determine the number
of inspections and the level of inspection reliability in order to minimize the sum of inspection-related costs. The practical contribution of this work lies in that it expands the ability of the designer of inspection systems to deal with cases where there is very little or no information regarding the reliability of the inspection operations.אַחֵר הכותבים -
Ben-Gal I. and (2002), Sequential DOE via Dynamic Programming,
IIE Transactions on Quality and Reliability, 34 (12), 1087-1100.
The paper considers a sequential Design Of Experiments (DOE) scheme. Our objective is to maximize both information and economic measures over a feasible set of experiments. Optimal DOE strategies are developed by introducing information criteria based on measures adopted from information theory. The evolution of acquired information along various stages of experimentation
is analyzed for linear models with a Gaussian noise term. We show that for particular cases, although the amount…The paper considers a sequential Design Of Experiments (DOE) scheme. Our objective is to maximize both information and economic measures over a feasible set of experiments. Optimal DOE strategies are developed by introducing information criteria based on measures adopted from information theory. The evolution of acquired information along various stages of experimentation
is analyzed for linear models with a Gaussian noise term. We show that for particular cases, although the amount of
information is unbounded, the desired rate of acquiring information decreases with the number of experiments. This observation implies that at a certain point in time it is no longer efficient to continue experimenting. Accordingly, we investigate methods of stochastic dynamic programming under imperfect state information as appropriate means to obtain optimal experimentation policies. We propose cost-to-go functions that model the trade-off between the cost of additional experiments and the benefit of incremental information. We formulate a general stochastic dynamic programming framework for design of experiments and illustrate it by analytic and numerical implementation examples.אַחֵר הכותבים -
An Application of Information Theory and Error-Correcting Codes to Fractional Factorial Experiments.
Journal of Statistical Planing and Inference. 92/1-2, 267-282
The objective of design of experiments (DOE) is addressed by introducing an information
optimality criterion, which is based on concepts adopted from information theory. In particular,
experiments are specied to maximize the information in the system responses about estimators of
the system parameters. It is shown that one has to maintain certain resolution of the design matrix
to maximize the information, obtainable by a design, about a system described by a linear model
with…The objective of design of experiments (DOE) is addressed by introducing an information
optimality criterion, which is based on concepts adopted from information theory. In particular,
experiments are specied to maximize the information in the system responses about estimators of
the system parameters. It is shown that one has to maintain certain resolution of the design matrix
to maximize the information, obtainable by a design, about a system described by a linear model
with interactions. The correspondence between error-correcting codes and fractional factorial
experiments provides a method to attain the required resolution with a smaller fractional factorial
experiment by increasing the number of levels associated with each factor – a result that in the
context of experimental design seems counterintuitive. In particular, the Gilbert–Varshamov and
the Singleton bounds are employed to obtain bounds on the size of the fractional experiment.
Analytical approximations and numerical results are given and illustrated by examples.
אַחֵר הכותבים -
Design of Control and Monitoring Rules for State Dependent Processes
The International Journal for Manufacturing Science and Production, Vol. 3, Nos. 2-4, p. 85-93.
אַחֵר הכותבים -
On the Uncertainties of Decentralized Controllers in a Transfer Production Line.
IIE Transactions on Design and Manufacturing, 32, 953-961
In this paper, an information theoretic approach is applied to analyze the performance of a decentralized control system. The control system plays the role of a correcting device which decreases the uncertainties associated with state variables of a production line by applying an appropriate ``correcting signal'' for each deviation from the target. In particular, a distributed feedback control
policy is considered to govern a transfer production line, which consists of machines and buers…In this paper, an information theoretic approach is applied to analyze the performance of a decentralized control system. The control system plays the role of a correcting device which decreases the uncertainties associated with state variables of a production line by applying an appropriate ``correcting signal'' for each deviation from the target. In particular, a distributed feedback control
policy is considered to govern a transfer production line, which consists of machines and buers and processes a single part type in response to a stochastic demand. It is shown how the uncertainty of the demand propagates dynamically into the production system, causing uncertainties associated with buer levels and machine production rates. The paper proposes upper estimates for these uncertainties as functions of the demand variance, parameters of the distributed controllers and some physical properties of the production line. The bounds are based on dynamic entropy measures of the system state and the control variables. Some practical implications into the area of decentralized controller design are proposed, an information-economical analysis is presented and a numerical study is performed.אַחֵר הכותבים -
Ben-Gal I., (2002), The Ergonomic Design of Workstation Using Rapid Prototyping and Response Surface Methodology
IIE Transactions on Design and Manufacturing, 34(4), 375-391
-
Comparing the diversity of information by word-of-mouth vs. web spread
EPL European Physical Society.
Many studies have explored spreading and diffusion through complex networks. The following study examines a specific case of spreading of opinions in modern society through two spreading schemes – defined as being either through ‘word-of-mouth’ (WOM), or through online search engines (WEB). We apply both modelling and real experimental results and compare the opinions people adopt through an exposure to their friend`s opinions, as opposed to the opinions they adopt when using a search engine…
Many studies have explored spreading and diffusion through complex networks. The following study examines a specific case of spreading of opinions in modern society through two spreading schemes – defined as being either through ‘word-of-mouth’ (WOM), or through online search engines (WEB). We apply both modelling and real experimental results and compare the opinions people adopt through an exposure to their friend`s opinions, as opposed to the opinions they adopt when using a search engine based on the PageRank algorithm. A simulated study shows that when members in a population adopt decisions through the use of the WEB scheme, the population ends up with a few dominant views, while other views are barely expressed. In contrast, when members adopt decisions based on the WOM scheme, there is a far more diverse distribution of opinions in that population. The simulative results are further supported by an online experiment which finds that people searching information through a search engine end up with far more homogenous opinions as compared to those asking their friends.
אַחֵר הכותבים -
Numerous publications
Proffesional journals
Irad wrote and edited six books, has published more than 80 scientific papers and patents, received numerous best papers awards and supervised more than 40 graduate students.
For Irad's selected publications list please see http://www.eng.tau.ac.il/~bengal/publications.html -
Probabilistic Sequential Methodology for Designing a Factorial System with Multiple Responses
International Journal of Production Research, Vol. 37(12) 2703-2724.
This paper addresses the problem of optimizing a factorial system with multiple
responses. A heuristic termed probabilistic sequential methodology (PSM) is proposed.
The PSMidenti®es those designs that maximize the likelihood of satisfying
a given set of functional requirements. It is based on sequential experimentation,
statistical inference and a probabilistic local search. The PSM comprises three
main steps: (1) screening and estimating the main location and dispersion e…This paper addresses the problem of optimizing a factorial system with multiple
responses. A heuristic termed probabilistic sequential methodology (PSM) is proposed.
The PSMidenti®es those designs that maximize the likelihood of satisfying
a given set of functional requirements. It is based on sequential experimentation,
statistical inference and a probabilistic local search. The PSM comprises three
main steps: (1) screening and estimating the main location and dispersion e ects
by applying fractional factorial experiments (FFE) techniques; (2) based on these
e ects, establishing probabilistic measures for di erent combinations of factorlevels;
and (3) constructing a set of candidate designs from which the best solution
is selected by applying a heuristic local search. The PSM is attractive when the
exact analytic relationship between factor-level combinations and the system’s
responses is unknown; when the system involves qualitative factors; and when
the number of experiments is limited. The PSM is illustrated by a detailed case
study of a Flexible Manufacturing Cell (FMC) design.
אַחֵר הכותבים -
Robust Eco-Design: A New Application for Quality Engineering
IIE Transactions, Vol. 40 (10), p. 907 - 918
The method of robust design has long been used for the design of systems that are insensitive to noises. In this paper it is demonstrated how this approach can be used to obtain a robust eco-design (ecological design). In a case study, robust design principles are applied to the design of a factory smokestack, using the Gaussian Plume Model (GPM). The GPM is a well-known model for describing
pollutant dispersal from a point source, subject to various atmospheric conditions. In this research,…The method of robust design has long been used for the design of systems that are insensitive to noises. In this paper it is demonstrated how this approach can be used to obtain a robust eco-design (ecological design). In a case study, robust design principles are applied to the design of a factory smokestack, using the Gaussian Plume Model (GPM). The GPM is a well-known model for describing
pollutant dispersal from a point source, subject to various atmospheric conditions. In this research, the mean-square-error (MSE) of the accumulated and the maximum pollution values around a given target are defined as the performance measures and used to adjust the design parameters. Both analytical and numerical approaches are used to evaluate the MSE measures over the design space. It is demonstrated how to use the non-linearity in the GPM to reach a low MSE value that produces a cheaper design configuration. The differences between the manufacturer viewpoint and the environmentalist viewpoint with respect to the considered eco-design problem are discussed and analyzed.אַחֵר הכותבים
ארגונים
-
CB4 ("See Before")
-
-
Stanford University
-
-
Tel Aviv University
-
עוד פעילות על ידי Irad
-
After nearly a decade I am concluding my role as CEO of the Israel-America Chamber of Commerce and will focus on areas dear to my heart, healthcare…
After nearly a decade I am concluding my role as CEO of the Israel-America Chamber of Commerce and will focus on areas dear to my heart, healthcare…
נוסף לייק על ידי Irad Ben-Gal
-
ד״ר Maayan Cohen מדברת בפודקאסט (המרתק תמיד!) המצב הפוסט אנושי, ומספרת על מספר מחקרים שהובילה במעבדה שלנו כפוסט-דוק על התפיסות האתיות העולות מהשטח עם…
ד״ר Maayan Cohen מדברת בפודקאסט (המרתק תמיד!) המצב הפוסט אנושי, ומספרת על מספר מחקרים שהובילה במעבדה שלנו כפוסט-דוק על התפיסות האתיות העולות מהשטח עם…
נוסף לייק על ידי Irad Ben-Gal
-
Today we were thrilled to collaborate with Drive TLV on this insightful event, "AI at Scale: Real-World Challenges & Solutions in Mobility", where…
Today we were thrilled to collaborate with Drive TLV on this insightful event, "AI at Scale: Real-World Challenges & Solutions in Mobility", where…
נוסף לייק על ידי Irad Ben-Gal
-
So much needed for data scientists and analytical officers that want to focus on modeling and performance, instead of technical overhead and…
So much needed for data scientists and analytical officers that want to focus on modeling and performance, instead of technical overhead and…
שותף על ידי Irad Ben-Gal
-
I’m thrilled to officially step into the role of CEO at Syte as part of Pereg Ventures’ investment in this incredible company. Syte has already…
I’m thrilled to officially step into the role of CEO at Syte as part of Pereg Ventures’ investment in this incredible company. Syte has already…
נוסף לייק על ידי Irad Ben-Gal
-
This week Bria won the Design Partner of the Year Award for the EMEA region🏆 It’s an honor to receive this award from the team at AWS. Our Chief…
This week Bria won the Design Partner of the Year Award for the EMEA region🏆 It’s an honor to receive this award from the team at AWS. Our Chief…
נוסף לייק על ידי Irad Ben-Gal
-
The company AAI, founded a year ago, is still in stealth mode, but today we are doing a "partial unstealthing".. A year and a half ago I gave a…
The company AAI, founded a year ago, is still in stealth mode, but today we are doing a "partial unstealthing".. A year and a half ago I gave a…
נוסף לייק על ידי Irad Ben-Gal
-
We're proud of our partnership with Hive and the ability to reveal valuable insights about our society through advanced technology. This…
We're proud of our partnership with Hive and the ability to reveal valuable insights about our society through advanced technology. This…
נוסף לייק על ידי Irad Ben-Gal
-
Over the past 14 months we've seen unbelievable situations develop where Israelis and Jews are attacked and cancelled on campuses. But it never…
Over the past 14 months we've seen unbelievable situations develop where Israelis and Jews are attacked and cancelled on campuses. But it never…
נוסף לייק על ידי Irad Ben-Gal
-
After class, I sat in my seat, staring at the wine and the bag (inside: a hot cocoa, a card). Pleasant surprises from my dear students. I thought…
After class, I sat in my seat, staring at the wine and the bag (inside: a hot cocoa, a card). Pleasant surprises from my dear students. I thought…
נוסף לייק על ידי Irad Ben-Gal
פרופילים דומים אחרים
אנשים אחרים ששמם Irad Ben-Gal
1 שם Irad Ben-Gal אחר נמצא על לינקדאין
הצג/י אנשים אחרים בשם Irad Ben-Gal