Table of Contents Table of Contents
Previous Page  356 / 561 Next Page
Information
Show Menu
Previous Page 356 / 561 Next Page
Page Background

INFORMS Nashville – 2016

356

3 - Facilitate Fit Revelation In a DistributionChannel

Lin Hao, University of Notre Dame, 351 Mendoza College Of

Business, Notre Dame, IN, 46556, United States,

lhao@nd.edu

,

Yong Tan

We investigate a retailer’s and a supplier’s incentive to facilitate fit revelation, i.e.,

facilitate consumer learning of their true product fit, under two popular channel

pricing models, agency pricing model and wholesale pricing model.

TD66

Mockingbird 2- Omni

Technometrics Invited Session: Recent Statistical

Methods for Analyzing Big Data

Sponsored: Quality, Statistics and Reliability

Sponsored Session

Chair: Peihua Qiu, Professor and Chair, University of Florida, 2004

Mowry Road, Gainesville, FL, 32611, United States,

pqiu@phhp.ufl.edu

1 - Discovering The Nature Of Variation In Nonlinear Profile Data

Daniel Apley, Northwestern University,

apley@northwestern.edu

Most prior work on profile data in the quality control literature has focused on

monitoring to detect sudden changes in the characteristics of the profiles, relative

to an in-control sample of profiles. In contrast, we develop an approach for

exploratory analysis of a sample of profiles to discover the nature of any profile-

to-profile variation present over the sample via manifold learning. Instead of

analyzing parameter variation in some prespecified parametric profile model, our

focus is on discovering and visualizing an appropriate characterization or

parameterization of the variation, as a tool to facilitate discovery (and ultimately

elimination) of its root causes.

2 - A Bootstrap Metropolis-hastings Algorithm For Bayesian Analysis

Of Big Data

Faming Liang, University of Florida,

faliang@ufl.edu

MCMC methods have proven to be a powerful tool for analyzing data of complex

structures. However, their computer-intensive nature precludes their use for big

data analysis. We propose the bootstrap Metropolis-Hastings (BMH) algorithm,

which provides a general framework for how to tame powerful MCMC methods

to be used for big data analysis; that is to replace the full data log-likelihood by a

Monte Carlo average of the log-likelihoods that are calculated in parallel from

multiple bootstrap samples. The BMH algorithm possesses an embarrassingly

parallel structure and avoids repeated scans of the full dataset in iterations, and is

thus feasible for big data problems.

3 - Online Updating Of Statistical Inference In The Big Data Setting

Elizabeth D. Schifano, University of Connecticut, Storrs, CT,

United States,

elizabeth.schifano@uconn.edu,

Ming-Hui Chen,

Chun Wang, Jing Wu, Jun Yan, Yuping Zhang

We present statistical methods for big data arising from online analytical

processing, where data arrive in streams and require fast analysis without

storage/access to historical data. In particular, we develop computationally

efficient, minimally storage-intensive iterative estimating algorithms and

statistical inferences for linear models and estimating equations that update as

new data arrive. We propose goodness of fit tests, a new estimator within the

estimating equation setting, and a modification to incorporate new variables

midway through the data stream. We demonstrate the effectiveness of our

procedures through theoretical and empirical analysis, as well as in application.

TD67

Mockingbird 3- Omni

QTQM Invited Session

Sponsored: Quality, Statistics and Reliability

Sponsored Session

Chair: Jing Li, Arizona State University, Arizona State University,

Tempe, AZ, 85287, United States,

jing.li.8@asu.edu

1 - A Mixed-effect Model For Analyzing Experiments With

Multistage Processes

Kaibo Wang, Tsinghua University,

kbwang@tsinghua.edu.cn

In industrial practice, most products are produced by processes that involve

multiple stages. In this paper, through an analysis of the error transmission

mechanism, we propose a mixed-effect model for analyzing experiments with

multistage processes. Based on an analysis of simulated and real experimental

data, we find that different conclusions about factor significance may be drawn if

the data are analyzed differently. In addition, the mixed-effect model can help

separate errors at different stages and hence provide more information on process

improvement.

2 - Setup Adjustment For Asymmetric Cost Functions Under

Unknown Process Parameters

Arda Vanli, Florida State University, 2525 Pottsdamer St,

Tallahassee, FL, FL, 32310, United States,

oavanli@eng.fsu.edu

,

Zilong Lian, Enrique Del Castillo”

We present a bayesian approach for the optimal control of a machine that can

experience setup errors assuming an asymmetric off-target cost function. It is

assumed that the setup error cannot be observed directly due to presence of

measurement and part-to-part errors and the variance of this error is not known

a priori. The setup error can be compensated by performing sequential

adjustments of the process mean based on observations of the parts produced. We

show how the proposed method converges to the optimal (known variance)

trajectory, recovering from a possibly biased initial variance estimate. Simulations

results are presented for constant asymmetric and quadratic asymmetric cost

functions.

3 - Bounded Loss Functions And The Characteristic Function

Inversion Method For Computing Expected Loss

Matthias Tan, City University of Hong Kong,

matthtan@cityu.edu.hk

In robust parameter design, the quadratic loss function is commonly used.

However, this loss function is not always realistic. We propose a general class of

bounded loss functions that are cumulative distribution functions and probability

density functions. New loss functions are investigated and they are shown to

behave differently from the quadratic loss. For models linear in the noise factors,

we give a numerical method based on characteristic functions inversion for

computing expected loss. The method is very quick and accurate. It is applicable

as long as the distributions chosen to represent the loss function and variation in

the noise factors have tractable characteristic functions.

4 - Quasi-feedforward And Feedback Control For Random Step Shift

Disturbance Models

Lihui Shi, Senior Data Scientist, Centerfield Corporation, El

Segundo, CA, 90245, United States,

shilihui@uw.edu

Lihui Shi, Senior Data Scientist, University of Washington,

Seattle, WA, 98195, United States,

shilihui@uw.edu

Process monitoring and process adjustment strategies are two important parts of

the process improvement methods, and they should be integrated together in

stead of separated. Integrated moving average (IMA) model is the most common

disturbance model, and step shift model is one type of more complicated one that

often exists in real applications. We investigate the IMA background disturbance

subject to random step shifts with a certain probability. We propose a feedback

control with a quasi-feedforward control by monitoring the output errors. It is

very robust against parameter misspecifications. We also investigate the type I and

type II errors in process adjustment on this disturbance model.

TD68

Mockingbird 4- Omni

Reliability Modeling and Optimization in Early Product

Design Stages

Sponsored: Quality, Statistics and Reliability

Sponsored Session

Chair: Zhaojun Li, Western New England University, Springfield, MA,

United States,

zhaojun.li@wne.edu

1 - Assessing Failure Dependency In A Complex System

Rong Pan, Arizona State University,

Rong.Pan@asu.edu

,

Petek Yontay

In this talk we will discuss a Bayesian network model for assessing system

reliability of a complex system. Coupling with Bayesian inference methods , the

posterior distributions of the conditional probabilities in a BN model can be

estimated by combining failure information and expert opinions at both system

and component levels.

2 - A Multi-objective Approach For Multi-stage Reliability

Growth Planning By Considering The Timing Of New

Technologies Introduction

Steven LI, Western New England University,

zhaojun.li@wne.edu,

Mohammad Sadegh Mobin, Hossein Cheraghi

This paper proposes a new multi-stage reliability growth planning model which

optimizes and balances the product development cost, time, and the product

reliability. The number of test units, test time, and the percentage of introduced

new technologies are major decision variables. Different reliability growth rates

are considered for each sub-system in each stage. An integrated approach is

developed to optimize the problem, which starts with a multi-objective

evolutionary algorithm to find a set of Pareto optimal solutions followed by the

application of clustering tools to cluster the solutions. The clustered solutions are

further ranked using a multiple criteria decision making tool.

TD66