|Registration Deadline:||March 29, 2001 over 13 years ago|
|To apply for Funding you must register by:||December 19, 2000 about 14 years ago|
- Ingrid Daubechies (Duke University)
Nonlinear Estimation and Classification Schedule For a tentative list of invited speakers and instructions on submitting a talk to the contributed portion of the program, please consult the main page of the conference:
Contributed presentations (talks and posters) are invited for original work related to the theme of the workshop. Submissions are to be in the form of an extended abstract, consisting of not more than 3 pages. Graduate students, new researchers, minorities and women are strongly encouraged to apply. Extended abstracts will be accepted in PostScript or PDF formats and should be mailed to email@example.com. When submitting papers, please also indicate whether travel funds are necessary for attendance. The deadline for extended abstracts is November 27, 2000. Notification of acceptance will be sent out on January 15, 2000.
Overview Researchers in many disciplines face the formidable task of analyzing massive amounts of high-dimensional and highly-structured data. This is due in part to recent advances in data collection and computing technologies. As a result, fundamental statistical research is being undertaken in a variety of different fields. Driven by the complexity of these new problems, and fueled by the explosion of available computer power, highly adaptive, non-linear procedures are now essential components of modern "data analysis," a term that we liberally interpret to include speech and pattern recognition, classification, data compression and signal processing. The development of new, flexible methods combines advances from many sources, including approximation theory, numerical analysis, machine learning, signal processing and statistics. The proposed workshop intends to bring together eminent experts from these fields in order to exchange ideas and forge directions for the future. It also intends to introduce the research topics to graduate students by providing travel support and by requiring the last speaker of each session to give an overview of the field. A Brief Survey of the Area Three ingredients are common to the new class of non-linear methods:
- Approximation spaces (wavelets, splines, neural networks, trees) or other simple probabilistic structures as building blocks for representations of statistical objects;
- Algorithms for identifying the "best" representation or combining several "promising" candidates; and
- Statistical frameworks to judge between competing representations.
A key component in recent non-linear modeling efforts is the design of efficient representations for (complex) objects such as curves, surfaces and images. Here, efficiency often relates to the number of descriptors (i.e. basis functions or general predictors). This definition has parallels with recent results in approximation theory on n-term expansions and best basis algorithms. Similarly, in machine learning, the focus is on repeatedly applying weak-learners or relatively simple representations like the sigmoid function used in (artificial) neural networks or very shallow trees, known as stumps. On the other hand, new tools for image and texture analysis draw their representations from the human visual system, where efficiency might be evaluated in biological terms. Finally, graphical or hidden Markov models represent a distribution involving complex dependencies by connecting simple (conjugate) distributions in a hierarchical manner. In statistical problems, the search for the single "best" representation among a (potentially vast) number of competitors is really a problem in model selection. Applying these ideas in the context of approximation spaces, for example, has led to new wavelet shrinkage schemes and new methods for free-knot splines. Various selection criteria have been suggested in these problems, each with differing strengths and weaknesses, but none dominating all the others over a wide range of problems. Even with an agreed selection criterion, difficulties can occur in identifying the "best" representation as it is often a single element in an extremely dense candidate set which cannot be exhaustively searched. Efficient optimization methods need to be devised to overcome this problem. Rather than focusing on a single model, recent empirical and theoretical results have shown that we can achieve improved predictive performance by combining several competing representations. The Bayesian paradigm naturally leads to a framework for averaging models, and has been applied extensively in the context of approximation spaces. From a theoretical standpoint, these techniques depend quite strongly on the underlying prior assumptions, and understanding their performance is an area of active research. Other techniques for combining models have arisen in the machine learning literature, most notably boosting and bagging, and have proved to be extremely effective in practice. When working with massive amounts of data, a test or hold-out data set has been commonly used for judging competing procedures. The question of which evaluation criterion or loss function to use still remains. Some loss functions are more sensible than others in a particular problem, although the squared loss has been the most conventional. With the current computing power, it is believed possible to use more sensible loss functions, for example, the absolute loss for image processing. This, however, needs a consented effort of the research community to deviate from the norm. The proposed workshop is an ideal place to reach such a consensus. Organizing Committee
- David Denison (Imperial College)
- Mark Hansen (Bell Labs)
- Chris Holmes (Imperial College)
- Robert Kohn (Univ. of New South Wales)
- Bani Mallick (Texas A&M)
- Martin Tanner (Northwestern)
- Bin Yu (UC Berkeley)
FUNDING APPLICATION DEADLINE: February 5, 2001 Funded in part by the National Security Agency and the Department of the Army Research Office.
To apply for funding, you must register by the funding application deadline displayed above.
Students, recent Ph.D.'s, women, and members of underrepresented minorities are particularly encouraged to apply. Funding awards are typically made 6 weeks before the workshop begins. Requests received after the funding deadline are considered only if additional funds become available.
MSRI has preferred rates at the Rose Garden Inn, depending on room availability. Reservations may be made by calling 1-800-992-9005 OR directly on their website. Click on Corporate at the bottom of the screen and when prompted enter code MATH (this code is not case sensitive). By using this code a new calendar will appear and will show the MSRI rate on all room types available.
MSRI has preferred rates at the Hotel Durant. Reservations may be made by calling 1-800-238-7268. When making reservations, guests must request the MSRI preferred rate. If you are making your reservations on line, please go to this link and enter the promo/corporate code 123MSRI. Our preferred rate is $139 per night for a Deluxe Queen/King, based on availability.
MSRI has preferred rates of $149 - $189 plus tax at the Hotel Shattuck Plaza, depending on room availability. Guests can either call the hotel's main line at 510-845-7300 and ask for the MSRI- Mathematical Science Research Inst. discount; or go to www.hotelshattuckplaza.com and click Book Now. Once on the reservation page, click “Promo/Corporate Code“ and input the code: msri.
MSRI has preferred rates of $110 - $140 at the Berkeley Lab Guest House, depending on room availability. Reservations may be made by calling 510-495-8000 or directly on their website. Select “I am an individual traveler affiliated with MSRI”.
Additional lodging options may be found on our short term housing page.