We are testing a new system for linking grants to scientists.
The funding information displayed below comes from the NIH Research Portfolio Online Reporting Tools and the NSF Award Database.The grant data on this page is limited to grants awarded in the United States and is thus partial. It can nonetheless be used to understand how funding patterns influence mentorship networks and vice-versa, which has deep implications on how research is done.
You can help! If you notice any innacuracies, please sign in and mark grants as correct or incorrect matches.
High-probability grants
According to our matching algorithm, Richard A. Olshen is the likely recipient of the following grants.Years | Recipients | Code | Title / Keywords | Matching score |
---|---|---|---|---|
1985 — 1987 | Olshen, Richard Allen | R01Activity Code Description: To support a discrete, specified, circumscribed project to be performed by the named investigator(s) in an area representing his or her specific interest and competencies. |
Biostatistics: Modeling and Inference @ University of California San Diego The research proposed involves five different topics (i) the extension of tree-structured, recursive partitioning methods to the analysis of censored survival data from clinical trials or laboratory experiments; (ii) aggregated Markov models (in neurophysiology) that are crucial to the understanding of the acetylcholine receptor; (iii) the application of multivariate methods to the understanding of ocular micro-circulation and the early diagnosis of diabetes; (iv) binary regression with errors in variables and its use in evaluating burn care facilities; (v) estimation in pharmacokinetic models which arise in high dose intracavitary chemotherapy as it is administered at the UC San Diego Cancer Center. |
0.954 |
1988 | Olshen, Richard Allen | R01Activity Code Description: To support a discrete, specified, circumscribed project to be performed by the named investigator(s) in an area representing his or her specific interest and competencies. |
Biostatistics: Modeling and Inference @ University of California San Diego The research we propose is all connected with extremely computer intensive statistics applied to cancer research, other medical problems and physiology. Olshen proposes studying problems in tree-structured methods and sample reuse techniques. In particular, he proposes study and implementation of a technique for producing unbiased estimates within the terminal nodes of a classification tree, and study of stratification and redesign of sample reuse methods. Investigations of asymptotic results on tree-structured methods that incorporate adaptiveness of splitting rules and rates of convergence for the algorithms are also proposed. Fredkin and Rice propose development of methodology for restoration of noisy single channel patch clamp data. A variety of issues arise in this regard: proper handling of filtering via deconvolution, data-based smoothing parameter estimates, extension of algorithms and computer code to handle multi conductance level channels, and special methods for dealing with flickering. Special attention will be given to the development of efficient computer code. Abramson proposes further study of a recursive technique for use in conditional analysis of models in which there are a large number of effects (for example, baseline measurements) that are not of primary interest and a small number of fixed effects that are. He proposes development of polished software for general distribution. |
0.954 |
1989 — 1990 | Olshen, Richard Allen | R01Activity Code Description: To support a discrete, specified, circumscribed project to be performed by the named investigator(s) in an area representing his or her specific interest and competencies. |
Biostatistics--Modeling and Inference @ University of California San Diego The research we propose is all connected with extremely computer intensive statistics applied to cancer research, other medical problems and physiology. Olshen proposes studying problems in tree-structured methods and sample reuse techniques. In particular, he proposes study and implementation of a technique for producing unbiased estimates within the terminal nodes of a classification tree, and study of stratification and redesign of sample reuse methods. Investigations of asymptotic results on tree-structured methods that incorporate adaptiveness of splitting rules and rates of convergence for the algorithms are also proposed. Fredkin and Rice propose development of methodology for restoration of noisy single channel patch clamp data. A variety of issues arise in this regard: proper handling of filtering via deconvolution, data-based smoothing parameter estimates, extension of algorithms and computer code to handle multi conductance level channels, and special methods for dealing with flickering. Special attention will be given to the development of efficient computer code. Abramson proposes further study of a recursive technique for use in conditional analysis of models in which there are a large number of effects (for example, baseline measurements) that are not of primary interest and a small number of fixed effects that are. He proposes development of polished software for general distribution. |
0.954 |
1991 — 1994 | Gray, Robert [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Image Compression Using Vector Quantization and Decision Trees @ Stanford University Research will be conducted on variable rate tree-structured vector quantizers for image compression applications. The emphasis will be on tree-structured codes designed using extensions and variations of the tree design techniques of Breiman, Friedman, Olshen and Stone, but a variety of other vector quantization structures will be considered in combination with the tree- structured codes. In particular, both finite state and predictive vector quantizers will be considered for incorporating memory in the coding and two-dimensional subsampling techniques will be considered as a means of increasing effective vector size, improving prediction accuracy and providing a natural data structure. The dual use of trees for classification and compression in finite-state vector quantizers will be explored, as will be the combination of compression of image sequences such as video and multimodal images of a common object or view, as in multispectral and color imaging. Experiments will be conducted with medical images, computer image data consisting of mixed video, imagery, and graphics, and satellite imagery, especially multispectral. |
0.915 |
1991 — 1995 | Friedman, Jerome Gray, Robert (co-PI) [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Tree-Structured Statistical Methods @ Stanford University The Principal Investigator will study statistical methods. These involve the optimal choices of sequences of binary questions for problems of classification, probability class estimation, regression, and survival analysis. Much of the computer- intensive research that is planned involves new capabilities and significant improvements of the algorithms and techniques given by Breiman, Friedman, Olshen, and Stone in Classification and Regression Trees and its associated CART software. Some computations will be done in an innovative distributed mode on a system of workstations. The investigators plan applications to test data sets, particularly a large data set concerning the quick diagnosis of heart attack. Details are given for an application to the enhancement of certain magnetic resonance images that are important to making prognoses for patients with lymphoma. (Support is not requested for the conduct of clinical trials; indeed, the data do not arise from conventional clinical trials.) The statistics research involved is of importance to a wide variety of applications, some of them medical. |
1 |
1993 — 1998 | Gray, Robert [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Tree-Structured Image Compression and Classification @ Stanford University Gray Tree-structured vector quantization is an approach to image compression that applies ideas from statistical clustering algorithms and tree-structured classification and regression algorithms to produce compression codes that trade off bit rate and average distortion in a near optimal fashion. This research is examining the explicit combination of these two forms of signal processing, compression and classification, into single tree-structured algorithms that permit a trade off between traditional distortion measures, such as squared error, with measures of classification accuracy such as Bayes risk. The intent is to produce codes with implicit classification information, that is, for which the stored or communicated compressed image incorporates classification information without further signal processing. Such systems can provide direct low level classification or provide an efficient front end to more sophisticated full-frame recognition algorithms. Vector quanitization algorithms for relatively large block sizes are also being developed with an emphasis on multiresolution compression algorithms. In order to improve the promising performance found in preliminary studies or combined compression and classification, it will be necessary to use larger block sizes or, equivalently, more context. Multiresolution or hierarchial quantizers provide a simple and effective means of accomplishing this. Other related issues are being explored, including improved prediction methods for predictive vector quantization and image sequence coding. |
0.915 |
1997 — 2000 | Gray, Robert [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Compression, Classification and Image Segmentation @ Stanford University Data compression aims at producing an efficient representation for storage or transmission in order to communicate or store the smallest possible number of bits while retaining high quality in the reconstructed image. Statistical classification and segmentation are both concerned with labeling parts of an image as being of a certain type or class, such as tumor or normal tissue in a medical image or text, graphics, or photographs in a document page. Classification and segmentation can be used to assist human users of images, as in highlighting suspicious tissue in a medical image or defects in an imaged circuit, or they can be used for the automatic extraction of features in intelligent browsing of large databases. All three operations have in common the goal of efficiently representing an image as a whole, by optimally trading off quality and cost or distortion, an din decomposing the image into useful components. Most of the systems involving these operations, however, focus on each separate operation rather than on combining them into common algorithms with multiple goals. This project is devoted to the development of theory and algorithms for jointly performing these operations and related signal processing such as the estimation of probability distributions or models from observed data. The approach involves optimizing single combined systems subject to possibly conflicting combined goals, such as maximizing signal-to-noise ratios and minimizing bit rates and Bayes risk. Theory will be reinforced by simulations for both artificial sources, where theoretical performance bounds can be used for comparison, and real-world examples drawn from medical and document applications. |
0.915 |
2000 — 2004 | Gray, Robert [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Gauss Mixture Quantization For Image Compression and Segmentation @ Stanford University The research is concerned with techniques from statistical signal processing and information theory as they apply to communication systems with multiple goals. Such systems arise in multimedia communications networks like the Internet. The decomposition of data streams into different types is critical to finding information desired by a user among vast available sources, and it can also provide methods for displaying, rendering, printing, or playing the received signal that take advantage of its particular structure. Signal processing and coding theory have provided powerful mathematical models of information sources and algorithms by which these sources can be communicated and processed. Typically systems are designed as a collection of separate, unrelated, components. This can result in much less than optimal overall performance. Furthermore, it can hamper theoretical understanding of the fundamental limits on achievable performance.We treat the simultaneous design of mathematical models that account at once for information sources, data compression, and signal processing and apply to extracting information from the received data. Our emphasis is on image communication and processing. Because, the techniques draw heavily from demonstrably successful methods in speech coding and recognition they are natural for both signal types, individually or together. |
0.915 |
2003 — 2007 | Gray, Robert [⬀] Olshen, Richard |
N/AActivity Code Description: No activity code was retrieved: click on the grant title for more information |
Quantization For Signal Compression, Classification, and Mixture Modeling @ Stanford University Abstract |
0.915 |
2011 — 2014 | Olshen, Richard Allen | U19Activity Code Description: To support a research program of multiple projects directed toward a specific major objective, basic theme or program goal, requiring a broadly based, multidisciplinary and often long-term approach. A cooperative agreement research program generally involves the organized efforts of large groups, members of which are conducting research projects designed to elucidate the various aspects of a specific objective. Substantial Federal programmatic staff involvement is intended to assist investigators during performance of the research activities, as defined in the terms and conditions of award. The investigators have primary authorities and responsibilities to define research objectives and approaches, and to plan, conduct, analyze, and publish results, interpretations and conclusions of their studies. Each research project is usually under the leadership of an established investigator in an area representing his/her special interest and competencies. Each project supported through this mechanism should contribute to or be directly related to the common theme of the total research effort. The award can provide support for certain basic shared resources, including clinical components, which facilitate the total research effort. These scientifically meritorious projects should demonstrate an essential element of unity and interdependence. |
@ Stanford University The Biostatistics Core of this project will be led by Professor Richard Olshen, who will be joined in supportive efforts by Professors Bradley Efron and Lu Tian. There will be high level statistical consulting on all projects and all investigators of SHIMR by these investigators. All data made available to these individuals will be anonymized compliant with HIPPA rules. Open source computer programs written in the popular R language http://www.r-project.org/ will be made available to SHIMR investigators. Stanford's Data Coordinating Center (DCC) is the umbrella organization that will supervise writing these open source computer programs. In most instances the programs will call existing routines, available for downloading from CRAN http://cran.r proiect.org/, though occasionally we will create the ingredient routines ourselves. |
1 |
2012 — 2016 | Olshen, Richard Allen Sabatti, Chiara (co-PI) [⬀] Tibshirani, Rob |
T32Activity Code Description: To enable institutions to make National Research Service Awards to individuals selected by them for predoctoral and postdoctoral research training in specified shortage areas. |
Interdisciplinary Training Program in Biostatistics and Personalized Medicine @ Stanford University DESCRIPTION (provided by applicant): This proposal outlines a new pre-doctoral training program in biostatistics that emphasizes applications to genomics and personalized medicine. Primary faculty members are from the Departments of Health Research & Policy and Statistics at Stanford University. A group of affiliate faculty from biochemistry, genetics, cardiology and other areas provide the necessary breadth for interdisciplinary research. The trainees in the program pursue a PhD in the Department of Statistics, with a concentration in Biostatistics; this option currently not available at Stanford. In addition to fulfilling the requirements of the PhD i Statistics, the proposed program will include additional course work in biomedical sciences, mentored research experienced in a collaborative setting, and training in the responsible conduct of research. |
1 |