ALBERT

All Library Books, journals and Electronic Records Telegrafenberg

Your email was sent successfully. Check your inbox.

An error occurred while sending the email. Please try again.

Proceed reservation?

Export
Filter
  • Articles  (75)
  • Computational Methods, Massively Parallel (Deep) Sequencing, Genomics  (31)
  • Agribusiness  (25)
  • Gravity, Geodesy and Tides  (19)
  • Oxford University Press  (75)
  • 2010-2014  (75)
  • 1980-1984
  • 1950-1954
  • 1
    Publication Date: 2014-12-17
    Description: Measurements of ground deformation can be used to identify and interpret geophysical processes occurring at volcanoes. Most studies rely on a single geodetic technique, or fit a geophysical model to the results of multiple geodetic techniques. Here we present a methodology that combines GPS, Total Station measurements and InSAR into a single reference frame to produce an integrated 3-D geodetic velocity surface without any prior geophysical assumptions. The methodology consists of five steps: design of the network, acquisition and processing of the data, spatial integration of the measurements, time series computation and finally the integration of spatial and temporal measurements. The most significant improvements of this method are (1) the reduction of the required field time, (2) the unambiguous detection of outliers, (3) an increased measurement accuracy and (4) the construction of a 3-D geodetic velocity field. We apply this methodology to ongoing motion on Arenal's western flank. Integration of multiple measurement techniques at Arenal volcano revealed a deformation field that is more complex than that described by individual geodetic techniques, yet remains consistent with previous studies. This approach can be applied to volcano monitoring worldwide and has the potential to be extended to incorporate other geodetic techniques and to study transient deformation.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 2
    Publication Date: 2014-11-09
    Description: In autumn 2012, the new release 05 (RL05) of monthly geopotencial spherical harmonics Stokes coefficients (SC) from Gravity Recovery and Climate Experiment (GRACE) mission was published. This release reduces the noise in high degree and order SC, but they still need to be filtered. One of the most common filtering processing is the combination of decorrelation and Gaussian filters. Both of them are parameters dependent and must be tuned by the users. Previous studies have analyzed the parameters choice for the RL05 GRACE data for oceanic applications, and for RL04 data for global application. This study updates the latter for RL05 data extending the statistics analysis. The choice of the parameters of the decorrelation filter has been optimized to: (1) balance the noise reduction and the geophysical signal attenuation produced by the filtering process; (2) minimize the differences between GRACE and model-based data and (3) maximize the ratio of variability between continents and oceans. The Gaussian filter has been optimized following the latter criteria. Besides, an anisotropic filter, the fan filter, has been analyzed as an alternative to the Gauss filter, producing better statistics.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 3
    Publication Date: 2012-12-28
    Description: We investigate the impact of decentralised decision-making on product quality. Comparing a cooperative and an investor-owned firm suggests that members of the cooperative have an incentive to produce too much and to free-ride on quality. Whether or not cooperatives deliver higher quality products depends on the way in which the quality of the final product is determined from the quality levels of the inputs delivered (quality aggregation) as well as the number of members of the cooperative. Empirical evidence on the Austrian wine market suggests that wines produced by cooperatives tend to be of significantly lower quality, ceteris paribus .
    Keywords: D22 - Firm Behavior: Empirical Analysis, D23 - Organizational Behavior ; Transaction Costs ; Property Rights, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 4
    Publication Date: 2013-02-26
    Description: The survival of agricultural marketing co-operatives depends on their capability of satisfying and maintaining their base of farmer members. Hypotheses regarding these two success factors are developed in neoclassical economics and transaction cost economics. They are tested with a survey of 321 members of marketing co-operatives specialising in fresh fruits and vegetables. Our results show support for both perspectives. Price paid to farmers is important for their satisfaction with the co-operative. Farmers' perceptions of transaction costs are even more important.
    Keywords: D22 - Firm Behavior: Empirical Analysis, D23 - Organizational Behavior ; Transaction Costs ; Property Rights, P13 - Cooperative Enterprises, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 5
    Publication Date: 2013-02-26
    Description: This paper explores the demand and willingness to pay (WTP) for value-added services to chicken. Since the demand for such services are likely to be highly segmented and often applies only to a market niche, models based on assumptions of homogeneity among consumers are likely to be inappropriate. For this reason, this paper combines discrete and continuous mixing distributions to concurrently identify the size of the niche market and the heterogeneity among consumers within the market niche. Failing to account for the niche market nature of value-added services is shown to have implications for predictions of WTP, demand and total revenue.
    Keywords: C25 - Discrete Regression and Qualitative Choice Models, D12 - Consumer Economics: Empirical Analysis, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 6
    Publication Date: 2012-10-10
    Description: Several bioinformatics methods have been proposed for the detection and characterization of genomic structural variation (SV) from ultra high-throughput genome resequencing data. Recent surveys show that comprehensive detection of SV events of different types between an individual resequenced genome and a reference sequence is best achieved through the combination of methods based on different principles (split mapping, reassembly, read depth, insert size, etc.). The improvement of individual predictors is thus an important objective. In this study, we propose a new method that combines deviations from expected library insert sizes and additional information from local patterns of read mapping and uses supervised learning to predict the position and nature of structural variants. We show that our approach provides greatly increased sensitivity with respect to other tools based on paired end read mapping at no cost in specificity, and it makes reliable predictions of very short insertions and deletions in repetitive and low-complexity genomic contexts that can confound tools based on split mapping of reads.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 7
    Publication Date: 2012-04-15
    Description: Exome sequencing strategy is promising for finding novel mutations of human monogenic disorders. However, pinpointing the casual mutation in a small number of samples is still a big challenge. Here, we propose a three-level filtration and prioritization framework to identify the casual mutation(s) in exome sequencing studies. This efficient and comprehensive framework successfully narrowed down whole exome variants to very small numbers of candidate variants in the proof-of-concept examples. The proposed framework, implemented in a user-friendly software package, named KGGSeq ( http://statgenpro.psychiatry.hku.hk/kggseq ), will play a very useful role in exome sequencing-based discovery of human Mendelian disease genes.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 8
    Publication Date: 2012-07-22
    Description: Small RNAs (sRNAs) are a class of short (20–25 nt) non-coding RNAs that play important regulatory roles in gene expression. An essential first step in understanding their function is to confidently identify sRNA targets. In plants, several classes of sRNAs such as microRNAs (miRNAs) and trans-acting small interfering RNAs have been shown to bind with near-perfect complementarity to their messenger RNA (mRNA) targets, generally leading to cleavage of the mRNA. Recently, a high-throughput technique known as Parallel Analysis of RNA Ends (PARE) has made it possible to sequence mRNA cleavage products on a large-scale. Computational methods now exist to use these data to find targets of conserved and newly identified miRNAs. Due to speed limitations such methods rely on the user knowing which sRNA sequences are likely to target a transcript. By limiting the search to a tiny subset of sRNAs it is likely that many other sRNA/mRNA interactions will be missed. Here, we describe a new software tool called PAREsnip that allows users to search for potential targets of all sRNAs obtained from high-throughput sequencing experiments. By searching for targets of a complete ‘sRNAome’ we can facilitate large-scale identification of sRNA targets, allowing us to discover regulatory interaction networks.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 9
    Publication Date: 2012-09-13
    Description: Gene fusions are common driver events in leukaemias and solid tumours; here we present FusionAnalyser, a tool dedicated to the identification of driver fusion rearrangements in human cancer through the analysis of paired-end high-throughput transcriptome sequencing data. We initially tested FusionAnalyser by using a set of in silico randomly generated sequencing data from 20 known human translocations occurring in cancer and subsequently using transcriptome data from three chronic and three acute myeloid leukaemia samples. in all the cases our tool was invariably able to detect the presence of the correct driver fusion event(s) with high specificity. In one of the acute myeloid leukaemia samples, FusionAnalyser identified a novel, cryptic, in-frame ETS2–ERG fusion. A fully event-driven graphical interface and a flexible filtering system allow complex analyses to be run in the absence of any a priori programming or scripting knowledge. Therefore, we propose FusionAnalyser as an efficient and robust graphical tool for the identification of functional rearrangements in the context of high-throughput transcriptome sequencing data.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 10
    Publication Date: 2012-09-13
    Description: The use of a priori knowledge in the alignment of targeted sequencing data is investigated using computational experiments. Adapting a Needleman–Wunsch algorithm to incorporate the genomic position information from the targeted capture, we demonstrate that alignment can be done to just the target region of interest. When in addition use is made of direct string comparison, an improvement of up to a factor of 8 in alignment speed compared to the fastest conventional aligner (Bowtie) is obtained. This results in a total alignment time in targeted sequencing of around 7 min for aligning approximately 56 million captured reads. For conventional aligners such as Bowtie, BWA or MAQ, alignment to just the target region is not feasible as experiments show that this leads to an additional 88% SNP calls, the vast majority of which are false positives (~92%).
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 11
    Publication Date: 2012-06-28
    Description: We introduce Grinder ( http://sourceforge.net/projects/biogrinder/ ), an open-source bioinformatic tool to simulate amplicon and shotgun (genomic, metagenomic, transcriptomic and metatranscriptomic) datasets from reference sequences. This is the first tool to simulate amplicon datasets (e.g. 16S rRNA) widely used by microbial ecologists. Grinder can create sequence libraries with a specific community structure, α and β diversities and experimental biases (e.g. chimeras, gene copy number variation) for commonly used sequencing platforms. This versatility allows the creation of simple to complex read datasets necessary for hypothesis testing when developing bioinformatic software, benchmarking existing tools or designing sequence-based experiments. Grinder is particularly useful for simulating clinical or environmental microbial communities and complements the use of in vitro mock communities.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 12
    Publication Date: 2012-06-06
    Description: The most crucial step in data processing from high-throughput sequencing applications is the accurate and sensitive alignment of the sequencing reads to reference genomes or transcriptomes. The accurate detection of insertions and deletions (indels) and errors introduced by the sequencing platform or by misreading of modified nucleotides is essential for the quantitative processing of the RNA-based sequencing (RNA-Seq) datasets and for the identification of genetic variations and modification patterns. We developed a new, fast and accurate algorithm for nucleic acid sequence analysis, FANSe, with adjustable mismatch allowance settings and ability to handle indels to accurately and quantitatively map millions of reads to small or large reference genomes. It is a seed-based algorithm which uses the whole read information for mapping and high sensitivity and low ambiguity are achieved by using short and non-overlapping reads. Furthermore, FANSe uses hotspot score to prioritize the processing of highly possible matches and implements modified Smith–Watermann refinement with reduced scoring matrix to accelerate the calculation without compromising its sensitivity. The FANSe algorithm stably processes datasets from various sequencing platforms, masked or unmasked and small or large genomes. It shows a remarkable coverage of low-abundance mRNAs which is important for quantitative processing of RNA-Seq datasets.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 13
    Publication Date: 2012-04-24
    Description: Ultra-deep RNA sequencing has become a powerful approach for genome-wide analysis of pre-mRNA alternative splicing. We develop MATS (multivariate analysis of transcript splicing), a Bayesian statistical framework for flexible hypothesis testing of differential alternative splicing patterns on RNA-Seq data. MATS uses a multivariate uniform prior to model the between-sample correlation in exon splicing patterns, and a Markov chain Monte Carlo (MCMC) method coupled with a simulation-based adaptive sampling procedure to calculate the P -value and false discovery rate (FDR) of differential alternative splicing. Importantly, the MATS approach is applicable to almost any type of null hypotheses of interest, providing the flexibility to identify differential alternative splicing events that match a given user-defined pattern. We evaluated the performance of MATS using simulated and real RNA-Seq data sets. In the RNA-Seq analysis of alternative splicing events regulated by the epithelial-specific splicing factor ESRP1, we obtained a high RT–PCR validation rate of 86% for differential exon skipping events with a MATS FDR of 〈10%. Additionally, over the full list of RT–PCR tested exons, the MATS FDR estimates matched well with the experimental validation rate. Our results demonstrate that MATS is an effective and flexible approach for detecting differential alternative splicing from RNA-Seq data.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 14
    Publication Date: 2012-05-13
    Description: Numerous algorithms have been developed to analyze ChIP-Seq data. However, the complexity of analyzing diverse patterns of ChIP-Seq signals, especially for epigenetic marks, still calls for the development of new algorithms and objective comparisons of existing methods. We developed Qeseq, an algorithm to detect regions of increased ChIP read density relative to background. Qeseq employs critical novel elements, such as iterative recalibration and neighbor joining of reads to identify enriched regions of any length. To objectively assess its performance relative to other 14 ChIP-Seq peak finders, we designed a novel protocol based on Validation Discriminant Analysis (VDA) to optimally select validation sites and generated two validation datasets, which are the most comprehensive to date for algorithmic benchmarking of key epigenetic marks. In addition, we systematically explored a total of 315 diverse parameter configurations from these algorithms and found that typically optimal parameters in one dataset do not generalize to other datasets. Nevertheless, default parameters show the most stable performance, suggesting that they should be used. This study also provides a reproducible and generalizable methodology for unbiased comparative analysis of high-throughput sequencing tools that can facilitate future algorithmic development.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 15
    Publication Date: 2012-05-13
    Description: The informational content of RNA sequencing is currently far from being completely explored. Most of the analyses focus on processing tables of counts or finding isoform deconvolution via exon junctions. This article presents a comparison of several techniques that can be used to estimate differential expression of exons or small genomic regions of expression, based on their coverage function shapes. The problem is defined as finding the differentially expressed exons between two samples using local expression profile normalization and statistical measures to spot the differences between two profile shapes. Initial experiments have been done using synthetic data, and real data modified with synthetically created differential patterns. Then, 160 pipelines (5 types of generator x 4 normalizations x 8 difference measures) are compared. As a result, the best analysis pipelines are selected based on linearity of the differential expression estimation and the area under the ROC curve. These platform-independent techniques have been implemented in the Bioconductor package rnaSeqMap. They point out the exons with differential expression or internal splicing, even if the counts of reads may not show this. The areas of application include significant difference searches, splicing identification algorithms and finding suitable regions for QPCR primers.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 16
    Publication Date: 2012-05-13
    Description: The rapid expansion in the quantity and quality of RNA-Seq data requires the development of sophisticated high-performance bioinformatics tools capable of rapidly transforming this data into meaningful information that is easily interpretable by biologists. Currently available analysis tools are often not easily installed by the general biologist and most of them lack inherent parallel processing capabilities widely recognized as an essential feature of next-generation bioinformatics tools. We present here a user-friendly and fully automated R NA- S eq a nalysis p ipeline (R-SAP) with built-in multi-threading capability to analyze and quantitate high-throughput RNA-Seq datasets. R-SAP follows a hierarchical decision making procedure to accurately characterize various classes of transcripts and achieves a near linear decrease in data processing time as a result of increased multi-threading. In addition, RNA expression level estimates obtained using R-SAP display high concordance with levels measured by microarrays.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 17
    Publication Date: 2012-05-23
    Description: Deciphering the structure of gene regulatory networks across the tree of life remains one of the major challenges in postgenomic biology. We present a novel ChIP-seq workflow for the archaea using the model organism Halobacterium salinarum sp. NRC-1 and demonstrate its application for mapping the genome-wide binding sites of natively expressed transcription factors. This end-to-end pipeline is the first protocol for ChIP-seq in archaea, with methods and tools for each stage from gene tagging to data analysis and biological discovery. Genome-wide binding sites for transcription factors with many binding sites (TfbD) are identified with sensitivity, while retaining specificity in the identification the smaller regulons (bacteriorhodopsin-activator protein). Chromosomal tagging of target proteins with a compact epitope facilitates a standardized and cost-effective workflow that is compatible with high-throughput immunoprecipitation of natively expressed transcription factors. The Pique package, an open-source bioinformatics method, is presented for identification of binding events. Relative to ChIP-Chip and qPCR, this workflow offers a robust catalog of protein–DNA binding events with improved spatial resolution and significantly decreased cost. While this study focuses on the application of ChIP-seq in H. salinarum sp. NRC-1, our workflow can also be adapted for use in other archaea and bacteria with basic genetic tools.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 18
    Publication Date: 2012-05-23
    Description: A flexible statistical framework is developed for the analysis of read counts from RNA-Seq gene expression studies. It provides the ability to analyse complex experiments involving multiple treatment conditions and blocking variables while still taking full account of biological variation. Biological variation between RNA samples is estimated separately from the technical variation associated with sequencing technologies. Novel empirical Bayes methods allow each gene to have its own specific variability, even when there are relatively few biological replicates from which to estimate such variability. The pipeline is implemented in the edgeR package of the Bioconductor project. A case study analysis of carcinoma data demonstrates the ability of generalized linear model methods (GLMs) to detect differential expression in a paired design, and even to detect tumour-specific expression changes. The case study demonstrates the need to allow for gene-specific variability, rather than assuming a common dispersion across genes or a fixed relationship between abundance and variability. Genewise dispersions de-prioritize genes with inconsistent results and allow the main analysis to focus on changes that are consistent between biological replicates. Parallel computational approaches are developed to make non-linear model fitting faster and more reliable, making the application of GLMs to genomic data more convenient and practical. Simulations demonstrate the ability of adjusted profile likelihood estimators to return accurate estimators of biological variability in complex situations. When variation is gene-specific, empirical Bayes estimators provide an advantageous compromise between the extremes of assuming common dispersion or separate genewise dispersion. The methods developed here can also be applied to count data arising from DNA-Seq applications, including ChIP-Seq for epigenetic marks and DNA methylation analyses.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 19
    Publication Date: 2011-11-24
    Description: This paper presents an empirical investigation of the link between intangible expenses of French wine companies and their financial performance. A flexible moment-based approach is used to analyse the impact of tangible and intangible expenses on the mean, variance and skewness of profit. Econometric evidence shows that a high level of intangible expenses has a positive impact on performance by increasing the expected profit and reducing variance risk. A lower level of intangible expenses reduces risk and mean of profit of corporations. This study provides insights on the use of intangible expenses as a risk management tool.
    Keywords: G32 - Financing Policy ; Financial Risk and Risk Management ; Capital and Ownership Structure, Q12 - Micro Analysis of Farm Firms, Farm Households, and Farm Input Markets, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 20
    Publication Date: 2011-11-24
    Description: Geographical origin labels are important information and marketing tools and have recently become a central component of EU agricultural promotion. We consider demand in a non-EU export market for two distinct label types: country of origin (COO) and geographical indications (GIs). Additionally, two types of GIs, ‘protected designations of origin’ (PDOs) and ‘protected GIs’ (PGIs) are considered. Empirical findings indicate consumers’ willingness to pay varies with the oil's COO and is greater for GIs than for non-GIs from a given country. Weaker evidence that consumers value PDOs more than PGIs is also found.
    Keywords: C25 - Discrete Regression and Qualitative Choice Models, D12 - Consumer Economics: Empirical Analysis, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 21
    Publication Date: 2012-03-08
    Description: We determine the circumstances when the absence of public listing, often believed to be a disadvantage, makes a cooperative the unique efficient governance structure. This is established in a multi-task principal–agent model, capturing that cooperatives are not publicly listed and their CEOs have to bring the downstream enterprise to value as well as to serve upstream member interests. Not having a public listing prevents the CEO from choosing the level of the downstream activities too high. Cooperatives are uniquely efficient when the upstream marginal product multiplied with a function increasing in the strength of the chain complementarities is higher than the downstream marginal product.
    Keywords: D21 - Firm Behavior, L23 - Organization of Production, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 22
    Publication Date: 2012-03-08
    Description: Branded food manufacturers vindicate the use of excess production capacities to justify their production of retailers’ brands. We study the distributor's and food manufacturer's private label (PL) strategy for production within a framework featuring endogenous store brand quality, bargaining power, possible differences in production technology and potential capacity constraints for the branded manufacturer. Depending on the structure of capacity constraint (applying to both products or to the PL only), we find that the retailer may prefer to choose an independent firm for the production of the store brand whereas the branded manufacturer is chosen in the case of excess capacity.
    Keywords: L11 - Production, Pricing, and Market Structure ; Size Distribution of Firms, L13 - Oligopoly and Other Imperfect Markets, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 23
    Publication Date: 2012-02-28
    Description: ChIP-seq is increasingly used to characterize transcription factor binding and chromatin marks at a genomic scale. Various tools are now available to extract binding motifs from peak data sets. However, most approaches are only available as command-line programs, or via a website but with size restrictions. We present peak-motifs , a computational pipeline that discovers motifs in peak sequences, compares them with databases, exports putative binding sites for visualization in the UCSC genome browser and generates an extensive report suited for both naive and expert users. It relies on time- and memory-efficient algorithms enabling the treatment of several thousand peaks within minutes. Regarding time efficiency, peak-motifs outperforms all comparable tools by several orders of magnitude. We demonstrate its accuracy by analyzing data sets ranging from 4000 to 1 28 000 peaks for 12 embryonic stem cell-specific transcription factors. In all cases, the program finds the expected motifs and returns additional motifs potentially bound by cofactors. We further apply peak-motifs to discover tissue-specific motifs in peak collections for the p300 transcriptional co-activator. To our knowledge, peak-motifs is the only tool that performs a complete motif analysis and offers a user-friendly web interface without any restriction on sequence size or number of peaks.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 24
    Publication Date: 2014-03-13
    Description: Genetic disorders can be detected by prenatal diagnosis using Chorionic Villus Sampling, but the 1:100 chance to result in miscarriage restricts the use to fetuses that are suspected to have an aberration. Detection of trisomy 21 cases noninvasively is now possible owing to the upswing of next-generation sequencing (NGS) because a small percentage of fetal DNA is present in maternal plasma. However, detecting other trisomies and smaller aberrations can only be realized using high-coverage NGS, making it too expensive for routine practice. We present a method, WISECONDOR (WIthin-SamplE COpy Number aberration DetectOR), which detects small aberrations using low-coverage NGS. The increased detection resolution was achieved by comparing read counts within the tested sample of each genomic region with regions on other chromosomes that behave similarly in control samples. This within-sample comparison avoids the need to re-sequence control samples. WISECONDOR correctly identified all T13, T18 and T21 cases while coverages were as low as 0.15–1.66. No false positives were identified. Moreover, WISECONDOR also identified smaller aberrations, down to 20 Mb, such as del(13)(q12.3q14.3), +i(12)(p10) and i(18)(q10). This shows that prevalent fetal copy number aberrations can be detected accurately and affordably by shallow sequencing maternal plasma. WISECONDOR is available at bioinformatics.tudelft.nl/wisecondor.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 25
    Publication Date: 2014-05-01
    Description: Alternative splicing is the main mechanism governing protein diversity. The recent developments in RNA-Seq technology have enabled the study of the global impact and regulation of this biological process. However, the lack of standardized protocols constitutes a major bottleneck in the analysis of alternative splicing. This is particularly important for the identification of exon–exon junctions, which is a critical step in any analysis workflow. Here we performed a systematic benchmarking of alignment tools to dissect the impact of design and method on the mapping, detection and quantification of splice junctions from multi-exon reads. Accordingly, we devised a novel pipeline based on TopHat2 combined with a splice junction detection algorithm, which we have named FineSplice. FineSplice allows effective elimination of spurious junction hits arising from artefactual alignments, achieving up to 99% precision in both real and simulated data sets and yielding superior F 1 scores under most tested conditions. The proposed strategy conjugates an efficient mapping solution with a semi-supervised anomaly detection scheme to filter out false positives and allows reliable estimation of expressed junctions from the alignment output. Ultimately this provides more accurate information to identify meaningful splicing patterns. FineSplice is freely available at https://sourceforge.net/p/finesplice/ .
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 26
    Publication Date: 2014-03-21
    Description: This is the first paper to analyze liquidity costs in agricultural futures markets based on the observed bid-ask spread (BAS) faced by market participants. The results reveal a highly liquid corn market that mostly offers order execution at minimum cost. The BAS responds negatively to volume and positively to price volatility, but also affects volume traded and price volatility. While statistically significant, these responses on a cents/bushel or a percentage basis are generally small. Liquidity costs are also virtually impervious to short-term changes in demand for spreading and trend-following trader activity, as well as differences from day-of-the-week changes in market activity. Much larger cents/bushel and percentage changes in BAS occur during commodity index trader roll periods and on USDA report release days. The roll period findings indicate a sunshine trading effect, while announcement effects identify the importance of unexpected information and adverse selection on order execution costs. Overall, our research demonstrates that the transition to electronic trading in the corn futures market has led to low and stable liquidity costs, despite the market turbulence in 2008–2009.
    Keywords: C36- Instrumental Variables (IV) Estimation, G12 - Asset Pricing ; Trading volume ; Bond Interest Rates, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 27
    Publication Date: 2014-03-21
    Description: To successfully market new products in a social network it is essential to identify influential individuals whose product recommendations influence the consumption choices of their peers. In this study, we use spatial econometric methods to determine how individuals revise their preferences for product attributes when exposed to product recommendations from peers, and how different individuals who vary in their degree of network connectedness exert influence on the product choices of others. We find evidence that consumers look to others for guidance from peers in their preference for subjective, taste-specific parameters, but tend not to respond to peer price choices. Our spatial methods allow us to empirically determine the influence exerted by individual members on the consumption choices of other members of the social network. We find that connected members of the social network are not always the most influential in revising the consumption choices of others. Our estimates reveal that network proximity explains only 8.8% of influence.
    Keywords: D12 - Consumer Economics: Empirical Analysis, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 28
    Publication Date: 2014-02-28
    Description: Combinatorial interactions among transcription factors (TFs) are critical for integrating diverse intrinsic and extrinsic signals, fine-tuning regulatory output and increasing the robustness and plasticity of regulatory systems. Current knowledge about combinatorial regulation is rather limited due to the lack of suitable experimental technologies and bioinformatics tools. The rapid accumulation of ChIP-Seq data has provided genome-wide occupancy maps for a large number of TFs and chromatin modification marks for identifying enhancers without knowing individual TF binding sites. Integration of the two data types has not been researched extensively, resulting in underused data and missed opportunities. We describe a novel method for discovering frequent combinatorial occupancy patterns by multiple TFs at enhancers. Our method is based on probabilistic item set mining and takes into account uncertainty in both types of ChIP-Seq data. By joint analysis of 108 TFs in four human cell types, we found that cell–type-specific interactions among TFs are abundant and that the majority of enhancers have flexible architecture. We show that several families of transposable elements disproportionally overlap with enhancers with combinatorial patterns, suggesting that these transposable element families play an important role in the evolution of combinatorial regulation.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 29
    Publication Date: 2014-04-03
    Description: Recent advances in high-throughput sequencing (HTS) technologies and computing capacity have produced unprecedented amounts of genomic data that have unraveled the genetics of phenotypic variability in several species. However, operating and integrating current software tools for data analysis still require important investments in highly skilled personnel. Developing accurate, efficient and user-friendly software packages for HTS data analysis will lead to a more rapid discovery of genomic elements relevant to medical, agricultural and industrial applications. We therefore developed Next-Generation Sequencing Eclipse Plug-in (NGSEP), a new software tool for integrated, efficient and user-friendly detection of single nucleotide variants (SNVs), indels and copy number variants (CNVs). NGSEP includes modules for read alignment, sorting, merging, functional annotation of variants, filtering and quality statistics. Analysis of sequencing experiments in yeast, rice and human samples shows that NGSEP has superior accuracy and efficiency, compared with currently available packages for variants detection. We also show that only a comprehensive and accurate identification of repeat regions and CNVs allows researchers to properly separate SNVs from differences between copies of repeat elements. We expect that NGSEP will become a strong support tool to empower the analysis of sequencing data in a wide range of research projects on different species.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 30
    Publication Date: 2014-04-05
    Description: Randomized controlled trials (RCTs) in the social sciences are typically not double-blind, so participants know they are "treated" and will adjust their behavior accordingly. Such effort responses complicate the assessment of impact. To gauge the potential magnitude of effort responses we implement a conventional RCT and double-blind trial in rural Tanzania, and randomly allocate modern and traditional cowpea seed varieties to a sample of farmers. Effort responses can be quantitatively important—for our case they explain the entire "treatment effect on the treated" as measured in a conventional economic RCT. Specifically, harvests are the same for people who know they received the modern seeds and for people who did not know what type of seeds they got; however, people who knew they had received the traditional seeds did much worse. Importantly, we also find that most of the behavioral response is unobserved by the analyst, or at least not readily captured using coarse, standard controls.
    Keywords: Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 31
    Publication Date: 2012-03-29
    Description: Next-generation sequencing (NGS) technologies-based transcriptomic profiling method often called RNA-seq has been widely used to study global gene expression, alternative exon usage, new exon discovery, novel transcriptional isoforms and genomic sequence variations. However, this technique also poses many biological and informatics challenges to extracting meaningful biological information. The RNA-seq data analysis is built on the foundation of high quality initial genome localization and alignment information for RNA-seq sequences. Toward this goal, we have developed RNASEQR to accurately and effectively map millions of RNA-seq sequences. We have systematically compared RNASEQR with four of the most widely used tools using a simulated data set created from the Consensus CDS project and two experimental RNA-seq data sets generated from a human glioblastoma patient. Our results showed that RNASEQR yields more accurate estimates for gene expression, complete gene structures and new transcript isoforms, as well as more accurate detection of single nucleotide variants (SNVs). RNASEQR analyzes raw data from RNA-seq experiments effectively and outputs results in a manner that is compatible with a wide variety of specialized downstream analyses on desktop computers.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 32
    Publication Date: 2014-10-23
    Description: The paper in question by Van Camp and co-authors [MVC] challenges previous work showing that ground gravity data arising from hydrology can provide a consistent signal for the comparison with satellite gravity data. The data sets used are similar to those used previously, that is, the gravity field as measured by the GRACE satellites versus ground-based data from superconducting gravimeters (SGs) over the same continental area, in this case Central Europe. One of the main impediments in this paper is the presentation that is frequently confusing and misleading as to what the data analysis really shows, for example, the irregular treatment of annual components that are first subtracted then reappear in the analysis. More importantly, we disagree on specific points. Two calculations are included in our comment to illustrate where we believe that the processing in [MVC] paper is deficient. The first deals with their erroneous treatment of the global hydrology using a truncated spherical harmonic approach which explains almost a factor 2 error in their computation of the loading. The second shows the effect of making the wrong assumption in the GRACE/hydrology/surface gravity comparison by inverting the whole of the hydrology loading for underground stations. We also challenge their claims that empirical orthogonal function techniques cannot be done in the presence of periodic components, and that SG data cannot be corrected for comparisons with GRACE data. The main conclusion of their paper, that there is little coherence between ground gravity stations and this invalidates GRACE comparisons, is therefore questionable. There is nothing in [MVC] that contradicts any of the previous papers that have shown clearly a strong relation between seasonal signals obtained from both ground gravity and GRACE satellite data.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 33
    Publication Date: 2014-10-23
    Description: The influence of changes in surface ice-mass redistribution and associated viscoelastic response of the Earth, known as glacial isostatic adjustment (GIA), on the Earth's rotational dynamics has long been known. Equally important is the effect of the changes in the rotational dynamics on the viscoelastic deformation of the Earth. This signal, known as the rotational feedback, or more precisely, the rotational feedback on the sea level equation, has been mathematically described by the sea level equation extended for the term that is proportional to perturbation in the centrifugal potential and the second-degree tidal Love number. The perturbation in the centrifugal force due to changes in the Earth's rotational dynamics enters not only into the sea level equation, but also into the conservation law of linear momentum such that the internal viscoelastic force, the perturbation in the gravitational force and the perturbation in the centrifugal force are in balance. Adding the centrifugal-force perturbation to the linear-momentum balance creates an additional rotational feedback on the viscoelastic deformations of the Earth. We term this feedback mechanism, which is studied in this paper, as the rotational feedback on the linear-momentum balance. We extend both the time-domain method for modelling the GIA response of laterally heterogeneous earth models developed by Martinec and the traditional Laplace-domain method for modelling the GIA-induced rotational response to surface loading by considering the rotational feedback on linear-momentum balance. The correctness of the mathematical extensions of the methods is validated numerically by comparing the polar-motion response to the GIA process and the rotationally induced degree 2 and order 1 spherical harmonic component of the surface vertical displacement and gravity field. We present the difference between the case where the rotational feedback on linear-momentum balance is considered against that where it is not. Numerical simulations show that the resulting difference in radial displacement and sea level change between these situations since the Last Glacial Maximum reaches values of ±25 and ±1.8 m, respectively. Furthermore, the surface deformation pattern is modified by up to 10 per cent in areas of former or ongoing glaciation, but by up to 50 per cent at the bottom of the southern Indian ocean. This also results in the movement of coastlines during the last deglaciation to differ between the two cases due to the difference in the ocean loading, which is seen for instance in the area around Hudson Bay, Canada and along the Chinese, Australian or Argentinian coastlines.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 34
    Publication Date: 2014-09-20
    Description: During megathrust earthquakes, great ruptures are accompanied by large scale mass redistribution inside the solid Earth and by ocean mass redistribution due to bathymetry changes. These large scale mass displacements can be detected using the monthly gravity maps of the GRACE satellite mission. In recent years it has become increasingly common to use the long wavelength changes in the Earth's gravity field observed by GRACE to infer seismic source properties for large megathrust earthquakes. An important advantage of space gravimetry is that it is independent from the availability of land for its measurements. This is relevant for observation of megathrust earthquakes, which occur mostly offshore, such as the $M_{\text{w}} \sim 9$ 2004 Sumatra–Andaman, 2010 Maule (Chile) and 2011 Tohoku-Oki (Japan) events. In Broerse et al. , we examined the effect of the presence of an ocean above the rupture on long wavelength gravity changes and showed it to be of the first order. Here we revisit the implementation of an ocean layer through the sea level equation and compare the results with approximated methods that have been used in the literature. One of the simplifications usually lies in the assumption of a globally uniform ocean layer. We show that especially in the case of the 2010 Maule earthquake, due to the closeness of the South American continent, the uniform ocean assumption is not valid and causes errors up to 57 per cent for modelled peak geoid height changes (expressed at a spherical harmonic truncation degree of 40). In addition, we show that when a large amount of slip occurs close to the trench, horizontal motions of the ocean floor play a mayor role in the ocean contribution to gravity changes. Using a slip model of the 2011 Tohoku-Oki earthquake that places the majority of slip close to the surface, the peak value in geoid height change increases by 50 per cent due to horizontal ocean floor motion. Furthermore, we test the influence of the maximum spherical harmonic degree at which the sea level equation is performed for sea level changes occurring along coastlines, which shows to be important for relative sea level changes occurring along the shore. Finally, we demonstrate that ocean floor loading, self-gravitation of water and conservation of water mass are of second order importance for coseismic gravity changes. When GRACE observations are used to determine earthquake parameters such as seismic moment or source depth, the uniform ocean layer method introduces large biases, depending on the location of the rupture with respect to the continent. The same holds for interpreting shallow slip when horizontal motions are not properly accounted for in the ocean contribution. In both cases the depth at which slip occurs will be underestimated.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 35
    Publication Date: 2014-09-06
    Description: This article investigates the announcement effects of major USDA reports using intraday Chicago Board of Trade corn futures prices and trading volume from the electronic trading platform for July 2009 to May 2012. Focusing on intraday market reactions, we analyze the extent to which new information impacts and is rapidly reflected in prices. Results show that USDA reports contain substantial information for market participants. Strongest price reactions to the releases are found immediately after the market opens, and market reactions persist for approximately ten minutes. The electronic corn futures market quickly incorporates this new public information, and little evidence exists to support systematic under- or overreactions in prices. Other more subtle reactions occur in the last trading session before USDA announcements as traders adjust their market exposure in anticipation of the release.
    Keywords: D80 - General, G14 - Information and Market Efficiency ; Event Studies, Q11 - Aggregate Supply and Demand Analysis ; Prices, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 2040-5790
    Electronic ISSN: 2040-5804
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 36
    Publication Date: 2014-09-06
    Description: Increases in crop yields and changing cropping patterns have placed stress on agribusiness handling and storage facilities. The objective of this research is to gain insight into the relationship between safety culture and safety performance, and to identify the determinants of safety culture in agribusinesses. The research suggests that investments in labor inputs such as increased training, consistent discipline, and recognition of safety achievements all increase safety culture. Furthermore, improvements in employee perceptions of safety culture have a positive impact on reducing employee injuries. Congress has recently funded nine centers to work on occupational health and safety research in agriculture, fisheries, and forestry.
    Keywords: J43 - Agricultural Labor Markets, L66 - Food ; Beverages ; Cosmetics ; Tobacco ; Wine and Spirits, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 2040-5790
    Electronic ISSN: 2040-5804
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 37
    Publication Date: 2014-09-07
    Description: Long-term volcanic subsidence provides insight into intereruptive processes, which comprise the longest portion of the eruptive cycle. Ground-based geodetic surveys of Medicine Lake Volcano (MLV), northern CA, document subsidence at rates of ~–10 mm yr –1 between 1954 and 2004. The long observation period plus the duration and stable magnitude of this signal presents an ideal opportunity to study long-term volcanic deformation, but this first requires accurate knowledge of the geometry and magnitude of the source. Best-fitting analytical source models to past levelling and GPS data sets show conflicting source parameters—primarily the model depth. To overcome this, we combine multiple tracks of InSAR data, each with a different look angle, to improve upon the spatial resolution of ground-based measurements. We compare the results from InSAR to those of past geodetic studies, extending the geodetic record to 2011 and demonstrating that subsidence at MLV continues at ~–10 mm yr –1 . Using geophysical inversions, we obtain the best-fitting analytical source model—a sill located at 9–10 km depth beneath the caldera. This model geometry is similar to those of past studies, providing a good fit to the high spatial density of InSAR measurements, while accounting for the high ratio of vertical to horizontal deformation derived from InSAR and recorded by existing levelling and GPS data sets. We discuss possible causes of subsidence and show that this model supports the hypothesis that deformation at MLV is driven by tectonic extension, gravitational loading, plus a component of volume loss at depth, most likely due to cooling and crystallization within the intrusive complex that underlies the edifice. Past InSAR surveys at MLV, and throughout the Cascades, are of variable success due to dense vegetation, snow cover and atmospheric artefacts. In this study, we demonstrate how InSAR may be successfully used in this setting by applying a suite of multitemporal analysis methods that account for atmospheric and orbital noise sources. These methods include: a stacking strategy based upon the noise characteristics of each data set; pixelwise rate-map formation (-RATE) and persistent scatterer InSAR (StaMPS).
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 38
    Publication Date: 2014-09-11
    Description: In the literature, the inverted coseismic slip models from seismological and geodetic data for the 2011 Tohoku-Oki earthquake portray significant discrepancies, in particular regarding the intensity and the distribution of the rupture near the trench. For a megathrust earthquake, it is difficult to discern the slip along the shallow part of the fault from the geodetic data, which are often acquired on land. In this paper, we discuss the uncertainties in the slip distribution inversion using the geodetic data for the 2011 Tohoku earthquake and the Fully Bayesian Inversion method. These uncertainties are due to the prior information regarding the boundary conditions at the edges of the fault, the dip subduction angle and the smoothing operator. Using continuous GPS data from the Japan Island, the results for the rigid and free boundary conditions show that they produce remarkably different slip distributions at shallow depths, with the latter producing a large slip exceeding 30 m near the surface. These results indicate that the smoothing operator (gradient or Laplacian schemes) does not severely affect the slip pattern. To better invert the coseismic slip, we then introduce the ocean bottom GPS (OB-GPS) data, which improve the resolution of the shallow part of the fault. We obtain a near-trench slip greater than 40 m that reaches the Earth's surface, regardless of which boundary condition is used. Additionally, we show that using a mean dip angle for the fault as derived from subduction models is adequate if the goal is to invert for the general features of the slip pattern of this megathrust event.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 39
    Publication Date: 2014-12-13
    Description: In a well-functioning futures market, the futures price at expiration equals the price of the underlying asset. This condition failed to hold in grain markets for most of 2005-2010, calling into question the ability of these markets to perform their price discovery and risk management functions. During this period, futures contracts expired up to 35% above the cash grain price. We develop a dynamic rational expectations model of commodity storage that explains how these recent convergence failures were generated by the institutional structure of the delivery system. When delivery occurs on a grain futures contract, the firm on the short side of the market provides a delivery instrument (a warehouse receipt or shipping certificate) to the firm on the long side of the market. The firm taking delivery may hold the delivery instrument indefinitely, providing it pays a daily storage rate. The futures exchange sets the maximum allowable storage rate at a fixed value. We show that non-convergence arises in equilibrium when the market price of physical grain storage exceeds the maximum storage rate on delivery instruments. We call the difference between the price of carrying physical grain and the maximum storage rate the wedge , and demonstrate theoretically and empirically that the magnitude of the non-convergence equals the expected present discounted value of a function of future wedges.
    Keywords: G13 - Contingent Pricing ; Futures Pricing, G14 - Information and Market Efficiency ; Event Studies, Q11 - Aggregate Supply and Demand Analysis ; Prices, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 40
    Publication Date: 2014-10-16
    Description: Complications arise in the interpretation of gravity fields because of interference from systematic degradations, such as boundary blurring and distortion. The major sources of these degradations are the various systematic errors that inevitably occur during gravity field data acquisition, discretization and geophysical forward modelling. To address this problem, we evaluate deconvolution method that aim to detect the clear horizontal boundaries of anomalous sources by the suppression of systematic errors. A convolution-based multilayer projection model, based on the classical 3-D gravity field forward model, is innovatively derived to model the systematic error degradation. Our deconvolution algorithm is specifically designed based on this multilayer projection model, in which three types of systematic error are defined. The degradations of the different systematic errors are considered in the deconvolution algorithm. As the primary source of degradation, the convolution-based systematic error is the main object of the multilayer projection model. Both the random systematic error and the projection systematic error are shown to form an integral part of the multilayer projection model, and the mixed norm regularization method and the primal-dual optimization method are therefore employed to control these errors and stabilize the deconvolution solution. We herein analyse the parameter identification and convergence of the proposed algorithms, and synthetic and field data sets are both used to illustrate their effectiveness. Additional synthetic examples are specifically designed to analyse the effects of the projection systematic error, which is caused by the uncertainty associated with the estimation of the impulse response function.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 41
    Publication Date: 2014-06-21
    Description: We propose to test if gravimetry can prove useful in discriminating different models of long-term deep crustal processes in the case of the Taiwan mountain belt. We discuss two existing tectonic models that differ in the deep processes proposed to sustain the long-term growth of the orogen. One model assumes underplating of the uppermost Eurasian crust with subduction of the deeper part of the crust into the mantle. The other one suggests the accretion of the whole Eurasian crust above crustal-scale ramps, the lower crust being accreted into the collisional orogen. We compute the temporal gravity changes caused only by long-term rock mass transfers at depth for each of them. We show that the underplating model implies a rate of gravity change of –6 x 10 –2 μGal yr –1 , a value that increases to 2 x 10 –2 μGal yr –1 if crustal subduction is neglected. If the accretion of the whole Eurasian crust occurs, a rate of 7 x 10 –2 μGal yr –1 is obtained. The two models tested differ both in signal amplitude and spatial distribution. The yearly gravity changes expected by long-term deep crustal mass processes in Taiwan are two orders of magnitude below the present-day uncertainty of land-based gravity measurements. Assuming that these annually averaged long-term gravity changes will linearly accumulate with ongoing mountain building, multidecadal time-series are needed to identify comparable rates of gravity change. However, as gravity is sensitive to any mass redistribution, effects of short-term processes such as seismicity and surface mass transfers (erosion, sedimentation, ground-water) may prevent from detecting any long-term deep signal. This study indicates that temporal gravity is not appropriate for deciphering the long-term deep crustal processes involved in the Taiwan mountain belt.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 42
    Publication Date: 2014-06-21
    Description: The computation of quasi-static deformation for axisymmetric viscoelastic structures on a gravitating spherical earth is addressed using the spectral element method (SEM). A 2-D spectral element domain is defined with respect to spherical coordinates of radius and angular distance from a pole of symmetry, and 3-D viscoelastic structure is assumed to be azimuthally symmetric with respect to this pole. A point dislocation source that is periodic in azimuth is implemented with a truncated sequence of azimuthal order numbers. Viscoelasticity is limited to linear rheologies and is implemented with the correspondence principle in the Laplace transform domain. This leads to a series of decoupled 2-D problems which are solved with the SEM. Inverse Laplace transform of the independent 2-D solutions leads to the time-domain solution of the 3-D equations of quasi-static equilibrium imposed on a 2-D structure. The numerical procedure is verified through comparison with analytic solutions for finite faults embedded in a laterally homogeneous viscoelastic structure. This methodology is applicable to situations where the predominant structure varies in one horizontal direction, such as a structural contrast across (or parallel to) a long strike-slip fault.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 43
    Publication Date: 2014-06-21
    Description: The geodetic rates for the gravity variation and vertical uplift in polar regions subject to past and present-day ice-mass changes (PDIMCs) provide important insight into the rheological structure of the Earth. We provide an update of the rates observed at Ny-Ålesund, Svalbard. To do so, we extract and remove the significant seasonal content from the observations. The rate of gravity variations, derived from absolute and relative gravity measurements, is –1.39 ± 0.11 μGal yr –1 . The rate of vertical displacements is estimated using GPS and tide gauge measurements. We obtain 7.94 ± 0.21 and 8.29 ± 1.60 mm yr –1 , respectively. We compare the extracted signal with that predicted by GLDAS/Noah and ERA-interim hydrology models. We find that the seasonal gravity variations are well-represented by local hydrology changes contained in the ERA-interim model. The phase of seasonal vertical displacements are due to non-local continental hydrology and non-tidal ocean loading. However, a large part of the amplitude of the seasonal vertical displacements remains unexplained. The geodetic rates are used to investigate the asthenosphere viscosity and lithosphere/asthenosphere thicknesses. We first correct the updated geodetic rates for those induced by PDIMCs in Svalbard, using published results, and the sea level change due to the melting of the major ice reservoirs. We show that the latter are at the level of the geodetic rate uncertainties and are responsible for rates of gravity variations and vertical displacements of –0.29 ± 0.03 μGal yr –1 and 1.11 ± 0.10 mm yr –1 , respectively. To account for the late Pleistocene deglaciation, we use the global ice evolution model ICE-3G. The Little Ice Age (LIA) deglaciation in Svalbard is modelled using a disc load model with a simple linear temporal evolution. The geodetic rates at Ny-Ålesund induced by the past deglaciations depend on the viscosity structure of the Earth. We find that viscous relaxation time due to the LIA deglaciation in Svalbard is more than 60 times shorter than that due to the Pleistocene deglaciation. We also find that the response to past and PDIMCs of an Earth model with asthenosphere viscosities ranging between 1.0 and 5.5 x 10 18 Pa s and lithosphere (resp. asthenosphere) thicknesses ranging between 50 and 100 km (resp. 120 and 170 km) can explain the rates derived from geodetic observations.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 44
    Publication Date: 2014-11-16
    Description: The 2 principle and the unbiased predictive risk estimator are used to determine optimal regularization parameters in the context of 3-D focusing gravity inversion with the minimum support stabilizer. At each iteration of the focusing inversion the minimum support stabilizer is determined and then the fidelity term is updated using the standard form transformation. Solution of the resulting Tikhonov functional is found efficiently using the singular value decomposition of the transformed model matrix, which also provides for efficient determination of the updated regularization parameter each step. Experimental 3-D simulations using synthetic data of a dipping dike and a cube anomaly demonstrate that both parameter estimation techniques outperform the Morozov discrepancy principle for determining the regularization parameter. Smaller relative errors of the reconstructed models are obtained with fewer iterations. Data acquired over the Gotvand dam site in the south-west of Iran are used to validate use of the methods for inversion of practical data and provide good estimates of anomalous structures within the subsurface.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 45
    Publication Date: 2014-11-19
    Description: Global navigation satellite systems (GNSSs) have revealed that a mega-thrust earthquake that occurs in an island-arc trench system causes post-seismic crustal deformation. Such crustal deformation data have been interpreted by combining three mechanisms: afterslip, poroelastic rebound and viscoelastic relaxation. It is seismologically important to determine the contribution of each mechanism because it provides frictional properties between the plate boundaries and viscosity estimates in the asthenosphere which are necessary to evaluate the stress behaviour during earthquake cycles. However, the observation sites of GNSS are mostly deployed over land and can detect only a small part of the large-scale deformation, which precludes a clear separation of the mechanisms. To extend the spatial coverage of the deformation area, recent studies started to use satellite gravity data that can detect long-wavelength deformations over the ocean. To date, compared with theoretical models for calculating the post-seismic crustal deformation, a few models have been proposed to interpret the corresponding gravity variations. Previous approaches have adopted approximations for the effects of compressibility, sphericity and self-gravitation when computing gravity changes. In this study, a new spectral-finite element approach is presented to consider the effects of material compressibility for Burgers viscoelastic earth model with a laterally heterogeneous viscosity distribution. After the basic principles are explained, it is applied to the 2004 Sumatra–Andaman earthquake. For this event, post-seismic deformation mechanisms are still a controversial topic. Using the developed approach, it is shown that the spatial patterns of gravity change generated by the above three mechanisms clearly differ from one another. A comparison of the theoretical simulation results with the satellite gravity data obtained from the Gravity Recovery and Climate Experiment reveals that both afterslip and viscoelastic relaxation are occurring. Considering the spatial patterns in satellite gravity fields is an effective method for investigating post-seismic deformation mechanisms.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 46
    Publication Date: 2014-09-02
    Description: This article discusses the current state of contract theory and its usefulness for conceptualizing issues related to agricultural contracting. The paper will explore the limitations of existing theory for applied work, and what methodological improvements are needed to enhance the usefulness of the theory to agricultural economists. One pervasive problem is that the economic literature on contracts is rather fragmented and the various methodological strands are narrow in their focus. As such, agricultural economists should engage in methodological research to develop applied contracting models that can capture higher-order features of real-world agricultural contracts while delivering generalizable comparative statics predictions because contracting continues to expand along the entire modern food marketing channel. In the latter part of this article, a simple model is developed to illustrate how classic methodological approaches can be combined with recent developments in contract and game theory to construct applied theory models that are useful for capturing some important features of agricultural contracts.
    Keywords: D43 - Oligopoly and Other Forms of Market Imperfection, D82 - Asymmetric and Private Information, D86 - Economics of Contract: Theory, Q12 - Micro Analysis of Farm Firms, Farm Households, and Farm Input Markets, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 47
    Publication Date: 2014-09-02
    Description: Information and communications technologies have spread rapidly in developing countries. We investigate the impact of mobile phones on traders' search behavior in Niger by constructing a theoretical model of search in which traders engage in sequential search for the optimal sales price. Using a trader panel dataset spanning 2005–2007, we find empirical support for the model in that the duration of mobile phone coverage is associated with increased search activity. This effect evolves dynamically over time and is stronger for larger traders, who engage in arbitrage over longer distances. Results provide empirical evidence for the observed linkages between mobile telephony and price dispersion.
    Keywords: O10 - General, O30 - General, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 48
    Publication Date: 2014-09-02
    Description: Conventionally, overall gene expressions from microarrays are used to infer gene networks, but it is challenging to account splicing isoforms. High-throughput RNA Sequencing has made splice variant profiling practical. However, its true merit in quantifying splicing isoforms and isoform-specific exon expressions is not well explored in inferring gene networks. This study demonstrates SpliceNet, a method to infer isoform-specific co-expression networks from exon-level RNA-Seq data, using large dimensional trace. It goes beyond differentially expressed genes and infers splicing isoform network changes between normal and diseased samples. It eases the sample size bottleneck; evaluations on simulated data and lung cancer-specific ERBB2 and MAPK signaling pathways, with varying number of samples, evince the merit in handling high exon to sample size ratio datasets. Inferred network rewiring of well established Bcl-x and EGFR centered networks from lung adenocarcinoma expression data is in good agreement with literature. Gene level evaluations demonstrate a substantial performance of SpliceNet over canonical correlation analysis, a method that is currently applied to exon level RNA-Seq data. SpliceNet can also be applied to exon array data. SpliceNet is distributed as an R package available at http://www.jjwanglab.org/SpliceNet .
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 49
    Publication Date: 2014-09-02
    Description: We present a new approach to automatic training of a eukaryotic ab initio gene finding algorithm. With the advent of Next-Generation Sequencing, automatic training has become paramount, allowing genome annotation pipelines to keep pace with the speed of genome sequencing. Earlier we developed GeneMark-ES, currently the only gene finding algorithm for eukaryotic genomes that performs automatic training in unsupervised ab initio mode. The new algorithm, GeneMark-ET augments GeneMark-ES with a novel method that integrates RNA-Seq read alignments into the self-training procedure. Use of ‘assembled’ RNA-Seq transcripts is far from trivial; significant error rate of assembly was revealed in recent assessments. We demonstrated in computational experiments that the proposed method of incorporation of ‘unassembled’ RNA-Seq reads improves the accuracy of gene prediction; particularly, for the 1.3 GB genome of Aedes aegypti the mean value of prediction Sensitivity and Specificity at the gene level increased over GeneMark-ES by 24.5%. In the current surge of genomic data when the need for accurate sequence annotation is higher than ever, GeneMark-ET will be a valuable addition to the narrow arsenal of automatic gene prediction tools.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 50
    Publication Date: 2014-08-07
    Description: Some of the major geothermal anomalies in central Europe are linked to tectonic structures within the top of crystalline basement, which modify strongly the top of this basement. Their assessment is a major challenge in exploration geophysics. Gravity has been proven to be suitable for the detection of mainly large scale lithological and structural inhomogeneities. Indeed, it is well known and proven by different wells that, for example, in northern Switzerland extended negative anomalies are linked to such structures. Due to depth limitation of wells, there vertical extension is often unknown. In this study, we have investigated the potential of gravity for the geometrical characterization of such basement structures. Our approach consists in the combination of the series of Butterworth filters, geological modelling and best-fitting between observed and computed residual anomalies. In this respect, filters of variable wavelength are applied to observed and computed gravity data. The geological model is discretized into a finite element mesh. Near-surface anomalies and the effect of the sedimentary cover were eliminated using cut-off wavelength of 10 km and geological and seismic information. We analysed the potential of preferential Butterworth filtering in a sensitivity study and applied the above mentioned approach to part of the Swiss molasses basin. Sensitivity analyses reveal that such sets of residual anomalies represents a pseudo-tomography revealing the distribution of different structures with depth. This finding allows for interpreting negative anomalies in terms of 3-D volumes. Best-fitting then permits determination of the most likely 3-D geometries of such basement structures. Our model fits both, geological observations and gravity: among 10 deep boreholes in the studied area, six reach the respective units and confirm our distribution of the negative (and positive) anomalies.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 51
    Publication Date: 2014-08-15
    Description: Next-generation sequencing (NGS) technologies enable new insights into the diversity of virus populations within their hosts. Diversity estimation is currently restricted to single-nucleotide variants or to local fragments of no more than a few hundred nucleotides defined by the length of sequence reads. To study complex heterogeneous virus populations comprehensively, novel methods are required that allow for complete reconstruction of the individual viral haplotypes. Here, we show that assembly of whole viral genomes of ~8600 nucleotides length is feasible from mixtures of heterogeneous HIV-1 strains derived from defined combinations of cloned virus strains and from clinical samples of an HIV-1 superinfected individual. Haplotype reconstruction was achieved using optimized experimental protocols and computational methods for amplification, sequencing and assembly. We comparatively assessed the performance of the three NGS platforms 454 Life Sciences/Roche, Illumina and Pacific Biosciences for this task. Our results prove and delineate the feasibility of NGS-based full-length viral haplotype reconstruction and provide new tools for studying evolution and pathogenesis of viruses.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 52
    Publication Date: 2012-12-14
    Description: Insertion sequences (ISs) are simple transposable elements present in most bacterial and archaeal genomes and play an important role in genomic evolution. The recent expansion of sequenced genomes offers the opportunity to study ISs comprehensively, but this requires efficient and accurate tools for IS annotation. We have developed an open-source program called OASIS, or Optimized Annotation System for Insertion Sequences, which automatically annotates ISs within sequenced genomes. OASIS annotations of 1737 bacterial and archaeal genomes offered an unprecedented opportunity to examine IS evolution. At a broad scale, we found that most IS families are quite widespread; however, they are not present randomly across taxa. This may indicate differential loss, barriers to exchange and/or insufficient time to equilibrate across clades. The number of ISs increases with genome length, but there is both tremendous variation and no increase in IS density for genomes 〉2 Mb. At the finer scale of recently diverged genomes, the proportion of shared IS content falls sharply, suggesting loss and/or emergence of barriers to successful cross-infection occurs rapidly. Surprisingly, even after controlling for 16S rRNA sequence divergence, the same ISs were more likely to be shared between genomes labeled as the same species rather than as different species.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 53
    Publication Date: 2012-12-14
    Description: The study of cell-population heterogeneity in a range of biological systems, from viruses to bacterial isolates to tumor samples, has been transformed by recent advances in sequencing throughput. While the high-coverage afforded can be used, in principle, to identify very rare variants in a population, existing ad hoc approaches frequently fail to distinguish true variants from sequencing errors. We report a method (LoFreq) that models sequencing run-specific error rates to accurately call variants occurring in 〈0.05% of a population. Using simulated and real datasets (viral, bacterial and human), we show that LoFreq has near-perfect specificity, with significantly improved sensitivity compared with existing methods and can efficiently analyze deep Illumina sequencing datasets without resorting to approximations or heuristics. We also present experimental validation for LoFreq on two different platforms (Fluidigm and Sequenom) and its application to call rare somatic variants from exome sequencing datasets for gastric cancer. Source code and executables for LoFreq are freely available at http://sourceforge.net/projects/lofreq/ .
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 54
    Publication Date: 2012-05-29
    Description: Using stated preference data from Kentucky and Ohio, USA, we estimate consumer willingness-to-pay for varieties of a processed food product (blackberry jam) that are differentiated with respect to their local production labelling and a series of other value-added claims. Results show that consumers were willing to pay more for the product indicating locally produced, produced in their state or in a well-identified multi-state region. Consumers were willing to purchase organic products, although there might be some confusion as to the meaning of the organic logo. Our results also supported the notion that consumers are willing to support small family farms.
    Keywords: Q11 - Aggregate Supply and Demand Analysis ; Prices, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 55
    Publication Date: 2012-05-29
    Description: This paper studies the determinants and consequences of heterogeneous market participation among Polish dairy farmers using a unique data set on supply chain characteristics and individuals with different market relationships. It investigates factors that cause households not to participate in the market and then estimates farm orientation effects on revenues, using semi-parametric methods. The key finding is that farms maintaining commercial dairy business were better off than those who ceased milk sales. However, detailed analysis shows that this difference could be attributed to supply chain modernisation and becomes insignificant once subsistence farmers are compared to commercial farms supplying the traditional marketing channel.
    Keywords: D22 - Firm Behavior: Empirical Analysis, Q12 - Micro Analysis of Farm Firms, Farm Households, and Farm Input Markets, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 56
    Publication Date: 2013-05-29
    Description: Deep transcriptome sequencing (RNA-Seq) has become a vital tool for studying the state of cells in the context of varying environments, genotypes and other factors. RNA-Seq profiling data enable identification of novel isoforms, quantification of known isoforms and detection of changes in transcriptional or RNA-processing activity. Existing approaches to detect differential isoform abundance between samples either require a complete isoform annotation or fall short in providing statistically robust and calibrated significance estimates. Here, we propose a suite of statistical tests to address these open needs: a parametric test that uses known isoform annotations to detect changes in relative isoform abundance and a non-parametric test that detects differential read coverages and can be applied when isoform annotations are not available. Both methods account for the discrete nature of read counts and the inherent biological variability. We demonstrate that these tests compare favorably to previous methods, both in terms of accuracy and statistical calibrations. We use these techniques to analyze RNA-Seq libraries from Arabidopsis thaliana and Drosophila melanogaster. The identified differential RNA processing events were consistent with RT–qPCR measurements and previous studies. The proposed toolkit is available from http://bioweb.me/rdiff and enables in-depth analyses of transcriptomes, with or without available isoform annotation.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 57
    Publication Date: 2013-11-22
    Description: India and China have the largest farm-household populations in the world—populations that are also among the poorest. Among the many factors that affect farm livelihoods, access to credit has been identified as a significant barrier preventing the escape from poverty. While there has been significant research on credit constraints in developing countries, there is surprisingly little information pertaining to the actual impacts of credit constraints on household well-being. The objective of this paper is to investigate the impacts of credit constraints on various factors affecting farm households, such as physical and human capital formation, agricultural inputs applications, consumption smoothing, and wage-seeking behavior using direct elicitation. This paper contributes to the literature and policy debates by comparing the effects of credit constraints in China and India as surveyed in 2008–2009. The analytical results and data demonstrate that binding credit constraints adversely affect a broad range of production and livelihood choices. We empirically show that credit constraints negatively affect food consumption, farm input applications, and health and educational attainments.
    Keywords: O13 - Agriculture ; Natural Resources ; Energy ; Environment ; Other Primary Products, O17 - Formal and Informal Sectors ; Shadow Economy ; Institutional Arrangements, Q12 - Micro Analysis of Farm Firms, Farm Households, and Farm Input Markets, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 2040-5790
    Electronic ISSN: 2040-5804
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 58
    Publication Date: 2013-11-21
    Description: The absence of a quality control (QC) system is a major weakness for the comparative analysis of genome-wide profiles generated by next-generation sequencing (NGS). This concerns particularly genome binding/occupancy profiling assays like chromatin immunoprecipitation (ChIP-seq) but also related enrichment-based studies like methylated DNA immunoprecipitation/methylated DNA binding domain sequencing, global run on sequencing or RNA-seq. Importantly, QC assessment may significantly improve multidimensional comparisons that have great promise for extracting information from combinatorial analyses of the global profiles established for chromatin modifications, the bindings of epigenetic and chromatin-modifying enzymes/machineries, RNA polymerases and transcription factors and total, nascent or ribosome-bound RNAs. Here we present an approach that associates global and local QC indicators to ChIP-seq data sets as well as to a variety of enrichment-based studies by NGS. This QC system was used to certify 〉5600 publicly available data sets, hosted in a database for data mining and comparative QC analyses.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 59
    Publication Date: 2013-10-19
    Description: Identifying variants using high-throughput sequencing data is currently a challenge because true biological variants can be indistinguishable from technical artifacts. One source of technical artifact results from incorrectly aligning experimentally observed sequences to their true genomic origin (‘mismapping’) and inferring differences in mismapped sequences to be true variants. We developed BlackOPs, an open-source tool that simulates experimental RNA-seq and DNA whole exome sequences derived from the reference genome, aligns these sequences by custom parameters, detects variants and outputs a blacklist of positions and alleles caused by mismapping. Blacklists contain thousands of artifact variants that are indistinguishable from true variants and, for a given sample, are expected to be almost completely false positives. We show that these blacklist positions are specific to the alignment algorithm and read length used, and BlackOPs allows users to generate a blacklist specific to their experimental setup. We queried the dbSNP and COSMIC variant databases and found numerous variants indistinguishable from mapping errors. We demonstrate how filtering against blacklist positions reduces the number of potential false variants using an RNA-seq glioblastoma cell line data set. In summary, accounting for mapping-caused variants tuned to experimental setups reduces false positives and, therefore, improves genome characterization by high-throughput sequencing.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 60
    Publication Date: 2013-10-19
    Description: The 3D chromatin structure modeling by chromatin interactions derived from Hi-C experiments is significantly challenged by the intrinsic sequencing biases in these experiments. Conventional modeling methods only focus on the bias among different chromatin regions within the same experiment but neglect the bias arising from different experimental sequencing depth. We now show that the regional interaction bias is tightly coupled with the sequencing depth, and we further identify a chromatin structure parameter as the inherent characteristics of Hi-C derived data for chromatin regions. Then we present an approach for chromatin structure prediction capable of relaxing both kinds of sequencing biases by using this identified parameter. This method is validated by intra and inter cell-line comparisons among various chromatin regions for four human cell-lines (K562, GM12878, IMR90 and H1hESC), which shows that the openness of chromatin region is well correlated with chromatin function. This method has been executed by an automatic pipeline (AutoChrom3D) and thus can be conveniently used.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 61
    Publication Date: 2014-06-22
    Description: On 2008 October 5, a magnitude 6.6 earthquake struck the eastern termination of the intermontane Alai valley between the southern Tien Shan and the northern Pamir of Kyrgyzstan. The shallow thrust earthquake occurred in the footwall of the Main Pamir thrust, where the Pamir orogen is colliding with the southern Tien Shan mountains. We measure the coseismic surface displacements using SAR (Synthetic Aperture RADAR) data; the results show clear gradients in the vertical and horizontal directions along a complex pattern of surface ruptures and active faults. To integrate and to interpret these observations in the context of the regional tectonics, we complement the SAR data analysis with seismological data and geological field observations. While the main moment release of the Nura earthquake appears to be on the Pamir Frontal thrust, the main surface displacements and surface rupture occurred in the footwall along the NE–SW striking Irkeshtam fault. With InSAR data from ascending and descending tracks along with pixel offset measurements, we model the Nura earthquake source as a segmented rupture. One fault segment corresponds to high-angle brittle faulting at the Pamir Frontal thrust and two more fault segments show moderate-angle and low-friction thrusting at the Irkeshtam fault. Our integrated analysis of the coseismic deformation argues for rupture segmentation and strain partitioning associated to the earthquake. It possibly activated an orogenic wedge in the easternmost segment of the Pamir-Alai collision zone. Further, the style of the segmentation may be associated with the presence of Palaeogene evaporites.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 62
    Publication Date: 2014-06-28
    Description: The terrestrial reference frame is a cornerstone for modern geodesy and its applications for a wide range of Earth sciences. The underlying assumption for establishing a terrestrial reference frame is that the motion of the solid Earth's figure centre relative to the mass centre of the Earth system on a multidecadal timescale is linear. However, past international terrestrial reference frames (ITRFs) showed unexpected accelerated motion in their translation parameters. Based on this underlying assumption, the inconsistency of relative origin motions of the ITRFs has been attributed to data reduction imperfection. We investigated the impact of surface mass loading from atmosphere, ocean, snow, soil moisture, ice sheet, glacier and sea level from 1983 to 2008 on the geocentre variations. The resultant geocentre time-series display notable trend acceleration from 1998 onward, in particular in the z -component. This effect is primarily driven by the hydrological mass redistribution in the continents (soil moisture, snow, ice sheet and glacier). The acceleration is statistically significant at the 99 per cent confidence level as determined using the Mann–Kendall test, and it is highly correlated with the satellite laser ranging determined translation series. Our study, based on independent geophysical and hydrological models, demonstrates that, in addition to systematic errors from analysis procedures, the observed non-linearity of the Earth-system behaviour at interannual timescales is physically driven and is able to explain 42 per cent of the disparity between the origins of ITRF2000 and ITRF2005, as well as the high level of consistency between the ITRF2005 and ITRF2008 origins.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 63
    Publication Date: 2014-07-23
    Description: We study the effect of alleviating the information asymmetry regarding product quality that is widespread in contracts between agricultural producers and buyers in developing countries. Opportunistic buyers may underreport quality levels to farmers to reduce the price that they have to pay. In response, farmers may curb investment, thereby negatively affecting farm productivity. In an experiment, we entitle randomly selected smallholder dairy farmers in Vietnam, who are contracted by a large company, to independently verify milk testing results. Results indicate that treatment farmers use 12% more inputs, and they also increase their output significantly. Some wider research and policy implications are discussed.
    Keywords: C93 - Field Experiments, D82 - Asymmetric and Private Information, O13 - Agriculture ; Natural Resources ; Energy ; Environment ; Other Primary Products, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 64
    Publication Date: 2014-07-29
    Description: This paper presents a novel mathematical reformulation of the theory of the free wobble/nutation of an axisymmetric reference earth model in hydrostatic equilibrium, using the linear momentum description. The new features of this work consist in the use of (i) Clairaut coordinates (rather than spherical polars), (ii) standard spherical harmonics (rather than generalized spherical surface harmonics), (iii) linear operators (rather than J-square symbols) to represent the effects of rotational and ellipticity coupling between dependent variables of different harmonic degree and (iv) a set of dependent variables all of which are continuous across material boundaries. The resulting infinite system of coupled ordinary differential equations is given explicitly, for an elastic solid mantle and inner core, an inviscid outer core and no magnetic field. The formulation is done to second order in the Earth's ellipticity. To this order it is shown that for wobble modes (in which the lowest harmonic in the displacement field is degree 1 toroidal, with azimuthal order m  = ±1), it is sufficient to truncate the chain of coupled displacement fields at the toroidal harmonic of degree 5 in the solid parts of the earth model. In the liquid core, however, the harmonic expansion of displacement can in principle continue to indefinitely high degree at this order of accuracy. The full equations are shown to yield correct results in three simple cases amenable to analytic solution: a general earth model in rigid rotation, the tiltover mode in a homogeneous solid earth model and the tiltover and Chandler periods for an incompressible homogeneous solid earth model. Numerical results, from programmes based on this formulation, are presented in part II of this paper.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 65
    Publication Date: 2014-07-29
    Description: Numerical solutions are presented for the formulation of the linear momentum description of Earth's dynamics using Clairaut coordinates. We have developed a number of methods to integrate the equations of motion, including starting at the Earth's centre of mass, starting at finite radius and separating the displacement associated with the primary rigid rotation. We include rotation and ellipticity to second order up to spherical harmonic T $_5^m$ , starting with the primary displacement T $_1^m$ with m  = ±1. We are able to confirm many of the previous results for models PREM (with no surface ocean) and 1066A, both in their original form and with neutrally stratified liquid cores. Our period search ranges from the near-seismic band [0.1 sidereal days (sd)] to 3500 sd, within which we have identified the four well-known wobble-nutation modes: the Free Core Nutation (retrograde) at –456 sd, the Free Inner Core Nutation (FICN, prograde) at 468 sd, the Chandler Wobble (prograde) at 402 sd, and the Inner Core Wobble (ICW, prograde) at about 2842 sd (7.8 yr) for neutral PREM. The latter value varies significantly with earth model and integration method. In addition we have verified to high accuracy the tilt-over mode at 1 sd within a factor 10 –6 . In an exhaustive search we found no additional near-diurnal wobble modes that could be identified as nutations. We show that the eigenfunctions for the as-yet-unidentified ICW are extremely sensitive to the details of the earth model, especially the core stability profile and there is no well-defined sense of its wobble relative to the mantle. Calculations are also done for a range of models derived from PREM with homogeneous layers, as well as with incompressible cores. For this kind of model the ICW ceases to have just a simple IC rigid motion when the fluid compressibility is either unchanged or multiplied by a factor 10; in this case the outer core exhibits oscillations that arise from an unstable fluid density stratification. For the FICN our results for the truncation at harmonic T 5 show less change from the T 3 truncation than a similar result reported elsewhere. Finally, we give a thorough discussion of the complete spectrum of the characteristic determinant including the location of poles and non-wobble gravity modes, and discuss in general the dynamics of the inviscid core at periods short compared to those involved in the geodynamo.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 66
    Publication Date: 2014-07-23
    Description: We address the question of how the strength of protection for geographical indications (GIs) affects the GI industry's promotion incentives, equilibrium market outcomes, and the distribution of welfare. Geographical indication producers engage in informative advertising by associating their true quality premium (relative to a substitute product) with a specific label emphasizing the GI's geographic origin. The extent to which the names/words of the GI label can be used and/or imitated by competing products—which depends on the strength of GI protection—determines how informative the GI promotion messages can be. Consumers’ heterogeneous preferences (vis-à-vis the GI quality premium) are modeled in a vertically differentiated framework. Both the GI industry and the substitute product industry are assumed to be competitive (with free entry). The model is calibrated and solved for alternative parameter values. Results show that producers of the GI and of the lower-quality substitute good have divergent interests: GI producers are better off with full protection, whereas the substitute good's producers prefer intermediate levels of protection (but they never prefer zero protection because they benefit indirectly if the GI producers’ incentives to promote are preserved). For consumers and aggregate welfare, the preferred level of protection depends on the model's parameters, with an intermediate level of protection being optimal in many circumstances.
    Keywords: D23 - Organizational Behavior ; Transaction Costs ; Property Rights, L15 - Information and Product Quality ; Standardization and Compatibility, M37 - Advertising, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 67
    Publication Date: 2014-07-23
    Description: Retail-price pass-through is one of the most important issues facing manufacturers of consumer packaged goods. Although retailers tend to pass higher wholesale prices through to consumers quickly and completely, they often do not pass on trade promotions. Currently, asymmetric pass-through is commonly thought to result from retailers’ exercise of market power. Alternatively, it may be because of consumer search behavior and retailers’ competitive response. We test this theory using a panel threshold asymmetric error-correction model applied to wholesale and retail scanner data for ready-to-eat cereal for a number of retailers in the Los Angeles metropolitan market. We find that consumer search behavior contributes significantly to imperfect pass-through. By allowing pass-through to depend on market power and consumer search costs, we find results that are contrary to the conventional wisdom. Namely, market power causes retail prices to fall quickly and rise slowly, whereas consumer search behavior causes retail prices to rise quickly and fall slowly—precisely the "rockets and feathers" phenomenon.
    Keywords: C35 - Discrete Regression and Qualitative Choice Models, L13 - Oligopoly and Other Imperfect Markets, L66 - Food ; Beverages ; Cosmetics ; Tobacco ; Wine and Spirits, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0002-9092
    Electronic ISSN: 1467-8276
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 68
    Publication Date: 2014-07-23
    Description: Considering the drawback of existing global weighted mean temperature model, this paper uses 2006–2012 NCEP reanalysis data to establish global empirical model for mapping zenith wet delays onto precipitable water—GTm_N, takes the influence of half-year periodicity of Tm into account when modelling and estimate the initial phase of each cycle. In order to evaluate the precision of GTm_N, we use three different Tm data sets from the NCEP during 2013, 650 radiosonde stations and COSMIC occultation in 2011 to test this model. The results show that GTm_N has higher precision in both ocean and continental area in every moment of every day. The accuracy of GTm_N is higher than Bevis formulas and GTm_II models. In addition, the actual surface temperature is not required in GTm_N model, and it will have wide application in GPS meteorology.
    Keywords: Gravity, Geodesy and Tides
    Print ISSN: 0956-540X
    Electronic ISSN: 1365-246X
    Topics: Geosciences
    Published by Oxford University Press on behalf of The Deutsche Geophysikalische Gesellschaft (DGG) and the Royal Astronomical Society (RAS).
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 69
    Publication Date: 2014-08-01
    Description: The challenge presented by high-throughput sequencing necessitates the development of novel tools for accurate alignment of reads to reference sequences. Current approaches focus on using heuristics to map reads quickly to large genomes, rather than generating highly accurate alignments in coding regions. Such approaches are, thus, unsuited for applications such as amplicon-based analysis and the realignment phase of exome sequencing and RNA-seq, where accurate and biologically relevant alignment of coding regions is critical. To facilitate such analyses, we have developed a novel tool, RAMICS, that is tailored to mapping large numbers of sequence reads to short lengths (〈10 000 bp) of coding DNA. RAMICS utilizes profile hidden Markov models to discover the open reading frame of each sequence and aligns to the reference sequence in a biologically relevant manner, distinguishing between genuine codon-sized indels and frameshift mutations. This approach facilitates the generation of highly accurate alignments, accounting for the error biases of the sequencing machine used to generate reads, particularly at homopolymer regions. Performance improvements are gained through the use of graphics processing units, which increase the speed of mapping through parallelization. RAMICS substantially outperforms all other mapping approaches tested in terms of alignment quality while maintaining highly competitive speed performance.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 70
    Publication Date: 2012-08-08
    Description: Determining the taxonomic lineage of DNA sequences is an important step in metagenomic analysis. Short DNA fragments from next-generation sequencing projects and microbes that lack close relatives in reference sequenced genome databases pose significant problems to taxonomic attribution methods. Our new classification algorithm, RITA (Rapid Identification of Taxonomic Assignments), uses the agreement between composition and homology to accurately classify sequences as short as 50 nt in length by assigning them to different classification groups with varying degrees of confidence. RITA is much faster than the hybrid PhymmBL approach when comparable homology search algorithms are used, and achieves slightly better accuracy than PhymmBL on an artificial metagenome. RITA can also incorporate prior knowledge about taxonomic distributions to increase the accuracy of assignments in data sets with varying degrees of taxonomic novelty, and classified sequences with higher precision than the current best rank-flexible classifier. The accuracy on short reads can be increased by exploiting paired-end information, if available, which we demonstrate on a recently published bovine rumen data set. Finally, we develop a variant of RITA that incorporates accelerated homology search techniques, and generate predictions on a set of human gut metagenomes that were previously assigned to different ‘enterotypes’. RITA is freely available in Web server and standalone versions.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 71
    Publication Date: 2012-08-03
    Description: This study investigates the choice of quality by producer organisations (POs) in charge of defining product specifications for geographical indications. The model assumes that the PO chooses the quality level that maximises joint producer profits in anticipation of the competitive equilibrium that arises once quality is set. Using a fairly general variant of the vertical differentiation model and a flexible specification of production costs, we show that the PO has an incentive to supply quality in excess of the socially optimal level.
    Keywords: L15 - Information and Product Quality ; Standardization and Compatibility, L44 - Antitrust Policy and Public Enterprises, Nonprofit Institutions, and Professional Organization, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 72
    Publication Date: 2012-12-28
    Description: This paper investigates household expenditures on food away from home by type of facility. A system of expenditures at full-service restaurants, fast-food restaurants and other facilities are estimated with a multivariate sample selection procedure, for three types of households by household composition. Statistical significance of error correlations suggests endogeneity of sample selectivity and justifies estimation of the equations in a system to improve statistical efficiency. Differentiated effects of economic and demographic factors are found on expenditures at different facilities and across household types. Findings on the roles of economic and socio-demographic factors can inform marketing strategies and policy deliberations.
    Keywords: C31 - Cross-Sectional Models ; Spatial Models ; Treatment Effect Models, D12 - Consumer Economics: Empirical Analysis, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 73
    Publication Date: 2013-12-24
    Description: The investigation of hypothetical bias in choice experiments (CEs) has typically been conducted in the laboratory with only minimal field experiments. Moreover, it is common practice to provide an initial endowment (money or coupons) to respondents. In this research, we employed a between-subject CE in a supermarket with three different treatments (i.e. hypothetical, hypothetical with cheap talk script and real). With the ‘real’ treatment, we required respondents to use their own money when making a payment. The proposed incentive-compatible mechanism mimics as close as possible the everyday purchasing situation. Results confirm the presence of hypothetical bias and the mixed effectiveness of a cheap talk script.
    Keywords: D12 - Consumer Economics: Empirical Analysis, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 74
    Publication Date: 2014-02-11
    Description: The advances of high-throughput sequencing offer an unprecedented opportunity to study genetic variation. This is challenged by the difficulty of resolving variant calls in repetitive DNA regions. We present a Bayesian method to estimate repeat-length variation from paired-end sequence read data. The method makes variant calls based on deviations in sequence fragment sizes, allowing the analysis of repeats at lengths of relevance to a range of phenotypes. We demonstrate the method’s ability to detect and quantify changes in repeat lengths from short read genomic sequence data across genotypes. We use the method to estimate repeat variation among 12 strains of Arabidopsis thaliana and demonstrate experimentally that our method compares favourably against existing methods. Using this method, we have identified all repeats across the genome, which are likely to be polymorphic. In addition, our predicted polymorphic repeats also included the only known repeat expansion in A. thaliana , suggesting an ability to discover potential unstable repeats.
    Keywords: Computational Methods, Massively Parallel (Deep) Sequencing, Genomics
    Print ISSN: 0305-1048
    Electronic ISSN: 1362-4962
    Topics: Biology
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
  • 75
    Publication Date: 2013-12-24
    Description: Using experimental auctions carried out on apples in different European countries, this paper contributes to the assessment of consumer willingness to pay for the reduction of pesticides. We study several systems of good agricultural practices, possibly signalled to consumers, ranging from Integrated Pest Management certifications to organic production methods. The results show a relatively homogeneous behaviour of European consumers and reveal that improving the information on pesticide reduction may have unintended consequences. Results also suggest that taste characteristics and reference to a Protected Denomination of Origin should not be overlooked.
    Keywords: C91 - Laboratory, Individual Behavior, D12 - Consumer Economics: Empirical Analysis, Q01 - Sustainable Development, Q13 - Agricultural Markets and Marketing ; Cooperatives ; Agribusiness
    Print ISSN: 0165-1587
    Electronic ISSN: 1464-3618
    Topics: Agriculture, Forestry, Horticulture, Fishery, Domestic Science, Nutrition , Economics
    Location Call Number Expected Availability
    BibTip Others were also interested in ...
Close ⊗
This website uses cookies and the analysis tool Matomo. More information can be found here...