Revistas
Autores:
Bakr, S.; Brennan, K.; Mukherjee, P.; et al.
Revista:
CELL REPORTS. METHODS
ISSN:
2667-2375
Año:
2023
Vol.:
3
N°:
1
Págs.:
100392
Despite the abundance of multimodal data, suitable statistical models that can improve our understanding of diseases with genetic underpinnings are challenging to develop. Here, we present SparseGMM, a statistical approach for gene regulatory network discovery. SparseGMM uses latent variable modeling with sparsity constraints to learn Gaussian mixtures from multiomic data. By combining coexpression patterns with a Bayesian framework, SparseGMM quantitatively measures confidence in regulators and uncertainty in target gene assignment by computing gene entropy. We apply SparseGMM to liver cancer and normal liver tissue data and evaluate discovered gene modules in an independent single-cell RNA sequencing (scRNA-seq) dataset. SparseGMM identifies PROCR as a regulator of angiogenesis and PDCD1LG2 and HNF4A as regu-lators of immune response and blood coagulation in cancer. Furthermore, we show that more genes have significantly higher entropy in cancer compared with normal liver. Among high-entropy genes are key multifunctional components shared by critical pathways, including p53 and estrogen signaling.
Revista:
GLIA
ISSN:
0894-1491
Año:
2023
Vol.:
71
N°:
3
Págs.:
571 - 587
Inflammation is a common feature in neurodegenerative diseases that contributes to neuronal loss. Previously, we demonstrated that the basal inflammatory tone differed between brain regions and, consequently, the reaction generated to a pro-inflammatory stimulus was different. In this study, we assessed the innate immune reaction in the midbrain and in the striatum using an experimental model of Parkinson's disease. An adeno-associated virus serotype 9 expressing the alpha-synuclein and mCherry genes or the mCherry gene was administered into the substantia nigra. Myeloid cells (CD11b(+)) and astrocytes (ACSA2(+)) were purified from the midbrain and striatum for bulk RNA sequencing. In the parkinsonian midbrain, CD11b(+) cells presented a unique anti-inflammatory transcriptomic profile that differed from degenerative microglia signatures described in experimental models for other neurodegenerative conditions. By contrast, striatal CD11b(+) cells showed a pro-inflammatory state and were similar to disease-associated microglia. In the midbrain, a prominent increase of infiltrated monocytes/macrophages was observed and, together with microglia, participated actively in the phagocytosis of dopaminergic neuronal bodies. Although striatal microglia presented a phagocytic transcriptomic profile, morphology and cell density was preserved and no active phagocytosis was detected. Interestingly, astrocytes presented a pro-inflammatory fingerprint in the midbrain and a low number of differentially displayed transcripts in the striatum. During alpha-synuclein-dependent degeneration, microglia and astrocytes experience context-dependent activation states with a different contribution to the inflammatory reaction. Our results point towards the relevance of selecting appropriate cell targets to design neuroprotective strategies aimed to modulate the innate immune system during the active phase of dopaminergic degeneration.
Revista:
ELIFE
ISSN:
2050-084X
Año:
2023
Vol.:
12
Págs.:
e79363
Early hematopoiesis is a continuous process in which hematopoietic stem and progenitor cells (HSPCs) gradually differentiate toward specific lineages. Aging and myeloid malignant transformation are characterized by changes in the composition and regulation of HSPCs. In this study, we used single-cell RNA sequencing (scRNA-seq) to characterize an enriched population of human HSPCs obtained from young and elderly healthy individuals.
Based on their transcriptional profile, we identified changes in the proportions of progenitor compartments during aging, and differences in their functionality, as evidenced by gene set enrichment analysis. Trajectory inference revealed that altered gene expression dynamics accompanied cell differentiation, which could explain aging-associated changes in hematopoiesis. Next, we focused on key regulators of transcription by constructing gene regulatory networks (GRNs) and detected regulons that were specifically active in elderly individuals. Using previous findings in healthy cells as a reference, we analyzed scRNA-seq data obtained from patients with myelodysplastic syndrome (MDS) and detected specific alterations of the expression dynamics of genes involved in erythroid differentiation in all patients with MDS such as TRIB2. In addition, the comparison between transcriptional programs and GRNs regulating normal HSPCs and MDS HSPCs allowed identification of regulons that were specifically active in MDS cases such as SMAD1, HOXA6, POU2F2, and RUNX1 suggesting a role of these transcription factors (TFs) in the pathogenesis of the disease.
In summary, we demonstrate that the combination of single-cell technologies with computational analysis tools enable the study of a variety of cellular mechanisms involved in complex biological systems such as early hematopoiesis and can be used to dissect perturbed differentiation trajectories associated with perturbations such as aging and malignant transformation. Furthermore, the identification of abnormal regulatory mechanisms associated with myeloid malignancies could be exploited for personalized therapeutic approaches in individual patients.
Revista:
SCIENCE ADVANCES
ISSN:
2375-2548
Año:
2022
Vol.:
8
N°:
39
Págs.:
eabo0514
Identification of new markers associated with long-term efficacy in patients treated with CAR T cells is a current medical need, particularly in diseases such as multiple myeloma. In this study, we address the impact of CAR density on the functionality of BCMA CAR T cells. Functional and transcriptional studies demonstrate that CAR T cells with high expression of the CAR construct show an increased tonic signaling with up-regulation of exhaustion markers and increased in vitro cytotoxicity but a decrease in in vivo BM infiltration. Characterization of gene regulatory networks using scRNA-seq identified regulons associated to activation and exhaustion up-regulated in CARHigh T cells, providing mechanistic insights behind differential functionality of these cells. Last, we demonstrate that patients treated with CAR T cell products enriched in CARHigh T cells show a significantly worse clinical response in several hematological malignancies. In summary, our work demonstrates that CAR density plays an important role in CAR T activity with notable impact on clinical response.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2022
Vol.:
39
N°:
9
Págs.:
2488 - 2495
Motivation An important step in the transcriptomic analysis of individual cells involves manually determining the cellular identities. To ease this labor-intensive annotation of cell-types, there has been a growing interest in automated cell annotation, which can be achieved by training classification algorithms on previously annotated datasets. Existing pipelines employ dataset integration methods to remove potential batch effects between source (annotated) and target (unannotated) datasets. However, the integration and classification steps are usually independent of each other and performed by different tools. We propose JIND (joint integration and discrimination for automated single-cell annotation), a neural-network-based framework for automated cell-type identification that performs integration in a space suitably chosen to facilitate cell classification. To account for batch effects, JIND performs a novel asymmetric alignment in which unseen cells are mapped onto the previously learned latent space, avoiding the need of retraining the classification model for new datasets. JIND also learns cell-type-specific confidence thresholds to identify cells that cannot be reliably classified. Results We show on several batched datasets that the joint approach to integration and classification of JIND outperforms in accuracy existing pipelines, and a smaller fraction of cells is rejected as unlabeled as a result of the cell-specific confidence thresholds. Moreover, we investigate cells misclassified by JIND and provide evidence suggesting that they could be due to outliers in the annotated datasets or errors in the original approach used for annotation of the target batch.
Revista:
COMMUNICATIONS BIOLOGY
ISSN:
2399-3642
Año:
2022
Vol.:
5
N°:
1
Págs.:
351
Single-cell RNA-Sequencing has the potential to provide deep biological insights by revealing complex regulatory interactions across diverse cell phenotypes at single-cell resolution. However, current single-cell gene regulatory network inference methods produce a single regulatory network per input dataset, limiting their capability to uncover complex regulatory relationships across related cell phenotypes. We present SimiC, a single-cell gene regulatory inference framework that overcomes this limitation by jointly inferring distinct, but related, gene regulatory dynamics per phenotype. We show that SimiC uncovers key regulatory dynamics missed by previously proposed methods across a range of systems, both model and non-model alike. In particular, SimiC was able to uncover CAR T cell dynamics after tumor recognition and key regulatory patterns on a regenerating liver, and was able to implicate glial cells in the generation of distinct behavioral states in honeybees. SimiC hence establishes a new approach to quantitating regulatory architectures between distinct cellular phenotypes, with far-reaching implications for systems biology.
Revista:
NATURE COMMUNICATIONS
ISSN:
2041-1723
Año:
2022
Vol.:
13
N°:
1
Págs.:
7619
Myelodysplastic syndromes (MDS) are hematopoietic stem cell (HSC) malignancies characterized by ineffective hematopoiesis, with increased incidence in older individuals. Here we analyze the transcriptome of human HSCs purified from young and older healthy adults, as well as MDS patients, identifying transcriptional alterations following different patterns of expression. While aging-associated lesions seem to predispose HSCs to myeloid transformation, disease-specific alterations may trigger MDS development. Among MDS-specific lesions, we detect the upregulation of the transcription factor DNA Damage Inducible Transcript 3 (DDIT3). Overexpression of DDIT3 in human healthy HSCs induces an MDS-like transcriptional state, and dyserythropoiesis, an effect associated with a failure in the activation of transcriptional programs required for normal erythroid differentiation. Moreover, DDIT3 knockdown in CD34+ cells from MDS patients with anemia is able to restore erythropoiesis. These results identify DDIT3 as a driver of dyserythropoiesis, and a potential therapeutic target to restore the inefficient erythroid differentiation characterizing MDS patients. © 2022, The Author(s).
Revista:
PROCEEDINGS OF THE IEEE
ISSN:
0018-9219
Año:
2021
Vol.:
109
N°:
9
Págs.:
1607 - 1622
The development and progress of high-throughput sequencing technologies have transformed the sequencing of DNA from a scientific research challenge to practice. With the release of the latest generation of sequencing machines, the cost of sequencing a whole human genome has dropped to less than $600. Such achievements open the door to personalized medicine, where it is expected that genomic information of patients will be analyzed as a standard practice. However, the associated costs, related to storing, transmitting, and processing the large volumes of data, are already comparable to the costs of sequencing. To support the design of new and interoperable solutions for the representation, compression, and management of genomic sequencing data, the Moving Picture Experts Group (MPEG) jointly with working group 5 of ISO/TC276 "Biotechnology" has started to produce the ISO/IEC 23092 series, known as MPEG-G. MPEG-G does not only offer higher levels of compression compared with the state of the art but it also provides new functionalities, such as built-in support for random access in the compressed domain, support for data protection mechanisms, flexible storage, and streaming capabilities. MPEG-G only specifies the decoding syntax of compressed bitstreams, as well as a file format and a transport format. This allows for the development of new encoding solutions with higher degrees of optimization while maintaining compatibility with any existing MPEG-G decoder.
Revista:
GENOME RESEARCH
ISSN:
1088-9051
Año:
2021
Vol.:
31
N°:
4
Págs.:
576 - 591
The liver is a multifunctional organ critical for carrying out numerous biosynthetic, metabolic, and detoxification functions. Owing to its detoxification roles, the liver is frequently exposed to many hepatotoxins, resulting in tissue damage and cell death. Accordingly, it has evolved a unique ability to regenerate in response to a wide range of physical and toxic injuries (Diehl and Chute 2013), and mammalian livers can replenish up to 70% of The adult liver has an exceptional ability to regenerate, but how it maintains its specialized functions during regeneration is unclear. Here, we used partial hepatectomy (PHx) in tandem with single-cell transcriptomics to track cellular transitions and heterogeneities of ?22,000 liver cells through the initiation, progression, and termination phases of mouse liver regeneration. Our results uncovered that, following PHx, a subset of hepatocytes transiently reactivates an early-postnatal-like gene expression program to proliferate, while a distinct population of metabolically hyperactive cells appears to compensate for any temporary deficits in liver function. Cumulative EdU labeling and immunostaining of metabolic, portal, and central vein?specific markers revealed that hepatocyte proliferation after PHx initiates in the midlobular region before proceeding toward the periportal and pericentral areas. We further demonstrate that portal and central vein proximal hepatocytes retain their metabolically active state to preserve essential liver functions while midlobular cells proliferate nearby. Through combined analysis of gene regulatory networks and cell?cell interaction maps, we found that regenerating hepatocytes redeploy key developmental regulons, which are guided by extensive ligand-receptor-mediated signaling events between hepatocytes and nonparenchymal cells. Altogether, our study offers a detailed blueprint of the intercellular crosstalk and cellular reprogramming that balances the metabolic and proliferative requirements of a regenerating liver.
Autores:
Voges, J.; Paridaens, T.; Müntefering, F.; et al.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
MOTIVATION: In an effort to provide a response to the ever-expanding generation of genomic data, the International Organization for Standardization (ISO) is designing a new solution for the representation, compression and management of genomic sequencing data: the MPEG-G standard. This paper discusses the first implementation of an MPEG-G compliant entropy codec: GABAC. GABAC combines proven coding technologies, such as context-adaptive binary arithmetic coding, binarization schemes and transformations, into a straightforward solution for the compression of sequencing data.
RESULTS: We demonstrate that GABAC outperforms well-established (entropy) codecs in a significant set of cases and thus can serve as an extension for existing genomic compression solutions, such as CRAM.
AVAILABILITY: The GABAC library is written in C++. We also provide a command line application which exercises all features provided by the library. GABAC can be downloaded from https://github.com/mitogen/gabac.
SUPPLEMENTARY INFORMATION: Supplementary data, including a complete list of the funding mechanisms and acknowledgements, are available at Bioinformatics online.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2020
Vol.:
36
N°:
18
Págs.:
4810 - 4812
Motivation: Sequencing data are often summarized at different annotation levels for further analysis, generally using the general feature format (GFF) or its descendants, gene transfer format (GTF) and GFF3. Existing utilities for accessing these files, like gffutils and gffread, do not focus on reducing the storage space, significantly increasing it in some cases. We propose GPress, a framework for querying GFF files in a compressed form. GPress can also incorporate and compress expression files from both bulk and single-cell RNA-Seq experiments, supporting simultaneous queries on both the GFF and expression files. In brief, GPress applies transformations to the data which are then compressed with the general lossless compressor BSC. To support queries, GPress compresses the data in blocks and creates several index tables for fast retrieval. Results: We tested GPress on several GFF files of different organisms, and showed that it achieves on average a 61% reduction in size with respect to gzip (the current de facto compressor for GFF files) while being able to retrieve all annotations for a given identifier or a range of coordinates in a few seconds (when run in a common laptop). In contrast, gffutils provides faster retrieval but doubles the size of the GFF files. When additionally linking an expression file, we show that GPress can reduce its size by more than 68% when compared to gzip (for both bulk and single-cell RNA-Seq experiments), while still retrieving the information within seconds. Finally, applying BSC to the data streams generated by GPress instead of to the original file shows a size reduction of more than 44% on average.
Revista:
ANNUAL REVIEW OF BIOMEDICAL DATA SCIENCE
ISSN:
2574-3414
Año:
2019
Vol.:
2
Págs.:
19 - 37
Recently, there has been growing interest in genome sequencing, driven by advances in sequencing technology, in terms of both efficiency and affordability. These developments have allowed many to envision whole-genome sequencing as an invaluable tool for both personalized medical care and public health. As a result, increasingly large and ubiquitous genomic data sets are being generated. This poses a significant challenge for the storage and transmission of these data. Already, it is more expensive to store genomic data for a decade than it is to obtain the data in the first place. This situation calls for efficient representations of genomic information. In this review, we emphasize the need for designing specialized compressors tailored to genomic data and describe the main solutions already proposed. We also give general guidelines for storing these data and conclude with our thoughts on the future of genomic formats and compressors.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2019
Vol.:
35
N°:
15
Págs.:
2674 - 2676
Motivation
High-Throughput Sequencing technologies produce huge amounts of data in the form of short genomic reads, associated quality values and read identifiers. Because of the significant structure present in these FASTQ datasets, general-purpose compressors are unable to completely exploit much of the inherent redundancy. Although there has been a lot of work on designing FASTQ compressors, most of them lack in support of one or more crucial properties, such as support for variable length reads, scalability to high coverage datasets, pairing-preserving compression and lossless compression.
Results
In this work, we propose SPRING, a reference-free compressor for FASTQ files. SPRING supports a wide variety of compression modes and features, including lossless compression, pairing-preserving compression, lossy compression of quality values, long read compression and random access. SPRING achieves substantially better compression than existing tools, for example, SPRING compresses 195 GB of 25× whole genome human FASTQ from Illumina¿s NovaSeq sequencer to less than 7 GB, around 1.6× smaller than previous state-of-the-art FASTQ compressors. SPRING achieves this improvement while using comparable computational resources.
Revista:
BIOINFORMATICS
ISSN:
1367-4811
The affordability of DNA sequencing has led to the generation of unprecedented volumes of raw sequencing data. These data must be stored, processed, and transmitted, which poses significant challenges. To facilitate this effort, we introduce FaStore, a specialized compressor for FASTQ files. The proposed algorithm does not use any reference sequences for compression, and permits the user to choose from several lossy modes to improve the overall compression ratio, depending on the specific needs. We demonstrate through extensive simulations that FaStore achieves a significant improvement in compression ratio with respect to previously proposed algorithms for this task. In addition, we perform an analysis on the effect that the different lossy modes have on variant calling, the most widely used application for clinical decision making, especially important in the era of precision medicine. We show that lossy compression can offer significant compression gains, while preserving the essential genomic information and without affecting the variant calling performance.
Revista:
BRIEFINGS IN BIOINFORMATICS
ISSN:
1467-5463
Año:
2017
Vol.:
18
N°:
2
Págs.:
183 - 194
Recent advancements in sequencing technology have led to a drastic reduction in genome sequencing costs. This development has generated an unprecedented amount of data that must be stored, processed, and communicated. To facilitate this effort, compression of genomic files has been proposed. Specifically, lossy compression of quality scores is emerging as a natural candidate for reducing the growing costs of storage. A main goal of performing DNA sequencing in population studies and clinical settings is to identify genetic variation. Though the field agrees that smaller files are advantageous, the cost of lossy compression, in terms of variant discovery, is unclear.
Bioinformatic algorithms to identify SNPs and INDELs use base quality score information; here, we evaluate the effect of lossy compression of quality scores on SNP and INDEL detection. Specifically, we investigate how the output of the variant caller when using the original data differs from that obtained when quality scores are replaced by those generated by a lossy compressor. Using gold standard genomic datasets and simulated data, we are able to analyze how accurate the output of the variant calling is, both for the original data and that previously lossily compressed. We show that lossy compression can significantly alleviate the storage while maintaining variant calling performance comparable to that with the original data. Further, in some cases lossy compression can lead to variant calling performance tha
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2016
Vol.:
32
N°:
7
Págs.:
1115 - 1117
Motivation: Data compression is crucial in effective handling of genomic data. Among several recently published algorithms, ERGC seems to be surprisingly good, easily beating all of the competitors.
Results: We evaluated ERGC and the previously proposed algorithms GDC and iDoComp, which are the ones used in the original paper for comparison, on a wide data set including 12 assemblies of human genome (instead of only four of them in the original paper). ERGC wins only when one of the genomes (referential or target) contains mixed-cased letters (which is the case for only the two Korean genomes). In all other cases ERGC is on average an order of magnitude worse than GDC and iDoComp.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2016
Vol.:
32
N°:
17
Págs.:
i479 - i486
Motivation
The dramatic decrease in the cost of sequencing has resulted in the generation of huge amounts of genomic data, as evidenced by projects such as the UK10K and the Million Veteran Project, with the number of sequenced genomes ranging in the order of 10 K to 1 M. Due to the large redundancies among genomic sequences of individuals from the same species, most of the medical research deals with the variants in the sequences as compared with a reference sequence, rather than with the complete genomic sequences. Consequently, millions of genomes represented as variants are stored in databases. These databases are constantly updated and queried to extract information such as the common variants among individuals or groups of individuals. Previous algorithms for compression of this type of databases lack efficient random access capabilities, rendering querying the database for particular variants and/or individuals extremely inefficient, to the point where compression is often relinquished altogether.
Results
We present a new algorithm for this task, called GTRAC, that achieves significant compression ratios while allowing fast random access over the compressed database. For example, GTRAC is able to compress a Homo sapiens dataset containing 1092 samples in 1.1¿GB (compression ratio of 160), while allowing for decompression of specific samples in less than a second and decompression of specific variants in 17¿ms. GTRAC uses and adapts techniques from information theory,
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2015
Vol.:
31
N°:
19
Págs.:
3122 - 3129
Motivation
Recent advancements in sequencing technology have led to a drastic reduction in the cost of sequencing a genome. This has generated an unprecedented amount of genomic data that must be stored, processed and transmitted. To facilitate this effort, we propose a new lossy compressor for the quality values presented in genomic data files (e.g. FASTQ and SAM files), which comprise roughly half of the storage space (in the uncompressed domain). Lossy compression allows for compression of data beyond its lossless limit.
Results
The proposed algorithm QVZ exhibits better rate-distortion performance than the previously proposed algorithms, for several distortion metrics and for the lossless case. Moreover, it allows the user to define any quasi-convex distortion function to be minimized, a feature not supported by the previous algorithms. Finally, we show that QVZ-compressed data exhibit better performance in the genotyping than data compressed with previously proposed algorithms, in the sense that for a similar rate, a genotyping closer to that achieved with the original quality values is obtained.
Revista:
BIOINFORMATICS
ISSN:
1367-4803
Año:
2015
Vol.:
31
N°:
5
Págs.:
626 - 633
Motivation:With the release of the latest next-generation sequencing (NGS) machine, the HiSeq Xby Illumina, the cost of sequencing a Human has dropped to a mere $4000. Thus we are approach-ing a milestone in the sequencing history, known as the $1000 genome era, where the sequencingof individuals is affordable, opening the doors to effective personalized medicine. Massive gener-ation of genomic data, including assembled genomes, is expected in the following years. There iscrucial need for compression of genomes guaranteed of performing well simultaneously on differ-ent species, from simple bacteria to humans, which will ease their transmission, dissemination andanalysis. Further, most of the new genomes to be compressed will correspond to individuals ofa species from which a reference already exists on the database. Thus, it is natural to proposecompression schemes that assume and exploit the availability of such references.Results:We propose iDoComp, a compressor of assembled genomes presented in FASTA formatthat compresses an individual genome using a reference genome for both the compression and thedecompression. In terms of compression efficiency, iDoComp outperforms previously proposed al-gorithms in most of the studied cases, with comparable or better running time. For example, we ob-serve compression gains of up to 60% in several cases, includingH.sapiensdata, when comparingwith the best compression performance among the previously proposed algorithms.
Revista:
JOURNAL OF BIOINFORMATICS AND COMPUTATIONAL BIOLOGY
ISSN:
0219-7200
With the release of the latest Next-Generation Sequencing (NGS) machine, the HiSeq X by Illumina, the cost of sequencing the whole genome of a human is expected to drop to a mere $1000. This milestone in sequencing history marks the era of affordable sequencing of individuals and opens the doors to personalized medicine. In accord, unprecedented volumes of genomic data will require storage for processing. There will be dire need not only of compressing aligned data, but also of generating compressed files that can be fed directly to downstream applications to facilitate the analysis of and inference on the data. Several approaches to this challenge have been proposed in the literature; however, focus thus far has been on the low coverage regime and most of the suggested compressors are not based on effective modeling of the data.
We demonstrate the bene fit of data modeling for compressing aligned reads. Specifically, we show that, by working with data models designed for the aligned data, we can improve considerably over the best compression ratio achieved by previously proposed algorithms. Our results indicate that the pareto-optimal barrier for compression rate and speed claimed by Bon field and Mahoney (2013) [Bon field JK and Mahoneys MV, Compression of FASTQ and SAM format sequencing data, PLOS ONE, 8(3): e59190, 2013.] does not apply for high coverage aligned data. Furthermore, our improved compression ratio is achieved by splitting the data in a manner conducive to o
Revista:
IEEE COMMUNICATIONS LETTERS
ISSN:
1089-7798
Año:
2010
Vol.:
14
N°:
9
Págs.:
794 - 796
In this paper, we design a new energy allocation strategy for non-uniform binary memoryless sources encoded by Low-Density Parity-Check (LDPC) codes and sent over Additive White Gaussian Noise (AWGN) channels. The new approach estimates the a priori probabilities of the encoded symbols, and uses this information to allocate more energy to the transmitted symbols that occur less likely. It can be applied to systematic and non-systematic LDPC codes, improving in both cases the performance of previous LDPC based schemes using binary signaling. The decoder introduces the source non-uniformity and estimates the source symbols by applying the SPA (Sum Product Algorithm) over the factor graph describing the code.
Revista:
IEEE COMMUNICATIONS LETTERS
ISSN:
1089-7798
Año:
2010
Vol.:
14
N°:
4
Págs.:
336 - 338
This letter proposes a novel one-layer coding/shaping scheme with single-level codes and sigma-mapping for the bandwidth-limited regime. Specifically, we consider non-uniform memoryless sources sent over AWGN channels. At the transmitter, binary data are encoded by a Turbo code composed of two identical RSC (Recursive Systematic Convolutional) encoders. The encoded bits are randomly interleaved and modulated before entering the sigma-mapper. The modulation employed in this system follows the unequal energy allocation scheme first introduced in [1]. The receiver consists of an iterative demapping/decoding algorithm, which incorporates the a priori probabilities of the source symbols. To the authors' knowledge, work in this area has only been done for the power-limited regime. In particular, the authors in [2] proposed a scheme based on a Turbo code with RSC encoders and unequal energy allocation. Therefore, it is reasonable to compare the performance - with respect to the Shannon limit - of our proposed bandwidth-limited regime scheme with this former power-limited regime scheme. Simulation results show that our performance is as good or slightly better than that of the system in [2].