There are a total of 53 samples in this study. The count represents the number of “sample.ids” in the study
Specifically, in this study all assays for each sample have correlations above 0.80. Thus the assay level gene count data can be combined for each sample using roll up
This point forward the gene counts for the samples after the Lilly RNASeq rollup are investigated for outliers
| Sample.id | Sample warning reason | Additional Reason | sample.group_name | |
|---|---|---|---|---|
| 1 | G039-5 | greater than 30% mitochondrial content | Pretreatment | |
| 2 | G064-3 | greater than 30% mitochondrial content | During treatment of RAM monotherapy | |
| 3 | G064-4 | very biased base composition|greater than 80% mitochondrial content | PC outlier | During treatment of RAM monotherapy |
| 4 | G082-3 | PC outlier | During treatment of PTX plus RAM | |
| 5 | G082-4 | PC outlier | During treatment of PTX plus RAM | |
| 6 | G127-6 | very biased base composition|greater than 80% mitochondrial content | PC outlier | During treatment of PTX plus RAM |
| 7 | G132-2 | potential herpes virus infection|potential fungal contamination | PC outlier | PTX |
| 8 | G132-3 | PC outlier | During treatment of RAM monotherapy | |
| 9 | G132-6 | greater than 30% mitochondrial content | Post PTX RAM | |
| 10 | NCCHE-G057-2 | Patient is reported by investigator to be male but appears to be female | Pretreatment | |
| 11 | NCCHE-G057-3 | Patient is reported by investigator to be male but appears to be female | Post PTX RAM | |
| 12 | NCCHE-G057-5 | Patient is reported by investigator to be male but appears to be female | During treatment of RAM monotherapy | |
| 13 | NCCHE-G136 | PC outlier | Pretreatment | |
| 14 | NCCHE-G164-2 | PC outlier | During treatment of PTX plus RAM |
## character(0)
## Read the 52 x 50 data matrix successfully!
## Using no_dims = 2, perplexity = 2.000000, and theta = 0.500000
## Computing input similarities...
## Normalizing input...
## Building tree...
## - point 0 of 52
## Done in 0.00 seconds (sparsity = 0.161982)!
## Learning embedding...
## Iteration 50: error is 79.812387 (50 iterations in 0.01 seconds)
## Iteration 100: error is 78.237445 (50 iterations in 0.01 seconds)
## Iteration 150: error is 68.266009 (50 iterations in 0.00 seconds)
## Iteration 200: error is 72.253209 (50 iterations in 0.01 seconds)
## Iteration 250: error is 71.060945 (50 iterations in 0.00 seconds)
## Iteration 300: error is 3.046715 (50 iterations in 0.01 seconds)
## Iteration 350: error is 2.382132 (50 iterations in 0.01 seconds)
## Iteration 400: error is 1.912596 (50 iterations in 0.00 seconds)
## Iteration 450: error is 2.211615 (50 iterations in 0.01 seconds)
## Iteration 500: error is 2.010348 (50 iterations in 0.01 seconds)
## Fitting performed in 0.07 seconds.
## [1] "perplexity = 2, max_iter = 500, learning rate = 200"
## Read the 52 x 50 data matrix successfully!
## Using no_dims = 2, perplexity = 5.000000, and theta = 0.500000
## Computing input similarities...
## Normalizing input...
## Building tree...
## - point 0 of 52
## Done in 0.00 seconds (sparsity = 0.389793)!
## Learning embedding...
## Iteration 50: error is 66.111092 (50 iterations in 0.01 seconds)
## Iteration 100: error is 69.386717 (50 iterations in 0.00 seconds)
## Iteration 150: error is 65.174801 (50 iterations in 0.01 seconds)
## Iteration 200: error is 63.650356 (50 iterations in 0.01 seconds)
## Iteration 250: error is 69.414769 (50 iterations in 0.00 seconds)
## Iteration 300: error is 2.505615 (50 iterations in 0.01 seconds)
## Iteration 350: error is 2.130702 (50 iterations in 0.00 seconds)
## Iteration 400: error is 1.794287 (50 iterations in 0.01 seconds)
## Iteration 450: error is 1.830252 (50 iterations in 0.01 seconds)
## Iteration 500: error is 1.522245 (50 iterations in 0.00 seconds)
## Fitting performed in 0.06 seconds.
## [1] "perplexity = 5, max_iter = 500, learning rate = 200"
## Read the 52 x 50 data matrix successfully!
## Using no_dims = 2, perplexity = 10.000000, and theta = 0.500000
## Computing input similarities...
## Normalizing input...
## Building tree...
## - point 0 of 52
## Done in 0.01 seconds (sparsity = 0.723373)!
## Learning embedding...
## Iteration 50: error is 59.716732 (50 iterations in 0.00 seconds)
## Iteration 100: error is 58.419821 (50 iterations in 0.01 seconds)
## Iteration 150: error is 57.084640 (50 iterations in 0.01 seconds)
## Iteration 200: error is 59.159695 (50 iterations in 0.00 seconds)
## Iteration 250: error is 62.227891 (50 iterations in 0.01 seconds)
## Iteration 300: error is 1.441920 (50 iterations in 0.00 seconds)
## Iteration 350: error is 0.942671 (50 iterations in 0.01 seconds)
## Iteration 400: error is 0.900135 (50 iterations in 0.01 seconds)
## Iteration 450: error is 0.643039 (50 iterations in 0.00 seconds)
## Iteration 500: error is 0.429907 (50 iterations in 0.01 seconds)
## Fitting performed in 0.06 seconds.
## [1] "perplexity = 10, max_iter = 500, learning rate = 200"
## [[1]]
## character(0)
##
## [[2]]
## character(0)
##
## [[3]]
## character(0)