This repository has been archived by the owner on Sep 18, 2020. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 1
long success story
Grzegorz Mrukwa edited this page Feb 6, 2017
·
3 revisions
Covers full processing pipeline.
- Log in.
- No datasets are available: upload a dataset.
- Select uploaded dataset.
- Upload an optical image.
- Match digital data with optical image.
- Apply peak alignment using PAFFT method. *
- Select derived dataset with peaks aligned.
- Apply baseline removal. *
- Select derived dataset with baseline removed.
- Estimate and apply Gaussian Mixture Model. *
- Select derived GMM-modelled dataset.
- Perform DiviK with limit of 3 steps. *
- Select an automatically found ROI which visually best overlaps optically found tumor.
- Calculate coverage statistics.
- Create classifier distinguishing selected ROI and the rest of preparation. *
- Select GMM-modelled dataset.
- Perform DiviK compression. *
- Select compressed dataset.
- Load ROI from step 13.
- Use compressed data for classifier training. *
- Compare classifiers from steps 15 and 20.
- Upload second preparation.
- Repeat steps 3-9 for this preparation.
- Apply GMM model estimated for first preparation.
- Classify spectra using classifiers from steps 15 and 20 and score them.
- Mark tumor regions.
- Create classification report basing on automatically found regions and manually selected ROIs.
- Log out.
All steps marked with '*' involve scheduling the job, presenting some progress info and running it in the background thread. It would be perfect, to allow user to work on a dataset which is under construction and schedule this work after current calculations, e.g.:
- User wants to perform GMM modelling. This takes some time, while no one wants to wait until calculations end.
- To make user not leaving the tool, when GMM got scheduled, we create dataset entry (indicated as virtual dataset, non-existing yet).
- User selects this virtual dataset.
- No insight can be performed into this dataset, while user can schedule another analysis on the same data - e.g. DiviK
- User logs out and goes for a one week party / whatever.
- Calculations finish.
- User logs in after a break.
- Datasets after GMM and DiviK are fully operable, so user can select them and visualize the spectra / groups.