CS295J/Project Schedule:Steve: Difference between revisions

From VrlWiki
Jump to navigation Jump to search
Line 11: Line 11:
'''Nov 22''' -- <del>continued: Video captures of research assistants analyzing MRIs for QA (second session)</del> cont: Interviews (+recording) with domain scientists about QA workflow
'''Nov 22''' -- <del>continued: Video captures of research assistants analyzing MRIs for QA (second session)</del> cont: Interviews (+recording) with domain scientists about QA workflow


'''Nov 29''' -- Coding/analysis of <del>video captures</del> interview videos completed; will share findings  
'''Nov 29''' -- Coding/analysis of interview videos completed; will share findings  


'''Dec 6''' -- Figures for final paper created
'''Dec 6''' -- Figures for final paper created

Revision as of 02:03, 11 November 2011

How do analysts assess scan quality in MRI imaging? How does that inform tool design?

Project Schedule

Nov 1 -- Filling out the paper: Abstract, related work and intro drafted; outlines for results and methods. Update on data and contact with Win.

Nov 8 -- Test MRIs (hopefully 10) collected and organized; evaluation protocol completed

Nov 15 -- Video captures of research assistants analyzing MRIs for QA (first session) Interviews (+recording) with domain scientists about QA workflow

Nov 22 -- continued: Video captures of research assistants analyzing MRIs for QA (second session) cont: Interviews (+recording) with domain scientists about QA workflow

Nov 29 -- Coding/analysis of interview videos completed; will share findings

Dec 6 -- Figures for final paper created

Dec 13 -- Final draft, two-page extended abstract of research project

Report Drafts

Data and Resource Links

/research/graphics/users/steveg/data/brain_atlases
  • STILL NEED -- MRIs with artifacts that make them "poor quality"

Evaluation Protocol

Subjects: 4 research assistants in Win Gongvatana's group who have been previously trained in assessing scan quality.

Data: Each subject will receive a set of 10 T1-weighted MRI scans that have been shuffled in random order. 5 scans will be "normal quality" and taken from the component scans used to build UCLA's LONI Probabilistic Brain Atlas. The other 5 scans will contain artifacts that make them "poor quality". All will be of normal, human brains.

Method: We will use a think-aloud protocol and ask subjects to examine and score all scans in one session. Subjects will use whichever viewing tools they regularly use to assess scan quality. The observer will prompt each subject to explain what s/he is looking for in each image. Video of the subject will be recorded during this process, and the screen will be captured in order to do a post-hoc analysis of time spent per frame.

Questionnaire: After completing the tasks, subjects will be given a questionnaire with the following questions:

  1. High-quality images
    1. Please describe the properties or features of a scan that you would typically consider "high quality".
    2. Are these features obvious or subtle?
    3. How do you go about visually finding them?
  2. Low-quality images
    1. Please describe the properties or features of a scan that you would typically consider "poor quality".
    2. Are these features obvious or subtle?
    3. How do you go about visually finding them?
  3. Current QA process
    1. Which tools do you currently use to assess quality in MRI scans?
    2. How much time do you typically spend doing quality checks on an MRI scan?
    3. How many slices do you typically look at during this process? (Please give both the number and the percentage of total, if not all are examined.)
    4. Please estimate the ratio of "poor quality" scans to "high quality" scans you come across in your research.
    5. Are some artifacts more apparent in certain parts of the brain? If so, please explain.
    6. Do you ever just sample some slices from the scan rather than inspecting all of them? If so, please explain how many slices you usually sample and how you choose those samples.
  4. Improving the process
    1. Do you wish the QA process were faster to complete?
    2. Do you wish the QA process were easier to complete?
    3. What tools, real or imaginary, would you use to help your QA process if they were available?

Boards