Hacker News new | past | comments | ask | show | jobs | submit login

I think they are using those 3000 brain scans of people who progressed to Alzheimer's - questions arise: what part of that was the training data, what part was the test set, and did the sets overlap?



This is not a machine learning algorithm, so no training is required. The software processes every MRI scan (a full set acquired in one session) independently of other scans. The algorithm quantifies microscopic structural distortion at every voxel (3D pixel) of the MRI. Statistical models can be used on a population, but only to test the performance of the algorithm, not to train it.


I get you might not have officially trained a model, but similar model evaluation questions might apply to your diagnosis system.

For example, did you do any hand tuning of thresholds? Did you do this tuning while looking at the scans, or did you design it without looking at ANY of the mentioned 3000 scans - and are you are basing your 90% accuracy on diagnosing a sample which is how representative of the population at large?

Is there a further description of this 90% statistic - do you have precision / recall and some sort of description of possible biases in this 3000 sample set? Based on another answer, it sounds like half are alzheimer's and half are 'similar' but non-alzheimer's. I don't want to be flippant, but I am always sceptical of 'accuracy' in medical diagnostics where a simple diagnosis of 'false' is 95% accurate.

Edit: I see you answered this second section in another comment, nevermind.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: