A Microscope That Teaches Itself to Know Best Settings for Diagnostic Images for Malaria

Engineers at Duke University have developed a microscope that adapts its lighting angles, colors and patterns while teaching itself the optimal settings needed to complete a given diagnostic task.

In the initial proof-of-concept study, the microscope simultaneously developed a lighting pattern and classification system that allowed it to quickly identify red blood cells infected by the malaria parasite more accurately than trained physicians and other machine learning approaches.

The results appear online on November 19 in the journal Biomedical Optics Express.

"A standard microscope illuminates a sample with the same amount of light coming from all directions, and that lighting has been optimized for human eyes over hundreds of years," said Roarke Horstmeyer, assistant professor of biomedical engineering at Duke.

A microscope with a disc on top and bowl-shaped lighting apparatus on bottom"But computers can see things humans can't," Hortmeyer said. "So not only have we redesigned the hardware to provide a diverse range of lighting options, we've allowed the microscope to optimize the illumination for itself."

Rather than diffusing white light from below to evenly illuminate the slide, the engineers developed a bowl-shaped light source with LEDs embedded throughout its surface. This allows samples to be illuminated from different angles up to nearly 90 degrees with different colors, which essentially casts shadows and highlights different features of the sample depending on the pattern of LEDs used.

The researchers then fed the microscope hundreds of samples of malaria-infected red blood cells prepared as thin smears, in which the cell bodies remain whole and are ideally spread out in a single layer on a microscope slide. Using a type of machine learning algorithm called a convolutional neural network, the microscope learned which features of the sample were most important for diagnosing malaria and how best to highlight those features.

The algorithm eventually landed on a ring-shaped LED pattern of different colors coming from relatively high angles. While the resulting images are noisier than a regular microscope image, they highlight the malaria parasite in a bright spot and are correctly classified about 90 percent of the time. Trained physicians and other machine learning algorithms typically perform with about 75 percent accuracy.

"The patterns it's picking out are ring-like with different colors that are non-uniform and are not necessarily obvious," said Horstmeyer. "Even though the images are dimmer and noisier than what a clinician would create, the algorithm is saying it'll live with the noise, it just really wants to get the parasite highlighted to help it make a diagnosis."

Horstmeyer then sent the LED pattern and sorting algorithm to another collaborator's lab across the world to see if the results were translatable to different microscope setups. The other laboratory showed similar successes.

"Physicians have to look through a thousand cells to find a single malaria parasite," said Horstmeyer. "And because they have to zoom in so closely, they can only look at maybe a dozen at a time, and so reading a slide takes about 10 minutes. If they only had to look at a handful of cells that our microscope has already picked out in a matter of seconds, it would greatly speed up the process."

The researchers also showed that the microscope works well with thick blood smear preparations, in which the red blood cells form a highly non-uniform background and may be broken apart. For this preparation, the machine learning algorithm was successful 99 percent of the time.

Two rows of stained cells. The top taken by a normal microscope makes it difficult to spot the malaria parasite. The bottom images light up the malaria parasite. According to Horstmeyer, the improved accuracy is expected because the tested thick smears were more heavily stained than the thin smears and exhibited higher contrast. But they also take longer to prepare, and part of the motivation behind the project is to cut down on diagnosis times in low-resource settings where trained physicians are sparse and bottlenecks are the norm.

With this initial success in hand, Horstmeyer is continuing to develop both the microscope and machine learning algorithm.

A group of Duke engineering graduate students has formed a startup company SafineAI to miniaturize the reconfigurable LED microscope concept, which has already earned a $120,000 prize at a local pitch competition.

Meanwhile, Horstmeyer is working with a different machine learning algorithm to create a version of the microscope that can adjust its LED pattern to any specific slide it's trying to read.

"We're basically trying to impart some brains into the image acquisition process," said Horstmeyer. "We want the microscope to use all of its degrees of freedom. So instead of just dumbly taking images, it can play around with the focus and illumination to try to get a better idea of what's on the slide, just like a human would."

"Learned Sensing: Jointly Optimized Microscope Hardware for Accurate Image Classification," Alex Muthumbi, Amey Chaware, Kanghyun Kim, Kevin C. Zhou, Pavan Chandra Konda, Richard Chen, Benjamin Judkewitz, Andreas Erdmann, Barbara Kappes, and Roarke Horstmeyer. Biomedical Optics Express, Vol. 10, No. 11, 1 November 2019. DOI: 10.1364/BOE.10.006351

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.