New software creates dictionary for retrieving images
New software that responds to written questions by retrieving digital images has potentially broad application, ranging from helping radiologists compare mammograms to streamlining museum curators archiving of artwork, say the Penn State researchers who developed the technology.
Dr. James Z. Wang, assistant professor in Penn States School of Information Sciences and Technology and principal investigator, says the Automatic Linguistic Indexing of Pictures (ALIP) system first builds a pictorial dictionary, and then uses it for associating images with keywords. The new technology functions like a human expert who annotates or classifies terms.
“While the prototype is in its infancy, it has demonstrated great potential for use in biomedicine by reading x-rays and CT scans as well as in digital libraries, business, Web searches and the military,” said Wang, who holds the PNC Technologies Career Development Professorship at IST and also is a member of the Department of Computer Science and Engineering.
ALIP processes images the way people seem to. When we see a new kind of vehicle with two wheels, a seat and a handlebar, for instance, we recognize it as “a bicycle” from information about related images stored in our brains. ALIP has a similar bank of statistical models “learned” from analyzing image features.
The system is detailed in a paper, “Learning-based Linguistic Indexing of Pictures with 2-D MHMMs,” to be given today (Dec. 4) at the Association of Computing Machinerys (ACM) Multimedia Conference in Juan Les Pins, France. Co-author is Dr. Jia Li, Penn State assistant professor of statistics.
Unlike other content-based retrieval systems that compare features of visually similar images, ALIP uses verbal cues that range from simple concepts such as “flowers” and “mushrooms” to higher-level ones such as “rural” and “European.” ALIP also can classify images into a larger number of categories than other systems, thereby broadening the uses of image databases.
Other advantages include ALIPs abilities to be trained with a relatively large number of concepts simultaneously and with images that are not necessarily visually similar.
In one experiment, Wang and Li “trained” ALIP with 24,000 photographs found on 600 CD-ROMs, with each CD-ROM collection assigned keywords to describe its content. After “learning” these images, the computer then automatically created a dictionary of concepts such as “building,” “landscape,” and “European.” Statistical modeling enabled ALIP to automatically index new or unlearned images with the linguistic terms of the dictionary.
Wang tested that dictionary with 5,000 randomly selected images to see if the computer could provide meaningful keyword annotations for the new images. His conclusion: The more specific the query for an image, the higher the systems degree of accuracy in retrieving an appropriate image.
Wang and Li are using ALIP as part of a three-year National Science Foundation research project to develop digital imagery technologies for the preservation and cataloguing of Asian art and cultural heritages. This research aims to bypass or reduce the efforts in the labor-intensive creation and entry of manual descriptions or artwork.
Eventually, the system is expected to identify the discriminating features of Chinese landscape paintings and the distinguishing characteristics of paintings from different historical periods, Wang notes.
The researchers progress in the first year of that project is discussed in the paper, “Interdisciplinary Research to Advance Digital Imagery Indexing and Retrieval Technologies for Asian Art and Cultural Heritages.” The research will be presented on Dec. 6 at in a special session of ACMs Multimedia Conference in France.
Further research will be aimed at improving ALIPs accuracy and speed.
ALIPs reading of a beach scene with sailboats yielded the keyword annotations of “ocean,” “paradise,” “San Diego,” “Thailand,” “beach” and “fish.” Even though the computer was intelligent enough to recognize the high-level concept of “paradise,” additional research will focus on making the technology more accurate, so that San Diego and Thailand will not appear in the annotation of the same picture, Wang says.
“This system has the potential to change how we handle images in our daily life by giving us better and more access,” Wang says. Wang and Lis latest research builds on their earlier efforts at Stanford University. Sun Microsystems provided most of the equipment used in the project.
Media Contact
More Information:
http://www.psu.edu/All latest news from the category: Information Technology
Here you can find a summary of innovations in the fields of information and data processing and up-to-date developments on IT equipment and hardware.
This area covers topics such as IT services, IT architectures, IT management and telecommunications.
Newest articles
NASA: Mystery of life’s handedness deepens
The mystery of why life uses molecules with specific orientations has deepened with a NASA-funded discovery that RNA — a key molecule thought to have potentially held the instructions for…
What are the effects of historic lithium mining on water quality?
Study reveals low levels of common contaminants but high levels of other elements in waters associated with an abandoned lithium mine. Lithium ore and mining waste from a historic lithium…
Quantum-inspired design boosts efficiency of heat-to-electricity conversion
Rice engineers take unconventional route to improving thermophotovoltaic systems. Researchers at Rice University have found a new way to improve a key element of thermophotovoltaic (TPV) systems, which convert heat…