Catal Image Vision API Tests
Claudia A Engel
Last updated: January 16, 2019
Intro
We randomly sampled 1000 images out of the ~138k that are currently accessioned into SDR and generated labels using the Google Vision API [GV]1 and the Clarifai Predict API [CP].
Code to generate GV labels is here: https://github.com/sul-dlss-labs/google-vision-ai
Catal project is here: http://catalhoyuk.com
SUL AI studio: https://sites.google.com/stanford.edu/sul-ai-studio/
SUL AI studio talk is here.
For various reasons we had some attrition and ended up with 766 images. Results shown here are based on those.
Many thanks to Peter Mangiafico.↩
0.1 Comments, ideas in random order
It could be useful for discoverability if the images are embedded in a much larger dataset, but not sure about within the image dataset itself.
Instead of looking for positive identification turn it around and look for things that are NOT excavation or artifacts. For example, ‘buildig’ gets us images of the shelter, so allows us to weed out relevant images(?).
Combine several labels to separate people (man, woman, person, group, …) from buildings (soil, wall, dirt, …) from X-finds etc.
Use GV_OCR to identify images with whiteboards, though the rendered text is not very usable.