SlideShare uma empresa Scribd logo
1 de 52
Baixar para ler offline
Exploring Machine Learning (ML) for
Libraries & Archives: Present & Future
Bohyun Kim
Chief Technology Officer & Professor, University of Rhode Island, USA
Twitter: @bohyunkim | Web: http://www.bohyunkim.net/blog/
Bite-sized Internet Librarian International 2021, September 22, 2021.
Machine Learning (ML)
Image from : https://labelbox.com/image-segmentation-overview
Images from : https://towardsdatascience.com/what-is-machine-
learning-a-short-note-on-supervised-unsupervised-semi-
supervised-and-aed1573ae9bb
Images from: https://towardsdatascience.com/the-mostly-
complete-chart-of-neural-networks-explained-3fb6f2367464
Discussing ML for Libraries/Archives
Images from : https://medium.com/@dhartidhami/machine-learning-basics-model-cost-function-and-gradient-descent-79b69ff28091 ;
https://www.analyticsvidhya.com/blog/2020/02/cnn-vs-rnn-vs-mlp-analyzing-3-types-of-neural-networks-in-deep-learning/
Can ML be a useful tool for the work of
libraries & archives?
Image from : https://dancohen.org/2019/09/17/humane-ingenuity-3-ai-in-the-archives/
Q. What would practical application of ML for
libraries/archives look like?
Things I will be going over today:
Q. What would practical application of ML
for libraries/archives look like?
I. ML for Expediting archival processing
II. ML for Expanding descriptive
metadata
III. ML for Enhancing metadata
IV. ML for Full content extraction
Image from: https://www.analyticsvidhya.com/blog/2020/02/cnn-vs-rnn-vs-mlp-analyzing-3-types-of-neural-networks-in-deep-learning/
I. ML for Expediting Archival Processing
• “Processing” refers to the arrangement, description, and housing of
materials for storage and use.
• Unprocessed materials pose an impediment to researchers because
they cannot be easily identified, accessed, and used.
• Machine learning can expedite the processing of archival materials.
https://projectaida.org/
The AIDA Project Explored:
• Document Segmentation;
• Graphic Element Classification &
Text Extraction;
• Document Type Classification;
• Digitization Type Differentiation;
• Document Image Quality
Assessment; and
• Document Clustering.
Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics,
and Augmented Description: A Demonstration Project” (University of
Nebraska - Lincoln, January 10, 2020),
https://digitalcommons.unl.edu/libraryscience/396.
(a) Identify/Classify Graphical Content & Pull
Text with Image Zoning and Segmentation
• The AIDA team tried to identify and classify graphical content (e.g. figures,
illustrations, and cartoons) and extract text in historical newspaper page
images, using ML’s image zoning and segmentation (U-NeXt).
• 2 sets of newspaper images were used for ML model training/evaluation.
• One set from the Library of Congress’s Beyond Words project, which
crowdsourced the graphical content zone demarcation and the text
transcription of historical newspaper images.
• 1,532 images and corresponding data from graphical content zones; 1,226 images used
for training and 306 for evaluation.
• The other set from the Europeana Newspapers Project.
• A set of 481 page images already zoned and segmented, with the segments classed as
background, text, figure, separator or table.
Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University
of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
What is Computer Vision(CV)?
• Computer vision is a field of artificial intelligence (AI) that enables
computers and systems to derive meaningful information from digital
images, videos and other visual inputs — and take actions or make
recommendations based on that information.
Computer Vision Concepts
• Image filtering, matching, segmentation, and recognition
• Segmentation divides whole images into pixel groupings which can
then be labelled and classified.
Images from : https://ai.stanford.edu/~syyeung/cvweb/tutorial1.html
Tasks Performed by Computer Vision
• Image classification
• Object detection
• Object tracking
• Content-based image retrieval
• Semantic segmentation
• Instance segmentation
(b) Natural Image / Document Classification
• The AIDA project team applied natural image and document classification
methods in machine learning to a collection of minimally processed manuscript
materials to see whether handwritten, printed, and mixed (both handwritten and
printed) documents can be distinguished from one another.
• The team generated their own model by using the VGG method with 16
categories (VGG-16) pre-trained on the RVL_CDIP dataset.
• 999 images were used in total.
Figure 9. Prediction failure cases. In the left example, the model classified the document as handwritten rather than
Mixed. Note that the printed regions are very small compared to the handwritten content in the image. In the right
example, the model classified the document as printed rather than mixed. Here, the handwritten region is very small
compared to the printed region in the image.
Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University
of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
(c) Detecting Digitization Type
• The AIDA project team also applied machine learning
techniques to both classifying manuscript collection
images as digitized from original vs. from microform
and analyzing their image quality.
• 36,103 images were used as the ground truth
classification
Figure 13. A digital image of a photograph, digitized from the original item, and classified
by the classifier as having been digitized from a microform source.
Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University
of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
Benefits of ML for Archival Processing
• There is no doubt that being able to identify graphical content contained in
archival materials and to peruse the text in them would make browsing and
navigating those materials a lot more manageable to users. And if ML can process
archival materials this way, it would surely improve their discoverability and
usefulness.
• Automating the classification of materials as handwritten, printed, and mixed can
surely save archivists’ and librarians’ time spent on performing such classification
manually.
• Automatically assessing the quality of images and assigning their original source
type can help archivists and librarians identify materials at higher-risk and take
appropriate preservation measures.
Source: Bohyun Kim, “Machine Learning for Libraries and Archives.” Online Searcher 45(1): 39-41, 2021.
Two Challenges
• The results from these explorations indicate that these machine
learning methods are not ready to be immediately put to use in
archival processing.
• “Training a classifier to learn information from rare classes is very hard.”
• The project report highlights two challenges:
• (a) general scarcity of training data sets;
• (b) overall difficulty in obtaining or establishing ground truth (= the expected
result against which the performance of a machine learning model is
measured).
Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University
of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
Need for More Data and Ground Truth
• Without a sufficiently large amount of training data, a machine
learning model for libraries/archives cannot be created.
• Without ground-truth datasets, it is difficult to tell how well a
machine learning model for libraries/archives works.
II. ML for Expanding Descriptive Metadata
• Mass digitization projects created many digital objects of historical
materials, and many more born-digital items are being added to
archival collections.
• Unfortunately, libraries and archives lack resources to manually create
rich metadata for all digital objects in a timely manner.
• Lacking such metadata, those digital objects are not easily discovered,
identified, or navigated.
• ML can advance the discovery, navigation, and clustering of these
materials by generating more detailed metadata.
(a) Image Classifier ML Algorithm
for the Frick Collection
• The Frick Art Reference Library in New York together
with Stanford University, Cornell University, and the
University of Toronto developed an image classifier
algorithm that applies a local classification system
based on visual elements to the library’s digitized
Photoarchive.
• Tens of thousands of already-labeled images &
hundreds of thousands of images that require labeling.
• Hierarchical classification headings:
• “American School: Portraits: Men: Without hands
(Without hats): Head to left”
Process
• The Cornell/Toronto/Stanford team modified the popular ResNet152
network34—with 152 layers and six million parameters—and trained
it on a collection of 45,857 classified reproductions of American
paintings provided by the Frick Art Reference Library’s Photoarchive.
• The majority were portraits and represented 98 unique classification
headings.
• After training, the network was fed images from the unlabeled
portion of the library’s Photoarchive, and the network predicted a
classification heading for each one.
Source: Ellen Prokop et al., “AI and the Digitized Photoarchive: Promoting Access and Discoverability,” Art Documentation:
Journal of the Art Libraries Society of North America 40, no. 1 (March 1, 2021): 1–20, https://doi.org/10.1086/714604.
Vetting & Prospect
• These images were then annotated with the
predictions and shown to the library’s
photoarchivists through an application to vet the
predictions.
• The library staff vetted 8,661 images in the year
August 2019 - August 2020 and agreed with the
network on the classification headings applied to
5,829 (67 %) of the images.
• Yet even the incorrect predictions were, in
general, “almost correct” with only one tag
incorrect or missing.
• Holds the potential to improve the staff’s
efficiency in in metadata creation and image
retrieval.
(b) CAMPI (Computer-Aided Metadata
Generation for Photo archives Initiative)
https://www.cmu.edu/news/stories/archives/2020/october/computer-vision-archive.html
Expanding Item-Level Description Metadata
• The CAMPI project team built a prototype application that uses
computer vision (CV), in order to aid archivists and metadata editors
for Carnegie Mellon University Archives’ General Photograph
Collection (GPC) of +20,000 images with metadata at some hierarchy
(e.g. job; directory).
• This collection of photographic prints, 35mm negatives, and born
digital images was specifically chosen for the project, because:
• those images lack item-level description and
• are stored in and accessed through a shared network drive by archivists and
metadata editors, with no browsing interface.
Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie
Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer-
Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
Visual Similarity Search Powered by an “Off-
the-shelf” Image Search Model
• The CAMPI prototype application offered archivists and metadata
editors visual similarity search using CV techniques in ML.
• For this visual similarity search capability, a pre-trained “off-the-shelf”
image search model was used.
• The team used the second-to-last pooled layer of the InceptionV3
trained neural network as the feature source. For each image, this
produced an array of 512 numbers, allowing them to treat the entire
collection of images as a set of points in 512-dimensional space.
• Then, the team used an approximate nearest neighbor search
algorithm to create an index of that feature space, allowing it to be
queried.
The prototype allows staff to search similar
images to reduce repeated metadata
generation and remove noise from the eventual
public discovery interface.
Speeding Up Tagging
• The application provided a browsing and CV- assisted tagging
interface faceted by the existing job and directory metadata
associated with the photographs.
• Using this application, archivists and metadata editors were able to
start with a tag, find a seed photo, and retrieve similar photos via
visual similarity search that specifies the number of neighbors
requested.
• Then, they were able to quickly tag those retrieved photos with tags
from the seed photo when appropriate.
==---==
ML for Metadata Workflow
• In their white paper, the CAMPI team reported that of the more than
43,000 tagging decisions made, a little over 1 in 5 were able to be
automatically distributed across a set, saving metadata editors more
than 9,000 decisions.
• They also noted that the prototype application identified around 28% of
the collection as sets with duplicates.
• This prototype had the effect of rapidly streamlining and speeding up
item-level metadata creation.
• These results demonstrate how machine learning can be integrated into
the existing metadata creation and editing workflow at libraries and
archives.
Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie
Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer-
Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
Image Browsing/Filtering UI
• What is interesting about the CAMPI project is that even though the use of
computer vision was minimal in their prototype application, this project
produced a successful outcome because it combined machine learning with a
well-designed user interface that took advantage of the existing archival
structure, thereby effectively augmenting archivists’ and metadata experts’
existing work.
• At libraries and archives, a browsing and filtering interface is primarily created
to provide patrons with public access to archival materials.
• However, such an interface significantly increases the overall productivity of
metadata-related work, if it fits in well with the way the staff perform their
work.
Need for More Training Sets
• “There is a field-wide need for specialized computer vision training sets based
on historical, nonborn-digital photograph collections.”
Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie
Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer-
Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
III. ML for Enhancing Metadata
• The Audiovisual Metadata Platform Pilot Development (AMPPD)
Project is run by Indiana University Libraries, University of Texas at
Austin School of Information, New York Public Library, and AVP, a data
management solutions and consulting company.
• Started in 2018 and continuing.
• The AMPPD team intends to create automated metadata generation
mechanisms and integrate them with the human metadata
generation process.
• The goal is to generate and manage metadata for audiovisual
materials at scale for libraries and archives.
https://wiki.dlib.in
diana.edu/pages/v
iewpage.action?pa
geId=531699941
(a) Generate and Manage Rich Metadata for
Audiovisual Materials at Scale
• AMPPD Project’s Priority Tasks
• First: Detecting silence, speech, and music; speech-to-text /
speaker diarization; named entity recognition, such as people,
geographic locations, and topics;
• Next: Video and structured data OCR (metadata, transcript,
subject terms); detecting scene and shot; detecting music genre in
audiovisual materials.
Source: Shawn Avercamp and Julie Hardesty, “AI Is Such a Tool: Keeping Your Machine Learning Outputs in
Check” (Code4Lib 2020 Conference, Pittsburgh, PA, March 11, 2020), https://2020.code4lib.org/talks/AI-is-
such-a-tool-Keeping-your-machine-learning-outputs-in-check.
Annotating AV Materials
• It is easy to see that their first-priority items alone will annotate
audiovisual materials in such a way that library and archive users can
quickly identify the exact content they are looking for and better
grasp what kinds of content each audiovisual material contains.
• If the AMP can automatically and reliably extract broad subject terms
by recognizing named entities within the content, that will also
greatly enhance the discoverability and accessibility of those
materials.
ML for Better Discovery, Identification, and
Navigation of AV Materials
• Extracting the full content of digital AV objects as machine-accessible
and interpretable data is a big step towards enabling better discovery,
identification, and navigation, because it makes full-text query
possible.
(b) Summarization of Textual Documents
• The Catholic Pamphlets collection of over 5,500 PDFs at the University
of Notre Dame had sparse summaries. While half of the collection
had a MARC summary field, most of these summaries were a few
words at most.
• 3 NLP (Natural Language Processing) automated summarization
techniques of ML were tested to create more robust summaries by
combining the summaries with LC subject headings.
• 4 NLP algorithms in ML were used: Rapid Automatic Keyword
Extraction (RAKE); a raw TextRank implementation algorithm; Genism;
Bert Extractive Summarizer (BERT; uses DL)
Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International
Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
Process & Result
• Of the 5,500 texts, 51 had sufficient summaries already, and another
50 summaries were written.
• The automated summaries were generated after feeding the
pamphlets as .pdf files into an OCR pipeline.
• Computer-generated summaries of the pamphlets were compared
against human-written summaries.
• The BERT Extractive technique yielded the best score - an average
Rouge F1 score of 0.239. The Gensim python package implementation
of TextRank scored at .151. A hand-implemented TextRank algorithm
created summaries that scored at 0.144.
Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International
Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International
Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
Challenges in Summarization & Prospect
• Latin text and OCR errors have persisted despite choosing more
sophisticated algorithms and better text preprocessing.
• Better OCR correction & the exclusion of Latin text and sentence
fragments remain as challenges.
• “For collections in which their text does not need OCR and all appears
in English, the BERT Extractive Summarizer likely could provide
summaries with few enough errors that catalogers would be
comfortable adding the computed-generated summaries with
minimal human oversight.”
Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International
Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
IV. ML for Full Content Extraction
• Archival materials are given a set of metadata either at the item,
collection, or some other unit level, so that they can be searched
online.
• But their full content is not always readily available in digital form as
machine-Interpretable data. This hinders their discoverability and use,
such as running full-text or full-content searches and performing
other computational analyses.
• Machine learning can be utilized for content extraction.
ML for text, auditory, visual, and tabular data
• OCR (an ML method of image recognition) for text documents.
• Auditory data can be extracted using machine learning’s speech-
recognition method.
• From materials such as photographs and video recordings, visual
information can be extracted by image recognition techniques.
ML for text, auditory, visual, and tabular data
• Tabular data in digitized images are highly problematic since they
cannot be easily identified and utilized as structured data.
• But they can also be extracted using table detection, cell recognition,
and text extraction algorithms in machine learning.
• Extracting tabular data along with their data structure will make
historical data immediately useful to researchers.
Ryan Cordell, “Machine Learning + Libraries” (Library of Congress, July 14, 2020),
https://labs.loc.gov/static/labs/work/reports/Cordell-LOC-ML-report.pdf?loclr=blogsig.
Future Focuses for Libraries/Archives
• Need for more data & ground truth for ML algorithm training and
evaluation
• Exploration of pre-trained off-the-shelf ML models
• e.g. AWS Rekognition (aws.amazon.com/rekognition) and Google Cloud Vision
AI (cloud.google.com/vision).
• Need for continued digitization
Thank you!
Bohyun Kim
Chief Technology Officer & Professor, University of Rhode Island, USA
Twitter: @bohyunkim | Web: http://www.bohyunkim.net/blog/
Bite-sized Internet Librarian International 2021, September 22, 2021.

Mais conteúdo relacionado

Mais procurados

Z39.50: Information Retrieval protocol ppt
Z39.50: Information Retrieval protocol pptZ39.50: Information Retrieval protocol ppt
Z39.50: Information Retrieval protocol pptSUNILKUMARSINGH
 
Introduction to Metadata
Introduction to MetadataIntroduction to Metadata
Introduction to MetadataJenn Riley
 
Bibliographic Control and Oclc
Bibliographic Control and OclcBibliographic Control and Oclc
Bibliographic Control and OclcDenise Garofalo
 
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Web
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic WebRDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Web
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Webrobin fay
 
Canon of classification
Canon of classificationCanon of classification
Canon of classificationavid
 
Descriptive cataloging: Overview
Descriptive cataloging:  OverviewDescriptive cataloging:  Overview
Descriptive cataloging: OverviewJohan Koren
 
FRBR, FRAD and RDA I don't speak cataloging why should I care
FRBR, FRAD and RDA   I don't speak cataloging why should I careFRBR, FRAD and RDA   I don't speak cataloging why should I care
FRBR, FRAD and RDA I don't speak cataloging why should I careDeann Trebbe
 
RDA (Resource Description & Access)
RDA (Resource Description & Access)RDA (Resource Description & Access)
RDA (Resource Description & Access)Jennifer Joyner
 
Common communication format
Common communication formatCommon communication format
Common communication formatavid
 
Cataloging with RDA: An Overview
Cataloging with RDA: An OverviewCataloging with RDA: An Overview
Cataloging with RDA: An OverviewEmily Nimsakont
 
Introduction to subject cataloguing
Introduction to subject cataloguingIntroduction to subject cataloguing
Introduction to subject cataloguingLiah Shonhe
 
Library automation history Anandraj.L
Library automation history Anandraj.LLibrary automation history Anandraj.L
Library automation history Anandraj.Lanujessy
 

Mais procurados (20)

MARC21
MARC21MARC21
MARC21
 
ISBD
ISBDISBD
ISBD
 
Z39.50: Information Retrieval protocol ppt
Z39.50: Information Retrieval protocol pptZ39.50: Information Retrieval protocol ppt
Z39.50: Information Retrieval protocol ppt
 
Introduction to Metadata
Introduction to MetadataIntroduction to Metadata
Introduction to Metadata
 
Bibliographic Control and Oclc
Bibliographic Control and OclcBibliographic Control and Oclc
Bibliographic Control and Oclc
 
Canons of library classification
Canons of library classificationCanons of library classification
Canons of library classification
 
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Web
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic WebRDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Web
RDA Intro - AACR2 / MARC> RDA / FRBR / Semantic Web
 
Canon of classification
Canon of classificationCanon of classification
Canon of classification
 
Descriptive cataloging: Overview
Descriptive cataloging:  OverviewDescriptive cataloging:  Overview
Descriptive cataloging: Overview
 
FRBR, FRAD and RDA I don't speak cataloging why should I care
FRBR, FRAD and RDA   I don't speak cataloging why should I careFRBR, FRAD and RDA   I don't speak cataloging why should I care
FRBR, FRAD and RDA I don't speak cataloging why should I care
 
RDA (Resource Description & Access)
RDA (Resource Description & Access)RDA (Resource Description & Access)
RDA (Resource Description & Access)
 
CANONS OF CATALOGUING ppt
CANONS OF CATALOGUING pptCANONS OF CATALOGUING ppt
CANONS OF CATALOGUING ppt
 
Common communication format
Common communication formatCommon communication format
Common communication format
 
ISO 2709
ISO 2709ISO 2709
ISO 2709
 
UNISIST
UNISISTUNISIST
UNISIST
 
User Education: what is it and why is it important?
User Education: what is it and why is it important?User Education: what is it and why is it important?
User Education: what is it and why is it important?
 
Metadata
MetadataMetadata
Metadata
 
Cataloging with RDA: An Overview
Cataloging with RDA: An OverviewCataloging with RDA: An Overview
Cataloging with RDA: An Overview
 
Introduction to subject cataloguing
Introduction to subject cataloguingIntroduction to subject cataloguing
Introduction to subject cataloguing
 
Library automation history Anandraj.L
Library automation history Anandraj.LLibrary automation history Anandraj.L
Library automation history Anandraj.L
 

Semelhante a ML Expands Metadata for Archives

Leveraging social media for training object detectors
Leveraging social media for training object detectorsLeveraging social media for training object detectors
Leveraging social media for training object detectorsManish Kumar
 
Image Search: Then and Now
Image Search: Then and NowImage Search: Then and Now
Image Search: Then and NowSi Krishan
 
Main principles of Data Science and Machine Learning
Main principles of Data Science and Machine LearningMain principles of Data Science and Machine Learning
Main principles of Data Science and Machine LearningNikolay Karelin
 
H2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupH2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupSri Ambati
 
The deep learning technology on coco framework full report
The deep learning technology on coco framework full reportThe deep learning technology on coco framework full report
The deep learning technology on coco framework full reportJIEMS Akkalkuwa
 
Research Inventy : International Journal of Engineering and Science is publis...
Research Inventy : International Journal of Engineering and Science is publis...Research Inventy : International Journal of Engineering and Science is publis...
Research Inventy : International Journal of Engineering and Science is publis...researchinventy
 
Research Inventy: International Journal of Engineering and Science
Research Inventy: International Journal of Engineering and ScienceResearch Inventy: International Journal of Engineering and Science
Research Inventy: International Journal of Engineering and Scienceresearchinventy
 
Automated Metadata Annotation What Is And Is Not Possible With Machine Learning
Automated Metadata Annotation  What Is And Is Not Possible With Machine LearningAutomated Metadata Annotation  What Is And Is Not Possible With Machine Learning
Automated Metadata Annotation What Is And Is Not Possible With Machine LearningJim Webb
 
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...Editor IJMTER
 
CONTENT BASED IMAGE RETRIEVAL SYSTEM
CONTENT BASED IMAGE RETRIEVAL SYSTEMCONTENT BASED IMAGE RETRIEVAL SYSTEM
CONTENT BASED IMAGE RETRIEVAL SYSTEMVamsi IV
 
2019 cvpr paper_overview
2019 cvpr paper_overview2019 cvpr paper_overview
2019 cvpr paper_overviewLEE HOSEONG
 
2019 cvpr paper overview by Ho Seong Lee
2019 cvpr paper overview by Ho Seong Lee2019 cvpr paper overview by Ho Seong Lee
2019 cvpr paper overview by Ho Seong LeeMoazzem Hossain
 
Visual Information Retrieval: Advances, Challenges and Opportunities
Visual Information Retrieval: Advances, Challenges and OpportunitiesVisual Information Retrieval: Advances, Challenges and Opportunities
Visual Information Retrieval: Advances, Challenges and OpportunitiesOge Marques
 
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEY
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEYAPPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEY
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEYcscpconf
 
Applications of spatial features in cbir a survey
Applications of spatial features in cbir  a surveyApplications of spatial features in cbir  a survey
Applications of spatial features in cbir a surveycsandit
 
Interactive Machine Learning Appendix
Interactive  Machine Learning AppendixInteractive  Machine Learning Appendix
Interactive Machine Learning AppendixZitao Liu
 
Content-based image retrieval based on corel dataset using deep learning
Content-based image retrieval based on corel dataset using deep learningContent-based image retrieval based on corel dataset using deep learning
Content-based image retrieval based on corel dataset using deep learningIAESIJAI
 

Semelhante a ML Expands Metadata for Archives (20)

Leveraging social media for training object detectors
Leveraging social media for training object detectorsLeveraging social media for training object detectors
Leveraging social media for training object detectors
 
Image Search: Then and Now
Image Search: Then and NowImage Search: Then and Now
Image Search: Then and Now
 
Main principles of Data Science and Machine Learning
Main principles of Data Science and Machine LearningMain principles of Data Science and Machine Learning
Main principles of Data Science and Machine Learning
 
H2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User GroupH2O with Erin LeDell at Portland R User Group
H2O with Erin LeDell at Portland R User Group
 
Ts2 c topic
Ts2 c topicTs2 c topic
Ts2 c topic
 
Ts2 c topic (1)
Ts2 c topic (1)Ts2 c topic (1)
Ts2 c topic (1)
 
The deep learning technology on coco framework full report
The deep learning technology on coco framework full reportThe deep learning technology on coco framework full report
The deep learning technology on coco framework full report
 
Research Inventy : International Journal of Engineering and Science is publis...
Research Inventy : International Journal of Engineering and Science is publis...Research Inventy : International Journal of Engineering and Science is publis...
Research Inventy : International Journal of Engineering and Science is publis...
 
Research Inventy: International Journal of Engineering and Science
Research Inventy: International Journal of Engineering and ScienceResearch Inventy: International Journal of Engineering and Science
Research Inventy: International Journal of Engineering and Science
 
Automated Metadata Annotation What Is And Is Not Possible With Machine Learning
Automated Metadata Annotation  What Is And Is Not Possible With Machine LearningAutomated Metadata Annotation  What Is And Is Not Possible With Machine Learning
Automated Metadata Annotation What Is And Is Not Possible With Machine Learning
 
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
FACE EXPRESSION IDENTIFICATION USING IMAGE FEATURE CLUSTRING AND QUERY SCHEME...
 
CONTENT BASED IMAGE RETRIEVAL SYSTEM
CONTENT BASED IMAGE RETRIEVAL SYSTEMCONTENT BASED IMAGE RETRIEVAL SYSTEM
CONTENT BASED IMAGE RETRIEVAL SYSTEM
 
research paper
research paperresearch paper
research paper
 
2019 cvpr paper_overview
2019 cvpr paper_overview2019 cvpr paper_overview
2019 cvpr paper_overview
 
2019 cvpr paper overview by Ho Seong Lee
2019 cvpr paper overview by Ho Seong Lee2019 cvpr paper overview by Ho Seong Lee
2019 cvpr paper overview by Ho Seong Lee
 
Visual Information Retrieval: Advances, Challenges and Opportunities
Visual Information Retrieval: Advances, Challenges and OpportunitiesVisual Information Retrieval: Advances, Challenges and Opportunities
Visual Information Retrieval: Advances, Challenges and Opportunities
 
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEY
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEYAPPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEY
APPLICATIONS OF SPATIAL FEATURES IN CBIR : A SURVEY
 
Applications of spatial features in cbir a survey
Applications of spatial features in cbir  a surveyApplications of spatial features in cbir  a survey
Applications of spatial features in cbir a survey
 
Interactive Machine Learning Appendix
Interactive  Machine Learning AppendixInteractive  Machine Learning Appendix
Interactive Machine Learning Appendix
 
Content-based image retrieval based on corel dataset using deep learning
Content-based image retrieval based on corel dataset using deep learningContent-based image retrieval based on corel dataset using deep learning
Content-based image retrieval based on corel dataset using deep learning
 

Mais de Bohyun Kim

New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...
New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...
New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...Bohyun Kim
 
Practical Considerations for Open Infrastructure
Practical Considerations for Open InfrastructurePractical Considerations for Open Infrastructure
Practical Considerations for Open InfrastructureBohyun Kim
 
AI for Libraries
AI for LibrariesAI for Libraries
AI for LibrariesBohyun Kim
 
The Potential and Challenges of Today's AI
The Potential and Challenges of Today's AIThe Potential and Challenges of Today's AI
The Potential and Challenges of Today's AIBohyun Kim
 
Robots: What Could Go Wrong? What Could Go Right?
Robots: What Could Go Wrong? What Could Go Right? Robots: What Could Go Wrong? What Could Go Right?
Robots: What Could Go Wrong? What Could Go Right? Bohyun Kim
 
AI & Us: Are We Intelligent Machines?
AI & Us: Are We Intelligent Machines?AI & Us: Are We Intelligent Machines?
AI & Us: Are We Intelligent Machines?Bohyun Kim
 
Blockchain: The New Technology and Its Applications for Libraries
Blockchain: The New Technology and Its Applications for LibrariesBlockchain: The New Technology and Its Applications for Libraries
Blockchain: The New Technology and Its Applications for LibrariesBohyun Kim
 
Machine Intelligence and Moral Decision-Making
Machine Intelligence and Moral Decision-MakingMachine Intelligence and Moral Decision-Making
Machine Intelligence and Moral Decision-MakingBohyun Kim
 
Impact of Artificial Intelligence (AI) on Libraries
Impact of Artificial Intelligence (AI) on Libraries Impact of Artificial Intelligence (AI) on Libraries
Impact of Artificial Intelligence (AI) on Libraries Bohyun Kim
 
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...Bohyun Kim
 
Moving Forward with Digital Disruption: A Right Mindset
Moving Forward with Digital Disruption: A Right MindsetMoving Forward with Digital Disruption: A Right Mindset
Moving Forward with Digital Disruption: A Right MindsetBohyun Kim
 
 Blockchain Overview: Possibilities and Issues
 Blockchain Overview: Possibilities and Issues Blockchain Overview: Possibilities and Issues
 Blockchain Overview: Possibilities and IssuesBohyun Kim
 
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can Do
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can DoAI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can Do
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can DoBohyun Kim
 
A Pedagogical Approach to Web Scale Discovery User Interface
A Pedagogical Approach to Web Scale Discovery User InterfaceA Pedagogical Approach to Web Scale Discovery User Interface
A Pedagogical Approach to Web Scale Discovery User InterfaceBohyun Kim
 
From Virtual Reality to Blockchain: Current and Emerging Tech Trends
From Virtual Reality to Blockchain: Current and Emerging Tech TrendsFrom Virtual Reality to Blockchain: Current and Emerging Tech Trends
From Virtual Reality to Blockchain: Current and Emerging Tech TrendsBohyun Kim
 
Interdisciplinary Learning through Libraries on Artificial Intelligence
Interdisciplinary Learning through Libraries on Artificial IntelligenceInterdisciplinary Learning through Libraries on Artificial Intelligence
Interdisciplinary Learning through Libraries on Artificial IntelligenceBohyun Kim
 
Facing Change: Tweak or Transform?
Facing Change: Tweak or Transform?Facing Change: Tweak or Transform?
Facing Change: Tweak or Transform?Bohyun Kim
 
Innovating Together: the UX of Discovery
Innovating Together: the UX of DiscoveryInnovating Together: the UX of Discovery
Innovating Together: the UX of DiscoveryBohyun Kim
 
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated Workflow
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated WorkflowCleaning Up the Mess: Modernizing Your Dev Team’s Outdated Workflow
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated WorkflowBohyun Kim
 
Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...
 Pricing, Staff Workflow, & Application Development for 3D Printing Service: ... Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...
Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...Bohyun Kim
 

Mais de Bohyun Kim (20)

New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...
New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...
New Technologies of the Fourth Industrial Revolution: AI, IoT, Robotics, and ...
 
Practical Considerations for Open Infrastructure
Practical Considerations for Open InfrastructurePractical Considerations for Open Infrastructure
Practical Considerations for Open Infrastructure
 
AI for Libraries
AI for LibrariesAI for Libraries
AI for Libraries
 
The Potential and Challenges of Today's AI
The Potential and Challenges of Today's AIThe Potential and Challenges of Today's AI
The Potential and Challenges of Today's AI
 
Robots: What Could Go Wrong? What Could Go Right?
Robots: What Could Go Wrong? What Could Go Right? Robots: What Could Go Wrong? What Could Go Right?
Robots: What Could Go Wrong? What Could Go Right?
 
AI & Us: Are We Intelligent Machines?
AI & Us: Are We Intelligent Machines?AI & Us: Are We Intelligent Machines?
AI & Us: Are We Intelligent Machines?
 
Blockchain: The New Technology and Its Applications for Libraries
Blockchain: The New Technology and Its Applications for LibrariesBlockchain: The New Technology and Its Applications for Libraries
Blockchain: The New Technology and Its Applications for Libraries
 
Machine Intelligence and Moral Decision-Making
Machine Intelligence and Moral Decision-MakingMachine Intelligence and Moral Decision-Making
Machine Intelligence and Moral Decision-Making
 
Impact of Artificial Intelligence (AI) on Libraries
Impact of Artificial Intelligence (AI) on Libraries Impact of Artificial Intelligence (AI) on Libraries
Impact of Artificial Intelligence (AI) on Libraries
 
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...
Taking on a New Leadership Challenge: Student-Focused Learning in Artificial ...
 
Moving Forward with Digital Disruption: A Right Mindset
Moving Forward with Digital Disruption: A Right MindsetMoving Forward with Digital Disruption: A Right Mindset
Moving Forward with Digital Disruption: A Right Mindset
 
 Blockchain Overview: Possibilities and Issues
 Blockchain Overview: Possibilities and Issues Blockchain Overview: Possibilities and Issues
 Blockchain Overview: Possibilities and Issues
 
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can Do
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can DoAI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can Do
AI Lab at a Library? Why Artificial Intelligence Matters & What Libraries Can Do
 
A Pedagogical Approach to Web Scale Discovery User Interface
A Pedagogical Approach to Web Scale Discovery User InterfaceA Pedagogical Approach to Web Scale Discovery User Interface
A Pedagogical Approach to Web Scale Discovery User Interface
 
From Virtual Reality to Blockchain: Current and Emerging Tech Trends
From Virtual Reality to Blockchain: Current and Emerging Tech TrendsFrom Virtual Reality to Blockchain: Current and Emerging Tech Trends
From Virtual Reality to Blockchain: Current and Emerging Tech Trends
 
Interdisciplinary Learning through Libraries on Artificial Intelligence
Interdisciplinary Learning through Libraries on Artificial IntelligenceInterdisciplinary Learning through Libraries on Artificial Intelligence
Interdisciplinary Learning through Libraries on Artificial Intelligence
 
Facing Change: Tweak or Transform?
Facing Change: Tweak or Transform?Facing Change: Tweak or Transform?
Facing Change: Tweak or Transform?
 
Innovating Together: the UX of Discovery
Innovating Together: the UX of DiscoveryInnovating Together: the UX of Discovery
Innovating Together: the UX of Discovery
 
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated Workflow
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated WorkflowCleaning Up the Mess: Modernizing Your Dev Team’s Outdated Workflow
Cleaning Up the Mess: Modernizing Your Dev Team’s Outdated Workflow
 
Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...
 Pricing, Staff Workflow, & Application Development for 3D Printing Service: ... Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...
Pricing, Staff Workflow, & Application Development for 3D Printing Service: ...
 

Último

Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...gurkirankumar98700
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Igalia
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityPrincipled Technologies
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 

Último (20)

Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
Kalyanpur ) Call Girls in Lucknow Finest Escorts Service 🍸 8923113531 🎰 Avail...
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
Raspberry Pi 5: Challenges and Solutions in Bringing up an OpenGL/Vulkan Driv...
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 

ML Expands Metadata for Archives

  • 1. Exploring Machine Learning (ML) for Libraries & Archives: Present & Future Bohyun Kim Chief Technology Officer & Professor, University of Rhode Island, USA Twitter: @bohyunkim | Web: http://www.bohyunkim.net/blog/ Bite-sized Internet Librarian International 2021, September 22, 2021.
  • 2. Machine Learning (ML) Image from : https://labelbox.com/image-segmentation-overview
  • 3. Images from : https://towardsdatascience.com/what-is-machine- learning-a-short-note-on-supervised-unsupervised-semi- supervised-and-aed1573ae9bb
  • 5. Discussing ML for Libraries/Archives Images from : https://medium.com/@dhartidhami/machine-learning-basics-model-cost-function-and-gradient-descent-79b69ff28091 ; https://www.analyticsvidhya.com/blog/2020/02/cnn-vs-rnn-vs-mlp-analyzing-3-types-of-neural-networks-in-deep-learning/
  • 6. Can ML be a useful tool for the work of libraries & archives? Image from : https://dancohen.org/2019/09/17/humane-ingenuity-3-ai-in-the-archives/
  • 7. Q. What would practical application of ML for libraries/archives look like?
  • 8. Things I will be going over today: Q. What would practical application of ML for libraries/archives look like? I. ML for Expediting archival processing II. ML for Expanding descriptive metadata III. ML for Enhancing metadata IV. ML for Full content extraction Image from: https://www.analyticsvidhya.com/blog/2020/02/cnn-vs-rnn-vs-mlp-analyzing-3-types-of-neural-networks-in-deep-learning/
  • 9. I. ML for Expediting Archival Processing • “Processing” refers to the arrangement, description, and housing of materials for storage and use. • Unprocessed materials pose an impediment to researchers because they cannot be easily identified, accessed, and used. • Machine learning can expedite the processing of archival materials.
  • 11. The AIDA Project Explored: • Document Segmentation; • Graphic Element Classification & Text Extraction; • Document Type Classification; • Digitization Type Differentiation; • Document Image Quality Assessment; and • Document Clustering. Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
  • 12. (a) Identify/Classify Graphical Content & Pull Text with Image Zoning and Segmentation • The AIDA team tried to identify and classify graphical content (e.g. figures, illustrations, and cartoons) and extract text in historical newspaper page images, using ML’s image zoning and segmentation (U-NeXt). • 2 sets of newspaper images were used for ML model training/evaluation. • One set from the Library of Congress’s Beyond Words project, which crowdsourced the graphical content zone demarcation and the text transcription of historical newspaper images. • 1,532 images and corresponding data from graphical content zones; 1,226 images used for training and 306 for evaluation. • The other set from the Europeana Newspapers Project. • A set of 481 page images already zoned and segmented, with the segments classed as background, text, figure, separator or table.
  • 13. Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
  • 14. What is Computer Vision(CV)? • Computer vision is a field of artificial intelligence (AI) that enables computers and systems to derive meaningful information from digital images, videos and other visual inputs — and take actions or make recommendations based on that information.
  • 15. Computer Vision Concepts • Image filtering, matching, segmentation, and recognition • Segmentation divides whole images into pixel groupings which can then be labelled and classified. Images from : https://ai.stanford.edu/~syyeung/cvweb/tutorial1.html
  • 16. Tasks Performed by Computer Vision • Image classification • Object detection • Object tracking • Content-based image retrieval • Semantic segmentation • Instance segmentation
  • 17. (b) Natural Image / Document Classification • The AIDA project team applied natural image and document classification methods in machine learning to a collection of minimally processed manuscript materials to see whether handwritten, printed, and mixed (both handwritten and printed) documents can be distinguished from one another. • The team generated their own model by using the VGG method with 16 categories (VGG-16) pre-trained on the RVL_CDIP dataset. • 999 images were used in total.
  • 18. Figure 9. Prediction failure cases. In the left example, the model classified the document as handwritten rather than Mixed. Note that the printed regions are very small compared to the handwritten content in the image. In the right example, the model classified the document as printed rather than mixed. Here, the handwritten region is very small compared to the printed region in the image. Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
  • 19. (c) Detecting Digitization Type • The AIDA project team also applied machine learning techniques to both classifying manuscript collection images as digitized from original vs. from microform and analyzing their image quality. • 36,103 images were used as the ground truth classification Figure 13. A digital image of a photograph, digitized from the original item, and classified by the classifier as having been digitized from a microform source. Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
  • 20. Benefits of ML for Archival Processing • There is no doubt that being able to identify graphical content contained in archival materials and to peruse the text in them would make browsing and navigating those materials a lot more manageable to users. And if ML can process archival materials this way, it would surely improve their discoverability and usefulness. • Automating the classification of materials as handwritten, printed, and mixed can surely save archivists’ and librarians’ time spent on performing such classification manually. • Automatically assessing the quality of images and assigning their original source type can help archivists and librarians identify materials at higher-risk and take appropriate preservation measures. Source: Bohyun Kim, “Machine Learning for Libraries and Archives.” Online Searcher 45(1): 39-41, 2021.
  • 21. Two Challenges • The results from these explorations indicate that these machine learning methods are not ready to be immediately put to use in archival processing. • “Training a classifier to learn information from rare classes is very hard.” • The project report highlights two challenges: • (a) general scarcity of training data sets; • (b) overall difficulty in obtaining or establishing ground truth (= the expected result against which the performance of a machine learning model is measured). Source: Elizabeth Lorang et al., “Digital Libraries, Intelligent Data Analytics, and Augmented Description: A Demonstration Project” (University of Nebraska - Lincoln, January 10, 2020), https://digitalcommons.unl.edu/libraryscience/396.
  • 22. Need for More Data and Ground Truth • Without a sufficiently large amount of training data, a machine learning model for libraries/archives cannot be created. • Without ground-truth datasets, it is difficult to tell how well a machine learning model for libraries/archives works.
  • 23. II. ML for Expanding Descriptive Metadata • Mass digitization projects created many digital objects of historical materials, and many more born-digital items are being added to archival collections. • Unfortunately, libraries and archives lack resources to manually create rich metadata for all digital objects in a timely manner. • Lacking such metadata, those digital objects are not easily discovered, identified, or navigated. • ML can advance the discovery, navigation, and clustering of these materials by generating more detailed metadata.
  • 24. (a) Image Classifier ML Algorithm for the Frick Collection • The Frick Art Reference Library in New York together with Stanford University, Cornell University, and the University of Toronto developed an image classifier algorithm that applies a local classification system based on visual elements to the library’s digitized Photoarchive. • Tens of thousands of already-labeled images & hundreds of thousands of images that require labeling. • Hierarchical classification headings: • “American School: Portraits: Men: Without hands (Without hats): Head to left”
  • 25. Process • The Cornell/Toronto/Stanford team modified the popular ResNet152 network34—with 152 layers and six million parameters—and trained it on a collection of 45,857 classified reproductions of American paintings provided by the Frick Art Reference Library’s Photoarchive. • The majority were portraits and represented 98 unique classification headings. • After training, the network was fed images from the unlabeled portion of the library’s Photoarchive, and the network predicted a classification heading for each one. Source: Ellen Prokop et al., “AI and the Digitized Photoarchive: Promoting Access and Discoverability,” Art Documentation: Journal of the Art Libraries Society of North America 40, no. 1 (March 1, 2021): 1–20, https://doi.org/10.1086/714604.
  • 26. Vetting & Prospect • These images were then annotated with the predictions and shown to the library’s photoarchivists through an application to vet the predictions. • The library staff vetted 8,661 images in the year August 2019 - August 2020 and agreed with the network on the classification headings applied to 5,829 (67 %) of the images. • Yet even the incorrect predictions were, in general, “almost correct” with only one tag incorrect or missing. • Holds the potential to improve the staff’s efficiency in in metadata creation and image retrieval.
  • 27. (b) CAMPI (Computer-Aided Metadata Generation for Photo archives Initiative) https://www.cmu.edu/news/stories/archives/2020/october/computer-vision-archive.html
  • 28. Expanding Item-Level Description Metadata • The CAMPI project team built a prototype application that uses computer vision (CV), in order to aid archivists and metadata editors for Carnegie Mellon University Archives’ General Photograph Collection (GPC) of +20,000 images with metadata at some hierarchy (e.g. job; directory). • This collection of photographic prints, 35mm negatives, and born digital images was specifically chosen for the project, because: • those images lack item-level description and • are stored in and accessed through a shared network drive by archivists and metadata editors, with no browsing interface. Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer- Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
  • 29. Visual Similarity Search Powered by an “Off- the-shelf” Image Search Model • The CAMPI prototype application offered archivists and metadata editors visual similarity search using CV techniques in ML. • For this visual similarity search capability, a pre-trained “off-the-shelf” image search model was used. • The team used the second-to-last pooled layer of the InceptionV3 trained neural network as the feature source. For each image, this produced an array of 512 numbers, allowing them to treat the entire collection of images as a set of points in 512-dimensional space. • Then, the team used an approximate nearest neighbor search algorithm to create an index of that feature space, allowing it to be queried.
  • 30. The prototype allows staff to search similar images to reduce repeated metadata generation and remove noise from the eventual public discovery interface.
  • 31. Speeding Up Tagging • The application provided a browsing and CV- assisted tagging interface faceted by the existing job and directory metadata associated with the photographs. • Using this application, archivists and metadata editors were able to start with a tag, find a seed photo, and retrieve similar photos via visual similarity search that specifies the number of neighbors requested. • Then, they were able to quickly tag those retrieved photos with tags from the seed photo when appropriate.
  • 32.
  • 34.
  • 35. ML for Metadata Workflow • In their white paper, the CAMPI team reported that of the more than 43,000 tagging decisions made, a little over 1 in 5 were able to be automatically distributed across a set, saving metadata editors more than 9,000 decisions. • They also noted that the prototype application identified around 28% of the collection as sets with duplicates. • This prototype had the effect of rapidly streamlining and speeding up item-level metadata creation. • These results demonstrate how machine learning can be integrated into the existing metadata creation and editing workflow at libraries and archives. Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer- Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
  • 36. Image Browsing/Filtering UI • What is interesting about the CAMPI project is that even though the use of computer vision was minimal in their prototype application, this project produced a successful outcome because it combined machine learning with a well-designed user interface that took advantage of the existing archival structure, thereby effectively augmenting archivists’ and metadata experts’ existing work. • At libraries and archives, a browsing and filtering interface is primarily created to provide patrons with public access to archival materials. • However, such an interface significantly increases the overall productivity of metadata-related work, if it fits in well with the way the staff perform their work.
  • 37. Need for More Training Sets • “There is a field-wide need for specialized computer vision training sets based on historical, nonborn-digital photograph collections.” Source: Matthew Lincoln et al., “CAMPI: Computer-Aided Metadata Generation for Photo Archives Initiative” (Carnegie Mellon University Libraries, October 9, 2020), https://kilthub.cmu.edu/articles/preprint/CAMPI_Computer- Aided_Metadata_Generation_for_Photo_archives_Initiative/12791807.
  • 38. III. ML for Enhancing Metadata • The Audiovisual Metadata Platform Pilot Development (AMPPD) Project is run by Indiana University Libraries, University of Texas at Austin School of Information, New York Public Library, and AVP, a data management solutions and consulting company. • Started in 2018 and continuing. • The AMPPD team intends to create automated metadata generation mechanisms and integrate them with the human metadata generation process. • The goal is to generate and manage metadata for audiovisual materials at scale for libraries and archives.
  • 40. (a) Generate and Manage Rich Metadata for Audiovisual Materials at Scale • AMPPD Project’s Priority Tasks • First: Detecting silence, speech, and music; speech-to-text / speaker diarization; named entity recognition, such as people, geographic locations, and topics; • Next: Video and structured data OCR (metadata, transcript, subject terms); detecting scene and shot; detecting music genre in audiovisual materials. Source: Shawn Avercamp and Julie Hardesty, “AI Is Such a Tool: Keeping Your Machine Learning Outputs in Check” (Code4Lib 2020 Conference, Pittsburgh, PA, March 11, 2020), https://2020.code4lib.org/talks/AI-is- such-a-tool-Keeping-your-machine-learning-outputs-in-check.
  • 41. Annotating AV Materials • It is easy to see that their first-priority items alone will annotate audiovisual materials in such a way that library and archive users can quickly identify the exact content they are looking for and better grasp what kinds of content each audiovisual material contains. • If the AMP can automatically and reliably extract broad subject terms by recognizing named entities within the content, that will also greatly enhance the discoverability and accessibility of those materials.
  • 42. ML for Better Discovery, Identification, and Navigation of AV Materials • Extracting the full content of digital AV objects as machine-accessible and interpretable data is a big step towards enabling better discovery, identification, and navigation, because it makes full-text query possible.
  • 43. (b) Summarization of Textual Documents • The Catholic Pamphlets collection of over 5,500 PDFs at the University of Notre Dame had sparse summaries. While half of the collection had a MARC summary field, most of these summaries were a few words at most. • 3 NLP (Natural Language Processing) automated summarization techniques of ML were tested to create more robust summaries by combining the summaries with LC subject headings. • 4 NLP algorithms in ML were used: Rapid Automatic Keyword Extraction (RAKE); a raw TextRank implementation algorithm; Genism; Bert Extractive Summarizer (BERT; uses DL) Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
  • 44. Process & Result • Of the 5,500 texts, 51 had sufficient summaries already, and another 50 summaries were written. • The automated summaries were generated after feeding the pamphlets as .pdf files into an OCR pipeline. • Computer-generated summaries of the pamphlets were compared against human-written summaries. • The BERT Extractive technique yielded the best score - an average Rouge F1 score of 0.239. The Gensim python package implementation of TextRank scored at .151. A hand-implemented TextRank algorithm created summaries that scored at 0.144.
  • 45. Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
  • 46. Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
  • 47. Challenges in Summarization & Prospect • Latin text and OCR errors have persisted despite choosing more sophisticated algorithms and better text preprocessing. • Better OCR correction & the exclusion of Latin text and sentence fragments remain as challenges. • “For collections in which their text does not need OCR and all appears in English, the BERT Extractive Summarizer likely could provide summaries with few enough errors that catalogers would be comfortable adding the computed-generated summaries with minimal human oversight.” Source: Jeremiah Flannery, “Using NLP to Generate MARC Summary Fields for Notre Dame ’s Catholic Pamphlets,” International Journal of Librarianship 5, no. 1 (July 23, 2020): 20–35, https://doi.org/10.23974/ijol.2020.vol5.1.158.
  • 48. IV. ML for Full Content Extraction • Archival materials are given a set of metadata either at the item, collection, or some other unit level, so that they can be searched online. • But their full content is not always readily available in digital form as machine-Interpretable data. This hinders their discoverability and use, such as running full-text or full-content searches and performing other computational analyses. • Machine learning can be utilized for content extraction.
  • 49. ML for text, auditory, visual, and tabular data • OCR (an ML method of image recognition) for text documents. • Auditory data can be extracted using machine learning’s speech- recognition method. • From materials such as photographs and video recordings, visual information can be extracted by image recognition techniques.
  • 50. ML for text, auditory, visual, and tabular data • Tabular data in digitized images are highly problematic since they cannot be easily identified and utilized as structured data. • But they can also be extracted using table detection, cell recognition, and text extraction algorithms in machine learning. • Extracting tabular data along with their data structure will make historical data immediately useful to researchers. Ryan Cordell, “Machine Learning + Libraries” (Library of Congress, July 14, 2020), https://labs.loc.gov/static/labs/work/reports/Cordell-LOC-ML-report.pdf?loclr=blogsig.
  • 51. Future Focuses for Libraries/Archives • Need for more data & ground truth for ML algorithm training and evaluation • Exploration of pre-trained off-the-shelf ML models • e.g. AWS Rekognition (aws.amazon.com/rekognition) and Google Cloud Vision AI (cloud.google.com/vision). • Need for continued digitization
  • 52. Thank you! Bohyun Kim Chief Technology Officer & Professor, University of Rhode Island, USA Twitter: @bohyunkim | Web: http://www.bohyunkim.net/blog/ Bite-sized Internet Librarian International 2021, September 22, 2021.