This presentation introduces OWLIM semantic repository at DM2E project meeting, held in Vienna in November 2012. Ontotext entered the DM2E consortium as associated partner.
The Use of Big Data Techniques for Digital ArchivingSven Schlarb
These slides were used in a presentation at the "Our Digital Future - Multidisciplinary Perspectives on Long Term Data Preservation and Access" conference in Cambridge/UK in March 2016 in the session "Current and Future perspectives on technology for data preservation and sharing". They describe work in progress in the E-ARK project, which is co-funded by the European Commission and has as its main objective the creation of a scalable open source, digital archiving system offering efficent search and access content of very large digital object collections. The focus of this presentation lies on describing the core big data technologies (Apache Hadoop, Apache Hbase, and the document repository Lily developed by NGData), the architecture of the E-ARK integrated prototype implementation, and data mining use cases related to geographical data, named entitity extraction, and OLAP data analysis.
A Framework for Improved Access to Museum Databases in the Semantic WebMariana Damova, Ph.D
This paper presents a framework for processing Museum databases according to a set of interlinked ontologies, including CIDOC-CRM, and loading them in a reason-able view of the web of data, providing additional links to datasets from the LOD cloud. The infrastructure allows accessing the data via SPARQL queries and to verbalize the query results in natural language, the GF formalism, which allows access to 18 natural languages.
1) The document discusses a vision for linking CRIS (Current Research Information Systems) and OAR (Open Access Repositories) data as Linked Data to improve interoperability in scholarly communication.
2) It proposes using common vocabularies and assigning persistent URIs to entities like publications, authors, projects and organizations to reduce duplication and improve data sharing across domains.
3) Next steps include adopting the KE CRIS-OAR data model and vocabulary and exploring how Linked Data approaches could be used to link publications, research data, and CRIS and OAR domains as part of the OpenAIREplus project.
"Shared Web Information Systems for Heritage in Scotland and Wales – Flexibility in Partnership"
D. Thomas (RCAHMW) and P. McKeague (RCAHMS)
This presentation was given at the XXIV International CIPA Symposium, 2013 - 'Recording, Documentation and Cooperation for Cultural Heritage'.
The Royal Commissions on the Ancient and Historical Monuments of Scotland and Wales were established in 1908 to investigate and record the archaeological and built heritage of their respective countries. The organisations have grown organically over the succeeding century, steadily developing their inventories and collections as card and paper indexes. Computerisation followed in the late 1980s and early 1990s, with RCAHMS releasing Canmore, an online searchable database, in 1998. Following a review of service provision in Wales, RCAHMW entered into partnership with RCAHMS in 2003 to deliver a database for their national inventories and collections. The resultant partnership enables both organisations to develop at their own pace whilst delivering efficiencies through a common experience and a shared IT infrastructure. Through innovative solutions the partnership has also delivered benefits to the wider historic environment community, providing online portals to a range of datasets, ultimately raising public awareness and appreciation of the heritage around them. Now celebrating its 10th year, Shared Web Information Systems for Heritage, or more simply SWISH, continues to underpin the work of both organisations in presenting information about the historic environment to the public.
The document discusses translating OGC Observation and Measurements (O&M) and Geography Markup Language (GML) features into Resource Description Framework (RDF) to integrate geospatial data into the Web of Data. It presents the Geographic Feature Pipes (GFP) which provides a Java API and proxy service that translates O&M and GML features into RDF and allows querying merged geospatial features and sensor observations using SPARQL. An example use case finds navigable rivers for vessels by querying river water level observations and river features for depth.
The document summarizes the European Archival Records and Knowledge (E-ARK) project, which developed an OAIS-compliant system for fast creation, search, and access of archival information packages. It describes the key components and functionality of the E-ARK reference implementation, including tools for ingest, archival storage, data management, access, and data mining of archived content. Current pilots of the E-ARK system are being used by several national archives for large-scale archiving and access of records.
The document describes a Maps4Finland workshop on common geographic information programming examples. It provides an agenda that includes discussions of content from various Finnish organizations like the National Land Survey of Finland and Statistics Finland. It also gives examples of code samples for creating maps using APIs and standards like OpenLayers, WMS, and GeoJSON. Finally, it discusses handling and converting spatial and non-spatial content between different formats.
The Use of Big Data Techniques for Digital ArchivingSven Schlarb
These slides were used in a presentation at the "Our Digital Future - Multidisciplinary Perspectives on Long Term Data Preservation and Access" conference in Cambridge/UK in March 2016 in the session "Current and Future perspectives on technology for data preservation and sharing". They describe work in progress in the E-ARK project, which is co-funded by the European Commission and has as its main objective the creation of a scalable open source, digital archiving system offering efficent search and access content of very large digital object collections. The focus of this presentation lies on describing the core big data technologies (Apache Hadoop, Apache Hbase, and the document repository Lily developed by NGData), the architecture of the E-ARK integrated prototype implementation, and data mining use cases related to geographical data, named entitity extraction, and OLAP data analysis.
A Framework for Improved Access to Museum Databases in the Semantic WebMariana Damova, Ph.D
This paper presents a framework for processing Museum databases according to a set of interlinked ontologies, including CIDOC-CRM, and loading them in a reason-able view of the web of data, providing additional links to datasets from the LOD cloud. The infrastructure allows accessing the data via SPARQL queries and to verbalize the query results in natural language, the GF formalism, which allows access to 18 natural languages.
1) The document discusses a vision for linking CRIS (Current Research Information Systems) and OAR (Open Access Repositories) data as Linked Data to improve interoperability in scholarly communication.
2) It proposes using common vocabularies and assigning persistent URIs to entities like publications, authors, projects and organizations to reduce duplication and improve data sharing across domains.
3) Next steps include adopting the KE CRIS-OAR data model and vocabulary and exploring how Linked Data approaches could be used to link publications, research data, and CRIS and OAR domains as part of the OpenAIREplus project.
"Shared Web Information Systems for Heritage in Scotland and Wales – Flexibility in Partnership"
D. Thomas (RCAHMW) and P. McKeague (RCAHMS)
This presentation was given at the XXIV International CIPA Symposium, 2013 - 'Recording, Documentation and Cooperation for Cultural Heritage'.
The Royal Commissions on the Ancient and Historical Monuments of Scotland and Wales were established in 1908 to investigate and record the archaeological and built heritage of their respective countries. The organisations have grown organically over the succeeding century, steadily developing their inventories and collections as card and paper indexes. Computerisation followed in the late 1980s and early 1990s, with RCAHMS releasing Canmore, an online searchable database, in 1998. Following a review of service provision in Wales, RCAHMW entered into partnership with RCAHMS in 2003 to deliver a database for their national inventories and collections. The resultant partnership enables both organisations to develop at their own pace whilst delivering efficiencies through a common experience and a shared IT infrastructure. Through innovative solutions the partnership has also delivered benefits to the wider historic environment community, providing online portals to a range of datasets, ultimately raising public awareness and appreciation of the heritage around them. Now celebrating its 10th year, Shared Web Information Systems for Heritage, or more simply SWISH, continues to underpin the work of both organisations in presenting information about the historic environment to the public.
The document discusses translating OGC Observation and Measurements (O&M) and Geography Markup Language (GML) features into Resource Description Framework (RDF) to integrate geospatial data into the Web of Data. It presents the Geographic Feature Pipes (GFP) which provides a Java API and proxy service that translates O&M and GML features into RDF and allows querying merged geospatial features and sensor observations using SPARQL. An example use case finds navigable rivers for vessels by querying river water level observations and river features for depth.
The document summarizes the European Archival Records and Knowledge (E-ARK) project, which developed an OAIS-compliant system for fast creation, search, and access of archival information packages. It describes the key components and functionality of the E-ARK reference implementation, including tools for ingest, archival storage, data management, access, and data mining of archived content. Current pilots of the E-ARK system are being used by several national archives for large-scale archiving and access of records.
The document describes a Maps4Finland workshop on common geographic information programming examples. It provides an agenda that includes discussions of content from various Finnish organizations like the National Land Survey of Finland and Statistics Finland. It also gives examples of code samples for creating maps using APIs and standards like OpenLayers, WMS, and GeoJSON. Finally, it discusses handling and converting spatial and non-spatial content between different formats.
This document summarizes a presentation given by Mariana Damova on using semantic technologies and Europeana data. It discusses how Europeana data has been converted to RDF and loaded into the OWLIM semantic graph database. This allows linking Europeana data to other datasets to enable queries across multiple sources. Examples of queries over Europeana and other cultural heritage data are provided. Future work on projects like Europeana Creative is also mentioned.
This document discusses empathy-based facilitation for hospital/healthcare management. It promotes using visual tools and serious play to build empathy, safety, and trust among medical professionals. This allows professionals to openly share their perspectives and develop shared visions and solutions. The approach aims to motivate staff and improve organizational outcomes like staff retention, productivity and cost reductions. It encourages healthcare managers to try these empathy-based visual facilitation techniques.
This presentation discusses the value of inferred knowledge over LOD and presents a new version of FactForge, a reason-able view, the biggest body of heterogeneous generic knowledge on which inference is performed, showing examples of inferred statements across LOD datasets.
Vladimir Alexiev presented ResearchSpace, a virtual research environment (VRE) based on the CIDOC CRM ontology. ResearchSpace aims to provide tools and services to support collaborative research projects for cultural heritage scholars. It aggregates data from various sources using semantic technologies and the CIDOC CRM ontology, allows semantic search of the data based on fundamental relations, and includes features for data analysis, collaboration, and web publication. The presentation provided an overview of Ontotext, the company developing ResearchSpace, described some of ResearchSpace's key capabilities, and discussed how the CIDOC CRM is central to ResearchSpace's approach.
Presentation given by Dr. Dimitris Gavrilis
Digital Curation Unit - IMIS, Athena Research Center
LoCloud Conference
Sharing local cultural heritage online with LoCloud services Amersfoort, Netherlands
5 February 2016
MARC records for archived websites on the Archive of Tomorrow project / Mark ...CILIP MDG
The presenters will discuss the metadata components of the National Library of Scotland-led ‘Archive of Tomorrow’ project, an 18-month multi-institutional collaboration focusing on capturing health resources online. Metadata work to be discussed includes the creation of a crosswalk to transform metadata produced in The British Library’s web archiving platform (ACT) into functioning MARC records, as well as subsequent enhancement work. Enhancements tested on the project included augmenting ACT metadata to generate authorised LCNAF headings; extending metadata using Wikidata and VIAF, ISNI and LC reconciliation services; and evaluating the analysis of ‘automatic’ subject heading assignation at scale, experimenting with the National Library of Finland’s AI project ‘ANNIF’ as well as other bespoke approaches. In addition to outlining the development and status of this work, the presentation will touch on project challenges and limitations, and the presenters’ experiences getting to grips with new platforms while testing ANNIF.
In addition to discussing the technical elements of the work performed, other strands of the work relevant to conference themes - from performing authority control outside of traditional platforms to making progress with linked data - will be open for discussion/Q&A. Other areas of the project work suitable for incorporation in the presentation include:
• The incorporation of Content Advisories in records for websites that might contain sensitive content, relaying the findings of a literature review conducted by Mark and project Rights Officer Jasmine Hide.
• Our dependence on parallel projects elsewhere, with reference to development work at the BL, user communities online, and ANNIF and Wikidata use across the field.
• The dynamics of multi-institutional project work, in this case performed remotely by dedicated and seconded project staff, touching on learning new skills, reporting findings, and seeking additional support.
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Eventos was founded in 2010 and has completed several natural language processing and semantic technology projects. It has a team of 60 people experienced in areas like natural language processing, document clustering, classification, and semantic storage. The company's OntoQuad RDF store uses a quad-based model and has performed well on benchmarks. By 2014, Eventos plans to build an open data store for a media company and improve its inference engine and scalability.
This document summarizes a presentation given by Mariana Damova on using semantic technologies and Europeana data. It discusses how Europeana data has been converted to RDF and loaded into the OWLIM semantic graph database. This allows linking Europeana data to other datasets to enable queries across multiple sources. Examples of queries over Europeana and other cultural heritage data are provided. Future work on projects like Europeana Creative is also mentioned.
This document discusses empathy-based facilitation for hospital/healthcare management. It promotes using visual tools and serious play to build empathy, safety, and trust among medical professionals. This allows professionals to openly share their perspectives and develop shared visions and solutions. The approach aims to motivate staff and improve organizational outcomes like staff retention, productivity and cost reductions. It encourages healthcare managers to try these empathy-based visual facilitation techniques.
This presentation discusses the value of inferred knowledge over LOD and presents a new version of FactForge, a reason-able view, the biggest body of heterogeneous generic knowledge on which inference is performed, showing examples of inferred statements across LOD datasets.
Vladimir Alexiev presented ResearchSpace, a virtual research environment (VRE) based on the CIDOC CRM ontology. ResearchSpace aims to provide tools and services to support collaborative research projects for cultural heritage scholars. It aggregates data from various sources using semantic technologies and the CIDOC CRM ontology, allows semantic search of the data based on fundamental relations, and includes features for data analysis, collaboration, and web publication. The presentation provided an overview of Ontotext, the company developing ResearchSpace, described some of ResearchSpace's key capabilities, and discussed how the CIDOC CRM is central to ResearchSpace's approach.
Presentation given by Dr. Dimitris Gavrilis
Digital Curation Unit - IMIS, Athena Research Center
LoCloud Conference
Sharing local cultural heritage online with LoCloud services Amersfoort, Netherlands
5 February 2016
MARC records for archived websites on the Archive of Tomorrow project / Mark ...CILIP MDG
The presenters will discuss the metadata components of the National Library of Scotland-led ‘Archive of Tomorrow’ project, an 18-month multi-institutional collaboration focusing on capturing health resources online. Metadata work to be discussed includes the creation of a crosswalk to transform metadata produced in The British Library’s web archiving platform (ACT) into functioning MARC records, as well as subsequent enhancement work. Enhancements tested on the project included augmenting ACT metadata to generate authorised LCNAF headings; extending metadata using Wikidata and VIAF, ISNI and LC reconciliation services; and evaluating the analysis of ‘automatic’ subject heading assignation at scale, experimenting with the National Library of Finland’s AI project ‘ANNIF’ as well as other bespoke approaches. In addition to outlining the development and status of this work, the presentation will touch on project challenges and limitations, and the presenters’ experiences getting to grips with new platforms while testing ANNIF.
In addition to discussing the technical elements of the work performed, other strands of the work relevant to conference themes - from performing authority control outside of traditional platforms to making progress with linked data - will be open for discussion/Q&A. Other areas of the project work suitable for incorporation in the presentation include:
• The incorporation of Content Advisories in records for websites that might contain sensitive content, relaying the findings of a literature review conducted by Mark and project Rights Officer Jasmine Hide.
• Our dependence on parallel projects elsewhere, with reference to development work at the BL, user communities online, and ANNIF and Wikidata use across the field.
• The dynamics of multi-institutional project work, in this case performed remotely by dedicated and seconded project staff, touching on learning new skills, reporting findings, and seeking additional support.
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Eventos was founded in 2010 and has completed several natural language processing and semantic technology projects. It has a team of 60 people experienced in areas like natural language processing, document clustering, classification, and semantic storage. The company's OntoQuad RDF store uses a quad-based model and has performed well on benchmarks. By 2014, Eventos plans to build an open data store for a media company and improve its inference engine and scalability.
ESWC SS 2012 - Wednesday Tutorial Barry Norton: Building (Production) Semanti...eswcsummerschool
Ontotext is a leading semantic technology company that has developed OWLIM, a family of semantic repositories for storing and querying RDF and OWL data. OWLIM can handle large datasets, perform reasoning, and supports features like full text search, notifications, and geo-spatial querying. It has been used successfully in large-scale production systems like the BBC's World Cup website to power semantic search and dynamic content delivery using semantic web technologies.
An overview of The European Library. Olaf Janssen presenting during DRH 2005,...Olaf Janssen
The European Library provides access to the collections of European national libraries through a central portal. It began as a feasibility study in 2001-2004 involving 8 pilot libraries. Since 2004, it has operated as a service, allowing users to search across the collections of participating national libraries. Currently over 15 libraries participate, providing access to over 500,000 digitized objects and 11 million records. The portal uses SRU/Z39.50 protocols to query libraries' databases and presents integrated search results to users through a single interface. Participating libraries benefit from increased visibility and opportunities for new users to discover their collections.
ICOS: Integrated Carbon Observation System Open data to open our eyes to clim...Blue BRIDGE
Presentation by Harry Lankreijer, ICOS-Carbon Portal, Lund University, Sweden.
ICOS is a pan-European research infrastructure (RI) for observing and understanding the greenhouse gas (GHG) balance of Europe and its adjacent regions. The major task of ICOS is to collect and make available in a transparent manner, the high-quality observational data from its state-of-the-art measurement stations. These ICOS data – from atmosphere, ecosystem and ocean stations – will contribute to research aiming to describe and understand the present state of the global carbon cycle. The Carbon Portal will be the virtual data center that present the data products and make it available. This presentation will briefly present the work of ICOS and the Carbon Portal towards open data with FAIR principles. ICOS has an open data policy with free use, requesting the user to give appropriate credit (Creative Commons Attribution 4.0 ). The Carbon Portal is developing a data catalogue using an ontology based on a semantic metadata description. This will make it possible to integrate ICOS observations with data from other RI’s as well with data of global networks. For integration, the Carbon Portal is actively following the developments of international standards for eg. metadata and data citation.
The document discusses several projects related to open metadata and linked data including:
1. The AIM25 project which aggregates archive descriptions from 123 partners and aims to test the value of linked data.
2. The COMET project which is releasing a large subset of bibliographic records under an open license and working to convert them to linked open data.
3. The Jerome project which harvests and unifies data from several library systems, supplements it with open data, and provides fast search APIs.
The document discusses Nick Sheppard's work as a repository developer at Leeds Metropolitan University, including developing applications to manage and share open educational resources across multiple repositories using metadata standards and protocols like OAI-PMH and SWORD. It also describes projects to build tools that extract and generate metadata for discovering and accessing learning objects stored in different repositories.
This paper describes the creation of linked data for cultural heritage domain, using semantic technologies. The Gothenburg city museum data are described according to an ontological model combining a series of upper-level and domain specific ontologies, such as PROTON and CIDOC-CRM, triplified and interlinked with data from LOD, e.g. DBpedia. The implementation is done as a reason-able view of the web of data and the data are loaded in OWLIM semantic repositoyr.
The document discusses the Research and Education Space (RES) project, which aims to create a web-based platform called Acropolis that aggregates and interconnects cultural heritage resources from various institutions like the British Library, British Museum, BBC archive, and others. It describes Acropolis' technical approach of using crawlers, indexes, and APIs to make these resources searchable. It also outlines challenges around standardizing heterogeneous metadata, reliably linking entities, and usability issues regarding tools, licensing, and stakeholder engagement. The author is looking to provide guidance on publishing cultural data as linked open data to help address these challenges.
Mind the gap! Reflections on the state of repository data harvestingSimeon Warner
A 24x7 presentation at Open Repositories 2017 in Brisbane, Australia.
I start with an opinionated history of the evolution of repository data harvesting since the late 1990's to the present. A conclusion is that we are currently in danger of creating a repository environment with fewer cross-repository services than before, with the potential to reinforce the silos we hope to open. I suggest that the community needs to agree upon a new solution, and further suggest that solution should be ResourceSync.
Infrastructure - A necessary platform for user empowermentRICHES
Presentation at the conference ICLAM 2011 in New Delhi, 15th-17th of February, 2011
http://www.flickr.com/groups/1658954@N22/
http://www.flickr.com/photos/rokal/sets/72157626087508810/
Managing director of Klokan Technologies GmbH, a small Swiss company that develops innovative geo applications for cultural heritage institutions. The document discusses Old Maps Online, a project that provides an easy-to-use gateway for searching historical maps from libraries around the world. It allows users to search maps by geographic location on an interactive world map and view high resolution maps from contributing institutions with proper crediting back to the libraries. The project is open to additional map contributors and uses tools like BoundingBox and Georeferencer to help enrich map metadata.
The Open Archives Initiative Protocol for Metadata Harvesting and ePrints UKAndy Powell
UKOLN is a center of expertise in digital information management supported by various organizations. The document discusses the Open Archives Initiative Protocol for Metadata Harvesting (OAI-PMH), including its history and how it allows harvesting of metadata from data providers by service providers through a simple protocol. It also discusses the potential impact of OAI-PMH on institutions, libraries, and researchers.
Nikola Ikonomov, Boyan Simeonov, Jana Parvanova and Vladimir Alexiev. In Digital Presentation and Preservation of Cultural and Scientific Heritage (DiPP 2013), Veliko Tarnovo, Bulgaria, Sep 2013.
The document discusses how semantic technologies can be applied to cultural heritage applications. It provides an overview of semantic web concepts like linked open data and describes standards for representing cultural heritage data like Europeana Data Model. Examples of related European Union and Bulgarian projects involving cultural heritage and semantic technologies are also presented.
This document summarizes the highs and lows of library linked data projects. It discusses two UK projects that exposed library catalog and archive data as linked open data (LOD), including modeling the data as RDF, transforming it, and loading it into a triplestore. It highlights the benefits of LOD like easier data integration and discovery. However, it also notes challenges like steep learning curves, complexity of archival data, issues of scalability, provenance and licensing.
This presentation gives insight to the overall Horizon 2020 Program and more specifically for the period 2018-2020 with emphasis to ICT. Mariana Damova is the National Contact Point for Horizon 2020 ICT in Bulgaria
Geography of Letters - The Spirituality of Sofia in the Historic MemoryMariana Damova, Ph.D
Presentation of the project The Spirituality of Sofia in the Historic Memory at the Round table on the future perspectives for Digital humanities in SEE within the Summer School in Advanced Tools for Digital Humanities and IT
The document describes IndustryInform, a semantic-based search and recommendation service for business networking. It allows industrial enterprises to advertise themselves, helps potential clients and investors find matching businesses, and provides a data as a service facility (DaaS) through annual subscriptions or pay-per-query plans. The service uses semantic web technologies and linked data to power searches across a database of over 50 million information units about 300,000 companies in 7 countries. It has features like extended search, company/user registration, and results displayed in table or Google-like formats. The system was developed by Mozaika's Humanizing Technologies Lab and has an engineering team, graphic designer, and business/marketing team to manage it.
Mozaika is a research center and SME operating since 2013 in the areas of data science, natural language interfaces, and human insight. It provides consulting, R&D projects, and data as a service solutions tailored to human behavior. Mozaika has expertise in semantic technologies, cognitive systems, and multimodal interactivity. It has completed projects in business networking, human resources management, cultural heritage, education, and aerospace with clients and partners from both private companies and research organizations.
This document summarizes Mozaika, a research center focused on humanizing technologies. It discusses technologies that make emerging technologies more understandable and give people more control, including reducing data complexity through semantic technologies. It provides examples of Mozaika's projects involving skills matching, city experience summarization, satellite communications, linked open data, and e-publishing. The goal is for technology to better support and enhance humanity.
Communication channels for the european single digital marketMariana Damova, Ph.D
Presentation about the importance of tackling the multilinguality in the strategy agenda for the European Digital Single Market, and about the role of language technology and the European language technology community in solving this issue endorsed by public funding
This is a presentation targeted to leaders of cultural institutions in Bulgaria to inform them about the opportunities to publish cultural content in Bulgariana and in Europeana and about what would be their benefits for doing this.
NLIWoD ISWC 2014 - Multilingual Retrieval Interface for Structured data on th...Mariana Damova, Ph.D
This presentation described a Multilingual Retrieval Interface for Structured data on the Web, a talk given at NLIWoD workshop at ISWC 2014. The approach is based on Grammatical framework and semantic web and linked data technologies
This document discusses humanizing technologies and trends in developing technologies that are more human-centric. It provides examples of technologies being developed by Mozaika, a research center, to better integrate technologies into human lives in natural ways. Mozaika is working on projects involving summarization, skills matching, information management, publishing, and more using techniques like natural language processing, sentiment analysis, and semantic technologies.
Presentation held at a meeting in Bulgaria (Varna Regional Library) coorganized by Europeana, Bulgariana, Varna Regional library and BBIA about Europeana.
Multilingual Access to Cultural Heritage Content on the Semantic Web - Acl2013Mariana Damova, Ph.D
The document discusses building an ontology-based application to communicate museum content in multiple languages on the Semantic Web. It aims to make cultural heritage accessible to both humans and computers by generating natural language descriptions from semantic data. The application uses Grammatical Framework to linearly multiple museum datasets and ontologies into 15 languages. It addresses challenges in cross-linguistically representing classes, properties, word order, tense, and reference. The system was demonstrated to generate descriptions of paintings from the Louvre museum in English and French.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
Salesforce Integration for Bonterra Impact Management (fka Social Solutions A...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on integration of Salesforce with Bonterra Impact Management.
Interested in deploying an integration with Salesforce for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
2. Ontotext
– Top-5 provider of core Semantic Technology
– Established in year 2000; offices in Bulgaria, UK, USA
– Active both in research and commercial projects (FP7 funding for 10 years)
• 360° semantic technology – unique portfolio:
– Semantic Databases: high-performance RDF DBMS, scalable reasoning
– Semantic Search: text-mining (IE), metadata generation, Information Retrieval (IR)
– Web Mining: focused crawling, screen scraping, data fusion
– Linked Data Management and Data Integration
Good recognition in the SemTech community
– Ontotext pages are ranked #1 for “semantic annotation” and “semantic repository” at
GYM, #3 for “linked data management” at Google
Several joint ventures and subsidiaries
– Innovantage: leading online recruitment intelligence provider in UK
3. Ontotext Clients (selected)
British Broadcasting Corporation (BBC)
– Run its World Cup 2010 sites on top of OWLIM
– Since Mar’12 BBC Sports
– 2012 Olympics sections are driven
by OWLIM and a Concept Extraction service developed by Ontotext
Press Association (UK)
– Analysis of Sports news
– Concept extraction
– Linked data generation
Top-3 USA media (not allowed to name)
The National Archives (UK) contracted Ontotext to implement
semantic KB and semantic search for the Government Web Archive
British Museum (UK) Ontotext leads the development of Phase 3 of
ResearchSpace project on collaborative research in cultural heritage;
British Museum’s public SPARQL end-point is powered by OWLIM
de Bibliothek (Holland) aggregation of data from 150 library databases
4. Semantic Technologies
• Semantic technologies (RDF, LOD) allow for an unprecedented ease of
integration of heterogeneous data sources
– Already adopted in pharmaceuticals and publishing industries
– Cultural heritage is the next
BBC – when MySQL was replaced with OWLIM in their “Dynamic Semantic
Publishing” architecture, the BBC team observed considerable reduction of
complexity of database design, query specification, application
development, and query evaluation time. BBC World Cup 2010 dynamic
semantic publishing. Jem Rayfield, Senior Technical Architect BBC News
and Knowledge.
http://www.bbc.co.uk/blogs/bbcinternet/2010/07/bbc_world_cup_2010_dyna
mic_sem.html
6. Semantic Repository for RDFS and OWL
• OWLIM is a family of scalable semantic repositories
• OWLIM-Lite: in-memory, fastest, scales to ~100 million statements
• OWLIM-SE: file-based, sameAs & query optimizations, scales to 20 billion
statements
• OWLIM-Enterprise: replication cluster deployment for resilience and high
performance parallel query-answering
• OWLIM provides
– Management, integration and analysis of heterogeneous data
– Combined with light-weight, high-performance reasoning
– The inference is based on logical rule-entailment
– Full RDFS, OWL Horst, restricted OWL-Lite, OWL2-QL and OWL2 RL
– Custom semantics can be defined via rules and axiomatic triples
7. OWLIM in the Cultural Heritage Domain
Selected commercial projects
ResearchSpace project funded by the Andrew W. Mellon Foundation
Support for collaborative web-based research, information sharing and web publishing for
the cultural heritage scholarly community. An Ontotext-led international consortium.
The Polish Digital National Museum aggregates artifacts from over 70 contributing
cultural institutions in the Digital Libraries Federation PIONIER Network using OWLIM
repository of Ontotext
LODAC (Linked Open Data in Academia), Japan's National Institute of Informatics
aggregates various information across multiple Japanese resources as LOD. The system
uses 8 OWLIM nodes and aggregates 19 collections with 700 000 entities and 15M triples.
SemTech for Cultural Heritage project funded by ITCC
Semantic publishing of Bulgarian cultural heritage to Europeana Establishing a Bulgarian
technical aggregator for Europeana
Selected research projects
MOLTO FP7 project, a use case in cultural heritage for a semantic knowledge
representationinfrastructure for querying RDF and presenting query results, includes close
to 9K museum objects from two collections of The Gothenburg City
Charisma (Cultural Heritage Advanced Research Infrastructures) an EU-funded
integrating activity project, a consortium of 21 partners, metadata from 6 major European
cultural institutions has selected OWLIM repository of Ontotext
8. OWLIM PERFORMANCE
• OWLIM is a scalable, robust and efficient triple store
– Serving the two most important web-sites for the London Olympic Games
• Official Olympics website
• BBC Olympics website
– Performance highlights
• OWLIM loads the 100M and the 200M datasets almost twice as fast as the next best product
(17 min. for 100M)
• Best query performance among those repositories that can handle update and multi-client
query tasks (5,285 Query-mixes-per-hour, where a query mix contains 25 queries; e.g. about
100 queries/sec)
• OWLIM v5 is 43% faster than v.4.3 on the BSBM Explore and Update scenario
• OWLIM v5 requires between 25% and 70% less storage space
• OWL 2 RL-type languages have proven to be the only feasible approach for
reasoning with billion statements
10. owl:sameAs Optimization
a way to handle the equivalent statements by a single master node,
which has as an impact efficient and compact handling of inferred
statements resulting in 4-6 times more statements available to query
than the explicitly introduced ones
11. OWLIM Replication Cluster
• Distribution through data replication is used to ensure:
– Better handling of concurrent user requests
– Failover support
• How does it work?
– Every user request is pushed in a transaction queue
– Each data write request is are multiplexed to all repository instances
– Each read request is dispatched to one of the
instance only
– To ensure load-balancing, each
read requests is send to the
instance with smallest execution
queue at this point in time
12. Geo-spatial index
• Geo-spatial information concerns the geometry of points, shapes and distances relative to the
surface of the Earth (or any spherical object).
• When using OWLIM-SE all angles are in decimal degrees with the latitude ranging from -90 to
+90 degrees and the longitude ranging from -180 to +180 degrees.
• airports have a reference point given by latitude, longitude and altitude;
• political boundaries can be specified by polygons where each vertex is a 2-Dimensional
latitude/longitude pair.
13. RDF Rank
• OWLIM-SE includes a plug-in that allows for efficient
calculation of a modification of PageRank over RDF graphs
• Computation of rank values is fast, e.g.
– 400M LOD statements takes 310 sec (27 iteraions)
• Results are available through a system predicate
• Example: get the 100 most important nodes in the RDF graph
SELECT ?n {?n rank:hasRDFRank ?r}
ORDER BY DESC(?r) LIMIT 100
14. Define: nested repositories
”Nested repositories” represent a new data
management concept for RDF data:
• a mechanism for sharing data stored across
multiple repositories, where
• one of them contains a large body of
knowledge which gets embedded in other
repositories
• each containing more specific data, which are
being interlinked with the common body of
knowledge