Unlock the full potential of your data with Linked Data - applying web principles to data integration, we empower valuable insights, underpin data engineering and data science and enable innovative data solutions. 💡💻
Very interesting contribution by Miels Vander Sande at #semantics2023 on 'Managing the #metadata of a diverse digital media archive as a #knowledgegraph' #semanticweb #linkeddata #lod #metadata #iiif #oaipmh Slides: https://www.dropbox.com/scl/fi/kzcp6k4hvjltxnm5orxoa/20230922_semantics.pdf?rlkey=fqxl08mhkk1wwrmz478tz3nak&dl=0 @mielvds
Tangential. There was a recent good HN discussion on sequence diagrams: https://news.ycombinator.com/item?id=36342931
Recently when checking out #GaiaX Federated Catalogues (a #LinkedData spec totally unrelated to #Fediverse) I noted their use of #BPMN (Business Process Modeling Notation) - an OMG open standard - as a possible interesting way to diagram behavioral aspects of #ActivityPub extensions.
Wanting to manage and publish #sensor and #observational data then we’d love to chat - we’ve extensive experience in publishing such data at scale with previous projects with clients such as UKCEH, the Environment Agency and others plus our own Agora measurement store technology.
See: https://www.epimorphics.com/projects for more on those projects.
One of the most effective ways to do it is to make the chatbot refer to sources of facts. Tag every RAG document partial with an anchor, and tell the chatbot refer to them. Then you can show in a final presentation a link to the document part used.
In a document index you similarly want to include metadata for partials, so that a user can click "show the whole document" and navigate the information if they want to see where the content came from.
Calling on to #SemanticWeb #LinkedData and the knowledgeable folks alike for hints: how is the subjective nature of knowledge addressed in building knowledge graphs? Where are these questions being answered or probed?
Subjectivity seem to be less relevant in the applications having a narrow user base (e.g. a single company). Though it's more interesting in case of e.g. Wikidata corpus, where a subjective aspects might vary a lot and impact how the data is interpreted (culture, trust...).
- Had some really interesting deep dives in some client workshops,
- Completed some client project interviews,
- Completed a development sprint on one of our projects and,
- We’ve been highlighting last weeks #Hydrology #API blog post as we also encourage users of the service to let us know how they use it.
Unlock the full potential of your data with linked-data - applying web principles to data integration, we empower valuable insights, underpin data engineering and data science and enable innovative data solutions.
Have a look at our past and current projects at: https://www.epimorphics.com/projects
Virtuoso 08.03.3339 PAGO edition for the Microsoft Azure cloud is now live.
* #IndieAuth & #RelMeAuth auth protocols support for self-sovereign identity (#SSI)
* #ActivityPub & #ActivityStreams Support
* #OpenAI Web Service import and export
* Enhanced native support of both #JSON and #JSONLD
Interested in solutions of how to integrate non-LOD research data resources into #knowledgegraphs? Come to my talk on Knowledge Graph-based research data management solutions in the Linking RDM 1 session at #Cordi2023
presentation slides: https://docs.google.com/presentation/d/1-q0At17YRndmUY1abP7hgccXYR5pgv2AT5xVdHHBwas/edit?usp=sharing
#lod #linkeddata #researchdata #nfdi #nfdirocks @NFDI @nfdi4culture @NFDI4DS #NFDIMatWerk @fizise @tabea @sashabruns @heikef @enorouzi @jonatan
Discover how the University of New Hampshire uses #LinkedData harvesting to uncover the languages and nationalities that are often overlooked in translation by watching "Learning about the Invisible: Using Linked Data to Enhance Collection Analysis." Watch now: https://youtu.be/-xNGMjRPYig?si=j5U9XCZNUaAZ0l2E
#HigherEdTech #DigitalLibraries #DigitalCollections #DEIA #Metadata
and in #LinkedData Zenodo community 🧵
Trognitz presents Linked Open Data in archeology in "Links everywhere: Linked Open Data for Aegean seals with Wikidata"
https://doi.org/10.5281/zenodo.8190065 -- part of an exciting PhD thesis "Computer-Based Analysis of Multi-Sided Minoan and Mycenean Seals" that combines machine learning and archeology.
Also in #LinkedData Zenodo community 🧵
"Optimizing SPARQL Queries using Shape Statistics" is the dataset used in experiments at https://relweb.cs.aau.dk/rdfshapes/
..see also their EDBT 2021 paper at https://openproceedings.org/2021/conf/edbt/p202.pdf
and datasets at https://kashif-rabbani.github.io/sparql-optimization/
Gruber's presentation "FAIR Findspots: Making Data Reusable" presents the FAIR principles for coin data, which you may remember from my earlier posts.
Garijo "Semantics 2023 Knowledge Graph" describes the scholarly outputs of Semantics 2023 conference, using schema.org.
Dataset "Terminology management tools comparison" by Křemen et al has evaluation of tools like TermIt, WebProtege, SKOS Shuttle
.. see also paper https://www.semantic-web-journal.net/content/termit-managing-normative-thesauri
Virtuoso 08.03.3339 PAGO edition for the Amazon Web Services (#AWS) cloud is now live.
* #IndieAuth & #RelMeAuth auth protocols support for self-sovereign identity (#SSI)
* #ActivityPub & #ActivityStreams Support
* #OpenAI Web Service import and export
* Enhanced native support of both #JSON and #JSONLD
Today you can watch me as I fumble around the digital landscape, using the second-century CE author, Pausanias, as my guide. Join me in person or online at 15:00 BST with @kingsdh 's computational research group, by signing up here: https://kingsdh.net/. It will include live demos (provided I make my sacrifices to the presentation gods).
#map #digitalhumanities #classics #geography #mapping #linkeddata #wikidata
My least favourite RFC 2119 MUST in the existance of the internet is https://www.w3.org/TR/tabular-metadata/#top-level-properties. #csvw #data #w3 #linkedData
I wonder what your feelings are wrt the LD aspects of AS/AP #Fediverse and related to the toots above?
In this #TechTalk article, Dave takes us through: Tracking small numbers of stations, Frequency of checking, Tracking multiple stations, Rewriting history, Good practice and fair use, and some ideas about the future of the API
In previous #TechTalk articles Tom has described the Environment Agency's (https://www.linkedin.com/company/environment-agency) #Hydrology service and in "Accessing the continuous monitoring archive via #API" article, gives a great introductory tour of the API
As well as supporting interactive exploration of the data and manual download of #TimeSeries, the #Hydrology service provides for automatic download of data through a rich web (REST) #API. https://www.epimorphics.com/ea-hydrology-api
Our newest #TechTalk article: Tracking #EnvironmentalData using the Environment Agency (https://www.linkedin.com/company/environment-agency) #Hydrology service #API
The hydrology archive is a tremendous asset with records from tens of thousands of measurement stations across England, this article talks about how best to use the API https://www.epimorphics.com/ea-hydrology-api
If anyone would like to watch me stumbling around the contemporary digital landscape, using Pausanias as my guide, come along to KCL's Computational Humanities research meeting next Tuesday at 15.oo BST. Sign up (for in person or remote attendance) here: https://kingsdh.net/
#classics #digitalhumanities #map #linkeddata #wikidata #OpenData #GLAM #geospatial
Some news from OSM-Wikidata Map Framework:
:overpass: Advanced wikidata=*+P* patterns can now be shown with Overpass+WDQS
🌐 The map switched to @maplibre GL JS
There are three new instances (beyond https://etymology.dsantini.it and https://burial.dsantini.it ):
📐 https://architect.dsantini.it : Architects who designed buildings and structures
🎨 https://artist.dsantini.it : Artists who created artworks and memorials
🌏 https://osmwd.dsantini.it : All OSM features linked to Wikidata
Fun with admin: Trying to contribute 16 rather trivial data points to an external data set triggered several month of contract drafts going back and forth, and two legal departments failing to agree on a liability clause (where none of us see any actual liability issue). I would tend to drop the case, but it would be nice to have “official” URIs for these data points. #LinkedData #LinkedOpenData #DigitalHumanities
As people return from their summer holidays, we hope to publish some more #TechTalk articles on a variety of topics.
Take a look at our blog for some of our recent articles: https://www.epimorphics.com/blog/
See our project page: https://www.epimorphics.com/projects/safety-alerts
💬 We’d also love to hear if you are an #API User 💬
See our project page: https://www.epimorphics.com/projects/regulated-products-register
💬 We’d also love to hear if you are a #DataUser 💬
See our project page: https://www.epimorphics.com/projects/hydrology-explorer
💬 We’d also love to hear if you are an #API User
UK Centre for Ecology & Hydrology (UKCEH) - one of our fantastic clients is recruiting a Semantic Web Architect in their Environmental Data Science team (Lancaster)
https://ceh.wd3.myworkdayjobs.com/en-US/CEH_Careers/job/Lancaster/Semantic-Web-Architect_JR1015 - lots of interesting projects.
Oh, and if you're into that sort of thing the linked data is represented in the quite new, as in still in development, Records in Context model/ontology.
As well as good old Schema.org of course.
, but given example prevent me to do so 🤪
We are also looking for a #PostDoc / Senior Researcher for the interdisciplinary Leibniz Science Campus project "Digital Transformation of Research" on the effects of #AI on Research in different disciplines. This is a full position 100%, German TVöD-13 paygrade according to experience, limited to 4 years with the option of extension/tenure.
More information: https://www.fiz-karlsruhe.de/en/stellenanzeigen/postdocsenior-researcher-wmx-wissenschaftscampus-ditrare
#knowledgegraphs #machinelearning #deeplearning #llm #semanticweb #linkeddata @fizise #jobad
The participant lists are slowly filling up. Register at https://swib.org/swib23/registration.html to have the full choice between five workshops.
Mapping ActivityStreams2 to #clojure spec is not as trivial as we hoped. But after some blood, sweat and tears we can now show first results:
And yes, at least we've proven our spec definitions by unittests :-)
We are intersted in feedback & discussion.
Are there any #LinkedData experts around?
In my world, #GND stands for "Gemeinsame Normdatei":
As a JSON-LD noob, I think this is what JSON-LD is trying to solve:
- maintain consistency on what data represents what
- provide a specification what document belongs to what set of objects (via @`type`)
- provide a specification on how documents are related to other documents (via @`id`)
But I might be wrong.
I'm thinking out loud because I'm trying to figure out how to make sense of JSON-LD
Nog twee dagen om onze enquête over het gebruik van linked data bij overheden in te vullen. Je hulp is van harte welkom.
Ik heb een extensie gemaakt voor Visual Studio Code waarmee je eenvoudiger #LinkedData in Turtle (TTL) en TriG kunt maken door veelgebruikte vocabulaires automatisch in te vullen (a.k.a. "autocomplete"). Zo zitten bv. alle schema.org classes en properties er in, voorzien van de bijbehorende beschrijving. Zie https://marketplace.visualstudio.com/items?itemName=MarkLindeman.turtle-vocab-autocomplete
In reality of course, the Fediverse is built on a stack of technologies that are defined by open standards aka protocols.
@sl007 starts by showing the first 8 minutes of this video that explains the data format JSON-LD: https://digitalcourage.video/w/jUKxSj7SCVe9t5rZA1piwU
Werkt jouw #overheidsorganisatie met #begrippenlijsten, #linkeddata of allebei? Vul onze enquête ✔ in (uiterlijk 27 juli). Dit kost je maar een paar minuten. We willen een beeld krijgen van gebruik van linked data 🌐 bij overheden.
@ryanfb Thanks for posting this. I'm going to be part of an ERC grant documenting and exploring the experiences of slaves in antiquity (broadly conceived) — SLaVEagents, led by Kostas Vlassopoulos (Crete). We're going to publishing the prosopography as #linkeddata. Maybe worth chatting sometime about the sources? #dh2023
I guess by starting talking (or rather writing a question in the Google Doc) about Linked Open Usable Data #LOUD - that is now very much tied to #LinkedArt and #IIIF (https://linked.art/loud) - in the #Pelagios session, lots of very interesting points have then been discussed around what defines it (API, Community-led). ping @azaroth42
Tomorrow I'll be speaking at the #LD4_2023 conference on #LinkedData about « #Wikidata & the sum of all video games: putting the "linked" in video game metadata » − join me at 16:30 CEST (14:30 UTC) https://2023ld4conferenceonlinkedda.sched.com/event/2176f0bade85b75195aab96c6777c884 https://www.youtube.com/watch?v=pvrxTh8xfdY
This Wednesday evening at #DH2023 I am going to be standing next to our poster titled "Enabling Participatory Data Perspectives for Image Archives through a Linked Art Workflow".
The basis of this poster is a collaboration with the @uniofoxford and more specifically with Tanya Gray, Alwyn Collinson, & Kevin R. Page who designed a Python-based three-step workflow that can generate JSON-LD #LinkedArt resources.
The goal is to expose those resources later as a RESTful API as a means of conveying semantically enriched events and as a benchmark against other collections leveraging this model alongside our existing #IIIF APIs and our basic metadata API within the PIA research project. #loud #LinkedData #pia_sinergia
Conference Paper: https://zenodo.org/record/8107399
If you're at @dh2023graz and want to learn about the complexities of mapping (an ancient) text and the collaborations it has facilitated, join us at the Visualizing text panel, 11-12.30 on Wednesday.
You can check out our abstract here: https://docs.google.com/document/d/1fXa6An00CknHqhypIJEZM90p3rqtgSR5BsWAnAhYULA/edit.
"With ActivityPods, you have only one profile, one outbox, one inbox and one list of followers - all in a single place. Applications connect to your Pod to post activities, read the inbox and fetch data. And of course they can connect to any existing fediverse application!"
I think our data portability problem may have been solved for us!
🆕 blog! “LinkedIn supports Schema.org metadata”
I'm a big fan of machine-readable metadata. It's useful for programs which need to extract information from messy and complicated websites. It's always surprising where it turns up. For example, take this post of mine on LinkedIn. If you view the source, you'll see this scrap of linked data: <script type="applicati…
Sets up stage for understanding and appreciating the semantics associated with entity identifiers.
For instance, you ultimately end up with an entity relationship graph comprising a variety of handles that denote the same entity.
In a #SemanticWeb this is ground-zero, since relationship type semantics exist for this kind of identity reconciliation 😀
Please fill in the survey if you are involved in either of these communities or if you have heard of those communities but aren't (yet) involved. 😃
@J12t "JSON-LD has benefits that have nothing to do with semantic web stuff, such as namespacing."
The following is ground-zero fora #SemanticWeb constructed from structured data deployed using #LinkedData principles.
## JSON LD Start ##
"name": "John Smith",
"jobTitle": "Software Engineer",
"telephone": "(123) 456-7890",
## JSON LD End ##
Re Prospective Search, the focal point is variables in the body of a #SPARQL query that satisfy conditions expressed in relations.
A query solution can always optionally include document identifiers (a/k/a named graphs).
Semantically, the following holds true (albeit currently missing from the <head/> section) of https://mastodon.social/tags/linkeddata.
Here's a very basic sponge (crawl) of content associated with #LinkedData from across Mastodon.
 https://medium.com/virtuoso-blog/what-is-the-virtuoso-sponger-middleware-about-and-why-is-it-important-293546b32a32 -- #VirtuosoRDBMS sponger (crawler) overview
This technology can crawl and merge data across links that denote hashtags across Mastodon instances, Twitter etc.. What it can't do is construct a canonical solution from an infinite universe of unknowns.
@bobwyman @J12t Prospective Search , is something that can be handled in a #SemanticWeb because the semantics of links (used to construct said #Web) enable a crawler to be fashioned within constrains (e.g., timeouts) while being informed navigationally by the semantics of relations denoted by links, courtesy of #LinkedData principles.
The above can happen within the context of a query or as an ETL-style background crawl.
Der #BullingerBriefderWoche stammt heute von Jakob Meyer aus #Basel und wurde am 8. März 1538, also vor 485 Jahren in Frühneuhochdeutsch verfasst. Der Zeitstrahl zeigt, dass Bullinger und Meyer sporadisch zwischen 1532 und 1539 in Kontakt standen. #digitalhumanities Zu Meyer findet sich nicht viel, aber sein Sohn (nehme ich mal an) hat eine Seite auf Wikipedia: https://de.wikipedia.org/wiki/Wolfgang_Meyer_(Theologe) Es wäre doch interessant, wenn es gelänge, diese Informationen zu verknüpfen. #LinkedData Meyers Grossvater wahr denn auch Bürgermeister von Basel & ein Unterstützer des Basler Reformators #Oekolampad, von welchem wir ebenfalls einen Brief in der Bullinger-Korrespondenz finden: https://bullinger-digital.ch/letter/10063. Zum Meyer-Geschlecht gibt es auch einen Artikel auf e-periodica: https://e-periodica.ch/digbib/view?pid=bzg-002:1925:23#110 Wir müssen diese Daten vernetzen!
I don’t know how any crawler solves the problem as described .
If I am interested in #LinkedData across instances that l encounter explicitly or serendipitously, our crawler can follow objects of owl:sameAs relations to wherever they lead, as part of a *timeout* constrained query solution pipeline or an explicit ETL-based crawl job .
The difference from conventional crawling boils down to an operation informed by the semantics of an owl:sameAs relation.
JSON-LD has the benefit of being maintained by W3C, being an open standard, having significant amount of momentum, and serving a plethora of different use-cases, many of which haven’t been invented yet.
+ [links on the semantic web](http://dig.csail.mit.edu/breadcrumbs/node/62)
+ [give yourself a URI](http://dig.csail.mit.edu/breadcrumbs/node/71)
+ [backward and forward links in RDF just as important](http://dig.csail.mit.edu/breadcrumbs/node/72)
the problem with #LinkedData technologies is that none of them work.
"here's a semantic web index, it's like a google for linked data"
Uh, what? No it's not. where do I search?
The only thing I see is API docs, surely that can't be right.
OK let's at least take a look anyway
god almighty what on earth
"here's a tool for extracting RDF-like data from web pages"
OK, apparently it's a web service? can I try it out?
no, not like that, not by downloading and compiling anything
halfway down the page, ok a live instance
(not a live instance)
I would love to see one linked data tool that isn't wikidata that a) tells me what it does, b) shows me a demo, c) lets me use it without having to read a whole book or compile some ancient blob of Java
ok so re-reading #IPFS paper and there are a few things I think in retrospect are undesirable about the #MerkelDAG spec. it's hard to parse them out as separable ideas because they depend on one another, but the main thing I think is how it conflates the structure of a metadata graph, the content of the graph, and the notion of authorship/identity.
In (basic) IPFS, each node contains some data and some links. the data is some unspecified binary blob, the links are all references to hashes of other nodes, and then the hash of all that identifies the node. There are some abstractions like flattened trees that can represent n-depth links, but that's the gist. I'm refreshing myself, so correct me where I'm wrong.
This makes traversing the graph expensive from a naive (cacheless) state- you have to fetch each node and parse its links serially, and since there isn't a notion of authorship except when used to sign a node, you might have to do the resolution process across a lot of the network instead of being able to say "ah ok this is from this identity so I should ask their neighborhood first"
Since the links are untyped, and because of the need for serial resolution, you can't really "plan" queries and move the query logic to the "edges" (in a networking, rather than graph parlance) of the network - the network resolution logic handles all that.
This structure also makes it so you can't "talk about" a node. A node contains its links. The links are directional, so I could make some statement about a node by pointing to it, but I can't, as a third party make a link under my identity, separate from the author and content of the node, that points from some object to another. That makes the network more like a hard drive than a social space.
Further, since links aren't typed, you have to move that metadata inside the node. This makes you need to re-hash each node more than you need to, and since "keys" for identifying different fields in the node aren't themselves links, you can't have any notion of "schema" where a term can be reused. So there isn't really a facility for being able to do graph queries like "find me this type of data whose field has this value" which restricts a whole huge range of possibilities too long to list here. This also makes knowing what the binary data inside a node is potentially impossible without out of band info, depending on how it's encoded. #IPLD and #Multiformats are intended to solve this, post-hoc.
I'll stop there for now, and save what I think could be a different model for later, but I am thinking along the lines of merging with #LinkedData #Triplets , encoding the notion of authorship into links (so that links can have an "utterance" rather than "fact" ontological status), a notion of container/contained for explicit block formation and metadata separation, and formalizing the notion of orthogonal Merkel DAGs to change the points where the content addressing happens to be able to have "graph subunits" that allow for cycles at a "complete" scope but for the purposes of hashing have no cycles. very much #WIP, still at conceptual stage haven't started writing spec yet.
For #ActivityPub you can arguably say that that also sort of applies.
Anyway, I don't have very strong opinion on whether or not AP should be a #LinkedData standard or not.
What I do know is that LD has this eternal struggle to gain adoption, and in the areas where its power lies, its complexity comes peeking around the corner. Making it very unloved in the broader developer community.
In theory LD is fabulous, leading to a magical wonderful semantic web, and in practice.. 🤔
In the context of responding to @helge #FEP proposal I bumped into a paper by Samuel M. Smith about the role of #LinkedData wrt the #W3C #VerifiableCredentials spec and thought about your earlier musings when exploring #ActivityPub and how #JSONSchema might have been used.
Sam's TL;DR is:
> The VC standard appears to be an adoption vector for Linked Data, not the other way around.
Might find it an interesting read:
And he claims the same applies to W3C #DID.
#AntiFascist, #queer and pro #OpenData and #OpenAccess: “Sharing isn’t immoral — it’s a moral imperative”
Aaron Swartz, Guerrilla Open Access Manifesto, 2008.
My blog'ish website: