mastodontech.de ist einer von vielen unabhängigen Mastodon-Servern, mit dem du dich im Fediverse beteiligen kannst.
Offen für alle (über 16) und bereitgestellt von Markus'Blog

Serverstatistik:

1,5 Tsd.
aktive Profile

#code4lib

1 Beitrag1 Beteiligte*r0 Beiträge heute

Darf ich vorstellen: codeberg.org/weirdfox/mfs

Ein Tool, mit dem man Metadaten herunterladen kann. Gerade nur Publikationen über DOI von Crossref / DataCite / OpenAlex, wird aber in alle möglichen Richtungen erweitert. Nutze es um Rust zu lernen, ist aber ernst gemeint :)

Momentan eine sehr Alpha-Version verfügbar, von jeglicher produktiven Nutzung ist stark abzuraten.

Ich freue mich über Tipps, Ideen, Kritik!

Über neue Releases wird hier ebenfalls berichtet.

Codeberg.orgmfsMetadata fetcher (software)

The first alpha version of Skosmos 3 has been published! This release provides a peek into the upcoming next major version of the Skosmos publishing tool for SKOS controlled vocabularies.

The release features a reimplemented front-end with a fresh layout and improved accessibility, as well as many architectural improvements and modernization of the codebase.

github.com/NatLibFi/Skosmos/re

GitHubRelease Skosmos 3.0-alpha.1 · NatLibFi/SkosmosThis is the first alpha release of Skosmos 3.0 that provides an early peek into the upcoming final 3.0 release. It is mainly intended for Skosmos users who want to plan ahead for the 3.0 release. Y...
#Skosmos#SKOS#thesaurus

Version 1.3 of the automated subject subject indexing tool #Annif has been released!

This release introduces support for the EstNLTK analyzer for better Estonian lemmatization 🇪🇪, optimizations to the MLLM backend, as well as maintenance and bug fixes, including better file permissions in multi-user environments.

github.com/NatLibFi/Annif/rele

GitHubRelease Annif 1.3 · NatLibFi/AnnifThis release introduces a new EstNLTK analyzer, improves the performance of the MLLM backend and fixes minor bugs. The key enhancement of this release is the addition of a new analyzer for lemmatiz...

I've been fooling around with a #Python script for converting #MARC bibliographic records to #Markdown for use with a static website generator. Probably only has an audience of one (me), but it's not complicated and it solves an interesting problem for private #librarians: how can we get our collections out there when LibraryThing isn't enough but a full OPAC is too much?

github.com/dazzlepansy/marc2md

GitHubGitHub - dazzlepansy/marc2md: Convert MARC bibliographic records to MarkdownConvert MARC bibliographic records to Markdown. Contribute to dazzlepansy/marc2md development by creating an account on GitHub.
Fortgeführter Thread

The only named defendant in this complaint is an experienced library technologist and #code4lib community member, who has since filed a motion to dismiss (courtlistener.com/docket/68157) on April 15th, 2024, in which she attested she has no connection to the WorldCat scraping (courtlistener.com/docket/68157) done by Anna’s Archive.

Im not sure who at #Code4Lib to ping on this, the contact just points to the instance, but re: requests for advice on avoiding problems like this, you can both improve the quality of the pub and add a method to catch problems like this by a) requiring all datasets come with a manifest that describes everything in them, field by field, eg. See datasheets for datasets, and b) making it the authors responsibility that the reviewers are able to properly evaluate the work, including providing any necessary I/O code or alternate formats needed. If your reviewers cant use the data, neither can your readers.

journal.code4lib.org/articles/

arXiv.orgDatasheets for DatasetsThe machine learning community currently has no standardized process for documenting datasets, which can lead to severe consequences in high-stakes domains. To address this gap, we propose datasheets for datasets. In the electronics industry, every component, no matter how simple or complex, is accompanied with a datasheet that describes its operating characteristics, test results, recommended uses, and other information. By analogy, we propose that every dataset be accompanied with a datasheet that documents its motivation, composition, collection process, recommended uses, and so on. Datasheets for datasets will facilitate better communication between dataset creators and dataset consumers, and encourage the machine learning community to prioritize transparency and accountability.

Good morning, everyone.

On December 4, 2023, Code4Lib Journal released the third article in four years with major patron privacy and ethics issues, an article that included an actual data breach.

Please consider signing this open letter calling for the editorial board to make transparent and significant changes to align with library ethics, research ethics, and correct information-security practices - docs.google.com/document/d/1c-

:boost_ok:

Google DocsOpen Letter to the Code4Lib Journal Editorial Board