this post was submitted on 15 Jul 2024
1982 points (99.6% liked)

Technology

59087 readers
3267 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Switzerland mandates all software developed for the government be open sourced

Switzerland mandates software source code disclosure for public sector: A legal milestone

https://joinup.ec.europa.eu/collection/open-source-observatory-osor/news/new-open-source-law-switzerland

@technology@lemmy.world

#tech #libre

you are viewing a single comment's thread
view the rest of the comments
[–] vk6flab@lemmy.radio 98 points 3 months ago (3 children)

In contrast, abandoned open source software can be picked up and updated by whomever gets paid to, where abandoned closed source software needs to be reimplemented from scratch at great expense to the tax payer.

Not only that, open source software can be adopted by the community (who already paid for the development through their taxes) for their own purposes. Consider for example the productivity impact on business that starts using tools that it cannot afford to develop itself.

Office things like document management, workflow management, accounting, but also tools used in the science community, transport and logistics, anything that government does is represented in some other way in society.

This is a big deal and I hope that it will reverberate across the globe and become the new normal.

Whilst we're at it, consider the impact of open data, where government datasets are available to the community.

[–] Randelung@lemmy.world 16 points 3 months ago (1 children)

I'll gladly upload my stuff into some repo they allow me to. I've inquired about it in the past - I wrote a piece of sw that fills a requirement hole left by a widely used SCADA tool - but they outright forbid it. That was about a year ago.

My point is less about open source and more about how they have no clue how to handle their IP even now. It's a nice gesture at best (at least currently. Maybe there's more on the way).

[–] vk6flab@lemmy.radio 2 points 3 months ago (1 children)

Who is "they" in your statement?

If it's the company who is contracted by the government, it seems obvious (to me) that the requirements to make it open source provides the push to make it public.

If it's the government, then I don't understand your point.

[–] uis@lemm.ee 2 points 3 months ago* (last edited 3 months ago)

the Swiss government, namely ASTRA.

[–] ulterno@lemmy.kde.social 8 points 3 months ago* (last edited 3 months ago) (1 children)

Whilst we’re at it, consider the impact of open data, where government datasets are available to the community.

That sounds like it would be pretty useful to get better quality statistical research papers (well, I guess quality would depend more upon the researcher), doable by people without corporate backing.

Isn't it already available in a lot of cases?

[–] vk6flab@lemmy.radio 7 points 3 months ago (3 children)

Here's some of what's happening in my country, Australia:

Not sure where Tasmania and the ACT are at, but those links are the federal and most state government data portals.

Behind that is much variety of data, from land use to baby names and everything in-between.

The Australian Bureau of Statistics has its own site:

[–] Dave@lemmy.nz 6 points 3 months ago (1 children)

NZ as well: https://data.govt.nz

Though this it takes work for the different government departments to maintain. The team at data.govt.nz work with the different government departments to try to identify suitable data sources and get them into an update cycle, but there's definitely not all data that can be released on there.

[–] vk6flab@lemmy.radio 4 points 3 months ago (1 children)

Yeah, same kind of process in Oz.

AFAIK, it was triggered by doing an annual event called GovHack where people were encouraged to create "hacks" with government data. It included software developers like me, data mentors from many different government departments, people with an interest and several departments with questions.

[–] Dave@lemmy.nz 4 points 3 months ago (1 children)

I think NZ's is a similar story. GovHack is run in NZ as well, though I haven't personally been involved in an event.

[–] vk6flab@lemmy.radio 3 points 3 months ago (1 children)

A decade ago I participated in three and won several awards but was disappointed with the government response to all our collective efforts and stopped participating.

Specifically "not invented here" was prevalent as a response to projects that represented hundreds of man-hours of effort.

It was demoralising to say the least.

I'm not sure what the missing ingredient was, but two of our projects were directly related to government effort in relation to public transport and public housing. Neither went anywhere despite face to face presentations to senior stakeholders in the relevant departments.

The third was a search engine with a completely different approach to that in use by the popular engines.

[–] Dave@lemmy.nz 2 points 3 months ago (1 children)

That sucks. What was the novel search engine approach?

[–] vk6flab@lemmy.radio 4 points 3 months ago (1 children)

Using the idea of six degrees of separation to get to any person on the planet, I came up with the idea to use a word cloud that would represent the top N words in all documents.

When you click on a word, (say "alpha") the resulting word cloud would represent the top N words for all the documents with "alpha" in it.

As you click, bravo -> charlie, etc. the list of documents gets smaller and smaller, until just your required document remains.

This has several advantages, you don't need to distinguish between words and numbers or need to "understand" the meaning of a word or interpret the user intent.

More importantly, the user doesn't need to know the relevant words or vocabulary, since they're all represented in the UI.

Enhancements include allowing for negative words, as-in, exclude documents with this word.

[–] Dave@lemmy.nz 1 points 3 months ago (1 children)

Ah that sounds really interesting! Does it scale OK? I guess you could index at a word level and filter quite quickly for quick searches, but it seems you're going to have to store the full text of every website?

[–] vk6flab@lemmy.radio 2 points 3 months ago (1 children)

You store just the word count for each word on each URL.

The search is pretty trivial in database terms since you don't need to do any wildcard or like matching.

[–] Dave@lemmy.nz 1 points 3 months ago (1 children)

Ah of course!

I guess one of the things the Google originally solved was that the internet if full of crap and not all sites should have equal weighing. With AI spam sites these days, you'd probably also need a method of weighting results?

[–] vk6flab@lemmy.radio 2 points 3 months ago (1 children)

We never got that far to test that kind of issue and while I've been reimplementing it locally to search through employment advertising, I'm not at a point where I'd be able to test such a thing.

The original implementation used a data store written by another team member and it made the original project much too complicated.

Today I'd likely use duckdb to implement it. My local version uses text files for a proof of concept implementation.

[–] Dave@lemmy.nz 1 points 3 months ago

It sounds like a really cool project regardless!

[–] uis@lemm.ee 2 points 3 months ago

Whilst we're at it, consider the impact of open data, where government datasets are available to the community

*imagines Moscow* You still would need more trees and fix old rain drain system.