Der aktuelle MEDIEN INTERNET und RECHT-Newsletter wurde versandt.
Themen u.a.: #Scraping, Unzulässige #Werbung in Social-Media-Netzwerken, #Ladenöffnung an Sonntagen, #Preiswerbung für Photovoltaik-Produkte, muenchen.de u.a.m. ... 🤗
Anmelden? 📮 http://newsletter.medien-internet-und-recht.de
#scraping #werbung #ladenoffnung #preiswerbung
#Scraping - Bei der Geltendmachung datenschutzrechtlicher Ansprüche (#Schadenersatz, #Unterlassung und #Auskunft) wegen eines Scraping-Vorfalls auf einer Social-Media-Plattform ist eine #Wertfestsetzung in Höhe von insgesamt EUR 6.000,00 angemessen
👉 OLG Frankfurt a.M., http://miur.de/3304
#Datenschutzrecht #SocialMedia #Streitwert
📮MEDIEN INTERNET und RECHT Newsletter abonnieren? http://newsletter.medien-internet-und-recht.de
#scraping #schadenersatz #unterlassung #auskunft #wertfestsetzung #datenschutzrecht #socialmedia #streitwert
Edöb äussert sich zu "Data #Scraping" - inside-it.ch https://www.inside-it.ch/edoeb-aeussert-sich-zu-data-scraping-20230828 #Datenschutz #privacy #SocialMedia
#scraping #datenschutz #privacy #socialmedia
@brook have you considered excluding sunny.garden from #AI #scraping #openai?
robots.txt
User-agent: GPTBot
Disallow: /
It seems odd to disallow ai art yet leave published original art susceptible to ai scraping, for example.
This is a genuine question not a complaint, there's lots I dont know about this area. Thanks for your work 🙂
2,6 Millionen Datensätze von #Duolingo-Nutzern bei Have I Been Pwned | Security https://www.heise.de/news/2-6-Millionen-Datensaetze-von-Duolingo-Nutzern-bei-Have-I-Been-Pwned-9283391.html #haveibeenpwned #Datenschutz #privacy #DataLeak #Datenleck #Scraping #DataBreach #DataTheft
#duolingo #haveibeenpwned #datenschutz #privacy #dataleak #datenleck #scraping #databreach #datatheft
HackRead: Overcoming web scraping blocks: Best practices and considerations https://www.hackread.com/web-scraping-blocks-practices-considerations/ #DataScraping #WebScraping #Technology #javascript #Scraping #Python #HowTo
#datascraping #webscraping #technology #javascript #scraping #python #howto
HackRead: Overcoming web scraping blocks: Best practices and considerations https://www.hackread.com/web-scraping-blocks-practices-considerations/ #DataScraping #WebScraping #Technology #javascript #Scraping #Python #HowTo
#datascraping #webscraping #technology #javascript #scraping #python #howto
I'm actually not entirely against AIs #scraping the web.
Once the genie is out of the bottle, you can't put it back in. If there's some content out there that is freely accessible, and it can be used to make large models better, it will certainly be used - we shouldn't be too naive or ideological about that.
I've always supported total freedom of scraping for everyone. I've always supported a world were all the content on the Internet can also be parsed by machines (that was the entire idea behind the semantic web). Once public content is out there, we lose control over who accesses it and for what purposes - that's simply how the web works.
But if Google and Meta are suddenly in this "we ♥ scraping" mood, I'd expect them to stick to their words and allow bidirectional scraping at least.
As an AI geek, I'd love to train my models on large corpora of audio extracted from YouTube videos. Or what people post in public Facebook groups when particular events happen. Or how the price of a product fluctuates on Amazon as the result of several external factors.
But I can't legally do any of these things. Those platforms are sealed, their APIs are very limited by design, only a limited amount of researchers can access some of that data (after signing lengthy NDAs and agreeing that the mother company will decide if the research can be published), and they will have tons of frontend-only checks to ensure that only a human downloads that content - and that they watch a sufficient amount of ads in the process. Not only - the developers behind scraping software like youtube-dl also get regularly harassed by Google.
So how come should I tolerate a world where if you're big enough you can afford to scrape the shit out of everyone, and use that knowledge to become even bigger and more powerful, but nobody is allowed to do the same with your own content?
We urgently need regulation that creates a level playing field when it comes to automated access to online information.
Freedom of scraping means freedom of growing. We can't give this freedom only to those who are already big enough. That's an unfair economic system with insurmountable entry barriers.
We need to make web scraping a fundamental human right.
And large companies should be compelled with sharing their data without barriers to scrapers too, if they aren't willing to build proper APIs.
Until that happens, I'll keep scraping the shit out of those monopolists without feeling an inch of guilt.
It's not even funny how easy it is to #scrape a #website which implements super strong anti #scraping tactics. Wrote a script today which rotates IP and User agent on every request. It also simulates all the #cookie magic that the website implements to rate limit scrapers. #AWS can actually help you implement scrapers like a DDOS farm.
#scrape #website #scraping #cookie #aws
Der Datenschutz Talk: Schüler verpflichtet Land #NRW - #Datenschutz News KW 30/2023
Was ist in der KW 30 in der Datenschutzwelt passiert, was ist für Datenschutzbeauftragte interessant ?
Wir geben einen kurzen Überblick der aktuellen Themen:
Kein Löschanspruch: VGH München, Beschluss vom 29.06.2023,
Schadensersatz #Scraping:
#Facebook
LG Ravensburg, Urteil vom 13.06.2023,
Schüler gegen NRW 🤩
Bußgeld wegen Paparazzifotos in die Privatwohnung Prominenter
Österreicher wegen #Überwachung der Ehefrau verurteilt
Kinder: Datenschutzverstoß durch Filmaufnahmen
Empfehlungen & Lesetipps:
E-Mail Kommunikation sicher gestalten
Webseite der Episode: https://migosens.de/schuler-verpflichtet-land-nrw-datenschutz-news-kw-30-2023/
Mediendatei: https://migosens.de/podlove/file/781/s/feed/c/ddt/DDT228.mp3
#uberwachung #facebook #scraping #datenschutz #nrw
I just read that Google is scraping Google Docs for their AI "training database." Wondering if this is indeed the case.
Some text files in my Drive folders convert to using the Google Docs interface when I open them. So I was wondering where one can store stuff on the cloud that isn't beholden to such thievery.
Found this article: "Google Docs AI: Is it safe? I’m a novel writer and tech journalist — let’s talk" Rami Tabari
#writingcommunity #scraping #copyright #ai
#Scraping is so exhilarating, I was hitting a website so hard it went down
It was our own website, I took our own website down
If anyone asks I was just red teaming
Is it possible to scrape dynamic webpages with #GoogleAppsScript?
#dtv #scraping
#scraping #dtv #googleappsscript
📝 "Web Scraping for Data Scientists (With No Web Programming Background)"
👤 Valery C. Briz (@valerybriz)
🔗 https://dev.to/valerybriz/web-scraping-for-data-scientists-with-no-web-programming-background-1j3a
#pyladies #python #oldiebutgoodie #datascience #datamining #scraping
#pyladies #python #oldiebutgoodie #datascience #datamining #scraping
Streitwert in "Scraping"-Verfahren
Rechtskonflikt im digitalen Zeitalter: Die Bedeutung von #Scraping und #Datenschutz
Das vorgegebene #Urteil wirft ein Schlaglicht auf ein drängendes Thema des digitalen Zeitalters: den Datenschutz in Zusammenhang mit dem sogenannten Scraping.
Symbolfoto:Andrii Yalanskyi /Shutterstock
#scraping #Datenschutz #Urteil #internetrecht
Facebook Nutzer erhalten 500 EUR Schadensersatz wegen DSGVO Verstoß
Facebook Nutzer erhielten 500 EUR Schadensersatz wegen einem erneuten DSGVO Verstoß. Im Zeitraum von 2018 bis 2019 wurden die Daten von 533 Millionen Facebook-Accounts von unbekannten Angreifern kompromittiert. Dabe
https://dsgvo.watch/gerichtsurteile/facebook-nutzer-500euro-schadensersatz-landgericht-luebeck/
#Darknet #Datenschutz #DSGVO #Facebook #Gerichtsurteile #LandgerichtLbeck #Meta #Schadensersatz #Scraping
#darknet #datenschutz #dsgvo #facebook #gerichtsurteile #landgerichtlbeck #meta #schadensersatz #scraping
@rustybrick do you think #Google will just ignore Robots.txt? I mean they'd love to be able to train on everything.
Though I would love more controls on stopping AI from scraping without blocking my sites from Search
#google #seo #ai #scraping #robotstxt
Looks like #Google has finally succumbed to the dark side.
https://gizmodo.com/google-says-itll-scrape-everything-you-post-online-for-1850601486
#google #ai #scraping #fediverse #creators