How do you scrape stuff, I tried wget but it didn’t work. A login + a button press is required to see anything. I can do this manually if required. There are many different slides and videos though.
If you login in a browser, it’ll most likely give you a “session cookie” that you should be able to see in the developer tools. (If you’re using Firefox’s developer tools, it’d be under the “storage” section.) The name of the cookie will generally have the word “session” in the name. After logging in, that cookie identifies you to the server, letting the server know that “this particular request is from CucumberSalad” (or whatever your user is named on that service.) Wget probably hasn’t been working because the requests from wget don’t include that cookie like the requests from your browser do.
(Just looking at my developer tools while using Lemmy, it seems like the Lemmy web ui doesn’t use session cookies but rather a JSON web token in a cookie named “jwt”, but I think that cookie would suffice if I was trying to scrape the Lemmy web ui.)
Once you have the proper cookie name and value, you can have wget send the cookie to the server with each request by adding the flag
--header 'Cookie: <cookie name>=<cookie value>'
(but replace the values in angle brackets. Example:--header 'Cookie: JSESSIONID=ksdjflasjdfaskdjfaosidhwe'
.)Also, if you can provide more info as to what you’re trying to scrape, folks can probably help more. Hopefully what I’ve given you is enough to get you going, but it’s possible there might be more hurtles to jump to get it working.
Lovely comment tgankdfor tge help
What exactly are you trying to get?
A course like thing, with slides there are vidros but i dont care much about those. The url incrments depending on the slide but there are mini questions which stop the user from skipping ahead without them being answered.
Is a link really necessary?
No it’s optional like including an image
Oh ok