1
0
mirror of https://github.com/RipMeApp/ripme.git synced 2025-08-30 17:20:20 +02:00

Merge pull request #2053 from RipMeApp/fix-deprecated-url-ctors

Fix usage of deprecated URL constructors, which found a couple of issues
This commit is contained in:
soloturn
2025-01-02 11:53:39 +01:00
committed by GitHub
8 changed files with 52 additions and 32 deletions

View File

@@ -121,6 +121,7 @@ the following combinations of tags:
- testSlow runs tests with tag "slow".
- tests can be run by test class, or single test. Use "testAll" so it does
not matter if a test is tagged or not.
- tests can give the full stack of an assertion, exception, or error if you pass `--info` to the command
```bash
./gradlew test
@@ -129,6 +130,7 @@ the following combinations of tags:
./gradlew testSlow
./gradlew testAll --tests XhamsterRipperTest
./gradlew testAll --tests XhamsterRipperTest.testXhamster2Album
./gradlew testAll --tests ChanRipperTest --info
```
Please note that some tests may fail as sites change and our rippers

View File

@@ -60,11 +60,15 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
public Document getNextPage(Document doc) throws IOException, URISyntaxException {
return null;
}
protected abstract List<String> getURLsFromPage(Document page) throws UnsupportedEncodingException;
protected abstract List<String> getURLsFromPage(Document page) throws UnsupportedEncodingException, URISyntaxException;
protected List<String> getDescriptionsFromPage(Document doc) throws IOException {
throw new IOException("getDescriptionsFromPage not implemented"); // Do I do this or make an abstract function?
}
protected abstract void downloadURL(URL url, int index);
protected DownloadThreadPool getThreadPool() {
return null;
}
@@ -167,7 +171,7 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
for (String imageURL : imageURLs) {
index += 1;
LOGGER.debug("Found image url #" + index + ": '" + imageURL + "'");
downloadURL(new URL(imageURL), index);
downloadURL(new URI(imageURL).toURL(), index);
if (isStopped() || isThisATest()) {
break;
}
@@ -182,19 +186,26 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
if (isStopped() || isThisATest()) {
break;
}
textindex += 1;
LOGGER.debug("Getting description from " + textURL);
String[] tempDesc = getDescription(textURL,doc);
if (tempDesc != null) {
if (Utils.getConfigBoolean("file.overwrite", false) || !(new File(
URL url = new URI(textURL).toURL();
String filename = fileNameFromURL(url);
boolean fileExists = new File(
workingDir.getCanonicalPath()
+ ""
+ File.separator
+ getPrefix(index)
+ (tempDesc.length > 1 ? tempDesc[1] : fileNameFromURL(new URL(textURL)))
+ ".txt").exists())) {
+ (tempDesc.length > 1 ? tempDesc[1] : filename)
+ ".txt").exists();
if (Utils.getConfigBoolean("file.overwrite", false) || !fileExists) {
LOGGER.debug("Got description from " + textURL);
saveText(new URL(textURL), "", tempDesc[0], textindex, (tempDesc.length > 1 ? tempDesc[1] : fileNameFromURL(new URL(textURL))));
saveText(url, "", tempDesc[0], textindex, (tempDesc.length > 1 ? tempDesc[1] : filename));
sleep(descSleepTime());
} else {
LOGGER.debug("Description from " + textURL + " already exists.");

View File

@@ -8,6 +8,7 @@ import org.json.JSONObject;
import java.io.File;
import java.io.IOException;
import java.net.MalformedURLException;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.URL;
import java.nio.charset.StandardCharsets;
@@ -94,7 +95,7 @@ public abstract class AbstractJSONRipper extends AbstractRipper {
index += 1;
LOGGER.debug("Found image url #" + index+ ": " + imageURL);
downloadURL(new URL(imageURL), index);
downloadURL(new URI(imageURL).toURL(), index);
}
if (isStopped() || isThisATest()) {

View File

@@ -6,6 +6,7 @@ import com.rarchives.ripme.utils.Utils;
import com.rarchives.ripme.utils.RipUtils;
import java.io.IOException;
import java.net.MalformedURLException;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.URL;
import java.util.ArrayList;
@@ -208,7 +209,7 @@ public class ChanRipper extends AbstractHTMLRipper {
return false;
}
@Override
public List<String> getURLsFromPage(Document page) {
public List<String> getURLsFromPage(Document page) throws URISyntaxException {
List<String> imageURLs = new ArrayList<>();
Pattern p; Matcher m;
for (Element link : page.select("a")) {
@@ -254,8 +255,8 @@ public class ChanRipper extends AbstractHTMLRipper {
//Copied code from RedditRipper, getFilesFromURL should also implement stuff like flickr albums
URL originalURL;
try {
originalURL = new URL(href);
} catch (MalformedURLException e) {
originalURL = new URI(href).toURL();
} catch (MalformedURLException | URISyntaxException | IllegalArgumentException e) {
continue;
}

View File

@@ -10,6 +10,8 @@ import org.jsoup.nodes.Document;
import java.io.IOException;
import java.io.UnsupportedEncodingException;
import java.net.MalformedURLException;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.URL;
import java.net.URLEncoder;
import java.util.ArrayList;
@@ -27,10 +29,10 @@ public class LusciousRipper extends AbstractHTMLRipper {
}
@Override
public URL sanitizeURL(URL url) throws MalformedURLException {
public URL sanitizeURL(URL url) throws MalformedURLException, URISyntaxException{
String URLToReturn = url.toExternalForm();
URLToReturn = URLToReturn.replaceAll("https?://(?:www\\.)?luscious\\.", "https://old.luscious.");
URL san_url = new URL(URLToReturn);
URL san_url = new URI(URLToReturn).toURL();
LOGGER.info("sanitized URL is " + san_url.toExternalForm());
return san_url;
}

View File

@@ -56,6 +56,7 @@ public class NudeGalsRipper extends AbstractHTMLRipper {
for (Element thumb : thumbs) {
String link = thumb.attr("src").replaceAll("thumbs/th_", "");
String imgSrc = "http://nude-gals.com/" + link;
imgSrc = imgSrc.replaceAll(" ", "%20");
imageURLs.add(imgSrc);
}

View File

@@ -82,7 +82,7 @@ public class XvideosRipper extends AbstractSingleFileRipper {
String[] lines = e.html().split("\n");
for (String line : lines) {
if (line.contains("html5player.setVideoUrlHigh")) {
String videoURL = line.replaceAll("\t", "").replaceAll("html5player.setVideoUrlHigh\\(", "").replaceAll("\'", "").replaceAll("\\);", "");
String videoURL = line.strip().replaceAll("\t", "").replaceAll("html5player.setVideoUrlHigh\\(", "").replaceAll("\'", "").replaceAll("\\);", "");
results.add(videoURL);
}
}

View File

@@ -2,6 +2,8 @@ package com.rarchives.ripme.ripper.rippers.video;
import java.io.IOException;
import java.net.MalformedURLException;
import java.net.URI;
import java.net.URISyntaxException;
import java.net.URL;
import java.util.List;
import java.util.regex.Matcher;
@@ -51,7 +53,7 @@ public class MotherlessVideoRipper extends VideoRipper {
}
@Override
public void rip() throws IOException {
public void rip() throws IOException, URISyntaxException {
LOGGER.info(" Retrieving " + this.url);
String html = Http.url(this.url).get().toString();
if (html.contains("__fileurl = '")) {
@@ -62,7 +64,7 @@ public class MotherlessVideoRipper extends VideoRipper {
throw new IOException("Could not find video URL at " + url);
}
String vidUrl = vidUrls.get(0);
addURLToDownload(new URL(vidUrl), HOST + "_" + getGID(this.url));
addURLToDownload(new URI(vidUrl).toURL(), HOST + "_" + getGID(this.url));
waitForThreads();
}
}