mirror of
https://github.com/RipMeApp/ripme.git
synced 2025-08-30 17:20:20 +02:00
Merge pull request #2053 from RipMeApp/fix-deprecated-url-ctors
Fix usage of deprecated URL constructors, which found a couple of issues
This commit is contained in:
@@ -121,6 +121,7 @@ the following combinations of tags:
|
||||
- testSlow runs tests with tag "slow".
|
||||
- tests can be run by test class, or single test. Use "testAll" so it does
|
||||
not matter if a test is tagged or not.
|
||||
- tests can give the full stack of an assertion, exception, or error if you pass `--info` to the command
|
||||
|
||||
```bash
|
||||
./gradlew test
|
||||
@@ -129,6 +130,7 @@ the following combinations of tags:
|
||||
./gradlew testSlow
|
||||
./gradlew testAll --tests XhamsterRipperTest
|
||||
./gradlew testAll --tests XhamsterRipperTest.testXhamster2Album
|
||||
./gradlew testAll --tests ChanRipperTest --info
|
||||
```
|
||||
|
||||
Please note that some tests may fail as sites change and our rippers
|
||||
|
@@ -60,11 +60,15 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
|
||||
public Document getNextPage(Document doc) throws IOException, URISyntaxException {
|
||||
return null;
|
||||
}
|
||||
protected abstract List<String> getURLsFromPage(Document page) throws UnsupportedEncodingException;
|
||||
|
||||
protected abstract List<String> getURLsFromPage(Document page) throws UnsupportedEncodingException, URISyntaxException;
|
||||
|
||||
protected List<String> getDescriptionsFromPage(Document doc) throws IOException {
|
||||
throw new IOException("getDescriptionsFromPage not implemented"); // Do I do this or make an abstract function?
|
||||
}
|
||||
|
||||
protected abstract void downloadURL(URL url, int index);
|
||||
|
||||
protected DownloadThreadPool getThreadPool() {
|
||||
return null;
|
||||
}
|
||||
@@ -167,7 +171,7 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
|
||||
for (String imageURL : imageURLs) {
|
||||
index += 1;
|
||||
LOGGER.debug("Found image url #" + index + ": '" + imageURL + "'");
|
||||
downloadURL(new URL(imageURL), index);
|
||||
downloadURL(new URI(imageURL).toURL(), index);
|
||||
if (isStopped() || isThisATest()) {
|
||||
break;
|
||||
}
|
||||
@@ -182,19 +186,26 @@ public abstract class AbstractHTMLRipper extends AbstractRipper {
|
||||
if (isStopped() || isThisATest()) {
|
||||
break;
|
||||
}
|
||||
|
||||
textindex += 1;
|
||||
LOGGER.debug("Getting description from " + textURL);
|
||||
String[] tempDesc = getDescription(textURL,doc);
|
||||
|
||||
if (tempDesc != null) {
|
||||
if (Utils.getConfigBoolean("file.overwrite", false) || !(new File(
|
||||
URL url = new URI(textURL).toURL();
|
||||
String filename = fileNameFromURL(url);
|
||||
|
||||
boolean fileExists = new File(
|
||||
workingDir.getCanonicalPath()
|
||||
+ ""
|
||||
+ File.separator
|
||||
+ getPrefix(index)
|
||||
+ (tempDesc.length > 1 ? tempDesc[1] : fileNameFromURL(new URL(textURL)))
|
||||
+ ".txt").exists())) {
|
||||
+ (tempDesc.length > 1 ? tempDesc[1] : filename)
|
||||
+ ".txt").exists();
|
||||
|
||||
if (Utils.getConfigBoolean("file.overwrite", false) || !fileExists) {
|
||||
LOGGER.debug("Got description from " + textURL);
|
||||
saveText(new URL(textURL), "", tempDesc[0], textindex, (tempDesc.length > 1 ? tempDesc[1] : fileNameFromURL(new URL(textURL))));
|
||||
saveText(url, "", tempDesc[0], textindex, (tempDesc.length > 1 ? tempDesc[1] : filename));
|
||||
sleep(descSleepTime());
|
||||
} else {
|
||||
LOGGER.debug("Description from " + textURL + " already exists.");
|
||||
|
@@ -8,6 +8,7 @@ import org.json.JSONObject;
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.net.MalformedURLException;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
import java.net.URL;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
@@ -94,7 +95,7 @@ public abstract class AbstractJSONRipper extends AbstractRipper {
|
||||
|
||||
index += 1;
|
||||
LOGGER.debug("Found image url #" + index+ ": " + imageURL);
|
||||
downloadURL(new URL(imageURL), index);
|
||||
downloadURL(new URI(imageURL).toURL(), index);
|
||||
}
|
||||
|
||||
if (isStopped() || isThisATest()) {
|
||||
|
@@ -6,6 +6,7 @@ import com.rarchives.ripme.utils.Utils;
|
||||
import com.rarchives.ripme.utils.RipUtils;
|
||||
import java.io.IOException;
|
||||
import java.net.MalformedURLException;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
import java.net.URL;
|
||||
import java.util.ArrayList;
|
||||
@@ -208,7 +209,7 @@ public class ChanRipper extends AbstractHTMLRipper {
|
||||
return false;
|
||||
}
|
||||
@Override
|
||||
public List<String> getURLsFromPage(Document page) {
|
||||
public List<String> getURLsFromPage(Document page) throws URISyntaxException {
|
||||
List<String> imageURLs = new ArrayList<>();
|
||||
Pattern p; Matcher m;
|
||||
for (Element link : page.select("a")) {
|
||||
@@ -254,8 +255,8 @@ public class ChanRipper extends AbstractHTMLRipper {
|
||||
//Copied code from RedditRipper, getFilesFromURL should also implement stuff like flickr albums
|
||||
URL originalURL;
|
||||
try {
|
||||
originalURL = new URL(href);
|
||||
} catch (MalformedURLException e) {
|
||||
originalURL = new URI(href).toURL();
|
||||
} catch (MalformedURLException | URISyntaxException | IllegalArgumentException e) {
|
||||
continue;
|
||||
}
|
||||
|
||||
|
@@ -10,6 +10,8 @@ import org.jsoup.nodes.Document;
|
||||
import java.io.IOException;
|
||||
import java.io.UnsupportedEncodingException;
|
||||
import java.net.MalformedURLException;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
import java.net.URL;
|
||||
import java.net.URLEncoder;
|
||||
import java.util.ArrayList;
|
||||
@@ -27,10 +29,10 @@ public class LusciousRipper extends AbstractHTMLRipper {
|
||||
}
|
||||
|
||||
@Override
|
||||
public URL sanitizeURL(URL url) throws MalformedURLException {
|
||||
public URL sanitizeURL(URL url) throws MalformedURLException, URISyntaxException{
|
||||
String URLToReturn = url.toExternalForm();
|
||||
URLToReturn = URLToReturn.replaceAll("https?://(?:www\\.)?luscious\\.", "https://old.luscious.");
|
||||
URL san_url = new URL(URLToReturn);
|
||||
URL san_url = new URI(URLToReturn).toURL();
|
||||
LOGGER.info("sanitized URL is " + san_url.toExternalForm());
|
||||
return san_url;
|
||||
}
|
||||
|
@@ -56,6 +56,7 @@ public class NudeGalsRipper extends AbstractHTMLRipper {
|
||||
for (Element thumb : thumbs) {
|
||||
String link = thumb.attr("src").replaceAll("thumbs/th_", "");
|
||||
String imgSrc = "http://nude-gals.com/" + link;
|
||||
imgSrc = imgSrc.replaceAll(" ", "%20");
|
||||
imageURLs.add(imgSrc);
|
||||
}
|
||||
|
||||
|
@@ -82,7 +82,7 @@ public class XvideosRipper extends AbstractSingleFileRipper {
|
||||
String[] lines = e.html().split("\n");
|
||||
for (String line : lines) {
|
||||
if (line.contains("html5player.setVideoUrlHigh")) {
|
||||
String videoURL = line.replaceAll("\t", "").replaceAll("html5player.setVideoUrlHigh\\(", "").replaceAll("\'", "").replaceAll("\\);", "");
|
||||
String videoURL = line.strip().replaceAll("\t", "").replaceAll("html5player.setVideoUrlHigh\\(", "").replaceAll("\'", "").replaceAll("\\);", "");
|
||||
results.add(videoURL);
|
||||
}
|
||||
}
|
||||
|
@@ -2,6 +2,8 @@ package com.rarchives.ripme.ripper.rippers.video;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.MalformedURLException;
|
||||
import java.net.URI;
|
||||
import java.net.URISyntaxException;
|
||||
import java.net.URL;
|
||||
import java.util.List;
|
||||
import java.util.regex.Matcher;
|
||||
@@ -51,7 +53,7 @@ public class MotherlessVideoRipper extends VideoRipper {
|
||||
}
|
||||
|
||||
@Override
|
||||
public void rip() throws IOException {
|
||||
public void rip() throws IOException, URISyntaxException {
|
||||
LOGGER.info(" Retrieving " + this.url);
|
||||
String html = Http.url(this.url).get().toString();
|
||||
if (html.contains("__fileurl = '")) {
|
||||
@@ -62,7 +64,7 @@ public class MotherlessVideoRipper extends VideoRipper {
|
||||
throw new IOException("Could not find video URL at " + url);
|
||||
}
|
||||
String vidUrl = vidUrls.get(0);
|
||||
addURLToDownload(new URL(vidUrl), HOST + "_" + getGID(this.url));
|
||||
addURLToDownload(new URI(vidUrl).toURL(), HOST + "_" + getGID(this.url));
|
||||
waitForThreads();
|
||||
}
|
||||
}
|
Reference in New Issue
Block a user