IP | Country | PORT | ADDED |
---|---|---|---|
192.252.211.193 | us | 4145 | 58 minutes ago |
122.151.54.147 | au | 80 | 58 minutes ago |
62.182.204.81 | ru | 88 | 58 minutes ago |
185.93.89.146 | ir | 14567 | 58 minutes ago |
50.63.12.101 | us | 54885 | 58 minutes ago |
139.59.1.14 | in | 8080 | 58 minutes ago |
98.170.57.231 | us | 4145 | 58 minutes ago |
67.201.58.190 | us | 4145 | 58 minutes ago |
128.140.113.110 | de | 8080 | 58 minutes ago |
68.1.210.189 | us | 4145 | 58 minutes ago |
103.118.46.176 | kh | 8080 | 58 minutes ago |
72.211.46.124 | us | 4145 | 58 minutes ago |
80.228.235.6 | de | 80 | 58 minutes ago |
203.95.198.35 | kh | 8080 | 58 minutes ago |
79.110.202.184 | pl | 8081 | 58 minutes ago |
175.34.36.22 | au | 8888 | 58 minutes ago |
50.171.122.27 | us | 80 | 58 minutes ago |
72.195.34.59 | us | 4145 | 58 minutes ago |
192.252.215.2 | us | 4145 | 58 minutes ago |
87.120.103.205 | it | 8080 | 58 minutes ago |
Our proxies work perfectly with all popular tools for web scraping, automation, and anti-detect browsers. Load your proxies into your favorite software or use them in your scripts in just seconds:
Connection formats you know and trust: IP:port or IP:port@login:password.
Any programming language: Python, JavaScript, PHP, Java, and more.
Top automation and scraping tools: Scrapy, Selenium, Puppeteer, ZennoPoster, BAS, and many others.
Anti-detect browsers: Multilogin, GoLogin, Dolphin, AdsPower, and other popular solutions.
Looking for full automation and proxy management?
Take advantage of our user-friendly PapaProxy API: purchase proxies, renew plans, update IP lists, manage IP bindings, and export ready-to-use lists — all in just a few clicks, no hassle.
PapaProxy offers the simplicity and flexibility that both beginners and experienced developers will appreciate.
And 500+ more tools and coding languages to explore
All modern Smart TVs allow you to use proxies to connect to the Internet or local network (both on Android and Tizen OS). You have to go to the device settings, open "Network" tab (can be named as "Ethernet"), and then in "Advanced settings" to activate the proxy, if necessary - specify its settings.
To scrape JSON data using RxJava in a Java application, you can use the RxJava library along with an HTTP client library to make requests. Below is an example using RxJava2 and OkHttp to scrape JSON data from a URL asynchronously.
Add Dependencies
Add the following dependencies to your project:
io.reactivex.rxjava2
rxjava
2.x.y
com.squareup.okhttp3
okhttp
4.x.y
Write the Code:
import io.reactivex.Observable;
import io.reactivex.schedulers.Schedulers;
import okhttp3.OkHttpClient;
import okhttp3.Request;
import okhttp3.Response;
public class JsonScrapingExample {
public static void main(String[] args) {
String url = "https://api.example.com/data"; // Replace with your JSON API URL
// Create an Observable that emits a single item (the URL)
Observable.just(url)
.observeOn(Schedulers.io()) // Specify the IO thread for network operations
.map(JsonScrapingExample::fetchJson)
.subscribe(
jsonData -> {
// Process the JSON data (replace this with your scraping logic)
System.out.println("Scraped JSON data: " + jsonData);
},
Throwable::printStackTrace
);
}
// Function to fetch JSON data using OkHttp
private static String fetchJson(String url) throws Exception {
OkHttpClient client = new OkHttpClient();
Request request = new Request.Builder()
.url(url)
.build();
try (Response response = client.newCall(request).execute()) {
if (!response.isSuccessful()) {
throw new Exception("Failed to fetch JSON. HTTP Code: " + response.code());
}
// Return the JSON data as a string
return response.body().string();
}
}
}
url
variable with the actual URL of the JSON API you want to scrape.fetchJson
function uses OkHttp
to make an HTTP request and fetch the JSON data.Run the Code:
This example uses RxJava's Observable
to create an asynchronous stream of events. The observeOn(Schedulers.io())
part specifies that the network operation (fetchJson
) should run on the IO thread to avoid blocking the main thread.
Make sure to handle exceptions appropriately and adjust the code based on the structure of the JSON API you are working with.
JSON scraping typically involves extracting data from a JSON response obtained from an API. When you mention doing JSON scraping sequentially, it could mean processing items in the JSON response one after another. Below is a simple example in Python that demonstrates sequential processing of JSON data:
import requests
def fetch_data(url):
response = requests.get(url)
return response.json()
def process_item(item):
# Replace this with your actual processing logic
print("Processing item:", item)
def scrape_sequentially(api_url):
data = fetch_data(api_url)
# Assuming the JSON response is a list of items
if isinstance(data, list):
for item in data:
process_item(item)
else:
print("Invalid JSON format. Expected a list of items.")
# Replace 'https://example.com/api/data' with the actual API URL
api_url = 'https://example.com/api/data'
scrape_sequentially(api_url)
In this example:
fetch_data
function sends a GET request to the specified API URL and returns the JSON response.process_item
function represents the logic you want to apply to each item in the JSON response.scrape_sequentially
function fetches the JSON data, checks if it's a list, and then iterates through each item, applying the processing logic sequentially.Make sure to replace the placeholder URL 'https://example.com/api/data'
with the actual URL of the API you want to scrape.
A DNS server is a remote computer that receives a domain request from a user device. And it converts it into an IP address. Sometimes it is through the DNS-server that ISPs block sites. And DNS-proxy, respectively, allows you to bypass these restrictions completely.
Parsing is the collection of all information. Accordingly, parsing a site is copying all of its source code as presented. You can use it to edit the site further or to analyze it for security purposes.
What else…