reqvsralger
The Go library "req" is a simple and easy-to-use library for making HTTP requests in Go. It is designed to make working with HTTP requests as simple as possible, by providing a clean and consistent API for handling various types of requests, including GET, POST, PUT, and DELETE.
One of the key features of req is its support for handling JSON data. The library automatically serializes and deserializes JSON data, making it easy to work with JSON data in your Go applications. Additionally, it supports multipart file uploads and automatic decompression of gzip and deflate encoded responses.
req also includes a number of convenience functions for working with common HTTP request types, such as sending GET and POST requests, handling redirects, and setting headers and query parameters. The library can also be easily extended with custom middleware and request handlers.
Overall, req is a powerful and flexible library that makes it easy to work with HTTP requests in Go. It is well-documented and actively maintained, making it a great choice for any Go project that needs to work with HTTP requests.
ralger is a small web scraping framework for R based on rvest and xml2.
It's goal to simplify basic web scraping and it provides a convenient and easy to use API.
It offers functions for retrieving pages, parsing HTML using CSS selectors, automatic table parsing and auto link, title, image and paragraph extraction.
Example Use
req.DevMode() // Use Client.DevMode to enable debugging details
// There are 2 ways to use req (like many other http clients)
// First way is to create a persistent session client:
client := req.C(). // defaults like timeout and headers can be set for the whole session
SetUserAgent("my-custom-client").
SetTimeout(5 * time.Second)
// defaults can be overriden and extended in each request
resp, err := client.R(). // Use R() to create a request and set with chainable request settings.
SetHeader("Accept", "application/vnd.github.v3+json").
SetPathParam("username", "imroc").
SetQueryParam("page", "1").
SetResult(&result). // Unmarshal response into struct automatically if status code >= 200 and <= 299.
SetError(&errMsg). // Unmarshal response into struct automatically if status code >= 400.
EnableDump(). // Enable dump at request level to help troubleshoot, log content only when an unexpected exception occurs.
Get("https://api.github.com/users/{username}/repos")
// Alternatively, it can be used as is without establishing a client
resp := client.Get("https://api.github.com/users/{username}/repos"). // Create a GET request with specified URL.
SetHeader("Accept", "application/vnd.github.v3+json").
SetPathParam("username", "imroc").
SetQueryParam("page", "1").
SetResult(&result).
SetError(&errMsg).
EnableDump().
Do() // Send request with Do.
library("ralger")
url <- "http://www.shanghairanking.com/rankings/arwu/2021"
# retrieve HTML and select elements using CSS selectors:
best_uni <- scrap(link = url, node = "a span", clean = TRUE)
head(best_uni, 5)
#> [1] "Harvard University"
#> [2] "Stanford University"
#> [3] "University of Cambridge"
#> [4] "Massachusetts Institute of Technology (MIT)"
#> [5] "University of California, Berkeley"
# ralger can also parse HTML attributes
attributes <- attribute_scrap(
link = "https://ropensci.org/",
node = "a", # the a tag
attr = "class" # getting the class attribute
)
head(attributes, 10) # NA values are a tags without a class attribute
#> [1] "navbar-brand logo" "nav-link" NA
#> [4] NA NA "nav-link"
#> [7] NA "nav-link" NA
#> [10] NA
#
# ralger can automatically scrape tables:
data <- table_scrap(link ="https://www.boxofficemojo.com/chart/top_lifetime_gross/?area=XWW")
head(data)
#> # A tibble: 6 × 4
#> Rank Title `Lifetime Gross` Year
#> <int> <chr> <chr> <int>
#> 1 1 Avatar $2,847,397,339 2009
#> 2 2 Avengers: Endgame $2,797,501,328 2019
#> 3 3 Titanic $2,201,647,264 1997
#> 4 4 Star Wars: Episode VII - The Force Awakens $2,069,521,700 2015
#> 5 5 Avengers: Infinity War $2,048,359,754 2018
#> 6 6 Spider-Man: No Way Home $1,901,216,740 2021