Skip to content

curl-cffivsguzzle

MIT 34 2 1,751
594.9 thousand (month) Feb 23 2022 0.7.1(2024-07-13 09:07:25 ago)
23,447 13 34 MIT
Nov 14 2011 536.5 thousand (month) 7.10.0(2025-08-23 22:36:01 ago)

Curl-cffi is a Python library for implementing curl-impersonate which is a HTTP client that appears as one of popular web browsers like: - Google Chrome - Microsoft Edge - Safari - Firefox Unlike requests and httpx which are native Python libraries, curl-cffi uses cURL and inherits it's powerful features like extensive HTTP protocol support and detection patches for TLS and HTTP fingerprinting.

Using curl-cffi web scrapers can bypass TLS and HTTP fingerprinting.

Guzzle is a PHP HTTP client library that makes it easy to send HTTP requests and trivial to integrate with web services. It allows you to send HTTP/1.1 requests with various methods like GET, POST, PUT, DELETE, and others.

Guzzle also supports sending both synchronous and asynchronous requests, caching, and even has built-in support for OAuth 1.0a. Additionally, it can handle different HTTP errors and handle redirects automatically. It also has built-in support for serializing and deserializing data using formats like JSON and XML, as well as sending multipart file uploads.

Overall Guzzle is an easy to use and powerful library for working with HTTP in PHP.

Highlights


bypasshttp2tls-fingerprinthttp-fingerprintsyncasync

Example Use


curl-cffi can be accessed as low-level curl client as well as an easy high-level HTTP client: ```python from curl_cffi import requests response = requests.get('https://httpbin.org/json') print(response.json()) # or using sessions session = requests.Session() response = session.get('https://httpbin.org/json') # also supports async requests using asyncio import asyncio from curl_cffi.requests import AsyncSession urls = [ "http://httpbin.org/html", "http://httpbin.org/html", "http://httpbin.org/html", ] async with AsyncSession() as s: tasks = [] for url in urls: task = s.get(url) tasks.append(task) # scrape concurrently: responses = await asyncio.gather(*tasks) # also supports websocket connections from curl_cffi.requests import Session, WebSocket def on_message(ws: WebSocket, message): print(message) with Session() as s: ws = s.ws_connect( "wss://api.gemini.com/v1/marketdata/BTCUSD", on_message=on_message, ) ws.run_forever() ```
```php use GuzzleHttp\Client; // Create a client session: $client = new Client(); // can also set session details like headers $client = new Client([ 'headers' => [ 'User-Agent' => 'webscraping.fyi', ] ]); // GET request: $response = $client->get('http://httpbin.org/get'); // print all details var_dump($response); // or the important bits printf("status: %s\n", $response->getStatusCode()); printf("headers: %s\n", json_encode($response->getHeaders(), JSON_PRETTY_PRINT)); printf("body: %s", $response->getBody()->getContents()); // POST request $response = $client->post( 'https://httpbin.org/post', // for JSON use json argument: ['json' => ['query' => 'foobar', 'page' => 2]] // or formdata use form_params: // ['form_params' => ['query' => 'foobar', 'page' => 2]] ); // For ASYNC requests getAsync function can be used: $promise1 = $client->getAsync('https://httpbin.org/get'); $promise2 = $client->getAsync('https://httpbin.org/get?foo=bar'); // await it: $results = Promise\unwrap([$promise1, $promise2]); foreach ($results as $result) { echo $result->getBody(); } // or add promise callback Promise\each([$promise1, $promise2], function ($response, $index, $callable) { echo $response->getBody(); }); ```

Alternatives / Similar


Was this page helpful?