# curl_cffi **Repository Path**: guanzs/curl_cffi ## Basic Information - **Project Name**: curl_cffi - **Description**: 模拟浏览器指纹的 HTTP 客户端。这是一个用 Python 写的 HTTP 客户端库,可以模拟浏览器 TLS、JA3 和 HTTP/2 指纹。它开箱即用、速度快,并且支持 WebSocket 和异步。 - **Primary Language**: Unknown - **License**: MIT - **Default Branch**: dependabot/github_actions/dot-github/workflows/actions/download-artifact-4.1.7 - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 1 - **Created**: 2024-09-27 - **Last Updated**: 2025-06-29 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # curl_cffi ![PyPI - Downloads](https://img.shields.io/pypi/dm/curl-cffi) ![PyPI - Python Version](https://img.shields.io/pypi/pyversions/curl_cffi) [![PyPI version](https://badge.fury.io/py/curl-cffi.svg)](https://badge.fury.io/py/curl-cffi) [![Generic badge](https://img.shields.io/badge/Telegram%20Group-join-blue?logo=telegram)](https://t.me/+lL9n33eZp480MGM1) [Documentation](https://curl-cffi.readthedocs.io) Python binding for [curl-impersonate](https://github.com/lwthiker/curl-impersonate) via [cffi](https://cffi.readthedocs.io/en/latest/). Unlike other pure python http clients like `httpx` or `requests`, `curl_cffi` can impersonate browsers' TLS/JA3 and HTTP/2 fingerprints. If you are blocked by some website for no obvious reason, you can give `curl_cffi` a try. The fingerprints in 0.6 on Windows are all wrong, you should update to 0.7 if you are on Windows. Sorry for the inconvenience. Only Python 3.8 and above are supported. Python 3.7 has reached its end of life. ------ Scrapfly.io [Scrapfly](https://scrapfly.io/?utm_source=github&utm_medium=sponsoring&utm_campaign=curl_cffi) is an enterprise-grade solution providing Web Scraping API that aims to simplify the scraping process by managing everything: real browser rendering, rotating proxies, and fingerprints (TLS, HTTP, browser) to bypass all major anti-bots. Scrapfly also unlocks the observability by providing an analytical dashboard and measuring the success rate/block rate in detail. Scrapfly is a good solution if you are looking for a cloud-managed solution for `curl_cffi`. If you are managing TLS/HTTP fingerprint by yourself with `curl_cffi`, they also maintain a [curl to python converter](https://scrapfly.io/web-scraping-tools/curl-python/curl_cffi). ------ ## Features - Supports JA3/TLS and http2 fingerprints impersonation, inlucding recent browsers and custome fingerprints. - Much faster than requests/httpx, on par with aiohttp/pycurl, see [benchmarks](https://github.com/lexiforest/curl_cffi/tree/main/benchmark). - Mimics requests API, no need to learn another one. - Pre-compiled, so you don't have to compile on your machine. - Supports `asyncio` with proxy rotation on each request. - Supports http 2.0, which requests does not. - Supports websocket. ||requests|aiohttp|httpx|pycurl|curl_cffi| |---|---|---|---|---|---| |http2|❌|❌|✅|✅|✅| |sync|✅|❌|✅|✅|✅| |async|❌|✅|✅|❌|✅| |websocket|❌|✅|❌|❌|✅| |fingerprints|❌|❌|❌|❌|✅| |speed|🐇|🐇🐇|🐇|🐇🐇|🐇🐇| ## Install pip install curl_cffi --upgrade This should work on Linux, macOS and Windows out of the box. If it does not work on you platform, you may need to compile and install `curl-impersonate` first and set some environment variables like `LD_LIBRARY_PATH`. To install beta releases: pip install curl_cffi --upgrade --pre To install unstable version from GitHub: git clone https://github.com/lexiforest/curl_cffi/ cd curl_cffi make preprocess pip install . ## Usage `curl_cffi` comes with a low-level `curl` API and a high-level `requests`-like API. ### requests-like ```python from curl_cffi import requests # Notice the impersonate parameter r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome") print(r.json()) # output: {..., "ja3n_hash": "aa56c057ad164ec4fdcb7a5a283be9fc", ...} # the js3n fingerprint should be the same as target browser # To keep using the latest browser version as `curl_cffi` updates, # simply set impersonate="chrome" without specifying a version. # Other similar values are: "safari" and "safari_ios" r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome") # To pin a specific version, use version numbers together. r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome124") # To impersonate other than browsers, bring your own ja3/akamai strings # See examples directory for details. r = requests.get("https://tls.browserleaks.com/json", ja3=..., akamai=...) # http/socks proxies are supported proxies = {"https": "http://localhost:3128"} r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome", proxies=proxies) proxies = {"https": "socks://localhost:3128"} r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome", proxies=proxies) ``` ### Sessions ```python s = requests.Session() # httpbin is a http test website, this endpoint makes the server set cookies s.get("https://httpbin.org/cookies/set/foo/bar") print(s.cookies) # ]> # retrieve cookies again to verify r = s.get("https://httpbin.org/cookies") print(r.json()) # {'cookies': {'foo': 'bar'}} ``` `curl_cffi` supports the same browser versions as supported by my [fork](https://github.com/lexiforest/curl-impersonate) of [curl-impersonate](https://github.com/lwthiker/curl-impersonate): However, only Chrome-like browsers are supported. Firefox support is tracked in [#59](https://github.com/lexiforest/curl_cffi/issues/59). Browser versions will be added **only** when their fingerprints change. If you see a version, e.g. chrome122, were skipped, you can simply impersonate it with your own headers and the previous version. If you are trying to impersonate a target other than a browser, use `ja3=...` and `akamai=...` to specify your own customized fingerprints. See the [docs on impersonatation](https://curl-cffi.readthedocs.io/en/latest/impersonate.html) for details. - chrome99 - chrome100 - chrome101 - chrome104 - chrome107 - chrome110 - chrome116 [1] - chrome119 [1] - chrome120 [1] - chrome123 [3] - chrome124 [3] - chrome99_android - edge99 - edge101 - safari15_3 [2] - safari15_5 [2] - safari17_0 [1] - safari17_2_ios [1] Notes: 1. Added in version `0.6.0`. 2. Fixed in version `0.6.0`, previous http2 fingerprints were [not correct](https://github.com/lwthiker/curl-impersonate/issues/215). 3. Added in version `0.7.0`. ### asyncio ```python from curl_cffi.requests import AsyncSession async with AsyncSession() as s: r = await s.get("https://example.com") ``` More concurrency: ```python import asyncio from curl_cffi.requests import AsyncSession urls = [ "https://google.com/", "https://facebook.com/", "https://twitter.com/", ] async with AsyncSession() as s: tasks = [] for url in urls: task = s.get(url) tasks.append(task) results = await asyncio.gather(*tasks) ``` ### WebSockets ```python from curl_cffi.requests import Session, WebSocket def on_message(ws: WebSocket, message): print(message) with Session() as s: ws = s.ws_connect( "wss://api.gemini.com/v1/marketdata/BTCUSD", on_message=on_message, ) ws.run_forever() ``` For low-level APIs, Scrapy integration and other advanced topics, see the [docs](https://curl-cffi.readthedocs.io) for more details. ## Acknowledgement - Originally forked from [multippt/python_curl_cffi](https://github.com/multippt/python_curl_cffi), which is under the MIT license. - Headers/Cookies files are copied from [httpx](https://github.com/encode/httpx/blob/master/httpx/_models.py), which is under the BSD license. - Asyncio support is inspired by Tornado's curl http client. - The WebSocket API is inspired by [websocket_client](https://github.com/websocket-client/websocket-client). ## [Sponsor] Bypass Cloudflare with API Yes Captcha! Yescaptcha is a proxy service that bypasses Cloudflare and uses the API interface to obtain verified cookies (e.g. `cf_clearance`). Click [here](https://yescaptcha.com/i/stfnIO) to register: https://yescaptcha.com/i/stfnIO ## [Sponsor] ScrapeNinja Scrape Ninja [ScrapeNinja](https://scrapeninja.net?utm_source=github&utm_medium=banner&utm_campaign=cffi) is a web scraping API with two engines: fast, with high performance and TLS fingerprint; and slower with a real browser under the hood. ScrapeNinja handles headless browsers, proxies, timeouts, retries, and helps with data extraction, so you can just get the data in JSON. Rotating proxies are available out of the box on all subscription plans. ## Sponsor Buy Me A Coffee