Metadata-Version: 2.2 Name: curl_cffi Version: 0.10.0 Summary: libcurl ffi bindings for Python, with impersonation support. Author-email: Lyonnet License: MIT License Project-URL: repository, https://github.com/lexiforest/curl_cffi Classifier: Development Status :: 4 - Beta Classifier: Intended Audience :: Developers Classifier: License :: OSI Approved :: MIT License Classifier: Programming Language :: Python :: 3 Classifier: Programming Language :: Python :: 3.9 Classifier: Programming Language :: Python :: 3.10 Classifier: Programming Language :: Python :: 3.11 Classifier: Programming Language :: Python :: 3.12 Classifier: Programming Language :: Python :: 3.13 Requires-Python: >=3.9 Description-Content-Type: text/markdown License-File: LICENSE Requires-Dist: cffi>=1.12.0 Requires-Dist: certifi>=2024.2.2 Provides-Extra: dev Requires-Dist: charset_normalizer<4.0,>=3.3.2; extra == "dev" Requires-Dist: coverage<7.0,>=6.4.1; extra == "dev" Requires-Dist: cryptography<43.0,>=42.0.5; extra == "dev" Requires-Dist: httpx==0.23.1; extra == "dev" Requires-Dist: mypy<2.0,>=1.9.0; extra == "dev" Requires-Dist: pytest<9.0,>=8.1.1; extra == "dev" Requires-Dist: pytest-asyncio<1.0,>=0.23.6; extra == "dev" Requires-Dist: pytest-trio<1.0,>=0.8.0; extra == "dev" Requires-Dist: ruff<1.0,>=0.3.5; extra == "dev" Requires-Dist: trio<1.0,>=0.25.0; extra == "dev" Requires-Dist: trustme<2.0,>=1.1.0; extra == "dev" Requires-Dist: uvicorn<1.0,>=0.29.0; extra == "dev" Requires-Dist: websockets<13.0,>=12.0; extra == "dev" Requires-Dist: typing_extensions; extra == "dev" Provides-Extra: build Requires-Dist: cibuildwheel; extra == "build" Requires-Dist: wheel; extra == "build" Provides-Extra: test Requires-Dist: charset_normalizer<4.0,>=3.3.2; extra == "test" Requires-Dist: cryptography<43.0,>=42.0.5; extra == "test" Requires-Dist: fastapi<1.0,==0.110.0; extra == "test" Requires-Dist: httpx==0.23.1; extra == "test" Requires-Dist: proxy.py<3.0,>=2.4.3; extra == "test" Requires-Dist: pytest<9.0,>=8.1.1; extra == "test" Requires-Dist: pytest-asyncio<1.0,>=0.23.6; extra == "test" Requires-Dist: pytest-trio<1.0,>=0.8.0; extra == "test" Requires-Dist: python-multipart<1.0,>=0.0.9; extra == "test" Requires-Dist: trio<1.0,>=0.25.0; extra == "test" Requires-Dist: trustme<2.0,>=1.1.0; extra == "test" Requires-Dist: uvicorn<1.0,>=0.29.0; extra == "test" Requires-Dist: websockets<13.0,>=12.0; extra == "test" Requires-Dist: typing_extensions; extra == "test" # curl_cffi ![PyPI - Downloads](https://img.shields.io/pypi/dm/curl-cffi) ![PyPI - Python Version](https://img.shields.io/pypi/pyversions/curl_cffi) [![PyPI version](https://badge.fury.io/py/curl-cffi.svg)](https://badge.fury.io/py/curl-cffi) [![Generic badge](https://img.shields.io/badge/Telegram%20Group-join-blue?logo=telegram)](https://t.me/+lL9n33eZp480MGM1) [![Generic badge](https://img.shields.io/badge/Discord-join-purple?logo=blue)](https://discord.gg/kJqMHHgdn2) [Documentation](https://curl-cffi.readthedocs.io) Python binding for [curl-impersonate fork](https://github.com/lexiforest/curl-impersonate) via [cffi](https://cffi.readthedocs.io/en/latest/). Unlike other pure python http clients like `httpx` or `requests`, `curl_cffi` can impersonate browsers' TLS/JA3 and HTTP/2 fingerprints. If you are blocked by some website for no obvious reason, you can give `curl_cffi` a try. Python 3.9 is the minimum supported version since v0.10. ## Sponsors

Maintenance of this project is made possible by all the contributors and sponsors. If you'd like to sponsor this project and have your avatar or company logo appear below click here. 💖

------ ProxyCurl Scrape public LinkedIn profile data at scale with [Proxycurl APIs](https://nubela.co/proxycurl/?utm_campaign=influencer_marketing&utm_source=github&utm_medium=social&utm_term=-&utm_content=lexiforest-curl_cffi). Built for developers, by developers. - GDPR, CCPA, SOC2 compliant - High rate limit (300 requests/min), Fast (APIs respond in ~2s), High accuracy - Fresh data - 88% of data is scraped real-time, other 12% is <29 days - Tons of data points returned per profile ------ SerpAPI Scrape Google and other search engines from [SerpApi](https://serpapi.com/)'s fast, easy, and complete API. 0.66s average response time (≤ 0.5s for Ludicrous Speed Max accounts), 99.95% SLAs, pay for successful responses only. ------ ### Bypass Cloudflare with API Yes Captcha! Yescaptcha is a proxy service that bypasses Cloudflare and uses the API interface to obtain verified cookies (e.g. `cf_clearance`). Click [here](https://yescaptcha.com/i/stfnIO) to register: https://yescaptcha.com/i/stfnIO ------ ## Features - Supports JA3/TLS and http2 fingerprints impersonation, including recent browsers and custome fingerprints. - Much faster than requests/httpx, on par with aiohttp/pycurl, see [benchmarks](https://github.com/lexiforest/curl_cffi/tree/main/benchmark). - Mimics requests API, no need to learn another one. - Pre-compiled, so you don't have to compile on your machine. - Supports `asyncio` with proxy rotation on each request. - Supports http 2.0, which requests does not. - Supports websocket. ||requests|aiohttp|httpx|pycurl|curl_cffi| |---|---|---|---|---|---| |http2|❌|❌|✅|✅|✅| |sync|✅|❌|✅|✅|✅| |async|❌|✅|✅|❌|✅| |websocket|❌|✅|❌|❌|✅| |fingerprints|❌|❌|❌|❌|✅| |speed|🐇|🐇🐇|🐇|🐇🐇|🐇🐇| ## Install pip install curl_cffi --upgrade This should work on Linux, macOS and Windows out of the box. If it does not work on you platform, you may need to compile and install `curl-impersonate` first and set some environment variables like `LD_LIBRARY_PATH`. To install beta releases: pip install curl_cffi --upgrade --pre To install unstable version from GitHub: git clone https://github.com/lexiforest/curl_cffi/ cd curl_cffi make preprocess pip install . ## Usage `curl_cffi` comes with a low-level `curl` API and a high-level `requests`-like API. ### requests-like v0.9: ```py from curl_cffi import requests r = requests.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome") ``` v0.10: ```python import curl_cffi # Notice the impersonate parameter r = curl_cffi.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome") print(r.json()) # output: {..., "ja3n_hash": "aa56c057ad164ec4fdcb7a5a283be9fc", ...} # the js3n fingerprint should be the same as target browser # To keep using the latest browser version as `curl_cffi` updates, # simply set impersonate="chrome" without specifying a version. # Other similar values are: "safari" and "safari_ios" r = curl_cffi.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome") # To pin a specific version, use version numbers together. r = curl_cffi.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome124") # To impersonate other than browsers, bring your own ja3/akamai strings # See examples directory for details. r = curl_cffi.get("https://tls.browserleaks.com/json", ja3=..., akamai=...) # http/socks proxies are supported proxies = {"https": "http://localhost:3128"} r = curl_cffi.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome", proxies=proxies) proxies = {"https": "socks://localhost:3128"} r = curl_cffi.get("https://tools.scrapfly.io/api/fp/ja3", impersonate="chrome", proxies=proxies) ``` ### Sessions v0.9: ```py from curl_cffi import requests s = requests.Session() ``` v0.10: ```python s = curl_cffi.Session() # httpbin is a http test website, this endpoint makes the server set cookies s.get("https://httpbin.org/cookies/set/foo/bar") print(s.cookies) # ]> # retrieve cookies again to verify r = s.get("https://httpbin.org/cookies") print(r.json()) # {'cookies': {'foo': 'bar'}} ``` `curl_cffi` supports the same browser versions as supported by my [fork](https://github.com/lexiforest/curl-impersonate) of [curl-impersonate](https://github.com/lwthiker/curl-impersonate): Browser versions will be added **only** when their fingerprints change. If you see a version, e.g. chrome122, were skipped, you can simply impersonate it with your own headers and the previous version. If you are trying to impersonate a target other than a browser, use `ja3=...` and `akamai=...` to specify your own customized fingerprints. See the [docs on impersonation](https://curl-cffi.readthedocs.io/en/latest/impersonate.html) for details. - chrome99 - chrome100 - chrome101 - chrome104 - chrome107 - chrome110 - chrome116 [1] - chrome119 [1] - chrome120 [1] - chrome123 [3] - chrome124 [3] - chrome131 [4] - chrome133a [5][6] - chrome99_android - chrome131_android [4] - edge99 - edge101 - safari15_3 [2] - safari15_5 [2] - safari17_0 [1] - safari17_2_ios [1] - safari18_0 [4] - safari18_0_ios [4] - firefox133 [5] - firefox135 [7] Notes: 1. Added in version `0.6.0`. 2. Fixed in version `0.6.0`, previous http2 fingerprints were [not correct](https://github.com/lwthiker/curl-impersonate/issues/215). 3. Added in version `0.7.0`. 4. Added in version `0.8.0`. 5. Added in version `0.9.0`. 6. The version postfix `-a`(e.g. `chrome133a`) means that this is an alternative version, i.e. the fingerprint has not been officially updated by browser, but has been observed because of A/B testing. 5. Added in version `0.10.0`. ### Asyncio ```python from curl_cffi import AsyncSession async with AsyncSession() as s: r = await s.get("https://example.com") ``` More concurrency: ```python import asyncio from curl_cffi import AsyncSession urls = [ "https://google.com/", "https://facebook.com/", "https://twitter.com/", ] async with AsyncSession() as s: tasks = [] for url in urls: task = s.get(url) tasks.append(task) results = await asyncio.gather(*tasks) ``` ### WebSockets ```python from curl_cffi import WebSocket def on_message(ws: WebSocket, message: str | bytes): print(message) ws = WebSocket(on_message=on_message) ws.run_forever("wss://api.gemini.com/v1/marketdata/BTCUSD") ``` For low-level APIs, Scrapy integration and other advanced topics, see the [docs](https://curl-cffi.readthedocs.io) for more details. ### Asyncio WebSockets ```python import asyncio from curl_cffi import AsyncSession async with AsyncSession() as s: ws = await s.ws_connect("wss://echo.websocket.org") await asyncio.gather(*[ws.send_str("Hello, World!") for _ in range(10)]) async for message in ws: print(message) ``` ## Acknowledgement - Originally forked from [multippt/python_curl_cffi](https://github.com/multippt/python_curl_cffi), which is under the MIT license. - Headers/Cookies files are copied from [httpx](https://github.com/encode/httpx/blob/master/httpx/_models.py), which is under the BSD license. - Asyncio support is inspired by Tornado's curl http client. - The synchronous WebSocket API is inspired by [websocket_client](https://github.com/websocket-client/websocket-client). - The asynchronous WebSocket API is inspired by [aiohttp](https://github.com/aio-libs/aiohttp).