370 likes | 485 Vues
Dive into the intricate world of HTTP header abuse with Zachary Wolff, a research engineer at LogRhythm Labs. This presentation uncovers the basics of HTTP headers, revealing vulnerabilities discovered through extensive testing across 1.6 million URLs. Learn about attack vectors, including malicious User-Agent strings and header manipulation, and see how to utilize tools like PyLobster for vulnerability assessment. Discover the hidden risks lurking in web server configurations and the necessity for enhanced header handling to defend against attacks. Join us in exploring this urgent security concern!
E N D
Adventures in Large Scale HTTP Header Abuse Zachary Wolff
About Me • SIEM Deployments, Research Engineer for LogRhythm(Labs) | 2+ years • Threat Research Analyst for Webroot Software | 2+ years
Lets Talk About HTTP Headers Browser Web Server
HTTP Headers Basics • Standard Fields for Request & Response defined in RFC- 2616 (RTFRFC) • GET / HTTP/1.0 is a legitimate request but may not return expected results • RFC sets no limits on size of header, field name, value or number of headers • Most webservers now impose their own limits : • IIS • v4 – 2MB • v5 – 128K-16KB* • V6 – 16 KB* • V7 – 16 KB* Apache v2.3 – 8KB* *Per Header Field
The Premise I want to break some logging applications!
Round 1: Begin Original Premise: GET Request returns 302, 200, (valid response) then send a second GET with a malicious User Agent string* to see if we can get 500 response • Crawler to collect URL’s • Python script to send attack/test UA String • Store results in SQLite3 DB • Profit!
Round 1: Results Data set: 400K URL’s • Lots of 500’s! • Lots of smaller, low traffic site, some bigger high traffic sites • Various different errors….
Round 1: Results Custom 500’s…
le yawn.. Non verbose IIS Errors…
Not as boring, generic apache 500 and the x.x.gov sites….?
Round 1: Conclusion What did we find? • Some SQL injectable 500’s • Possible application level DOS • Lots of websites are not expecting malicious Header requests… • Further exploration is warranted
the Question How extensive is the problem of improper HTTP header handling?
Round 2: Begin • Need a more effective way to identify vulnerabilities • Lets attack/audit more than just User-Agent Header • Expand beyond backtick, additional attack strings • Larger sample set, 1.6 Million URL’s • Must be able to store and access very large set of result data efficiently (Shodan is amazing)
Round 2: Vulnerability Identification 500’s are ok, but much to broad What is a good indication of a possible SQLi vulnerability? Run regular Expression against HTML.data response to match on, “you have an error in your sql syntax”
Round 2: Vulnerability Identification Improved error detection, basic SQLi & beyond *Thanks to @j0emccray for contributing to regEx list
Beyond RegEx based Error Detection Byte Anomaly Detection Added (--bad) Compare content-length of response data from original/clean GET to data from malicious GET. *Set margin of alert to 150 bytes above and 150 bytes below clean request, log results (including HTML response data) to file
Round 2: Additional Header Fields • Let’s test: Host, From*, X-Forwarded-For, Referer, User-Agent, Non existent Header • Smart Mode (-s) : Will look at all Header fields returned by the server and test those (minus whitelist of rarely dynamic Headers) • Cookies!
Cookie Support Cookie Support added. Server Sends us this: PyLobster Responds with this: And the server says?
Round 3: Updates Updated Testing Values: “,;,%00, %00’
Round 2: Design “I Improved the crawler to harvest 500K+ URL’s a day. You should put my picture in your whitepaper” Output additions (beyond SQLite): • Elasticsearch Indexing support added (fast, efficient, JSON to webinterface) • Flat File logging Mark Vankempen, LogRhythm Labs
More Improvments Added Footprint mode (-g) • Generate random(ish) Hash or value • Save to key.txt file in same directory as pylobster.py • Activate Footprint mode: ./pylobster.py –g pyLobsterwill now send your unique string/hash as a request like so: Then, Wait for it… Days, Weeks, Months Google/Bing/duckduckgo your hash/string to discover unprotected Log directories ;)
pyLobsters maiden voyage Ready Begin! pyLobsteris currently a single threaded tool so I divided my 1.6 Million URL’s into 78 unique lists and spawned 78 instances #!/bin/bash nohup python pyLobster.py -f a --bad -s -l -g & nohup python pyLobster.py -f b --bad -s -l -g & nohup python pyLobster.py -f c --bad -s -l -g & nohup python pyLobster.py -f d --bad -s -l -g & And so on……
PyLobster’s Maiden Voyage Results • Sending a null byte in your HTTP Headers will catch a fair bit of IDS attention ;) • Grepresponse HTML on regEx error match directory to find patterns & specific components/module/application/CMS vulnerabilities. (highest value finding: one vulnerable component can lead to many others, shared DB’s as well) • Various vulnerable components identified
Findings: Breakdown by RegEx # Out of 1.6 Million Unique URL’s, 14,500 Error RegEx’sMatched! • * *0,1 & 2 are MySQL errors, 18 & 19 are PHP
Findings: Error Breakdown by Test String Of the 14,500 Error RegEx’s Matched
Findings: Error breakdown by HTTP Header *Cookies: 1584
Findings: error #0, breakdown by header field Error #0: “you have an error in you SQL syntax"
Findings: Footprint Mode Footprint Mode 12/13/2012 02/25/2013
Foot Print Mode 3/27/2013
Findings: (--bad) Byte Anomaly Detection Results • Work to be done…. • grep over dir for [wordpress|joomla|error|pass.*=|cms|.*?|] • Sort response files by size for like errors • Sort by status code response & file size
Defending Against HTTP Header Attacks • Raise developer awareness that any dynamically handled Header values need to be seen as user input and processed accordingly • Audit your sites HTTP Header Processing (pyLobster on github, SQLmap now supports custom Header testing too. bingo!) • Proactively review/monitor your web logs
This: Creates this Log trail:
The End Thank you! @nopslip https://github.com/nopslip/pyLobster Zacharywolff@gmail.com