Files
litterbox/GrumpyCats/grumpycat.py
T
2025-05-16 07:26:48 -07:00

776 lines
30 KiB
Python

import argparse
import hashlib
import json
import logging
import os
import requests
import sys
from datetime import datetime
from pathlib import Path
from typing import Optional, Dict, List, Union, BinaryIO, Any, Tuple
from requests.adapters import HTTPAdapter, Retry
from urllib.parse import urljoin
class LitterBoxError(Exception):
"""Base exception for LitterBox client errors"""
pass
class LitterBoxAPIError(LitterBoxError):
"""Exception for API-related errors"""
def __init__(self, message: str, status_code: Optional[int] = None, response: Optional[Dict] = None):
super().__init__(message)
self.status_code = status_code
self.response = response
class LitterBoxClient:
"""A Python client for interacting with the LitterBox malware analysis sandbox API."""
def __init__(self,
base_url: str = "http://127.0.0.1:1337",
timeout: int = 120,
max_retries: int = 3,
verify_ssl: bool = True,
logger: Optional[logging.Logger] = None,
proxy_config: Optional[Dict] = None,
headers: Optional[Dict] = None):
"""Initialize the LitterBox client.
Args:
base_url: The base URL of the LitterBox server
timeout: Request timeout in seconds
max_retries: Maximum number of retries for failed requests
verify_ssl: Whether to verify SSL certificates
logger: Custom logger instance
proxy_config: Proxy configuration dictionary (e.g., {"http": "http://proxy:8080"})
headers: Additional headers to include in requests
"""
self.base_url = base_url.rstrip('/')
self.timeout = timeout
self.verify_ssl = verify_ssl
self.logger = logger or logging.getLogger(__name__)
self.proxy_config = proxy_config
self.headers = headers or {}
# Configure session with retries
self.session = requests.Session()
retry_strategy = Retry(
total=max_retries,
backoff_factor=0.5,
status_forcelist=[500, 502, 503, 504],
allowed_methods=["HEAD", "GET", "POST", "PUT", "DELETE", "OPTIONS", "TRACE"]
)
adapter = HTTPAdapter(max_retries=retry_strategy)
self.session.mount("http://", adapter)
self.session.mount("https://", adapter)
if proxy_config:
self.session.proxies.update(proxy_config)
if not verify_ssl:
self.session.verify = False
self.session.headers.update(self.headers)
def _make_request(self,
method: str,
endpoint: str,
**kwargs) -> requests.Response:
"""Make HTTP request with error handling.
Args:
method: HTTP method
endpoint: API endpoint
**kwargs: Additional arguments for requests
Returns:
Response object
Raises:
LitterBoxAPIError: If the API returns an error
LitterBoxError: For other client errors
"""
url = urljoin(self.base_url, endpoint)
try:
kwargs.setdefault('timeout', self.timeout)
response = self.session.request(method, url, **kwargs)
response.raise_for_status()
return response
except requests.exceptions.HTTPError as e:
try:
error_data = response.json()
except:
error_data = {'error': response.text}
raise LitterBoxAPIError(
f"API error: {error_data.get('error', 'Unknown error')}",
status_code=response.status_code,
response=error_data
)
except requests.exceptions.RequestException as e:
raise LitterBoxError(f"Request failed: {str(e)}")
def upload_file(self,
file_path: Union[str, Path, BinaryIO],
file_name: Optional[str] = None) -> Dict:
"""Upload a file for analysis.
Args:
file_path: Path to the file or file-like object
file_name: Optional name to use for the file
Returns:
Dict containing upload status and file information
"""
if isinstance(file_path, (str, Path)):
path = Path(file_path)
if not path.exists():
raise LitterBoxError(f"File not found: {path}")
files = {'file': (file_name or path.name, open(path, 'rb'), 'application/octet-stream')}
else:
if not file_name:
raise ValueError("file_name is required when uploading file-like objects")
files = {'file': (file_name, file_path, 'application/octet-stream')}
response = self._make_request('POST', '/upload', files=files)
return response.json()
def analyze_file(self,
target: str,
analysis_type: str,
cmd_args: Optional[List[str]] = None,
wait_for_completion: bool = True) -> Dict:
"""Run analysis on a file or process.
Args:
target: File hash or PID
analysis_type: 'static' or 'dynamic'
cmd_args: Optional command line arguments for dynamic analysis
wait_for_completion: Whether to wait for analysis completion
Returns:
Dict containing analysis results
"""
# Initial validations
if analysis_type not in ['static', 'dynamic']:
raise ValueError("analysis_type must be either 'static' or 'dynamic'")
# For dynamic analysis with PID, validate first
if analysis_type == 'dynamic' and target.isdigit():
self.validate_process(target)
# For static analysis, cannot use PID
if analysis_type == 'static' and target.isdigit():
raise ValueError("Cannot perform static analysis on PID")
# Prepare request
params = {'wait': '1' if wait_for_completion else '0'}
# Validate and prepare command line arguments
data = {}
if cmd_args is not None:
if not isinstance(cmd_args, list):
raise ValueError("Arguments must be provided as a list")
if not all(isinstance(arg, str) for arg in cmd_args):
raise ValueError("All arguments must be strings")
if any(any(char in arg for char in [';', '&', '|']) for arg in cmd_args):
raise ValueError("Invalid argument characters detected")
data['args'] = cmd_args
# Make the analysis request
response = self._make_request(
'POST',
f'/analyze/{analysis_type}/{target}',
params=params,
json=data
)
# Handle early termination case (202 status)
if response.status_code == 202:
result = response.json()
return {
'status': 'early_termination',
'error': result.get('error', {}).get('message', 'Process terminated early'),
'details': {
'termination_time': result.get('error', {}).get('termination_time'),
'init_time': result.get('error', {}).get('init_time'),
'message': result.get('error', {}).get('details')
}
}
# Handle normal response
result = response.json()
# If error occurred during analysis
if result.get('status') == 'error':
return {
'status': 'error',
'error': result.get('error', {}).get('message', 'Analysis failed'),
'details': result.get('error', {}).get('details')
}
# Success case
return {
'status': 'success',
'results': result.get('results', {})
}
"""Run analysis on a file or process.
Args:
target: File hash or PID
analysis_type: 'static' or 'dynamic'
cmd_args: Optional command line arguments for dynamic analysis
wait_for_completion: Whether to wait for analysis completion
Returns:
Dict containing:
- status: 'success', 'error', or 'early_termination'
- results: Analysis results (on success)
- error: Error message (on error)
- details: Additional error details (on error/early termination)
Raises:
ValueError: If analysis type is invalid
LitterBoxAPIError: If API returns an error
"""
if analysis_type not in ['static', 'dynamic']:
raise ValueError("analysis_type must be either 'static' or 'dynamic'")
# Validate command line arguments if provided
data = {}
if cmd_args is not None:
if not isinstance(cmd_args, list):
raise ValueError("Arguments must be provided as a list")
if not all(isinstance(arg, str) for arg in cmd_args):
raise ValueError("All arguments must be strings")
if any(any(char in arg for char in [';', '&', '|']) for arg in cmd_args):
raise ValueError("Invalid argument characters detected")
data['args'] = cmd_args
params = {'wait': '1' if wait_for_completion else '0'}
response = self._make_request(
'POST',
f'/analyze/{analysis_type}/{target}',
params=params,
json=data
)
if response.status_code == 202: # Early termination
return response.json()
result = response.json()
if result.get('status') == 'error':
error_info = result.get('error', {})
if isinstance(error_info, str):
error_info = {'message': error_info}
return {
'status': 'error',
'error': error_info.get('message', 'Analysis failed'),
'details': error_info.get('details')
}
return result
"""Run analysis on a file or process.
Args:
target: File hash or PID
analysis_type: 'static' or 'dynamic'
cmd_args: Optional command line arguments for dynamic analysis
wait_for_completion: Whether to wait for analysis completion
Returns:
Dict containing analysis results
"""
if analysis_type not in ['static', 'dynamic']:
raise ValueError("analysis_type must be either 'static' or 'dynamic'")
# For non-PID targets, verify the file exists first
if not str(target).isdigit() and verify_file:
try:
# Check if file exists by attempting to get its info
self._make_request('GET', f'/api/results/{target}/info')
except LitterBoxAPIError as e:
if e.status_code == 404:
raise LitterBoxError(f"File {target} not found or not yet available")
# Validate PID for dynamic analysis
if analysis_type == 'dynamic' and str(target).isdigit():
self.validate_process(target)
elif analysis_type == 'static' and str(target).isdigit():
raise ValueError("Cannot perform static analysis on PID")
params = {'wait': '1' if wait_for_completion else '0'}
data = {}
if cmd_args is not None:
# Validate command line arguments
if not isinstance(cmd_args, list):
raise ValueError("cmd_args must be provided as a list")
if not all(isinstance(arg, str) for arg in cmd_args):
raise ValueError("All arguments must be strings")
if any(any(char in arg for char in [';', '&', '|']) for arg in cmd_args):
raise ValueError("Invalid argument characters detected")
data['args'] = cmd_args
response = self._make_request(
'POST',
f'/analyze/{analysis_type}/{target}',
params=params,
json=data
)
result = response.json()
# Handle early termination case for dynamic analysis
if result.get('status') == 'early_termination':
return {
'status': 'early_termination',
'error': result.get('error', {}).get('message', 'Process terminated early'),
'details': {
'termination_time': result.get('error', {}).get('termination_time'),
'init_time': result.get('error', {}).get('init_time'),
'message': result.get('error', {}).get('details')
}
}
return result
def get_results(self,
target: str,
analysis_type: str) -> Dict:
"""Get results for a specific analysis.
Args:
target: File hash or PID
analysis_type: 'static', 'dynamic', or 'info'
Returns:
Dict containing analysis results
"""
if analysis_type not in ['static', 'dynamic', 'info']:
raise ValueError("analysis_type must be one of: 'static', 'dynamic', 'info'")
response = self._make_request(
'GET',
f'/api/results/{target}/{analysis_type}'
)
return response.json()
def get_files_summary(self) -> Dict:
"""Get summary of all analyzed files and processes.
Returns:
Dict containing analysis summaries
"""
response = self._make_request('GET', '/files')
return response.json()
def run_blender_scan(self) -> Dict:
"""Run a system-wide Blender scan.
Returns:
Dict containing scan results
"""
data = {"operation": "scan"}
response = self._make_request('POST', '/blender', json=data)
return response.json()
def compare_with_blender(self,
file_hash: str) -> Dict:
"""Compare a file's analysis results with current system state.
Args:
file_hash: Hash of the file to compare
Returns:
Dict containing comparison results
"""
params = {'hash': file_hash}
response = self._make_request('GET', '/blender', params=params)
return response.json()
def cleanup(self,
include_uploads: bool = True,
include_results: bool = True,
include_analysis: bool = True) -> Dict:
"""Clean up analysis artifacts and uploaded files.
Args:
include_uploads: Whether to clean upload directory
include_results: Whether to clean results directory
include_analysis: Whether to clean analysis artifacts
Returns:
Dict containing cleanup results
"""
data = {
'cleanup_uploads': include_uploads,
'cleanup_results': include_results,
'cleanup_analysis': include_analysis
}
response = self._make_request('POST', '/cleanup', json=data)
return response.json()
def check_health(self) -> Dict:
"""Check the health status of the LitterBox service.
Returns:
Dict containing health status information
"""
response = self._make_request('GET', '/health')
return response.json()
def delete_file(self,
file_hash: str) -> Dict:
"""Delete a file and its analysis results.
Args:
file_hash: Hash of the file to delete
Returns:
Dict containing deletion status
"""
response = self._make_request('DELETE', f'/file/{file_hash}')
return response.json()
def validate_process(self, pid: Union[str, int]) -> Dict:
"""Validate if a process ID exists and is accessible.
Args:
pid: Process ID to validate
Returns:
Dict containing validation status
"""
response = self._make_request('POST', f'/validate/{pid}')
return response.json()
def analyze_with_doppelganger(self,
analysis_type: str,
operation: str,
file_hash: Optional[str] = None,
folder_path: Optional[str] = None,
extensions: Optional[List[str]] = None,
threshold: int = 1) -> Dict:
"""Unified method for doppelganger analysis operations.
Args:
analysis_type: 'blender' or 'fuzzy'
operation: 'scan', 'create_db', or 'analyze'
file_hash: Hash of file to analyze (for analyze operation)
folder_path: Path to folder (for create_db operation)
extensions: List of file extensions (for create_db operation)
threshold: Similarity threshold (for fuzzy analysis)
Returns:
Dict containing analysis results
"""
if analysis_type not in ['blender', 'fuzzy']:
raise ValueError("analysis_type must be either 'blender' or 'fuzzy'")
if operation == 'scan' and analysis_type != 'blender':
raise ValueError("scan operation is only available for blender analysis")
# For GET requests (comparisons)
if file_hash and operation != 'analyze':
params = {'type': analysis_type, 'hash': file_hash}
response = self._make_request('GET', '/doppelganger', params=params)
return response.json()
# For POST requests
data = {
'type': analysis_type,
'operation': operation
}
if operation == 'create_db':
if not folder_path:
raise ValueError("folder_path is required for create_db operation")
data['folder_path'] = folder_path
if extensions:
data['extensions'] = extensions
elif operation == 'analyze':
if not file_hash:
raise ValueError("file_hash is required for analyze operation")
data['hash'] = file_hash
data['threshold'] = threshold
response = self._make_request('POST', '/doppelganger', json=data)
return response.json()
def run_system_scan(self) -> Dict:
"""Run a system-wide scan using doppelganger blender analysis."""
return self.analyze_with_doppelganger('blender', 'scan')
def compare_against_system(self, file_hash: str, analysis_type: str = 'blender') -> Dict:
"""Compare a file against system state using doppelganger."""
return self.analyze_with_doppelganger(analysis_type, 'compare', file_hash=file_hash)
def create_fuzzy_db(self, folder_path: str, extensions: Optional[List[str]] = None) -> Dict:
"""Create fuzzy hash database using doppelganger."""
return self.analyze_with_doppelganger('fuzzy', 'create_db',
folder_path=folder_path,
extensions=extensions)
def analyze_fuzzy(self, file_hash: str, threshold: int = 1) -> Dict:
"""Analyze a file using fuzzy hash comparison."""
return self.analyze_with_doppelganger('fuzzy', 'analyze',
file_hash=file_hash,
threshold=threshold)
def __enter__(self):
"""Support for context manager protocol."""
return self
def __exit__(self, exc_type, exc_val, exc_tb):
"""Clean up resources when exiting context manager."""
self.session.close()
def create_arg_parser():
"""Create and configure the argument parser for command line usage."""
parser = argparse.ArgumentParser(
description="LitterBox Malware Analysis Client",
formatter_class=argparse.RawDescriptionHelpFormatter,
epilog="""
Examples:
# Upload and analyze a file
%(prog)s upload malware.exe --analysis static dynamic
# Analyze a running process
%(prog)s analyze-pid 1234 --wait
# Run Doppelganger scan
%(prog)s doppelganger-scan --type blender
# Run Doppelganger analysis
%(prog)s doppelganger abc123def --type fuzzy
# Create fuzzy hash database
%(prog)s doppelganger-db --folder /path/to/files --extensions .exe .dll
# Get analysis results
%(prog)s results abc123def --type static
# Clean up analysis artifacts
%(prog)s cleanup --all
"""
)
# General options
parser.add_argument('--debug', action='store_true', help='Enable debug logging')
parser.add_argument('--url', default='http://127.0.0.1:1337', help='LitterBox server URL')
parser.add_argument('--timeout', type=int, default=30, help='Request timeout in seconds')
parser.add_argument('--no-verify-ssl', action='store_true', help='Disable SSL verification')
parser.add_argument('--proxy', help='Proxy URL (e.g., http://proxy:8080)')
subparsers = parser.add_subparsers(dest='command', help='Command to execute')
# Upload command
upload_parser = subparsers.add_parser('upload', help='Upload file for analysis')
upload_parser.add_argument('file', help='File to upload')
upload_parser.add_argument('--name', help='Custom name for the file')
upload_parser.add_argument('--analysis', nargs='+', choices=['static', 'dynamic'],
help='Run analysis after upload')
upload_parser.add_argument('--args', nargs='+', help='Command line arguments for dynamic analysis')
# Analyze PID command
analyze_pid_parser = subparsers.add_parser('analyze-pid', help='Analyze running process')
analyze_pid_parser.add_argument('pid', type=int, help='Process ID to analyze')
analyze_pid_parser.add_argument('--wait', action='store_true',
help='Wait for analysis completion')
analyze_pid_parser.add_argument('--args', nargs='+', help='Command line arguments')
# Results command
results_parser = subparsers.add_parser('results', help='Get analysis results')
results_parser.add_argument('target', help='File hash or PID')
results_parser.add_argument('--type', choices=['static', 'dynamic', 'info'],
required=True, help='Type of results to retrieve')
# Files summary command
subparsers.add_parser('files', help='Get summary of all analyzed files')
# Doppelganger scan command
doppelganger_scan_parser = subparsers.add_parser('doppelganger-scan', help='Run doppelganger system scan')
doppelganger_scan_parser.add_argument('--type', choices=['blender', 'fuzzy'], default='blender',
help='Type of scan to perform')
# Doppelganger analyze command
doppelganger_parser = subparsers.add_parser('doppelganger', help='Run doppelganger analysis')
doppelganger_parser.add_argument('hash', help='File hash to analyze')
doppelganger_parser.add_argument('--type', choices=['blender', 'fuzzy'], required=True,
help='Type of analysis to perform')
doppelganger_parser.add_argument('--threshold', type=int, default=1,
help='Similarity threshold for fuzzy analysis')
# Doppelganger database command
db_parser = subparsers.add_parser('doppelganger-db', help='Create doppelganger fuzzy database')
db_parser.add_argument('--folder', required=True, help='Folder path to process')
db_parser.add_argument('--extensions', nargs='+', help='File extensions to include')
# Cleanup command
cleanup_parser = subparsers.add_parser('cleanup', help='Clean up analysis artifacts')
cleanup_parser.add_argument('--all', action='store_true',
help='Clean all artifacts')
cleanup_parser.add_argument('--uploads', action='store_true',
help='Clean upload directory')
cleanup_parser.add_argument('--results', action='store_true',
help='Clean results directory')
cleanup_parser.add_argument('--analysis', action='store_true',
help='Clean analysis artifacts')
# Health check command
subparsers.add_parser('health', help='Check service health')
# Delete command
delete_parser = subparsers.add_parser('delete', help='Delete file and its results')
delete_parser.add_argument('hash', help='File hash to delete')
return parser
def main():
parser = create_arg_parser()
args = parser.parse_args()
# Configure logging
log_level = logging.DEBUG if args.debug else logging.INFO
logging.basicConfig(level=log_level,
format='%(asctime)s - %(levelname)s - %(message)s')
logger = logging.getLogger('litterbox')
# Create client with CLI arguments
client_kwargs = {
'base_url': args.url,
'timeout': args.timeout,
'verify_ssl': not args.no_verify_ssl,
'logger': logger,
}
if args.proxy:
client_kwargs['proxy_config'] = {
'http': args.proxy,
'https': args.proxy
}
try:
client = LitterBoxClient(**client_kwargs)
if args.command == 'upload':
# Upload file
result = client.upload_file(args.file, file_name=args.name)
file_hash = result['file_info']['md5']
print(f"File uploaded successfully. Hash: {file_hash}")
# Run requested analysis
if args.analysis:
for analysis_type in args.analysis:
print(f"Running {analysis_type} analysis...")
analysis_args = args.args if analysis_type == 'dynamic' else None
result = client.analyze_file(
file_hash,
analysis_type,
cmd_args=analysis_args,
wait_for_completion=True
)
if result.get('status') == 'early_termination':
print("Process terminated early:")
print(f"Error: {result.get('error')}")
print("Details:")
print(f" Termination time: {result['details'].get('termination_time')}")
print(f" Init time: {result['details'].get('init_time')}")
print(f" Message: {result['details'].get('message')}")
elif result.get('status') == 'error':
print(f"Analysis failed: {result.get('error')}")
if 'details' in result:
print(f"Details: {result['details']}")
else:
print(json.dumps(result, indent=2))
elif args.command == 'analyze-pid':
# Analyze process
print(f"Analyzing process {args.pid}...")
result = client.analyze_file(
str(args.pid),
'dynamic',
cmd_args=args.args,
wait_for_completion=args.wait
)
print(json.dumps(result, indent=2))
elif args.command == 'results':
result = client.get_results(args.target, args.type)
print(json.dumps(result, indent=2))
elif args.command == 'files':
result = client.get_files_summary()
print(json.dumps(result, indent=2))
elif args.command == 'doppelganger-scan':
print(f"Running doppelganger scan with type: {args.type}")
result = client.analyze_with_doppelganger(args.type, 'scan')
print(json.dumps(result, indent=2))
elif args.command == 'doppelganger':
print(f"Running doppelganger analysis with type: {args.type}")
result = client.analyze_with_doppelganger(
args.type,
'analyze',
file_hash=args.hash,
threshold=args.threshold
)
print(json.dumps(result, indent=2))
elif args.command == 'doppelganger-db':
print("Creating doppelganger fuzzy database...")
result = client.analyze_with_doppelganger(
'fuzzy',
'create_db',
folder_path=args.folder,
extensions=args.extensions
)
print(json.dumps(result, indent=2))
elif args.command == 'cleanup':
if args.all:
args.uploads = args.results = args.analysis = True
result = client.cleanup(
include_uploads=args.uploads,
include_results=args.results,
include_analysis=args.analysis
)
print(json.dumps(result, indent=2))
elif args.command == 'health':
result = client.check_health()
print(json.dumps(result, indent=2))
elif args.command == 'delete':
result = client.delete_file(args.hash)
print(json.dumps(result, indent=2))
else:
parser.print_help()
except LitterBoxAPIError as e:
logger.error(f"API Error (Status {e.status_code}): {str(e)}")
if args.debug:
logger.debug(f"Response data: {e.response}")
sys.exit(1)
except LitterBoxError as e:
logger.error(f"Client Error: {str(e)}")
sys.exit(1)
except Exception as e:
logger.error(f"Unexpected Error: {str(e)}")
if args.debug:
logger.exception("Detailed error information:")
sys.exit(1)
finally:
client.session.close()
if __name__ == "__main__":
try:
main()
except KeyboardInterrupt:
print("\nOperation cancelled by user")
sys.exit(130)