Androidmeda is an LLM tool designed to deobfuscate Android application code and find vulnerabilities by leveraging the power of Large Language Models (LLMs). It analyzes decompiled Android source code to identify obfuscated patterns, suggest more readable names for variables, methods, and classes, and ultimately provide a clearer understanding of an application's logic.
Android apps generally use Proguard for obfuscating the app and reversing app can be hard with meaning less variable and function names. LLMs seems to do a good job in understanding the context of the code and renaming variables, functions, adding comments etc. (to certain extent but embrace the unpredictibility of it too).
Androidmeda supports both LLM APIs (like OpenAI, Gemini, Anthropic) and local LLM inference via Ollama, offering flexibility and control over your analysis environment.
Apparently Androidmeda can be used to deobfuscate android malware as well. Some great articles covering it.
Please cite, If you use this software in your Research papers, articles etc.
@software{Androidmeda-Deobfuscate-android-app_2024,
author = {Agrawal, Vaibhav},
month = nov,
title = {{Androidmeda-Deobfuscate-android-app: Android app Vulnerability Scanner and Deobfuscator using LLM}},
url = {https://github.com/In3tinct/deobfuscate-android-app},
version = {1.0.0},
year = {2024}
}
- LLM-Powered Deobfuscation: Utilizes advanced LLMs to intelligently deobfuscate code elements and add helpful comments.
- Local Model Support: Seamless integration with local LLMs via Ollama for privacy-focused and cost-effective analysis.
- API Model Support: Compatibility with popular LLM APIs for broader model access.
- Vulnerability Reporting: Generates reports highlighting potential vulnerabilities found during the analysis.
Input - Takes decompiled code directory as an input.
Output -
-
- Vulnerability Report: A JSON file with name "vuln_report" summarizing identified vulnerabilities.
-
- Deobfuscated Code: If
--save_code truewas used, Deobfuscates each file for easier readability and save it in package directory structure for manual reviews. Additionally, labels any security issues seen in the generated code with #SECURITY-ISSUE.
- Deobfuscated Code: If
git clone https://github.com/In3tinct/Deobfuscate-android-app.git
pip3 install -r requirements.txt
Get the APK of the intended app. And You can use jadx https://github.com/skylot/jadx to decompile. It will create a "resources" and "sources" directory. "Sources" directory is where the decompiled .java files sit.
jadx androidapp.apk
a. Using Google/Anthropic/OpenAI LLM Models.
EXPORT API_KEY= "Your API Key"
You can get the API key for Google gemini, OpenAI chatgpt, Anthropic claude
To Run script with Public LLM APIs,
Gemini - python3 androidmeda.py --llm_provider google --llm_model gemini-1.5-flash -output_dir /tmp/ver/ -source_dir "input_dir1/ input_dir2/"
ChatGPT - python3 androidmeda.py --llm_provider openai --llm_model gpt-4.1 -output_dir /tmp/ver/ -source_dir "input_dir1/ input_dir2/"
b. Using Ollama open source models to run locally.
Follow steps here to download and run the model locally github.com/ollama/ollama
- Warning: Resource Intensity! LLM models are large (several GBs) and require significant RAM.
- 7B models (e.g.,
llama3,mistral): Minimum 8GB RAM, 16GB recommended. - 13B models: Minimum 16GB RAM, 32GB recommended.
- 33B models: Minimum 32GB RAM, 64GB recommended. Ensure your system (or WSL instance) has sufficient free memory.
- 7B models (e.g.,
To Run script with Ollama,
python3 androidmeda.py --llm_provider ollama --llm_model llama3.2 -output_dir /tmp/ver/ -source_dir "input_dir1/ input_dir2/"
Parameters -
-llm_provider is the LLM provider of the model. e.g. google, anthropic, openaI, ollama
-llm_model is the LLM model to use, Gemini, Claude, ChatGPT are supported. You can get the model variants from here. google openai anthropic
-output_dir is the output directory you want to save generated files.
-source_dir is the source directory for the decompiled code. You can send more than one directory separated by space as above.
-save_code (optional) Default is false. if set as True, it will deobfuscate the code and save in the output directory provided, otherwise only vuln_report file will be generated.
Important - Don't send the entire package at once which would contain libraries etc. Otherwise It may take forever to scan. Send the specific directories as input which contains app specific code. For example - if package directory looks like com/google/android/yourapp, send com/google/android/yourapp/receivers/
Decompiled code (Obfuscated)
Decompiled code (After processing with LLM)
Security Issues identified by LLM
See CONTRIBUTING.md for details.
We welcome contributions to Androidmeda! If you have suggestions for improvements, bug fixes, or new features, please follow these steps:
- Fork the repository.
- Create a new branch (
git checkout -b feature/your-feature-name). - Make your changes.
- Commit your changes (
git commit -m 'Add new feature'). - Push to the branch (
git push origin feature/your-feature-name). - Open a Pull Request.
Apache 2.0; see LICENSE for details.
This project is not an official Google project. It is not supported by Google and Google specifically disclaims all warranties as to its quality, merchantability, or fitness for a particular purpose. This is an experimental project and the owner of this project shall not be held liable for any actions performed using this tool. It is the sole responsibility of the end-user. Also, please review the TOS for Gemini API before using the tool. https://ai.google.dev/gemini-api/terms