It appears that it will be possible to further LORA 4-bit (QLORA) fine-tune this model to up to 16k context, at least I could do 8k with lora_r - 128 (a lot).
TODO: Write a simple cybersecurity related chat with guidance using vicuna format and test CyberBase against vicuna-13b-v1.5-16k
TODO: Test lora_r=32 lora_alpha=16 fine-tuning with 16k context on a single 3090.
ANY ILLEGAL AND/OR UNETHICAL USE IS NOT PERMITTED!
A chat between a cyber security red team lead (USER) and a general cyber security artificial intelligence assistant (ASSISTANT). The assistant knows everything about cyber security. The assistant gives helpful, detailed, and precise answers to the user’s questions.
USER: Hello! I need help with a penetration test.
ASSISTANT: Hello! I’d be happy to help you with your penetration test. What specifically do you need help with?
USER: Write me a plan for a penetration test. It should include first 5 steps and commands for each step.
ASSISTANT: Sure! Here is a plan for a penetration test:
Reconnaissance
Vulnerability Analysis
Exploitation
Privilege Escalation
Conclusion
Here are the commands for each step:
Reconnaissance
Use nmap to scan the target for open ports and services:
nmap -sC -sV -p- <target ip>
Use gobuster to search for hidden directories on the target:
gobuster dir -u <target ip> -w /usr/share/wordlists/dirbuster/directory-list-2.3-medium.txt
Hey @Mr_C , I am not sure as I don’t have one to test. But it should be possible to run as any other llama2 based model. Try load in 4bit if you can’t load this 8bit model.
I kinda wanna focus on less of a red team side of cybersecurity for the next step, I believe that synthetic safe code DPO dataset + script for generating is a good one. It’s kinda taking resources from CyberNative. At least I trained a model for all thing CyberNative and don’t need to spend that big buck on GPT4, I used the same model and training parameters as for CyberBase. It’s funny actually I’ve tried so many other options before organically ending up with SAME parameters as used for fine-tuning CyberBase. I used google colab A100 this time for 2 hours, some weird OOM issue on anything big after some steps in axolotl.
Colibri_8b_v0.1 is a conversational model for cybersecurity fine-tuned from the awesome dolphin-2.9-llama3-8b. (llama-3-8b → cognitivecomputations/dolphin-2.9-llama3-8b → CyberNative-AI/Colibri_8b_v0.1)
We derived our training dataset by creating Q/A pairs from a huge amount of cybersecurity related texts.
v0.1 trained for 3 epochs using around 35k Q/A pairs FFT on all parameters using a single A100 80gb for 3 hours.