Offensive AI Security Engineer – Red Team
Newark, CA
Full Time Senior-level / Expert USD 128K - 177K
Lucid Motors
Lucid is the future of sustainable mobility, designing electric cars that further reimagines the driving experience.Job Title: Offensive AI Security Engineer – Red Team
We are seeking an Offensive AI Security Engineer to join our AI Red Team within the Security Engineering team. This role focuses on adversarial machine learning (ML), AI-driven offensive security, and red teaming AI systems to uncover vulnerabilities in AI-powered automotive security models and vehicle platforms.
As part of Lucid’s Offensive AI Security team, you will attack, manipulate, and exploit AI/ML models to identify real-world threats and weaknesses in AI-driven security solutions. You will develop AI-enhanced security automation tools, perform LLM-based penetration testing, and integrate AI/ML attack techniques into offensive security operations.
Key Responsibilities:
AI Red Teaming & Adversarial Attack Development
- Design and execute adversarial attacks on AI-powered security systems.
- Conduct LLM-based penetration testing to uncover AI security flaws in vehicle cybersecurity applications.
- Identify attack surfaces in AI-driven perception systems (LiDAR, Radar, Cameras) and develop exploits against automotive AI models.
AI-Driven Offensive Security Automation
- Build tools and systems to analyze emerging AI threats in vehicle security and enterprise AI models.
- Develop AI-assisted security automation tools for:
- Reconnaissance – Automate vulnerability discovery using LLMs and RAG-based intelligence gathering.
- Exploitation – Use AI to generate attack payloads and automate offensive security operations.
- Fuzzing – Enhance automated fuzz testing with AI-driven input mutation strategies.
- Reverse Engineering – Apply LLM-assisted binary analysis for rapid security assessments.
- Build offensive and defensive AI security tools, integrating ML-driven automation into security assessments and exploit development.
ML-Driven Security Research & Exploitation
- Use ML to characterize a program, identifying security-critical functions and behavioral anomalies.
- Understand LLVM Intermediate Representation (LLVM IR) to analyze compiled AI/ML software for security weaknesses.
- Develop AI-driven techniques for:
- Threat Detection – Use ML to automate malware detection and anomaly recognition.
- Cryptographic Algorithm Classification – Identify cryptographic weaknesses in compiled binaries.
- Function Recognition – Use AI models to automate binary function analysis and decompilation.
- Vulnerability Discovery – Automate zero-day discovery using ML-based exploit prediction models.
- Evaluate ML security models for robustness, performance, and adversarial resilience.
Offensive AI Research & Red Teaming Strategy
- Research novel AI attack techniques and evaluate their impact on vehicle cybersecurity and enterprise AI security models.
- Collaborate with internal Red Teams, SOC analysts, and AI security researchers to refine AI-driven offensive security approaches.
- Stay ahead of emerging AI threats, tracking advancements in AI security, adversarial ML, and autonomous vehicle AI exploitation.
Required Skills & Qualifications:
AI Red Teaming & Offensive Security Expertise
✔ Hands-on experience with AI/ML exploitation, adversarial ML, and AI-driven pentesting.
✔ Strong background in attacking AI models, including LLMs, deep learning systems, and computer vision AI.
✔ Experience with LLM-based vulnerability analysis, prompt engineering attacks, and model evasion techniques.
✔ Proficiency in penetration testing, AI fuzzing, and red teaming AI-driven security applications.
Cybersecurity & AI Security Research
✔ Experience in exploiting AI-powered vehicle security mechanisms.
✔ Ability to analyze, test, and exploit AI models within embedded systems and cloud environments.
Programming & Security Tooling
✔ Strong Python, C, C++ skills for AI security research and offensive security automation.
✔ Proficiency with AI security frameworks (TensorFlow, PyTorch, Hugging Face, OpenAI APIs, LangChain).
✔ Hands-on experience with reverse engineering tools (Ghidra, IDA Pro, Binary Ninja, Qiling, Angr).
✔ Familiarity with AI fuzzing tools (AFL++, Honggfuzz, Syzkaller) and adversarial ML frameworks (CleverHans, Foolbox, ART).
Preferred Qualifications:
- Bachelor’s degree in Cybersecurity, AI/ML, Computer Science, or related technical field(s) is required. Master’s degree or higher education is preferred.
- Experience in offensive AI security, red teaming AI models, or AI-powered security automation.
- Prior work in LLM-based security analysis, AI-driven pentesting, and adversarial ML research.
By Submitting your application, you understand and agree that your personal data will be processed in accordance with our Candidate Privacy Notice. If you are a California resident, please refer to our California Candidate Privacy Notice.
To all recruitment agencies: Lucid Motors does not accept agency resumes. Please do not forward resumes to our careers alias or other Lucid Motors employees. Lucid Motors is not responsible for any fees related to unsolicited resumes.Tags: APIs Classification Computer Science Computer Vision Deep Learning Engineering LangChain Lidar LLMs Machine Learning ML models OpenAI Privacy Prompt engineering Python PyTorch Radar RAG Research Security TensorFlow Testing
Perks/benefits: 401(k) matching Career development Competitive pay Equity / stock options Health care Insurance
More jobs like this
Explore more career opportunities
Find even more open roles below ordered by popularity of job title or skills/products/technologies used.