
Đã đóng
Đã đăng vào
Thanh toán khi bàn giao
I’ve just released a production-grade chatbot and now need a seasoned red-teamer to see how far its guard-rails can be pushed. Your job is to run an end-to-end security assessment that zeroes in on prompt injection, jailbreak techniques, and any threat-detection angles relevant to LLM-powered systems. You’ll have direct access to the public interface (plus limited system-prompt details if that helps). Think like an adversary, chain prompts, exploit context windows, and surface every path that lets a user override hidden instructions, leak data, or trigger unsafe actions. When you reply, point me to past work that proves you’ve broken—or secured—similar chatbots. Solid examples of exploits or red-team reports are the single most important factor in my selection. Deliverables • Methodology or test plan outlining the attacks attempted • Step-by-step transcripts (and screenshots) of every successful jailbreak or prompt-injection exploit • Impact and severity rating for each finding • Clear, actionable fixes and guardrail improvements • Concise final report (PDF or Markdown) I’ll sign off once I can reproduce the issues from your steps and confirm that your mitigations close each gap.
Mã dự án: 40324931
10 đề xuất
Dự án từ xa
Hoạt động 20 ngày trước
Thiết lập ngân sách và thời gian
Nhận thanh toán cho công việc
Phác thảo đề xuất của bạn
Miễn phí đăng ký và cháo giá cho công việc
10 freelancer chào giá trung bình ₹6.871 INR cho công việc này

Hi there, I’ve reviewed your requirement for a full red-team assessment of your production chatbot, focusing on prompt injection, jailbreaks, and guardrail bypass techniques. I’ll approach this from an adversarial perspective, chaining prompts, exploiting context windows, and testing for data leakage, instruction override, and unsafe action triggers. The goal is to uncover every realistic attack path and validate how resilient your system is under pressure. I can provide a structured test plan covering injection vectors, role confusion, tool misuse, and memory/context exploits. You’ll receive detailed transcripts of successful attacks, severity ratings, and clear reproduction steps along with screenshots. Each finding will include actionable fixes such as prompt hardening, input/output filtering, and system design improvements to close vulnerabilities effectively. With experience testing and securing LLM workflows, I focus on practical, reproducible exploits and clear mitigation strategies. I’ll deliver a concise, well-documented report and ensure you can validate each issue and confirm the fixes. I’m ready to begin immediately and provide a thorough, professional assessment. Best regards, Muhammad Adil Portfolio: https://www.freelancer.com/u/webmasters486
₹12.000 INR trong 4 ngày
4,7
4,7

We at Offensium Vault Private Limited (ISO 27001:2022 & ISO 9001:2015) specialize in LLM security testing and red teaming and can perform a full adversarial assessment of your chatbot. Approach • Prompt injection & jailbreak testing using chained prompts, context manipulation, and role override attempts • Testing for data leakage, system prompt exposure, and unsafe action triggering • Evaluation of guardrails, moderation, and response filtering mechanisms • Simulation of real-world attacker behavior aligned with OWASP LLM Top 10 concepts Deliverables • Test methodology / attack plan • Step-by-step transcripts & screenshots of successful exploits • Severity ratings with impact analysis • Actionable fixes and guardrail improvements • Final report (PDF/Markdown) with reproducible findings Our team has experience identifying vulnerabilities in AI systems, SaaS platforms, and APIs, including advanced logic abuse and injection scenarios. We can start immediately and provide a structured red-team assessment to strengthen your chatbot’s security before wider exposure.
₹10.000 INR trong 7 ngày
3,6
3,6

Hello, I have a few queries regarding the chatbot red-teaming and security assessment. 1) Which specific LLM model is powering your chatbot? 2) Are there certain topics or data types that are strictly off-limits for the bot? 3) Do you have a preferred format for the vulnerability report? I will conduct a deep security audit of your chatbot using advanced prompt injection and jailbreak techniques to test your guard rails. I will use various adversarial methods to try and bypass hidden instructions, leak system prompts, or trigger restricted responses. My approach involves testing the context window limits and chaining complex queries to identify any logic flaws in the safety layer. I will document every successful exploit and provide a detailed report on how to harden your system against these specific threats. I have performed similar red-teaming for several AI-driven platforms, focusing on breaking multi-layered defense prompts and ensuring data privacy. Thanks, Nivedita
₹10.000 INR trong 7 ngày
2,1
2,1

I am a security researcher with a specialized focus on LLM Red Teaming and AI Safety. I recently participated in the Grey Swan Arena benchmark, where I worked on identifying vulnerabilities and testing the safety boundaries of large language models. This experience has given me a deep understanding of how to chain prompts and exploit context windows to bypass hidden instructions. My Approach for Your Chatbot: Adversarial Prompting: Using techniques like DAN, role-playing, and payload splitting to test jailbreak resistance. Indirect Prompt Injection: Testing if external data or user inputs can override the system prompt. Data Leakage: Attempting to extract the system prompt or any sensitive training data/instructions. Reporting: I will provide a Markdown/PDF report with step-by-step transcripts, severity ratings (Low to Critical), and specific remediation steps to strengthen your guardrails. I am currently an eJPT student with hands-on experience in penetration testing tools, which helps me think like a real-world attacker. I can deliver a comprehensive assessment that ensures your production-grade chatbot is truly secure. Looking forward to helping you secure your system
₹7.000 INR trong 7 ngày
0,0
0,0

Hello, your need for a thorough end-to-end security assessment of your chatbot, particularly focusing on prompt injection and jailbreak techniques, is exactly where I excel. I am Mubashir, a security specialist with 6+ years of experience in penetration testing and AI systems. I understand that you require a detailed analysis to identify vulnerabilities and ensure robust guardrails for your chatbot. 1. First, I will review the public interface and any system-prompt details you provide to understand the architecture and potential weaknesses. 2. Then, I will develop a test plan that outlines various prompt injection and jailbreak techniques tailored to your chatbot's unique structure. 3. After that, I will execute the planned attacks, documenting each step with transcripts and screenshots of successful exploits. 4. I will assess the impact and severity of each finding, followed by recommendations for actionable fixes and improvements to the guardrails. 5. Finally, I will compile a concise report in PDF or Markdown format detailing all findings and suggested mitigations. Even if you’re not ready to move forward yet, I’d be happy to provide a FREE detailed quotation and project proposal, including a suggested roadmap. This can serve as a valuable reference even if you choose to work with someone else. >>> My Work: https://www.freelancer.com/u/mubashir021/QA-Security-Engineer <<< Drop me a message and let's get this sorted. Mubashir
₹7.209,01 INR trong 7 ngày
0,0
0,0

Hello, I can test your chatbot to check for prompt injection, jailbreaks, and data leakage. I'll offer concise, repeatable results with implications and workable solutions.
₹2.000 INR trong 7 ngày
0,0
0,0

Expert AI Security & Prompt Injection Testing Specialist I specialize in AI security testing and prompt injection pentesting. My approach: 1. Comprehensive analysis of your AI chatbot for vulnerabilities 2. Methodical testing for prompt injection weaknesses 3. Detailed reporting with severity ratings 4. Practical remediation guidance My expertise: - AI Security Testing (OpenAI GPT models) - Prompt Injection techniques - OWASP AI Security guidelines - Risk assessment and prioritization My offer: - Price: ₹3,000 INR (56% below average) - Delivery: 7 days - Guarantee: 100% satisfaction or full refund I'll provide a free preliminary security assessment within 24 hours. Ready to secure your AI chatbot! Best regards, Valery - AI & Security Specialist
₹7.000 INR trong 7 ngày
0,0
0,0

Vadodara, India
Phương thức thanh toán đã xác thực
Thành viên từ thg 7 24, 2019
₹600-1500 INR
$30-250 USD
$10-30 USD
$30-250 USD
₹600-1500 INR
$25-50 AUD/ giờ
€8-30 EUR
$300-1000 USD
$10000-20000 USD
₹75000-150000 INR
€8-30 EUR
₹12500-37500 INR
€30-250 EUR
₹150000-250000 INR
$30-250 CAD
$30-250 USD
$10-30 CAD
₹1500-12500 INR
₹600-1500 INR
₹600-1500 INR
$3000-5000 USD
$30-250 USD
$8-15 USD/ giờ
£250-750 GBP
₹37500-75000 INR