- Sep 9, 2025
- Parsed from source:Sep 9, 2025
- Detected by Releasebot:Dec 23, 2025
Resemble AI and Carahsoft Partner to Bring Deepfake Simulation and Detection to the Public Sector
Resemble AI unveils its Generative AI Voice‑Based Deepfake Simulation Platform for the Public Sector, available via Carahsoft. Agencies can run live phishing simulations with hyper realistic voice cloning, get scoring and analytics, and accelerate deployments through government contracts.
Resemble AI Generative AI Voice-Based Deepfake Simulation Platform for Public Sector
AI-powered threats are no longer science fiction. From voice phishing scams to real-time social engineering attacks, deepfakes have entered the mainstream — and government agencies are among the most at risk.
That’s why we’re excited to announce a major milestone: Resemble AI has launched its Generative AI Voice-Based Deepfake Simulation Platform, now available to the Public Sector through Carahsoft Technology Corp., The Trusted Government IT Solutions Provider®.
This collaboration makes it possible for agencies across federal, state, and local levels to proactively defend against AI-driven threats.
Resemble AI’s platform equips agencies with critical tools and training, decreasing the risk of deepfake attacks. It provides agencies with a crucial protection layer and the ability to pinpoint organizational risks.
Michael Adams
Program Executive for AI Solutions,
CarahsoftWhy Deepfake Simulation Matters
Deepfake fraud has already cost organizations billions globally. Attackers are no longer limited to fake emails or static phishing pages. Instead, they’re cloning real voices, calling employees, and manipulating them into revealing sensitive information.
Traditional security training isn’t designed for this new reality. Slide decks and static phishing tests can’t replicate the stress of answering what sounds like your boss’s voice on the other end of a call.
That’s where simulation changes the game.
Resemble AI’s platform uses hyper-realistic voice cloning and adaptive AI to run live phishing scenarios. Imagine a spoofed customer call asking for credentials, or a WhatsApp message from a cloned executive voice. Employees are scored on their responses, while organizations get detailed analytics to highlight blind spots and areas of risk.Why Carahsoft Partnership Matters
As Resemble AI’s Master Government Aggregator®, Carahsoft will make this platform widely accessible through its extensive contract vehicles, including NASA SEWP V, ITES-SW2, NASPO ValuePoint, and OMNIA Partners.
This means agencies can adopt Resemble AI’s platform quickly, without procurement hurdles, and begin strengthening defenses against deepfakes immediately.Learn more about Resemble’s Deepfake Detection Platform
- Learn more about Resemble’s Deepfake Detection Platform
- Start a free trial
- Sep 4, 2025
- Parsed from source:Sep 4, 2025
- Detected by Releasebot:Dec 23, 2025
Introducing Chatterbox Multilingual: Open Source TTS for 23 Languages
Chatterbox Multilingual brings 23 languages, expressive control, and enterprise‑grade reliability to open‑source TTS, enabling natural voices across locales. The Pro tier adds fine‑tuning, sub‑200 ms latency, SLAs, and robust watermarking for safe, scalable voice apps.
Two months ago, we did something the industry said was impossible. Our team released Chatterbox—an open-source text-to-speech model that outperformed ElevenLabs with 63.75% user preference in blind evaluations.
The response was overwhelming. Within weeks, Chatterbox exploded to over 1 Million downloads on Hugging Face and above 11,000 GitHub stars. Developers built audiobook generators, game studios integrated it for NPC dialogue, and educators created language learning tools. DigitalOcean called it “significant progress in personalized Voice AI.” The community had spoken: open-source TTS had finally arrived.
But in every GitHub issue, Reddit thread, and Discord conversation, one request dominated: “When will Chatterbox support other languages?”
Today, we answer that call.Languages Supported
- Arabic (ar)
- Danish (da)
- German (de)
- Greek (el)
- English (en)
- Spanish (es)
- Finnish (fi)
- French (fr)
- Hebrew (he)
- Hindi (hi)
- Italian (it)
- Japanese (ja)
- Korean (ko)
- Malay (ms)
- Dutch (nl)
- Norwegian (no)
- Polish (pl)
- Portuguese (pt)
- Russian (ru)
- Swedish (sv)
- Swahili (sw)
- Turkish (tr)
- Chinese (zh)
State of the Art Voice Cloning in 23 Languages
Demand for multilingual TTS is skyrocketing. Brands, educators, and developers are no longer limited to English-only voice AI. Users expect apps and agents to sound human, speak in their native language, and deliver content with authentic tone and emotion.
Chatterbox Multilingual was built to meet that demand. It combines:- Breadth of languages: 23 supported languages from launch.
- Expressive control: Fine-tune delivery with emotion and intensity settings.
- Enterprise reliability: Ultra-stable inference and built-in watermarking.
Whether you’re designing a voice AI agent for customer support, a language-learning app, or a global gaming experience, Chatterbox Multilingual gives you the tools to build voices that resonate everywhere.
Chatterbox Multilingual Samples
Get Started from HuggingFace
Six lines of code to access Hollywood-quality voice synthesis in 23 languages.
Enterprise Level Voice Cloning with Chatterbox Multilingual Pro
While the open-source release brings world-class TTS to everyone, many enterprises need fine-tuned performance and strict service guarantees. That’s where Chatterbox Multilingual Pro comes in.
- 🎯 Custom fine-tuning: Train on your brand’s vocabulary, product names, and domain terms.
- ⚡ Sub-200 ms latency: Real-time streaming with global infrastructure.
- 📈 SLAs and reliability: Guaranteed uptime and throughput.
- 🔒 Advanced watermarking & detection: Enhanced protection against deepfakes.
For call centers, financial services, healthcare platforms, and any organization where accuracy and compliance matter, Chatterbox Multilingual Pro closes the last mile.
Get started with Chatterbox Multilingual for free
Create your Voice Clone
Original source Report a problem - Aug 27, 2025
- Parsed from source:Aug 27, 2025
- Detected by Releasebot:Dec 23, 2025
Introducing Telephony Optimized Deepfake Detection Model
Resemble AI unveils codec-aware inline in-call detection across G.711, G.729, AMR-WB and Opus with a major accuracy boost for synthetic speech in compressed streams. Enhanced reliability, codec‑level security, and enterprise scale for contact centers and telecom networks.
Resemble AI is raising the bar for inline in-call detection with new support for leading telephony codecs — G.711, G.729, AMR-WB, and Opus — combined with a significant accuracy breakthrough in detecting synthetic and manipulated speech across compressed audio streams.
Why it matters
Most voice fraud today rides over bandwidth-constrained codecs optimized for telephony. Detecting deepfakes or impersonation attacks in these conditions is notoriously challenging due to lossy compression, packetization, and bandwidth artifacts. Our latest release closes this gap, ensuring high-precision detection even on low-bitrate, real-world call traffic.
What’s new
- Codec-Aware Detection Models – Enhanced accuracy across narrowband and wideband codecs without requiring PCM-quality input.
- Improved Inline Reliability – Greater resilience to compression artifacts, jitter, and packet loss in SIP, SIPREC, and RTP streams.
- Enterprise & Carrier Scale – Optimized for high concurrency deployments in contact centers, SBC environments, and telco networks.
- Security Against Evasion – Models hardened against codec-level adversarial perturbations, closing common attack vectors.
The outcome
Enterprises and carriers can now deploy real-time, in-call detection with confidence that voice integrity checks remain accurate regardless of the codec in use. Whether protecting a financial services hotline, a carrier call path, or a government comms channel, detection remains reliable, fast, and inline.
About Resemble’s Deepfake Detection Model
Resemble AI’s Deepfake Detection Model is a multimodal system built to identify manipulated audio, images, and video with industry-leading accuracy. Trained on a wide range of generative model outputs, it can spot synthetic media even as new architectures and techniques emerge. The model operates in real time, making it ideal for securing customer interactions, financial transactions, and enterprise communications where trust is critical. By combining audio forensics, visual signal analysis, and cross-modal consistency checks, Resemble delivers a unified layer of protection against deepfakes, ensuring authenticity across every channel where fraud may appear.
Ready to deploy realtime Deepfake Detection?
Schedule Free Demo
Original source Report a problem - Jun 24, 2025
- Parsed from source:Jun 24, 2025
- Detected by Releasebot:Dec 23, 2025
Introducing Deepfake Security Awareness Training Platform to Reduce Gen AI-Based Threats
Resemble AI unveils a voice-based deepfake simulation platform for cyber defense, enabling realistic impersonation drills across calls, voicemails, and WhatsApp. In limited release, it scores risk per user and helps teams harden defenses before real attacks.
Today, Resemble AI is excited to introduce a groundbreaking approach to cybersecurity: a voice-based deepfake simulation platform designed to help organizations test and harden their defenses against AI-driven social engineering. Early adopters have already reported up to a 90% reduction in successful attacks.
While most security awareness training relies on static videos and generic phishing emails, Resemble AI’s platform brings simulations to life with hyper-realistic voice cloning and adaptive conversations. Think of it as red-teaming, but with AI-powered impersonation attempts sent straight to your employees — via phone calls, voicemails, and even WhatsApp messages.
These aren’t scripted prompts. Our platform uses agentic AI that maintains context and handles objections in real-time, powered by Resemble’s proprietary voice models and LLM integrations. From a spoofed CFO voicemail asking for wire transfers to a simulated customer demanding account access, employees experience real-world pressure, with none of the real-world fallout.
Every simulation generates a risk score from 0 to 100, personalized at the individual, team, and department level. Security teams can instantly identify blind spots and high-risk behavior. Whether it’s someone clicking a suspicious link or revealing sensitive info to a fake executive. The result? Targeted training where it’s actually needed, not across-the-board compliance theater.
This simulation platform is already in limited release, piloted by leaders in finance, healthcare, and customer service — sectors where deepfake-based fraud is both prevalent and devastating. In industries where frontline workers answer every call and make real-time decisions, deepfake protection can’t be theoretical. It has to be tested.
According to Resemble AI’s deepfake incident tracker, voice-based fraud has already caused over $2.6 billion in damages globally. Our new platform allows companies to take a proactive approach, actively simulating how those attacks would unfold—and giving them a chance to intervene before real money, data, or trust is lost.
“Today’s attackers aren’t playing by the old rules,” said Zohaib Ahmed, CEO and Co-founder of Resemble AI. “They’re using cloned voices, social cues, and urgency to manipulate in real time. We built this platform so companies can experience what that looks like inside their own walls—and build muscle memory before it counts.”
A Shift From Passive to Active Defense
Legacy tools like KnowBe4 and Cofense focus on passive learning. Resemble AI’s approach is active, adaptive, and voice-first. We don’t just test your weakest link, we help you reinforce it.
Recent deepfake scams, like the one targeting Accenture’s CEO in May, show how voice impersonation is becoming the go-to vector for attackers. Whether it’s a fake vendor, a spoofed HR exec, or a “friendly” internal request, these attacks often sound legitimate and slip past traditional training methods. Our simulations are designed to mimic that reality.
Built Ethically. Deployed Securely.
Resemble AI leads with ethical guardrails and secure design. Our platform integrates:
- Resemble Detect for deepfake detection
- Resemblyzer for speaker verification
- Resemble Watermark for forensic watermarking
We’re serious about safety.
Want to see how your team stacks up against voice-based deepfakes?
Request Access to the Simulation Platform
Original source Report a problem - Apr 30, 2025
- Parsed from source:Apr 30, 2025
- Detected by Releasebot:Dec 23, 2025
Democratizing Truth: Why We Built a WhatsApp Deepfake Detector Anyone Can Use
A new WhatsApp based deepfake detection service, 218-NO-FAKES, lets users send suspicious media to get immediate analysis. It supports images, video, and audio with 94% accuracy and requires no downloads or accounts. Real world tests show fast, actionable insights to verify authenticity.
The deepfake arms race has created a deeply unfair fight. On one side: increasingly powerful AI creation tools with slick interfaces anyone can use. On the other: detection capabilities hidden behind enterprise paywalls, academic jargon, and technical complexity.
It’s like we’ve given everyone flamethrowers while locking fire extinguishers in special high-security vaults that require advanced degrees to access.
We’ve had enough of this asymmetry, so we built something that flips the script: a deepfake detection service that works through WhatsApp. Text 218-NO-FAKES, send any suspicious media, and get an immediate analysis that doesn’t just tell you if something is manipulated, but how it was likely created.The Simplicity Imperative
When we started designing this service, we obsessed over one thing: removing friction. Every step, every requirement, every technical hurdle would mean fewer people using it.
So we opted for the most ubiquitous messaging platform on Earth. No downloads. No accounts. No subscriptions. No specialized knowledge required.
Just send media to 218-NO-FAKES.
Our detection technology works across modalities – analyzing images, video, and audio with 94% accuracy. But accuracy isn’t enough if the tools remain inaccessible.Real-World Verification in Action
We’ve been testing this service across a spectrum of everyday scenarios where the line between authentic and synthetic content gets blurry. Let me walk you through a few examples:
The Digital Receipt Runaround
We took a ChatGPT-generated receipt – the kind of thing someone might use to fake an expense report or establish a false alibi – and sent it to our detector. Within seconds, it flagged the telltale patterns of AI generation. The subtle pixel-level irregularities that human eyes miss? Our system caught them immediately. No more wondering if that invoice is legitimate.
The Forwarded Fear-Mongering
We’ve all been there – someone forwards an alarming image on iMessage that looks just credible enough to cause anxiety. Is that really a hurricane heading toward your city? Did that politician actually do that embarrassing thing? Our system lets you forward that content directly from your messaging apps to 218-NO-FAKES, providing clarity when you need it most.
The “Wait, Was That Really My Bank?” Check
Perhaps most chilling is how easily voice synthesis can now mimic people we trust. We tested a synthetic voicemail that sounded convincingly like it came from a financial institution. Our detector immediately highlighted the voice as synthetic, pointing out the micro-patterns that distinguish AI-generated speech from human voices.
Join the Verification Movement
Text 218-NO-FAKES on WhatsApp and try it yourself. Send that suspicious image. Forward that weird audio clip. Check that too-perfect video.
Original source Report a problem
The more people with verification tools, the less incentive there is to create malicious deepfakes in the first place. It’s a simple equation: when deception becomes pointless, it stops being profitable.
This is just the beginning. We’re already working on enhanced detection capabilities, additional platforms, and deeper educational resources.
Because in a world increasingly shaped by artificial intelligence, the power to discern truth shouldn’t be artificial, too.
It should belong to all of us. - Mar 5, 2025
- Parsed from source:Mar 5, 2025
- Detected by Releasebot:Dec 23, 2025
Voice Design: Transforming Text into Unlimited AI Voices
Introducing Voice Design, a groundbreaking feature that turns simple text prompts into distinct AI voices in seconds, unlocking limitless creative voices for creators, games, education, accessibility, marketing, and more.
The Power of Description
Today, we’re thrilled to unveil Voice Design, our most groundbreaking feature yet. Voice Design represents a fundamental shift in how creators approach voice generation by translating simple text descriptions into fully-realized AI voices in seconds.
Creating the perfect voice has traditionally been a complex process involving multiple steps, technical parameters, and often, compromise. Voice Design changes that paradigm entirely.
With Voice Design, you simply describe the voice you want in natural language—as if you were explaining it to a colleague:- “An old school pirate, speaking with a hearty growl, stretching vowels dramatically”
- “A calm podcast host with a deep, resonant voice and slight southern accent”
- “Energetic TV game show host with rapid delivery and enthusiastic tone”
Our advanced AI then analyzes this description, identifying key voice characteristics, tonal qualities, speech patterns, and stylistic elements to generate a voice that matches your imagination.
Unlimited Creative Possibilities
Voice Design isn’t just a tool—it’s a canvas for voice creativity without boundaries. Every description produces a unique voice, meaning your creative options are virtually limitless. By harnessing the power of natural language descriptions, we’ve transformed voice creation from a technical process to an intuitive, creative act accessible to everyone.
The core innovation of Voice Design lies in its ability to interpret nuanced descriptions and translate them into distinctive vocal characteristics—from subtle inflections to dramatic personality traits. Whether you need a voice that conveys authority, warmth, excitement, or any combination of qualities, it’s now just a description away.Transforming Industries Through Voice Innovation
The applications of Voice Design extend far beyond conventional use cases, revolutionizing workflows across numerous industries:
Content Creators
For animators, podcasters, and video producers, Voice Design revolutionizes the creative process by offering unprecedented flexibility in voice creation. Content creators can now experiment with different narrator styles for documentaries, develop distinct character voices for animations, and establish unique vocal identities for channels or series—all without the scheduling and budget constraints of traditional voice recording. This freedom to iterate quickly means creators can explore creative directions that would have been impractical before, ultimately leading to more distinctive and engaging content that resonates with audiences.
Game Developers
Game development teams face unique challenges when creating immersive worlds populated by dozens or even hundreds of characters. Voice Design transforms this aspect of game creation by enabling developers to rapidly prototype character voices during early development, test how different vocal styles affect emotional impact, and create distinctive voices for non-player characters throughout the game world. This capability not only streamlines the development process but also opens up new possibilities for creating richer, more diverse game environments where each character can have a truly unique vocal identity.
Educational Content
In educational settings, engagement is everything. Voice Design enables educators and e-learning developers to create varied narrator voices across different modules, keeping students engaged through long learning sessions. By tailoring voices to specific subject matter, age groups, or learning contexts, educational content becomes more accessible and compelling. Language learning applications particularly benefit from the ability to generate consistent pronunciation examples, while complex subjects can be explained in voices specifically designed to emphasize clarity and comprehension, ultimately enhancing learning outcomes.
Accessibility Solutions
Voice Design represents a significant advancement in creating truly inclusive audio experiences for diverse audiences. Developers can now design voices optimized for clarity and comprehension for visually impaired users, create voices that speak at adjustable paces for people with cognitive processing differences, and develop region-specific accents to make content more relatable across global markets. This level of customization means that accessibility is no longer about finding the closest acceptable option, but rather creating voices specifically designed to meet the unique needs of different user groups.
Advertising and Marketing
Brand voice has always been a critical but abstract concept. With Voice Design, marketers can literally create voices that embody brand personalities, turning conceptual brand attributes into audible experiences that resonate with audiences. Agencies can quickly prototype different vocal approaches for client presentations, ensure consistency across campaign elements, and develop seasonal variations that maintain brand recognition while conveying different emotional tones. This capability transforms how brands think about their audio identity, making voice a more central and strategic element of the marketing mix.
Entertainment and Performance Arts
The entertainment industry thrives on innovation, and Voice Design offers creators new tools for pushing creative boundaries. Directors and producers can conceptualize character voices before casting decisions, generate atmospheric background voices for immersive scenes, and experiment with unconventional vocal styles that might be difficult to achieve through traditional means. This technology doesn’t replace performers, but rather expands the creative palette available to artists, allowing them to explore new possibilities and refine their vision before moving into production.
Voice Design in Action
Here are some examples showcasing the versatility of Voice Design:
Example 1: Character Voices
Wise elderly wizard with a slightly raspy voice, speaking slowly and deliberately.
Example 2: Narrator Styles
A super deep male voice announcer with a dramatic performance that is unveiling Resemble AI's new Voice Design feature
Example 3: Creative Concepts
Robot assistant with a friendly, slightly mechanical voice that speaks with perfect diction
The Bigger Vision
Beyond these specific features, our long-term vision is to democratize voice creation completely. We believe that voice, like visual design, should be a creative medium accessible to everyone regardless of technical expertise.
We envision Voice Design becoming an essential tool in creative workflows—where generating the perfect voice is as simple as describing what you hear in your imagination. This vision extends to creating a marketplace where voice creators and consumers can collaborate, opening new possibilities for customization and personalization.
As AI voice technology evolves, we remain committed to responsible development—prioritizing transparency, quality, and creative empowerment while building tools that amplify human creativity rather than replace it.
We invite you to join us on this journey, share your feedback, and help shape the future of voice creation together.Try Voice Design
Original source Report a problem - Feb 25, 2025
- Parsed from source:Feb 25, 2025
- Detected by Releasebot:Dec 23, 2025
Rapid Voice Cloning 2.0: New Voice Cloning Model with Unmatched Accuracy
Resemble AI unveils Rapid Voice Clone 2.0, a breakthrough that lets you create and edit voices from just 20 seconds of audio with adjustable tone, gender and speed. Now on Google Cloud Marketplace, boosting accessibility, accuracy and scale for game, media and customer service use cases.
Create your own AI Voice
Today, Resemble AI announced its latest breakthrough model, Rapid Voice Clone 2.0, enabling users to generate high-quality voice content with just 20 seconds of audio. This powerful tool allows seamless voice generation, editing, and localization, empowering users to make instant modifications—whether it’s swapping words, fine-tuning tone, or adjusting delivery—without requiring re-recordings.
From crafting dynamic game characters to personalizing narration and translating speech, Rapid Voice Clone 2.0 adapts effortlessly to any scenario. Users can build voice clones with precise control over attributes like gender, emotion, and speed, ensuring more authentic and engaging experiences for audiences.
A Head-to-Head Comparison
To showcase the accuracy of our model, we conducted a direct comparison between Resemble AI and ElevenLabs, using identical 20-second audio clips.
The results were clear: while ElevenLabs produced a reasonable imitation of the original voice in a different language, it failed to retain key prosodic attributes, an issue common across many generative voice AI models.
In contrast, Resemble AI preserved the unique characteristics of the speaker’s voice, ensuring that the clone sounded natural and true to the original accent.
While both models generate similar phrases, Resemble AI captures the speaker’s true accent, making it significantly more natural and lifelike. In a blind survey, 85% of respondents preferred Resemble AI’s voice clones over competitors.Accent Integrity in Action
Resemble AI’s latest TTS model leads the market in accent preservation, ensuring that cloned voices retain the unique intonations, rhythm, and speech patterns of the original speaker. Unlike other models that flatten accents or introduce inconsistencies, Resemble AI’s technology maintains regional and cultural authenticity, allowing users to produce highly personalized voice experiences without compromise.
This breakthrough is especially critical for industries requiring authentic voice replication, including entertainment, global media, and customer service. Whether it’s capturing the warmth of a Southern drawl, the cadence of a British accent, or the tonality of an Indian speaker, Resemble AI ensures that voices remain faithful to their original speakers.
From Hollywood to gaming studios, leading brands trust Resemble AI to produce high-quality, diverse voices at scale.Red Games – Crayola Adventures
Red Games developed Crayola Adventures, an interactive and educational game that transforms traditional storytelling into a choose-your-own-adventure experience. A core element of this project was fostering player creativity and self-expression. By integrating Resemble AI’s dynamic voice cloning technology, the game enables players to personalize their characters and worlds, making each adventure truly their own.
TrueFan – Bollywood Celebrity Voice Cloning
TrueFan, a platform that connects Bollywood celebrities with fans through personalized video messages, leveraged Resemble AI to scale their production from 500 videos per month to 35,000, all while maintaining the authentic voice characteristics of Bollywood stars.
During a Mother’s Day campaign, Resemble AI powered AI-generated celebrity voiceovers, delivering hyper-personalized video messages that deeply resonated with fans. This innovation enabled TrueFan to scale operations while ensuring a highly personalized and emotional connection between celebrities and their audiences.Resemble AI Joins Google Cloud Marketplace
As of today, Resemble AI is available on Google Cloud Marketplace, making it easier for users to access our Rapid Voice Clone 2.0 solution and explore new possibilities in voice technology. We’re also excited to announce that Resemble AI is now part of the Google Cloud ISV Startup Springboard program.
“Bringing Rapid Voice Clone 2.0 to Google Cloud Marketplace will help customers quickly deploy, manage, and grow Rapid Voice Clone 2.0 on Google Cloud’s trusted, global infrastructure,” said Dai Vu, Managing Director, Marketplace & ISV GTM Programs at Google Cloud. “Resemble AI can now securely scale and support customers on their digital transformation journeys.”
With Google Cloud’s infrastructure, Resemble AI is achieving new levels in accuracy and authenticity of synthetic voices. With the integration of Google Cloud’s Gemini models, our AI-powered voices are becoming smarter, offering more natural conversations and better context understanding. This means faster, higher-quality voice production for narration, interactive experiences, and content localization—cutting production time from weeks to just hours while reducing costs.Ethics and Responsible Use
Looking ahead, we’re committed to refining our technology and expanding its use cases and with our availability on Google Cloud Marketplace we can help more people leverage AI voices. Whether it’s for creating more personalized content, improving the gaming experience, or exploring new industries, Resemble AI remains at the forefront of voice cloning technology.
Original source Report a problem
Resemble AI remains committed to pushing the boundaries of Generative AI while prioritizing AI security. Our dedication to trustworthy AI is evident through advancements like Detect-2B, our state-of-the-art deepfake detection model, and PerTH, our neural watermarking method that embeds imperceptible AI-generated watermarks for enhanced content integrity. As we continue to innovate, we remain steadfast in our mission to deliver authentic, secure, and cutting-edge AI-driven voice technology. - Oct 30, 2024
- Parsed from source:Oct 30, 2024
- Detected by Releasebot:Dec 23, 2025
Introducing State-of-the-Art in Multimodal Deepfake Detection
Resemble AI expands its detection platform to multimodal media, adding image and video analysis to the existing audio system. The unified, enterprise‑grade solution offers scalable, real-time protection with new detection algorithms for modern text-to-image and text-to-video generators.
Today, we present our research on Multimodal Deepfake Detection, expanding our industry-leading deepfake detection platform to support image and video analysis. Our approach builds on our established audio detection system to deliver comprehensive protection across all media types. Detect Multimodal maintains the same enterprise-grade scalability and real-time processing capabilities that our customers rely on, while introducing sophisticated new detection algorithms for images, and video.
Breaking New Ground in Multimodal Detection
The rapid advancement of generative AI technologies, particularly in text-to-image and text-to-video models, has created an urgent need for robust detection methods. We present a unified approach to synthetic media detection that addresses the challenges posed by modern AI generators such as DALL-E 3, Midjourney, FLUX, SORA, MovieGen, and others.
Resemble AI created a testing dataset that represents a comprehensive benchmark for evaluating synthetic image detection capabilities across modern AI generation methods. We call this “Modern Dataset” in table 1. Unlike many existing benchmarks that focus primarily on GAN-generated images or older models, our dataset incorporates content from the latest text-to-image models including DALL-E 3, Midjourney, and FLUX, ensuring relevance to current real-world challenges in synthetic content detection.
The dataset is meticulously curated to maintain high quality and balanced representation across different generation methods. We employ a careful sampling strategy, taking up to 100 samples from each source to prevent any single generation method from dominating the evaluation metrics. Real images are sourced from established datasets including the Kaggle DALL-E recognition dataset and ArtiFact dataset, providing a robust foundation for testing false positive rates. Importantly, we maintain strict separation between our training and testing data – while our training set utilizes the complete collections of these sources, our test set uses distinct samples that never appear in training.
A key distinguishing feature of our dataset is its focus on high-resolution, high-quality images that reflect real-world usage. We specifically excluded lower-resolution datasets (such as CIFAKE with its 32×32 images) to ensure our testing represents practical deployment scenarios.
Table 1. Modern Dataset incorporates content from the latest text-to-image models including DALL-E 3, Midjourney, and FLUX
Detecting Modern AI Text-to-Image Generators
Our image detection system leverages state-of-the-art neural networks to identify AI-generated content from all major image generation platforms, including DALL-E 3, Midjourney, Google’s Gemini, Grok, FLUX, and Stable Diffusion.
Capturing Nuances and Detecting Text-to-Video AI Models
The emergence of sophisticated text-to-video AI models like OpenAI’s SORA and Meta’s MovieGen has ushered in a new era of synthetic video content. In response, we’ve developed a comprehensive video analysis system that operates at multiple levels of sophistication. Our technology processes video content frame-by-frame, analyzing both temporal consistency and spatial artifacts to identify AI-generated footage with unprecedented accuracy.
What sets our video analysis apart is its ability to adapt to emerging generation techniques. As new text-to-video models emerge, our system’s architecture allows for rapid adaptation and continued accuracy in detection. This adaptability is crucial in maintaining effective protection against increasingly sophisticated video generation technologies.
Built for Enterprise Scale
Resemble Detect’s multimodal platform has been architected from the ground up to meet the demanding requirements of enterprise deployments. At its core, our system features an API-first design philosophy that enables seamless integration into existing enterprise workflows, whether you’re processing content through custom applications, content management systems, or security platforms. This flexibility allows organizations to implement detection capabilities exactly where they’re needed, without disrupting established processes.
Our scalable architecture adapts dynamically to processing demands, handling everything from individual file analysis to high-volume batch processing across all content types. The system’s distributed processing capabilities ensure consistent performance even under heavy loads, making it suitable for organizations dealing with massive content volumes across audio, image, and video formats. Whether you’re analyzing thousands of customer service calls, monitoring social media feeds, or validating assets in content management system, our platform maintains its speed and accuracy without compromise.
Security-conscious organizations will appreciate our comprehensive deployment options, including fully air-gapped installations for environments with the strictest security requirements. This flexibility extends to hybrid deployments that can balance security needs with operational efficiency. Our on-premise solutions provide the same advanced detection capabilities as our cloud-based offerings, ensuring that organizations never have to choose between data security and functionality.
Original source Report a problem - Sep 26, 2024
- Parsed from source:Sep 26, 2024
- Detected by Releasebot:Dec 23, 2025
Resemble AI Partners with Carahsoft to Bring Voice AI and Deepfake Detection to the Public Sector
Resemble AI teams with Carahsoft to bring real-time voice AI and deepfake detection to government agencies. The Detect suite—deepfake detection, watermarking, identity verification, and audio intelligence—appears available via Carahsoft contracts for rapid, secure deployment.
We are thrilled to announce our strategic partnership with Carahsoft Technology Corp., The Trusted Government IT Solutions Provider®. This collaboration marks a significant milestone in our mission to bring cutting-edge voice AI technology and advanced deepfake detection solutions to the public sector.
Empowering Government Agencies with Voice AI
As the landscape of digital communication evolves, so do the needs of government agencies. Our partnership with Carahsoft will enable public sector organizations to harness the power of our industry-leading real-time generative voice technology. This means agencies can now create realistic, ethical, and personalized voice content at scale, opening up new possibilities for streamlined operations and enhanced communication.
Comprehensive Protection with Our Detect Suite
At the core of this partnership is our commitment to security and content integrity. We’re excited to offer our complete Detect suite to government agencies through Carahsoft’s extensive network:
- Deepfake Detection: Our state-of-the-art AI models can identify manipulated audio content with high accuracy, helping agencies combat the growing threat of audio deepfakes.
- Neural Speech Watermarking: We offer invisible, inaudible watermarking for audio files, ensuring the authenticity and traceability of official communications.
- Identity Verification: Our voice biometrics technology allows for secure speaker identification, adding an extra layer of security to sensitive communications.
- Audio Intelligence: Going beyond mere transcription, our audio intelligence AI model provide deep insights into speech patterns, emotions, and context, enhancing decision-making processes.
Rapid Deployment and Secure Implementation
Understanding the unique needs of government agencies, we’ve designed our solutions to be quickly deployable, even in air-gapped or offline environments. This ensures that agencies can benefit from our technology while maintaining the highest levels of security and reliability.
A Shared Commitment to Ethics and Responsibility
From our inception, Resemble AI has placed ethics at the forefront of our development process. This aligns perfectly with the public sector’s need for responsible AI solutions. Our partnership with Carahsoft reinforces this commitment, ensuring that government agencies can leverage the power of voice AI and deepfake detection in a manner that is both innovative and ethically sound.
Availability and Support
Our solutions are now available through Carahsoft’s SEWP V, ITES-SW2, and OMNIA Partners contracts. This makes it easier than ever for public sector organizations to access and implement our technology.
For more information or to explore how Resemble AI’s solutions can benefit your agency, please contact the Resemble AI team at Carahsoft at (703) 921-4160 or [email protected].
We’re excited about the possibilities this partnership brings and look forward to working closely with Carahsoft and government agencies to create a safer, more efficient digital communication landscape for the public sector.
Original source Report a problem
This is the end. You've seen all the release notes in this feed!