Welcome to the Off-Shore Club

The #1 Social Engineering Project in the world since 2004 !

Important Notice:

✅UPGRADE YOUR ACCOUNT TODAY TO ACCESS ALL OFF-SHORE FORUMS✅

[New]Telegram Channel

In case our domain name changes, we advise you to subscribe to our new TG channel to always be aware of all events and updates -
https://t.me/rtmsechannel

OFF-SHORE Staff Announcement: 30% Bonus on ALL Wallet Deposit this week


For example, if you deposit $1000, your RTM Advertising Balance will be $1300 that can be used to purchase eligible products and service on forums or request withdrawal. The limit deposit to get the 30% bonus is $10,000 for a $3000 Marketplace wallet balance Bonus.

Deposit Now and claim 30% more balance ! - BTC/LTC/XMR


Always use a Mixer to keep Maximum anonimity ! - BTC to BTC or BTC to XMR

🗂️Keep in Mind Microsoft Won't Let You Use Its New AI Voice Tool

⚠️Always Remember to keep your identity safe by using a Zero-KYC Zero-AML like https://coinshift.money⚠️

Gold

_=*Croft*=_

Business Club
💰 Business Club
USDT(TRC-20)
$0.0
It's no secret that AI is getting pretty darn realistic: Companies like OpenAI are making tools that can replicate images, audio, and videos in ways that are becoming increasingly more difficult to identify as such on the fly. But while it's bad enough that some of these programs are available to the public already, it's concerning to hear about a tool that's so good, it's being kept from the rest of us.

Vall-E 2 can steal your voice​


As reported by TechSpot, Microsoft has created a new version of its "neural codec language model," Vall-E, appropriately now called Vall-E 2. Microsoft detailed Vall-E 2's advances in a blog post, highlighting some key milestones with this latest model. Chiefly, Vall-E 2 achieves "human parity," which seems to be a fancy way of saying, "Our model's outputs sound like real humans." Be afraid.

Vall-E 2 apparently achieves two key enhancements over Vall-E: The new model doesn't have an "infinite loop" issue the original had when processing repeating tokens. The new model accounts for repeating tokens, and thus is able to decode a sample that contains them. In addition, Vall-E 2 shortens the length of a given sequence by grouping codec codes, which Microsoft says both increases interference speed, and skips over issues that arise from modeling long sequences.

If that's all a bit technical, perhaps this won't be: Vall-E 2 improves upon Vall-E in "speech robustness, naturalness, and speaker similarity," and, according to Microsoft, is the first of its class to achieve human parity in these categories. In fact, the company says, "VALL-E 2 can generate accurate, natural speech in the exact voice of the original speaker, comparable to human performance."

It's not just theory​


You don't just have to read about Vall-E 2 to believe how good it is: Microsoft offers examples of how Vall-E 2 can take a sample recording of a voice, and replicate it when prompted with new text. The company also provided examples of the model completing a sentence after being given segments of a sample recording, in three, five, and 10-second chunks. This demonstrates the model's ability to take a very short example of a voice, and replicate it with text that doesn't appear in the original sample recording.

There are still plenty of the quirks you'd expect to find with any text-to-speech model (incorrect pronunciations, stuttered speech, etc.) but there's no doubt that the Vall-E 2 examples are not only often realistic, but match the voice of the original sample quite closely. It especially does well when given a longer recording of a voice: If given three seconds of a recording, the output is still impressive, but when given a five or, especially, a 10-second recording, the output can be remarkably realistic.

If you click through the examples yourself, check out how well Vall-E 2 matches the 10-second recording when reciting "My life has changed a lot" under "VCTK Samples." I don't have any experience with training AI systems, but to my ear, the model nails the raspy voice of the speaker in the sample, especially after receiving the full 10-second clip. It's jarring to hear the original speaker reading a certain sentence, then hear the model speak a new sentence in a voice that essentially matches the speaker's.

Vall-E 2's risks​


But if you're a bit freaked out by this whole thing, you aren't alone. Microsoft is aware its model could be dangerous if used maliciously: In an ethics statement at the bottom of the post, the company acknowledges that, while Vall-E 2 could be used for a variety of positive tasks, it could also be used to impersonate a specific person. Microsoft says the model is meant to be used with consenting users who understand their voice is being replicated, and that the model should have a protocol to check for consent before processing a request. That said, it doesn't seem like such a protocol actually exists right now, which is likely why Microsoft current has, "no plans to incorporate VALL-E 2 into a product or expand access to the public."

The examples here are based on voice samples the LibriSpeech and VCTK datasets, not from samples Microsoft recorded themselves. As such, as a outside observer, it isn't clear how this model would actually perform if given recordings of, say, President Biden, Elon Musk, or your boss. However, if we assume that Vall-E 2 can generate a realistic output when given a 10-second sample, imagine how realistic its output could be when fed with hours of samples. Couple that with a solid AI video model, and you have the perfect storm for generating misinformation, just in time for election seasons across the globe.
Full story here:
 

Create an account or login to comment

You must be a member in order to leave a comment

Create account

Create an account on our community. It's easy!

Log in

Already have an account? Log in here.

Friendly Disclaimer We do not host or store any files on our website except thread messages, most likely your DMCA content is being hosted on a third-party website and you need to contact them. Representatives of this site ("service") are not responsible for any content created by users and for accounts. The materials presented express only the opinions of their authors.
🚨 Do not get Ripped Off ! ⚖️ Deal with approved sellers or use RTM Escrow on Telegram
Gold
Mitalk.lat official Off Shore Club Chat


Gold

Panel Title #1

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.

Panel Title #2

Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat.
Top