Deeply Korean Read Speech Corpus - Audio AI & ML Training Data product image in hero

Deeply Korean Read Speech Corpus - Audio AI & ML Training Data

Deeply
No reviews yetBadge iconVerified Data Provider
#
xxxxxxxxxx
Xxxxxxxxx
xxxxxx
xxxxxxxxxx
Xxxxx
Xxxxxx
Xxxxxxxxxx
Xxxxxx
1 xxxxxxxxxx Xxxxxxxxx xxxxxx xxxxxxxxxx Xxxxx Xxxxxx Xxxxxxxxxx Xxxxxx
2 Xxxxxxxxx Xxxxxxxxxx xxxxxxxxx Xxxxxxxxx xxxxxxxxx Xxxxxxx xxxxxx Xxxxx
3 xxxxxxxxxx xxxxxx Xxxxxxxxxx xxxxxx Xxxxx Xxxxxx xxxxx xxxxxxxx
4 xxxxxxx Xxxxx Xxxxxxxx xxxxxxxxxx xxxxxx Xxxxxxxxx xxxxxx Xxxxxxxxx
5 Xxxxxxxxx xxxxxxxxxx Xxxxxx Xxxxx xxxxxx xxxxxxx xxxxxxx Xxxxx
6 xxxxxx Xxxxxxxxxx xxxxxxxx xxxxxx Xxxxx Xxxxxxx xxxxxx Xxxxxxxx
7 Xxxxxxx Xxxxx xxxxxx xxxxxxxxxx Xxxxx xxxxxxxxxx xxxxxxxxx Xxxxxxx
8 xxxxxxxx xxxxxxxx Xxxxxxxxxx Xxxxxxxx Xxxxxxxx xxxxxxxxx Xxxxxxxxxx Xxxxxx
9 Xxxxxxxxx xxxxx xxxxxxx xxxxxxxxx Xxxxxx Xxxxxxx Xxxxxxxxx xxxxxxxxx
10 xxxxxxxxx Xxxxx xxxxxxxx Xxxxxxx xxxxxxxxx Xxxxxxx xxxxx Xxxxxxx
... xxxxxxx Xxxxx xxxxxxxxxx Xxxxxxx Xxxxx xxxxxxxxxx Xxxxxx xxxxxx
Request Data Sample
Volume
190K
records
Data Quality
99%
Validity
Coverage
1
Country

Description

Pairs of Korean speakers reading a script with 3 distinct text sentiments, with 3 distinct voice sentiments, are recorded. The recordings took place in 3 different places, of which the level of reverberation differs. Every experiment is recorded at 3 distinct distances with 2 types of smartphones.
□ Recording contents A pair of adults reading scripts containing 3 distinct text sentiments(negative, neutral, positive) with 3 distinct voice sentiments(negative, neutral, positive). (Script: movie reviews(positive, negative), everyday conversation(neutral) □ Recording environments Anechoic Chamber (no reverb), Studio apartment (moderate reverb), Dance studio (high reverb) □ Device iPhone X (iOS), Samsung Galaxy S7 (Android) □ Distance from the source 0.4m, 2.0m, 4.0m □ Volume ~ 290 hours, ~ 190,000 utterances, ~ 107 GB □ Format wav(44100Hz, 16-bit, mono), or h5(16000Hz, 16-bit, mono) □ Language Korean □ Demographics 34 Korean adults, with 26% males and 74% females, and 47% are in 20s, 20.5% in 30s, 17.5% in 40s, 6% in 50s, 9% in 60s. The Read Speech dataset consists of 289.9 hours of audio clips of reading the scripts with 3 text sentiments with 3 voice sentiments recorded at 3 distinct places using 2 different smartphones running under different operating systems. The participants are encouraged to record repetitively in all 3 types of place (anechoic chamber, studio apartment, dance studio), and every recording is conducted systematically at 3 ordinal distances(0.4m, 2.0m, 4.0m) with 2 types of device(iPhone X and Galaxy S7). The type of text sentiments and voice sentiments is categorized as follows:
 ‘Negative text sentiment’, ‘neutral text sentiment’, ‘positive text sentiment’ indicates that the contents being vocalized are negative, neutral, and positive respectively. Specifically, for the negative and positive text sentiments, negative/positive movie reviews, containing degradations, criticisms or compliments, were used. And, for the neutral text sentiment, everyday conversations without typical emotions were used. ‘Negative voice sentiment’ indicates that the speaker vocalized the script with a negative tone of voice, for the sake of consistency, we instructed the speakers to vocalize as if they were angry. ‘Neutral voice sentiment’ indicates that the speaker vocalized the script with a neutral tone of voice, with any emotions involved. Finally, ‘positive voice sentiment’ indicates that the speakers vocalized the script with a positive tone of voice, especially as if they were happy. Each type of voice sentiment was vocalized regardless of the content of the script (text sentiment), for example, the speakers were also asked to vocalize the script positively even though the content was negative. The dataset also includes metadata such as a script(speech-to-text aligned), speaker, age, sex, noise, type of place, distance, and device. The impulse responses of each type of place are available upon request.

Country Coverage

Asia (1)
Korea (Republic of)

Volume

190,000 records
290 Hours of audio
107 GB

Pricing

Free sample available
Deeply has not published pricing information for this product yet. You can request detailed pricing information below.

Suitable Company Sizes

Small Business
Medium-sized Business
Enterprise

Quality

Self-reported by the provider
99%
Validity

Use Cases

Artificial Intelligence (AI)
Machine Learning (ML)
Sentiment Analysis
Automatic Speech Recognition
Room Acoustics

Categories

Related Products

10K images
9 countries covered
10 years of historical data
Collection of 10,000+ images of traffic scene from low view that are ready to use for optimizing the accuracy of computer vision models.
1B Monthly records
USA covered
Website visit data with URLs, categories, timestamps, and anonymized unique identifiers.
600 Hours of Recording
64 countries covered
We offer a comprehensive collection of audio data, amounting to over 600 hours of high-quality recordings. Our audio datasets are meticulously curated and de...
5B records
98% accuracy
USA covered
CrawlBee ML datasets are specially curated and cleansed to provide the highest quality training data for those looking to provide real-world answers to big p...

Frequently asked questions

What is Deeply Korean Read Speech Corpus - Audio AI & ML Training Data?

Pairs of Korean speakers reading a script with 3 distinct text sentiments, with 3 distinct voice sentiments, are recorded. The recordings took place in 3 different places, of which the level of reverberation differs. Every experiment is recorded at 3 distinct distances with 2 types of smartphones.

What is Deeply Korean Read Speech Corpus - Audio AI & ML Training Data used for?

This product has 5 key use cases. Deeply recommends using the data for Artificial Intelligence (AI), Machine Learning (ML), Sentiment Analysis, Automatic Speech Recognition, and Room Acoustics. Global businesses and organizations buy AI Training Data from Deeply to fuel their analytics and enrichment.

Who can use Deeply Korean Read Speech Corpus - Audio AI & ML Training Data?

This product is best suited if you’re a Small Business, Medium-sized Business, or Enterprise looking for AI Training Data. Get in touch with Deeply to see what their data can do for your business and find out which integrations they provide.

Which countries does Deeply Korean Read Speech Corpus - Audio AI & ML Training Data cover?

This product includes data covering 1 country like South Korea. Deeply is headquartered in Korea (Republic of).

How much does Deeply Korean Read Speech Corpus - Audio AI & ML Training Data cost?

Pricing information for Deeply Korean Read Speech Corpus - Audio AI & ML Training Data is available by getting in contact with Deeply. Connect with Deeply to get a quote and arrange custom pricing models based on your data requirements.

What is the data quality of Deeply Korean Read Speech Corpus - Audio AI & ML Training Data?

Deeply has reported that this product has the following quality and accuracy assurances: 99% Validity. You can compare and assess the data quality of Deeply using Datarade’s data marketplace.

What are similar products to Deeply Korean Read Speech Corpus - Audio AI & ML Training Data?

This product has 3 related products. These alternatives include Pixta AI Imagery Data Global 10,000 Stock Images Annotation and Labelling Services Provided Traffic scenes from low view for AI & ML, BIGDBM Website Visits Data With Industry/Context Categorization - Training Set for ML and AI, and WebAutomation Off the Shelf Datasets Audio Data for AI & ML Training 600+ Hours of Recording Speech Recognition, Natural Language Processing. You can compare the best AI Training Data providers and products via Datarade’s data marketplace and get the right data for your use case.

Pricing available upon request