by Yu Qin and Yi Yang, ACL 2019.
Predicting financial risk is an essential task in financial market. Prior research has shown that textual information in a firm's financial statement can be used to predict its stock's risk level. Nowadays, firm CEOs communicate information not only verbally through press releases and financial reports, but also nonverbally through investor meetings and earnings conference calls. There are anecdotal evidences that CEO's vocal features, such as emotions and voice tones, can reveal the firm's performance. However, how vocal features can be used to predict risk levels, and to what extent, is still unknown. To fill the gap, we obtain earnings call audio recordings and textual transcripts for S&P 500 companies in recent years. We propose a multimodal deep regression model (MDRM) that jointly model CEO's verbal (from text) and vocal (from audio) information in a conference call. Empirical results show that our model that jointly considers verbal and vocal features achieves significant and substantial prediction error reduction. We also discuss several interesting findings and the implications to financial markets. The processed earnings conference calls data (text and audio) are released for readers who are interested in reproducing the results or designing trading strategy.
This is the segmented earnings conference call dataset of S&P 500 companies in 2017.
Since the dataset includes text transcripts and corresponding audio recordings, it is too large to put on GitHub. We put a few examples in this repository, and upload our complete dataset on Google Drive, please check the link to download the complete dataset.
The whole dataset is split to five files, please download all the files and unzip them.
To unzip the dataset under Linux environment, please first use zip -s0 ACL19_Release.zip --out ACL19_Release_All.zip
to merge the files.
Then use unzip -q ACL19_Release_all.zip
to unzip the dataset.
If you use this data in your research, please cite the following paper:
Qin, Yu., & Yang, Yi. (2019). What You Say and How You Say It Matters: Predicting Stock Volatility Using Verbal and Vocal Cues. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics.
@inproceedings{qin-yang-2019-say,
title = "What You Say and How You Say It Matters: Predicting Stock Volatility Using Verbal and Vocal Cues",
author = "Qin, Yu and
Yang, Yi",
booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P19-1038",
pages = "390--401",}
Each folder in our dataset represents an earnings conference call; the folders are named as "CompanyName_Date".
In each folder, we put the processed text transcript and segmented audio recordings. As we described in our paper, we only select the most spoken executive (usually the CEO) to avoid interference.
In the processed text transcript, each line is a sentence of CEO, ordered by time.
For the segmented audio recordings, they are named as "Speaker_Paragraph_Sentence". Since in our Iterative Forced Alignment (IFA) processing, the whole audio recording is firstly segmented on paragraph level, then the sentence level. The order of segmented audio recordings as from "Speaker_1_1", "Speaker_1_2", "Speaker_1_m1" to "Speaker_n_mn", where mx represents the number of sentences of paragraph x, and n represents the number of paragraphs.
Each line in the text file corresponds to an audio file. And the total number of lines in a text file is the same as the total number of audio files for the same earnings conference call.
If you have any questions regarding the repo, please contact Yu Qin ([email protected]) or create an issue.