kunzhou9646 / emovox Goto Github PK
View Code? Open in Web Editor NEWThis is the implementation of the paper "Emotion Intensity and its Control for Emotional Voice Conversion".
This is the implementation of the paper "Emotion Intensity and its Control for Emotional Voice Conversion".
Thank you for sharing your code.
In "train_ser.py", line 199, '../transformer_ser/data_500_final.pkl' is loaded, but this file is not found. How do I make this file?
Hi, Mr Zhou,
I have read the paper, the idea of emotion density control is very attractive.
I am not Enligsh native, but I feel the samples of proposed in https://kunzhou9646.github.io/Emovox_demo/ are not so good, they are even less intelligable or natural compared with conterpart samples.
Is the hurt on speech quality a necessary cost when learning to be more emotional?
Hey! Thank you for sharing your work. I was wondering whether you could provide pretrained weights of the final model to check the quality of the conversion on our own samples?
I can not find the relevant python script, could you please help me? Thanks!
Hi Kun, I have read this paper and tried to train this network, but I meet some questions as follows:
Look forward your kind reply! Thank you:)
data_500_final.pkl 您好可以提供一下这个文件吗~感谢感谢!
Hello, Kun! Thanks a lot for the codes you provide. However, I have met some problems when I run the codes. When I used pre-processing.py to extract open-smile features, I couldn't find the files named IS09_emotion.conf and SMILExtract. I want to ask how to get these two files. Hope that you could help me, thanks!
Hi thanks for the code! I was struggling with the relative attributes training part. As is mentioned in #4 (comment), the path is set in the code and it seems nowhere I can find it in this repo.
I'm stuck at here mainly because I want to train your model on a different speaker. I would be better if you can share the csv files strengths.txt
related to the Mixed Emotion repo on another speaker (especially 0015). Thank you!
Hello, Thank you for sharing your work!
I have a question in the "step1. Learning relative attributes" part, pre-processing.py
.
In the process of extracting the acoustic features with openSMILE feature, which file is used in the following path?
opensmile_path = '/Users/kun/Desktop/workspace/open_smile/opensmile/build/progsrc/smilextract/SMILExtract'
config_path = '/Users/kun/Desktop/workspace/open_smile/opensmile/config/is09-13/IS09_emotion.conf'
I couldn't find a specific way to specify the path in the paper.
请问怎么样才可以下到opensmile提取音频特征的文件呢?
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.