Coder Social home page Coder Social logo

cone-of-silence's People

Contributors

mjenrungrot avatar vivjay30 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

cone-of-silence's Issues

dataset used for COS

Hey, I loved your work, I was trying to replicate it, to do that I was generating some synthetic dataset but got some errors and doubts.
As you mentioned the dataset used is VCTK, it has dataset in .flac format which is not recognized by the program, so did you guys did any preprocessing over the dataset?
And there is no data folder in the original dataset i,e. VCTK (mentioned in the command to generate synthetic dataset).
And can you share the dataset you guys used for training?

No directory 'mir_eval'

Hi, I'm trying to recreate your results, when I run the code I get error message saying there's no such module 'mir_eval', and I wasn't able to find it. perhaps you moved the files somewhere else or renamed the directory? the reference is from 'cos/helpers/eval_utils.py'

from mir_eval.separation import bss_eval_sources

Thanks

[bug] Duplicate un-normalize in train.py

Hi, vivjay

# Un-normalize
output_signal = output_signal * stds.unsqueeze(
3) + means.unsqueeze(3)
# Un-normalize
output_signal = unnormalize_input(output_signal, means, stds)
output_voices = output_signal[:, 0]
loss = model.loss(output_voices, label_voice_signals)

line 117 and line 121 are the same.
Un-normalize twice will make the test loss (validation) do not decrease.
much like the problem @zhangshengoo asked this #12 (comment)

Sample 4-channel audio files?

Hi, in your README, you mention that "We even provide a sample 4 channel file for you to run". Where is this file located?

Real Data

Hi,

Your paper mentions you recorded 3 hours of data into the Seeed 4 mic hat from VCTK corpus.

Would you be willing to make that available?
We'd like to duplicate the results, and we don't get the same level from training on just the synthetic data.

Thanks!!
Richard

Real-time

Hi. I do realize that you have already answered the question that whether the algorithm is real-time or not in #9 . I just want to know whether there is any way that we can do to make it operational in real-time?

Two-channel audio recordings?

From what I understand, the model will work for 4-channel and 6-channel wav files.

Does this model also work on 2-channel recordings? Is there a pretrained model for that?

Mean and STD of the signal peak

Hi Vivek,

Thanks for your awesome work.
What's the meaning for FG_VOL_MIN、FG_VOL_MAX、BG_VOL_MIN、BG_VOL_MAX in generate_dataset.py and how did you calculate these four values?

Best regards,
KenHuang

real-time?

Hi,

thanks for this library.
Is it possible to use this in a real-time scenario like conferences?

Best regards,
Dirk

The effect of model size on the overall performance

Hi Vivek,

Thanks for opensourcing this interesting project - nice work! I just have one question about the model size and the performance. I checked your Demucs model implementation and calculated the number of parameters, and with your default hyperparameter setting there are over 260M parameters. I'm not sure if this is the actual setting you used for training, but if so, this is really a huge amount of parameters as in other separation models the model sizes are typically smaller than 10M nowadays. I'm wondering whether you have done any experiments on how the performance will be if you shrink the model size, e.g. to the level of multi-channel Conv-TasNet or TAC reported in your paper. Thanks!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.