Comments (7)
For wiki given by author, it is single_label, so what I got is micro=sample=acc . Or do you have a more complete data for wiki?
from graphembedding.
For wiki given by author, it is single_label, so what I got is micro=sample=acc . Or do you have a more complete data for wiki?
here is the document of parameter average
of sklean.metirc.f1_score:
average : string, [None, ‘binary’ (default), ‘micro’, ‘macro’, ‘samples’, ‘weighted’]
This parameter is required for multiclass/multilabel targets.
- 'micro':
Calculate metrics globally by counting the total true positives, false negatives and false positives.- 'macro':
Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.- 'weighted':
Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters ‘macro’ to account for label imbalance; it can result in an F-score that is not between precision and recall.- 'samples':
Calculate metrics for each instance, and find their average (only meaningful for multilabel classification where this differs from accuracy_score).
So, I think it will get different results in a multiclass case.
from graphembedding.
@dawnranger That's good,I think you can open a pull request about the results on datasets and the codes to reproduce the results in a new folder.
from graphembedding.
@dawnranger
'samples':
Calculate metrics for each instance, and find their average (only meaningful for multilabel classification where this differs from accuracy_score).
Wiki is multiclass rather than multilabel, isn‘t it? Why there is a difference between sample and acc?
In addition, for flight data in your result, micro=sample=acc.
from graphembedding.
@dawnranger
'samples':
Calculate metrics for each instance, and find their average (only meaningful for multilabel classification where this differs from accuracy_score).
Wiki is multiclass rather than multilabel, isn‘t it? Why there is a difference between sample and acc?
In addition, for flight data in your result, micro=sample=acc.
I think you are right. I use shenweichen's code :
averages = ["micro", "macro", "samples", "weighted"]
results = {}
for average in averages:
results[average] = f1_score(Y, Y_, average=average)
results['acc'] = accuracy_score(Y,Y_)
and I got a warning with wiki dataset:
python3/lib/python3.6/site-packages/sklearn/metrics/classification.py:1135: UndefinedMetricWarning: F-score is ill-defined and being set to 0.0 in labels with no predicted samples.
As discussed in stackoverflow, the ill spliting of the train/test set might be blamed for this issue.
from graphembedding.
@dawnranger
Yes. I found that the classify.py is similar to scoring.py in deepwalk which is provided by writer https://github.com/phanein/deepwalk/blob/master/example_graphs/scoring.py
what I was confused is author did not provide the result and the origin of wiki.
In addition, I tried data BlogCatalog(multi-lable) as the node2vec paper mentioned, and I set parameter as the paper did(d=128, r=10, l=80, k=10. training percent=50%, p=q=0.25), but I got a 0.12(MacroF1), far from the result which author provided(0.2581).
So depressed...
from graphembedding.
hello, from these results, the accuracy does not seem to be high, what is the cause, is it a data problem?
from graphembedding.
Related Issues (20)
- Documentation: Node Embedding in relations to Role Discovery
- ReFeX (recursive features)
- ImportError: libcublas.so.9.0: cannot open shared object file: No such file or directory HOT 1
- __init__() got an unexpected keyword argument 'size' HOT 5
- 如何使用GPU加速 HOT 1
- 请问能否提供数据来源,想了解数据的背景和含义?
- 运行deepwalk_wiki.py时出现DLL load failed
- line loss计算问题 HOT 3
- 能否提供可执行环境要求,总是报错 HOT 2
- SDNE,'NoneType' object is not callable
- TypeError: __init__() got an unexpected keyword argument 'iter'
- 在创建图G时发生读取文件错误
- node embedding methods to apply to multi-graph problems HOT 1
- line训练时报错
- fatal error C1083 HOT 1
- 'charmap' codec can't decode HOT 1
- 请问如何用gpu进行训练呢?
- 【小问题】struc2vec超过16线程跑不了
- 【Question】NotImplementedError: run SDNE in Python3.8, tf2.11.0 HOT 1
- Wrong:SDNE
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from graphembedding.