Comments (13)
wherehows.db.username
is the connection info that used to connect to wherehows mysql database. It's in the wh_etl_job
table. https://github.com/linkedin/WhereHows/wiki/Set-Up-New-Metadata-ETL-Jobs#2-fill-in-all-the-configurations-needed-for-the-job-type
You can also refer to set up in the VM.
Just found we lack of an API to add that configuration. Will add that later.
from datahub.
It seems the correct table is wh_property and not wh_etl_job.
from datahub.
Oh yes, let me correct that
from datahub.
I am getting this same error but when I do a select * from wh_property,
wherehows.db.username wherehows N NULL
I have that property already inserted into my DB. Just not sure why it would be still having that error. Is it because I did a manual insert statements?
from datahub.
Manual insert should be fine, that's what I did here. Do you also see the error of 'KeyError: wherehows.db.username' ?
from datahub.
I am getting this: The file below is iostream its blank in the comment.
`
2016-04-06 15:21:54 ERROR Job Launcher:83 - Traceback (most recent call last):
File "", line 340, in
KeyError: wherehows.db.username
at org.python.core.Py.KeyError(Py.java:249)
at org.python.core.PyObject.__getitem__(PyObject.java:738)
at org.python.pycode._pyx1.f$0(<iostream>:356)
at org.python.pycode._pyx1.call_function(<iostream>)
at org.python.core.PyTableCode.call(PyTableCode.java:167)
at org.python.core.PyCode.call(PyCode.java:18)
at org.python.core.Py.runCode(Py.java:1386)
at org.python.util.PythonInterpreter.execfile(PythonInterpreter.java:296)
at org.python.util.PythonInterpreter.execfile(PythonInterpreter.java:291)
at metadata.etl.dataset.hdfs.HdfsMetadataEtl.load(HdfsMetadataEtl.java:160)
at metadata.etl.EtlJob.run(EtlJob.java:181)
at metadata.etl.Launcher.main(Launcher.java:77)`
from datahub.
Has there been any traction on this issue?
from datahub.
Follow the steps here : https://github.com/linkedin/WhereHows/wiki/Set-Up-New-Metadata-ETL-Jobs#2-fill-in-all-the-configurations-needed-for-the-job-type
Insert the wherehows.db.username (the mysql connection username) into wh_property can solve this problem.
from datahub.
As I mentioned in the above comment I have already inserted those parameters yet still receiving that error. jdbc.url, password, username, etc all filled in.
from datahub.
Looks like your error happened when java pass the parameters to jython.
Are you running on your local machine or in the VM we provided?
Are you executing it as a single job testing mode or executing it as production mode by start the play service?
Could you do these :
1 at this line : https://github.com/linkedin/WhereHows/blob/master/metadata-etl/src/main/resources/jython/HdfsLoad.py#L335
add print args
2 at this line : https://github.com/linkedin/WhereHows/blob/master/metadata-etl/src/main/java/metadata/etl/dataset/hdfs/HdfsMetadataEtl.java#L170
add System.out.println(interpreter.getSystemState().argv);
This will print out the parameter before and after call jython interpreter. So we can check whether the parameter have passed in jython successfully.
from datahub.
So pulled the latest code, added those lines and deployed. When I run the test mode job something different is happening:
[root@ip-206-40-95 backend-service-1.0-SNAPSHOT]# java -Djob=HADOOP_DATASET_METADATA_ETL -Dconfig=sample.properties -cp "lib/*" metadata.etl.Launcher
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/space/wherehows/backend-service-1.0-SNAPSHOT/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/space/wherehows/backend-service-1.0-SNAPSHOT/lib/ch.qos.logback.logback-classic-1.0.13.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
This is what my config looks like:
# HDFS properties
hdfs.cluster=cluster1
hdfs.private_key_location=/home/jmblnvr/.ssh/id_rsa
hdfs.remote.raw_metadata=.wherehows/hdfs_metadata_file
hdfs.remote.sample=.wherehows/hdfs_sample.dat
hdfs.remote.user=jmblnvr
hdfs.remote.machine=localhost
hdfs.remote.jar=.wherehows/schemaFetch.jar
hdfs.local.raw_metadata=/tmp/hdfs_metadata_file
hdfs.local.metadata=/tmp/hdfs_metadata.dat
hdfs.local.field_metadata=/tmp/hdfs_field_metadata
hdfs.local.sample=/tmp/hdfs_sample.dat
hdfs.white_list=/user,/src
hdfs.num_of_thread=2
hdfs.file_path_regex_source_map=[{".*":"Hdfs"}]
Last time it executed but this time it just shows those log4j things but nothing else.
from datahub.
This looks like a slf4j binding problem. It may caused by classpath configuration. Will check it later.
from datahub.
The output is in the /var/tmp/wherehows/wherehows.log file .
The log4j output was changed in this commit : a0b7cb9
You can add 'stdout' back so it will show up in the console.
from datahub.
Related Issues (20)
- Error prefixed fuzzy retrieval of the value of key in customProperties (including slashes)
- Lineage cannot be added to datajobs that are not configured with the datasets attribute.
- Glossary Term's Related Entities (Column) would show incorrect imformation in UI when it inherited by others HOT 2
- Not all Tags show up in the Global Search and Filter Page HOT 7
- Metadata Ingestion: 'Advanced -> Add extra pip libraries for an ingestion execution' are not installed in scheduled runs HOT 1
- sdfsdf
- No assets ingested during Business Glossary Ingestion HOT 1
- Sigma stateful ingestion is broken HOT 1
- [spark-lineage-beta] openlineage is too old to support spark 3.3 HOT 2
- Description for column is visible only in the sidebar
- description part of the display part cannot be displayed
- Starburst/Trino integration not working properly HOT 2
- CLI Ingestion getting stuck indefinitely if no response from hive source
- feat(Metabase): ingestion lacks ability to omit personal collections
- Search queries require all terms to match, or nothing is returned
- How do I recover data in es after it is deleted HOT 1
- [hive lineage] not support insert into table_name(filed1,filed2,filed3,filed4) HOT 1
- NullPointerException in SendMAEStep when performing indices restoration HOT 1
- [tableau] force_extraction_of_lineage_from_custom_sql_queries is not not permitted HOT 1
- When spark task is a long sql statement, It takes up the whole screen
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from datahub.