0 votes

Hello, I have used Dataiku last year and a few days back i restarted using it but everytime I try to run a recipe (whether it is prepare or join) the job always fails! It says activity was successful and I can see the output dataset but then it has errors in computing metrics and I am never sure if the recipe actually works.

I am using the tutorial 101 and the first prepare job already gives an error:

[2017/02/17-14:37:13.179] [FRT-38-FlowRunnable] [INFO] [dku.format] - Start uncompressed stream : /haiku_shirt_sales.csv
[2017/02/17-14:37:13.681] [FRT-38-FlowRunnable] [INFO] [dku.format] - after stream totalComp=2393867 totalUncomp=2393867 totalRec=29379
[2017/02/17-14:37:13.682] [FRT-38-FlowRunnable] [INFO] [dku.flow.stream] - Split runner done
[2017/02/17-14:37:13.683] [FRT-39-FlowRunnable] [INFO] [dku.flow.stream] - Parallel streamer done
[2017/02/17-14:37:13.683] [FRT-38-FlowRunnable] [INFO] [dku.format] - Extractor run done, totalCompressed=2393867 totalRecords=29379
[2017/02/17-14:37:13.684] [FRT-39-FlowRunnable] [INFO] [dku.flow.activity] - Run thread done for activity compute_haiku_shirt_sales_prepared_NP
[2017/02/17-14:37:13.688] [FRT-38-FlowRunnable] [INFO] [dku.flow.activity] - Run thread done for activity compute_haiku_shirt_sales_prepared_NP
[2017/02/17-14:37:13.780] [ActivityExecutor-34] [INFO] [dku.flow.activity] running compute_haiku_shirt_sales_prepared_NP - activity is finished
[2017/02/17-14:37:13.781] [ActivityExecutor-34] [INFO] [dku.flow.activity] running compute_haiku_shirt_sales_prepared_NP - Activity successful, computing metrics and running checks
[2017/02/17-14:37:13.802] [ActivityExecutor-34] [ERROR] [dku.flow.activity] running compute_haiku_shirt_sales_prepared_NP - Failed to run checks for activity
com.dataiku.common.server.APIError$APIErrorException: General error: "java.lang.ArrayIndexOutOfBoundsException"; SQL statement:
SELECT IFNULL(MAX(session_index), 0) + 1 FROM metrics_sessions [50000-176], caused by: ArrayIndexOutOfBoundsException: null
	at com.dataiku.common.apiclient.InternalAPIClient.handleJSONResp(InternalAPIClient.java:499)
	at com.dataiku.common.apiclient.InternalAPIClient.postFormToJSON(InternalAPIClient.java:251)
	at com.dataiku.dip.dataflow.kernel.slave.JobKernelAPIClient.runChecksForActivity(JobKernelAPIClient.java:88)
	at com.dataiku.dip.dataflow.jobrunner.ActivityRunner.waitForEnd(ActivityRunner.java:152)
	at com.dataiku.dip.dataflow.jobrunner.ActivityRunner.runActivity(ActivityRunner.java:519)
	at com.dataiku.dip.dataflow.jobrunner.JobRunner.runActivity(JobRunner.java:108)
	at com.dataiku.dip.dataflow.jobrunner.JobRunner.access$800(JobRunner.java:31)
	at com.dataiku.dip.dataflow.jobrunner.JobRunner$ActivityExecutorThread.run(JobRunner.java:294)
[2017/02/17-14:37:13.804] [ActivityExecutor-34] [INFO] [dku.flow.activity] running compute_haiku_shirt_sales_prepared_NP - Removing samples for DKU_TUTORIAL101STARTINGPROJECT_1.haiku_shirt_sales_prepared
[2017/02/17-14:37:13.808] [ActivityExecutor-34] [INFO] [dku.flow.activity] running compute_haiku_shirt_sales_prepared_NP - Done post-activity tasks
[2017/02/17-14:37:13.809] [ActivityExecutor-34] [INFO] [dku.flow.object.event.remote] running compute_haiku_shirt_sales_prepared_NP - sending sja to backend
[2017/02/17-14:37:13.836] [ActivityExecutor-34] [ERROR] [dku.flow.jobrunner] running compute_haiku_shirt_sales_prepared_NP - Failed to notify end of activity
com.dataiku.common.server.APIError$APIErrorException: General error: "java.lang.ArrayIndexOutOfBoundsException"; SQL statement:
SELECT IFNULL(MAX(session_index), 0) + 1 FROM metrics_sessions [50000-176], caused by: ArrayIndexOutOfBoundsException: null
	at com.dataiku.common.apiclient.InternalAPIClient.handleJSONResp(InternalAPIClient.java:499)
	at com.dataiku.common.apiclient.InternalAPIClient.postFormToJSON(InternalAPIClient.java:251)
	at com.dataiku.dip.dataflow.kernel.slave.JobKernelAPIClient.onActivityDone(JobKernelAPIClient.java:84)
	at com.dataiku.dip.dataflow.jobrunner.JobRunner$ActivityExecutorThread.run(JobRunner.java:314)

 

asked by anonymous

1 Answer

+2 votes
Best answer
can you try deleting the jobs_database.h2.* files in your DIP_HOME folder? It looks like a corrupted database. If this is a 4.0 update, then delete databases/jobs* from DIP_HOME
answered by
Hey! It worked, thank you for your help!
395 questions
426 answers
250 comments
253 users