alibaba / sparkcube Goto Github PK
View Code? Open in Web Editor NEWSparkCube is an open-source project for extremely fast OLAP data analysis. SparkCube is an extension of Apache Spark.
License: Apache License 2.0
SparkCube is an open-source project for extremely fast OLAP data analysis. SparkCube is an extension of Apache Spark.
License: Apache License 2.0
Why is the interface of CubeSource designed in two versions? But the js code is all in v1 version.
When I run this, I meet a issue about WEB UI, and I have check the code seems ok, any point I have missed?
java.lang.Exception: Could not find resource path for Web UI: com/alibaba/sparkcube/execution/ui/static
at org.apache.spark.ui.JettyUtils$.createStaticHandler(JettyUtils.scala:197)
at org.apache.spark.ui.WebUI.addStaticHandler(WebUI.scala:121)
at org.apache.spark.ui.SparkCubeTab.<init>(SparkCubeTab.scala:42)
at org.apache.spark.sql.CubeSharedState$$anonfun$3.apply(CubeSharedState.scala:47)
at org.apache.spark.sql.CubeSharedState$$anonfun$3.apply(CubeSharedState.scala:47)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.sql.CubeSharedState.<init>(CubeSharedState.scala:47)
at org.apache.spark.sql.CubeSharedState$.get(CubeSharedState.scala:63)
at com.alibaba.sparkcube.CubeManager.cubeCatalog(CubeManager.scala:213)
at com.alibaba.sparkcube.CubeManager.listAllCaches(CubeManager.scala:230)
at org.apache.spark.ui.SparkCubePage.render(SparkCubePage.scala:54)
at org.apache.spark.ui.WebUI$$anonfun$2.apply(WebUI.scala:84)
at org.apache.spark.ui.WebUI$$anonfun$2.apply(WebUI.scala:84)
at org.apache.spark.ui.JettyUtils$$anon$3.doGet(JettyUtils.scala:90)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:687)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.spark_project.jetty.servlet.ServletHolder.handle(ServletHolder.java:848)
at org.spark_project.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:584)
at org.spark_project.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1180)
at org.spark_project.jetty.servlet.ServletHandler.doScope(ServletHandler.java:512)
at org.spark_project.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1112)
at org.spark_project.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at org.spark_project.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:213)
at org.spark_project.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at org.spark_project.jetty.server.Server.handle(Server.java:539)
at org.spark_project.jetty.server.HttpChannel.handle(HttpChannel.java:333)
at org.spark_project.jetty.server.HttpConnection.onFillable(HttpConnection.java:251)
at org.spark_project.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:283)
at org.spark_project.jetty.io.FillInterest.fillable(FillInterest.java:108)
at org.spark_project.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
at org.spark_project.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303)
when to create raw cache,it show below error message:
Cannot persistent default.cube_test into hive metastore as table property keys may not start with 'spark.sql.': [spark.sql.cached.info.raw.enableRewrite, spark.sql.cached.info.raw.cacheColumns, spark.sql.cached.info.raw, spark.sql.cached.info.raw.zorderBy, spark.sql.cached.info.raw.storagePath, spark.sql.cached.info.raw.provider, spark.sql.cached.info.raw.lastUpdatedTime, spark.sql.cached.info.raw.partitionBy, spark.sql.cached];
when to create cube cache,it show below error message:
Cannot persistent default.cube_test into hive metastore as table property keys may not start with 'spark.sql.': [spark.sql.cached.info.cube, spark.sql.cached.info.cube.lastUpdatedTime, spark.sql.cached.info.cube.storagePath, spark.sql.cached.info.cube.provider, spark.sql.cached.info.cube.enableRewrite, spark.sql.cached, spark.sql.cached.info.cube.measures, spark.sql.cached.info.cube.partitionBy, spark.sql.cached.info.cube.dims];
Hi all,
I would like to know whether the new release is compatible with Spark 3 or not, and how to use the jar file of SparkCube
I tried to compile the code
There's a problem with the following dependency it's unavailable
<groupId>com.swoop</groupId>
<artifactId>spark-alchemy_2.11</artifactId>
<version>0.3.28</version>
Failed to collect dependencies at com.swoop:spark-alchemy_2.11:jar:0.3.28: Failed to read artifact descriptor for com.swoop:spark-alchemy_2.11:jar:0.3.28: Could not transfer artifact com.swoop:spark-alchemy_2.11:pom:0.3.28 from/to swoop-inc (https://dl.bintray.com/swoop-inc/maven/): Access denied to: https://dl.bintray.com/swoop-inc/maven/com/swoop/spark-alchemy_2.11/0.3.28/spark-alchemy_2.11-0.3.28.pom , ReasonPhrase:Forbidden
If I change the dependency in pom.xml to
spark-alchemy-test_2.12
1.0.1
I have errors (please see below)
private implicit def cacheIdToTableIdent(cacheIdentifier: CacheIdentifier): TableIdentifier = {
[WARNING] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/execution/PreCountDistinctTransformer.scala:20: object spark is not a member of package com.swoop.alchemy
[ERROR] import com.swoop.alchemy.spark.expressions.hll.HyperLogLogInitSimpleAgg
[ERROR] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/execution/PreCountDistinctTransformer.scala:44: not found: value HyperLogLogInitSimpleAgg
[ERROR] HyperLogLogInitSimpleAgg(childExpr, relativeSD)
[ERROR] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/optimizer/GenPlanFromCache.scala:22: object spark is not a member of package com.swoop.alchemy
[ERROR] import com.swoop.alchemy.spark.expressions.hll.{HyperLogLogCardinality, HyperLogLogMerge}
[ERROR] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/optimizer/GenPlanFromCache.scala:330: not found: value HyperLogLogCardinality
[ERROR] case _: CardinalityAfter => HyperLogLogCardinality(attrs.head)
[ERROR] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/optimizer/GenPlanFromCache.scala:417: not found: value HyperLogLogCardinality
[ERROR] Some(HyperLogLogCardinality(other))
[ERROR] ^
[ERROR] /home/rym/Downloads/SparkCube-0.3.0/src/main/scala/com/alibaba/sparkcube/optimizer/GenPlanFromCache.scala:509: not found: value HyperLogLogMerge
[ERROR] HyperLogLogMerge(args.head, hllpp.mutableAggBufferOffset, hllpp.inputAggBufferOffset)
[ERROR] ^
[WARNING] one warning found
[ERROR] 6 errors found
thank you
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.