Replies: 5 comments 5 replies
-
I think that scala-info objects are not the better choice just because of scala. For PySpark part calling Java code better than Scala too (py4j is not well suited for scala). I suggest have:
|
Beta Was this translation helpful? Give feedback.
-
Agree to Sem, The meta-part is the core and implementation of GraphAr format, unified with Java submodule would help to keep the format implementation alignment between libraries. |
Beta Was this translation helpful? Give feedback.
-
Which naming convention should we use, remain existing style or use normal Java naming convention? existing style:
normal naming convention in Java:
|
Beta Was this translation helpful? Give feedback.
-
We have:
With this case PySpark implementation will work on both regular PySpark and SparkConnect jobs. And also there is an option to use pure java (with provided Info) without Spark (do we really need it?) @Thespica @acezen FYI as a continue of the discussion in #401 |
Beta Was this translation helpful? Give feedback.
-
A point needs to be discussed here about immutability. We hope meta-info is immutable for ability of parallel and safety, but the So we need some immutable class to replace original collection like List, Map and etc. Here are two options I found:
I prefer the first now, but I'm not sure about that, so I need your perspective. Besides, we should return |
Beta Was this translation helpful? Give feedback.
-
Java and scala work on JVM and can call each other's code directly. We can let Java SDK and spark SDK use the same meta-info part so
that we can avoid duplicate work when updating.
For the meta-info part, there are two options:
I'm not sure about how to choose.
Related issue: #276
Beta Was this translation helpful? Give feedback.
All reactions