Replies: 6 comments 8 replies
-
Here I put forward an idea: I don't store multiple fields in the database like Flink deployment mode. I save it in the form of YAML. We can simply write it to the database after Base64 and save it. This can avoid database field expansion as the functions increase. What do you think about this? |
Beta Was this translation helpful? Give feedback.
-
Hello, your technical plan is very good, I have used spark before, can I complete this with you |
Beta Was this translation helpful? Give feedback.
-
+1 |
Beta Was this translation helpful? Give feedback.
-
hi: As mentioned in the background introduction of the proposal, currently StreamPark already provides support for flink jobs and has a large user base. However, support for Spark capabilities has not been initiated yet. StreamPark has never abandoned support for Spark jobs and I sincerely appreciate your drive for this PR. The initial plan for StreamPark support for Spark is as follows: On one hand, we will provide development framework for Spark jobs (this feature is not part of the current discussion). On the other hand, we will provide the deployment and management capabilities of Spark jobs. For this aspect, I suggest we divide it into several steps:
Regarding the deployment method of spark jobs, there are differences between the implementation on yarn and k8s. We can begin with a simplest deployment mode. For the job status handling, we can directly leverage Spark's API. This part ensures the minimal functionality (jobs can be added, started, and their status automatically tracked).
|
Beta Was this translation helpful? Give feedback.
-
This plan is great, allowing streampark to support both spark and flink tasks. Can I join this program and contribute a little bit? Thanks! |
Beta Was this translation helpful? Give feedback.
-
@tam-lab Of course, the community has already begun to adapt spark on yarn. You can leave a message under the corresponding issue. |
Beta Was this translation helpful? Give feedback.
-
Flink job deployment StreamPark has already supported it relatively well. Spark has not yet done so in this area. We are now proposing relevant designs. We welcome your discussion and will continue to collect relevant functional opinions and design suggestions.
Draft:Deploy Spark Job In StreamPark
Beta Was this translation helpful? Give feedback.
All reactions