The customization is basically get a spark 2.4.3 distribution installed in
the image and add the required jars in $SPARK_HOME/jars directory to access
Ceph/S3 buckets. the fix for the issue you mention is added in the code
since spark 2.3 iirc.
If the current image already has these things done, I can use the image as
is.
On Mon, Aug 19, 2019 at 3:31 PM Sherard Griffin <shgriffi(a)redhat.com> wrote:
Ricardo,
What are the customizations? Is this to fix retrieving the tables and
databases from Hive metastore? I've fixed the permissions on the
spark-cluster-image repo, although I'm not a fan of the name of it...
Should just be "spark".
Thanks,
Sherard
On Mon, Aug 19, 2019 at 12:02 PM Ricardo Martinelli de Oliveira <
rmartine(a)redhat.com> wrote:
> One side note about the spark image: Because Spark SQL Thrift server has
> an open issue in spark 2.2 that breaks thrift server we need to use spark
> 2.4. I already have the image built and I can push it when I have proper
> permissions to do it.
>
> As for Landon suggestion, I think it's a good idea and I can in advance
> create a quay.io/opendatahub/spark-cluster-image:2.4 tag if needed so we
> can reuse the same name but with a different tag.
>
> On Mon, Aug 19, 2019 at 11:10 AM Landon LaSmith <llasmith(a)redhat.com>
> wrote:
>
>> Do we want to store all of the ODH spark images in the same repository?
>> We already have quay.io/opendatahub/spark-cluster-image
>> <
https://quay.io/repository/opendatahub/spark-cluster-image?tab=tags>.
>> Should we deprecate that repo and create a new one to store both?
>>
>> On Mon, Aug 19, 2019 at 10:06 AM Ricardo Martinelli de Oliveira <
>> rmartine(a)redhat.com> wrote:
>>
>>> Thanks everyone for the feedback!
>>>
>>> As the winner is #2 (push the custom imago into quay.io operndatahub
>>> organization), who should I ask for permissions to push my image in this
>>> org?
>>>
>>> My quay.io username is the same as my kerberos user.
>>>
>>> On Mon, Aug 19, 2019 at 10:57 AM Landon LaSmith <llasmith(a)redhat.com>
>>> wrote:
>>>
>>>> Agree with #2
>>>>
>>>> On Mon, Aug 19, 2019 at 9:53 AM Václav Pavlín <vasek(a)redhat.com>
>>>> wrote:
>>>>
>>>>> I agree with #2 - ODH should work out of box, so we need to provide
>>>>> the image (which is a no for #1), and #3 sounds like an overkill
>>>>>
>>>>> Thanks,
>>>>> V.
>>>>>
>>>>> On Mon, Aug 19, 2019 at 3:43 PM Alex Corvin
<acorvin(a)redhat.com>
>>>>> wrote:
>>>>>
>>>>>> I think my vote is for #2. Option #1 will continue to be
supported
>>>>>> for groups that need it, but we can make it easier for people to
get up and
>>>>>> running by curating an official image.
>>>>>>
>>>>>>
>>>>>> On August 19, 2019 at 9:33:44 AM, Ricardo Martinelli de Oliveira
(
>>>>>> rmartine(a)redhat.com) wrote:
>>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I'm integrating Spark SQL Thrift server into ODH operator and
I need
>>>>>> to use a custom spark image (other than the RADAnalytics image)
with
>>>>>> additional jars to access Ceph/S3 buckets. Actually, both thrift
server and
>>>>>> the spark cluster will need this custom spark image in order to
access the
>>>>>> buckets.
>>>>>>
>>>>>> With that being said, I'd like to discuss some options to get
this
>>>>>> done. I am thinking about these options:
>>>>>>
>>>>>> 1) Let the customer specify the custom image in the yaml file
(this
>>>>>> is already possible)
>>>>>> 2) Create that custom spark image and publish on quay.io
>>>>>> opendarahub organization
>>>>>> 3) Add a buildconfig object and make operator create the custom
>>>>>> build and set the image location into the deploymentconfig
objects
>>>>>>
>>>>>> Although the third option automate everything and deliver the
whole
>>>>>> set with the custom image, there's this thing about
supporting custom
>>>>>> images within operators. We'd need to add a spark_version
variable where
>>>>>> the build could download the spark distribution corresponding to
that
>>>>>> version and the artifacts related and run the build. In the first
option,
>>>>>> we simply don't create the build objects and document that in
order to use
>>>>>> Thrift server in ODH operator, both spark cluster and thrift must
use a
>>>>>> custom spark image containing the jars needed to access Ceph/S3.
At last,
>>>>>> the middle term between both is option two, so we don't need
to worry about
>>>>>> delegate this task to the user or the operator.
>>>>>>
>>>>>> What do you think? What could be the best option for this
scenario?
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Ricardo Martinelli De Oliveira
>>>>>>
>>>>>> Data Engineer, AI CoE
>>>>>>
>>>>>> Red Hat Brazil <
https://www.redhat.com/>
>>>>>>
>>>>>> Av. Brigadeiro Faria Lima, 3900
>>>>>>
>>>>>> 8th floor
>>>>>>
>>>>>> rmartine(a)redhat.com T: +551135426125
>>>>>> M: +5511970696531
>>>>>> @redhatjobs <
https://twitter.com/redhatjobs> redhatjobs
>>>>>> <
https://www.facebook.com/redhatjobs> @redhatjobs
>>>>>> <
https://instagram.com/redhatjobs>
>>>>>> <
https://www.redhat.com/>
>>>>>> _______________________________________________
>>>>>> Contributors mailing list -- contributors(a)lists.opendatahub.io
>>>>>> To unsubscribe send an email to
>>>>>> contributors-leave(a)lists.opendatahub.io
>>>>>>
>>>>>> _______________________________________________
>>>>>> Contributors mailing list -- contributors(a)lists.opendatahub.io
>>>>>> To unsubscribe send an email to
>>>>>> contributors-leave(a)lists.opendatahub.io
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Open Data Hub, AI CoE, Office of CTO, Red Hat
>>>>> Brno, Czech Republic
>>>>> Phone: +420 739 666 824
>>>>>
>>>>> _______________________________________________
>>>>> Contributors mailing list -- contributors(a)lists.opendatahub.io
>>>>> To unsubscribe send an email to
>>>>> contributors-leave(a)lists.opendatahub.io
>>>>>
>>>>
>>>>
>>>> --
>>>> Landon LaSmith
>>>> Sr.Software Engineer
>>>> Red Hat, AI CoE - Data Hub
>>>>
>>>
>>>
>>> --
>>>
>>> Ricardo Martinelli De Oliveira
>>>
>>> Data Engineer, AI CoE
>>>
>>> Red Hat Brazil <
https://www.redhat.com/>
>>>
>>> Av. Brigadeiro Faria Lima, 3900
>>>
>>> 8th floor
>>>
>>> rmartine(a)redhat.com T: +551135426125
>>> M: +5511970696531
>>> @redhatjobs <
https://twitter.com/redhatjobs> redhatjobs
>>> <
https://www.facebook.com/redhatjobs> @redhatjobs
>>> <
https://instagram.com/redhatjobs>
>>> <
https://www.redhat.com/>
>>>
>>
>>
>> --
>> Landon LaSmith
>> Sr.Software Engineer
>> Red Hat, AI CoE - Data Hub
>>
>
>
> --
>
> Ricardo Martinelli De Oliveira
>
> Data Engineer, AI CoE
>
> Red Hat Brazil <
https://www.redhat.com/>
>
> Av. Brigadeiro Faria Lima, 3900
>
> 8th floor
>
> rmartine(a)redhat.com T: +551135426125
> M: +5511970696531
> @redhatjobs <
https://twitter.com/redhatjobs> redhatjobs
> <
https://www.facebook.com/redhatjobs> @redhatjobs
> <
https://instagram.com/redhatjobs>
> <
https://www.redhat.com/>
> _______________________________________________
> Contributors mailing list -- contributors(a)lists.opendatahub.io
> To unsubscribe send an email to contributors-leave(a)lists.opendatahub.io
>
--
Thanks,
Sherard Griffin
Av. Brigadeiro Faria Lima, 3900
8th floor
rmartine(a)redhat.com T: +551135426125
M: +5511970696531
@redhatjobs <