In [13]:
!pip install --user xlrd


Requirement already satisfied: xlrd in /gpfs/global_fs01/sym_shared/YPProdSpark/user/s1df-1767d8774d3251-73caa6cfaa60/.local/lib/python2.7/site-packages

In [3]:
import Common_Functions


fundDF = pd.read_excel(Common_Functions.getFileFromObjectStorage('MizuhoPOC', 'Funds.xlsm'),header=[0]).rename(index=str, columns={"ALADDIN": "ID"})
fundDF.head(5)


  File "<ipython-input-3-be4853edaa4a>", line 1
    import Mizuho-POC.Common_Functions
                 ^
SyntaxError: invalid syntax

In [14]:
import pandas as pd
from io import BytesIO
import requests
import json
import xlrd 

from pyspark.sql.functions import *
from datetime import datetime
from dateutil.parser import parse

from ingest.Connectors import Connectors

In [15]:
# The code was removed by DSX for sharing.

In [16]:
# The code was removed by DSX for sharing.

In [17]:
fundDF = pd.read_excel(getFileFromObjectStorage('MizuhoPOC', 'Funds.xlsm'),header=[0]).rename(index=str, columns={"ALADDIN": "ID"})
fundDF.head(5)


Out[17]:
ID
0 I-CJF
1 I-CGF
2 I-HFR
3 I-HMC
4 I-SABF1

In [18]:
spark = SparkSession.builder.getOrCreate()  

mhcbSparkDF = spark.createDataFrame(fundDF)


mhcbSparkDF.printSchema()


# Connection to Dash DB for writing the data
dashdbsaveoption = {
                     Connectors.DASHDB.HOST              : dashCredentials["host"],
                     Connectors.DASHDB.DATABASE          : dashCredentials["db"],
                     Connectors.DASHDB.USERNAME          : dashCredentials["username"],
                     Connectors.DASHDB.PASSWORD          : dashCredentials["password"],
                     Connectors.DASHDB.TARGET_TABLE_NAME : dashCredentials["tableName"],
                     Connectors.DASHDB.TARGET_WRITE_MODE : 'merge' 
}

mhcbDashDBDF = mhcbSparkDF.write.format("com.ibm.spark.discover").options(**dashdbsaveoption).save()


root
 |-- ID: string (nullable = true)


In [ ]: