In [13]:
!pip install --user xlrd
In [3]:
import Common_Functions
fundDF = pd.read_excel(Common_Functions.getFileFromObjectStorage('MizuhoPOC', 'Funds.xlsm'),header=[0]).rename(index=str, columns={"ALADDIN": "ID"})
fundDF.head(5)
In [14]:
import pandas as pd
from io import BytesIO
import requests
import json
import xlrd
from pyspark.sql.functions import *
from datetime import datetime
from dateutil.parser import parse
from ingest.Connectors import Connectors
In [15]:
# The code was removed by DSX for sharing.
In [16]:
# The code was removed by DSX for sharing.
In [17]:
fundDF = pd.read_excel(getFileFromObjectStorage('MizuhoPOC', 'Funds.xlsm'),header=[0]).rename(index=str, columns={"ALADDIN": "ID"})
fundDF.head(5)
Out[17]:
In [18]:
spark = SparkSession.builder.getOrCreate()
mhcbSparkDF = spark.createDataFrame(fundDF)
mhcbSparkDF.printSchema()
# Connection to Dash DB for writing the data
dashdbsaveoption = {
Connectors.DASHDB.HOST : dashCredentials["host"],
Connectors.DASHDB.DATABASE : dashCredentials["db"],
Connectors.DASHDB.USERNAME : dashCredentials["username"],
Connectors.DASHDB.PASSWORD : dashCredentials["password"],
Connectors.DASHDB.TARGET_TABLE_NAME : dashCredentials["tableName"],
Connectors.DASHDB.TARGET_WRITE_MODE : 'merge'
}
mhcbDashDBDF = mhcbSparkDF.write.format("com.ibm.spark.discover").options(**dashdbsaveoption).save()
In [ ]: