write.jdbc {SparkR}R Documentation

Save the content of SparkDataFrame to an external database table via JDBC.

Description

Save the content of the SparkDataFrame to an external database table via JDBC. Additional JDBC database connection properties can be set (...) You can find the JDBC-specific option and parameter documentation for writing tables via JDBC in https://spark.apache.org/docs/latest/sql-data-sources-jdbc.html#data-source-option Data Source Option in the version you use.

Usage

write.jdbc(x, url, tableName, mode = "error", ...)

## S4 method for signature 'SparkDataFrame,character,character'
write.jdbc(x, url, tableName, mode = "error", ...)

Arguments

x

a SparkDataFrame.

url

JDBC database url of the form jdbc:subprotocol:subname.

tableName

the name of the table in the external database.

mode

one of 'append', 'overwrite', 'error', 'errorifexists', 'ignore' save mode (it is 'error' by default)

...

additional JDBC database connection properties.

Details

Also, mode is used to specify the behavior of the save operation when data already exists in the data source. There are four modes:

Note

write.jdbc since 2.0.0

See Also

Other SparkDataFrame functions: SparkDataFrame-class, agg(), alias(), arrange(), as.data.frame(), attach,SparkDataFrame-method, broadcast(), cache(), checkpoint(), coalesce(), collect(), colnames(), coltypes(), createOrReplaceTempView(), crossJoin(), cube(), dapplyCollect(), dapply(), describe(), dim(), distinct(), dropDuplicates(), dropna(), drop(), dtypes(), exceptAll(), except(), explain(), filter(), first(), gapplyCollect(), gapply(), getNumPartitions(), group_by(), head(), hint(), histogram(), insertInto(), intersectAll(), intersect(), isLocal(), isStreaming(), join(), limit(), localCheckpoint(), merge(), mutate(), ncol(), nrow(), persist(), printSchema(), randomSplit(), rbind(), rename(), repartitionByRange(), repartition(), rollup(), sample(), saveAsTable(), schema(), selectExpr(), select(), showDF(), show(), storageLevel(), str(), subset(), summary(), take(), toJSON(), unionAll(), unionByName(), union(), unpersist(), withColumn(), withWatermark(), with(), write.df(), write.json(), write.orc(), write.parquet(), write.stream(), write.text()

Examples

## Not run: 
##D sparkR.session()
##D jdbcUrl <- "jdbc:mysql://localhost:3306/databasename"
##D write.jdbc(df, jdbcUrl, "table", user = "username", password = "password")
## End(Not run)

[Package SparkR version 3.2.3 Index]