|
1 |
| -= HadoopWriteConfigUtil |
| 1 | +# HadoopWriteConfigUtil |
2 | 2 |
|
3 |
| -`HadoopWriteConfigUtil[K, V]` is an <<contract, abstraction>> of <<implementations, writer configurers>>. |
| 3 | +`HadoopWriteConfigUtil[K, V]` is an [abstraction](#contract) of [writer configurers](#implementations) for [SparkHadoopWriter](SparkHadoopWriter.md) to [write a key-value RDD](SparkHadoopWriter.md#write) (for [RDD.saveAsNewAPIHadoopDataset](rdd/PairRDDFunctions.md#saveAsNewAPIHadoopDataset) and [RDD.saveAsHadoopDataset](rdd/PairRDDFunctions.md#saveAsHadoopDataset) operators). |
4 | 4 |
|
5 |
| -`HadoopWriteConfigUtil` is used for <<spark-internal-io-SparkHadoopWriter.md#, SparkHadoopWriter>> utility when requested to <<spark-internal-io-SparkHadoopWriter.md#write, write an RDD of key-value pairs>> (for rdd:PairRDDFunctions.md#saveAsNewAPIHadoopDataset[saveAsNewAPIHadoopDataset] and rdd:PairRDDFunctions.md#saveAsHadoopDataset[saveAsHadoopDataset] transformations). |
| 5 | +## Contract |
6 | 6 |
|
7 |
| -[[contract]] |
8 |
| -.HadoopWriteConfigUtil Contract |
9 |
| -[cols="30m,70",options="header",width="100%"] |
10 |
| -|=== |
11 |
| -| Method |
12 |
| -| Description |
| 7 | +### <span id="assertConf"> assertConf |
13 | 8 |
|
14 |
| -| assertConf |
15 |
| -a| [[assertConf]] |
16 |
| - |
17 |
| -[source, scala] |
18 |
| ----- |
| 9 | +```scala |
19 | 10 | assertConf(
|
20 | 11 | jobContext: JobContext,
|
21 | 12 | conf: SparkConf): Unit
|
22 |
| ----- |
| 13 | +``` |
23 | 14 |
|
24 |
| -| closeWriter |
25 |
| -a| [[closeWriter]] |
| 15 | +### <span id="closeWriter"> closeWriter |
26 | 16 |
|
27 |
| -[source, scala] |
28 |
| ----- |
| 17 | +```scala |
29 | 18 | closeWriter(
|
30 | 19 | taskContext: TaskAttemptContext): Unit
|
31 |
| ----- |
| 20 | +``` |
32 | 21 |
|
33 |
| -| createCommitter |
34 |
| -a| [[createCommitter]] |
| 22 | +### <span id="createCommitter"> createCommitter |
35 | 23 |
|
36 |
| -[source, scala] |
37 |
| ----- |
| 24 | +```scala |
38 | 25 | createCommitter(
|
39 | 26 | jobId: Int): HadoopMapReduceCommitProtocol
|
40 |
| ----- |
| 27 | +``` |
| 28 | + |
| 29 | +Creates a [HadoopMapReduceCommitProtocol](HadoopMapReduceCommitProtocol.md) committer |
| 30 | + |
| 31 | +Used when: |
41 | 32 |
|
42 |
| -| createJobContext |
43 |
| -a| [[createJobContext]] |
| 33 | +* `SparkHadoopWriter` is requested to [write data out](SparkHadoopWriter.md#write) |
44 | 34 |
|
45 |
| -[source, scala] |
46 |
| ----- |
| 35 | +### <span id="createJobContext"> createJobContext |
| 36 | + |
| 37 | +```scala |
47 | 38 | createJobContext(
|
48 | 39 | jobTrackerId: String,
|
49 | 40 | jobId: Int): JobContext
|
50 |
| ----- |
| 41 | +``` |
51 | 42 |
|
52 |
| -| createTaskAttemptContext |
53 |
| -a| [[createTaskAttemptContext]] |
| 43 | +### <span id="createTaskAttemptContext"> createTaskAttemptContext |
54 | 44 |
|
55 |
| -[source, scala] |
56 |
| ----- |
| 45 | +```scala |
57 | 46 | createTaskAttemptContext(
|
58 | 47 | jobTrackerId: String,
|
59 | 48 | jobId: Int,
|
60 | 49 | splitId: Int,
|
61 | 50 | taskAttemptId: Int): TaskAttemptContext
|
62 |
| ----- |
| 51 | +``` |
63 | 52 |
|
64 |
| -Creates a Hadoop https://hadoop.apache.org/docs/r2.7.3/api/org/apache/hadoop/mapreduce/TaskAttemptContext.html[TaskAttemptContext] |
| 53 | +Creates a Hadoop [TaskAttemptContext]({{ hadoop.api }}/org/apache/hadoop/mapreduce/TaskAttemptContext.html) |
65 | 54 |
|
66 |
| -| initOutputFormat |
67 |
| -a| [[initOutputFormat]] |
| 55 | +### <span id="initOutputFormat"> initOutputFormat |
68 | 56 |
|
69 |
| -[source, scala] |
70 |
| ----- |
| 57 | +```scala |
71 | 58 | initOutputFormat(
|
72 | 59 | jobContext: JobContext): Unit
|
73 |
| ----- |
| 60 | +``` |
74 | 61 |
|
75 |
| -| initWriter |
76 |
| -a| [[initWriter]] |
| 62 | +### <span id="initWriter"> initWriter |
77 | 63 |
|
78 |
| -[source, scala] |
79 |
| ----- |
| 64 | +```scala |
80 | 65 | initWriter(
|
81 | 66 | taskContext: TaskAttemptContext,
|
82 | 67 | splitId: Int): Unit
|
83 |
| ----- |
| 68 | +``` |
84 | 69 |
|
85 |
| -| write |
86 |
| -a| [[write]] |
| 70 | +### <span id="write"> write |
87 | 71 |
|
88 |
| -[source, scala] |
89 |
| ----- |
| 72 | +```scala |
90 | 73 | write(
|
91 | 74 | pair: (K, V)): Unit
|
92 |
| ----- |
| 75 | +``` |
93 | 76 |
|
94 | 77 | Writes out the key-value pair
|
95 | 78 |
|
96 |
| -Used when `SparkHadoopWriter` is requested to <<spark-internal-io-SparkHadoopWriter.md#executeTask, executeTask>> (while <<spark-internal-io-SparkHadoopWriter.md#write, writing out key-value pairs of a partition>>) |
97 |
| - |
98 |
| -|=== |
99 |
| - |
100 |
| -[[implementations]] |
101 |
| -.HadoopWriteConfigUtils |
102 |
| -[cols="30,70",options="header",width="100%"] |
103 |
| -|=== |
104 |
| -| HadoopWriteConfigUtil |
105 |
| -| Description |
| 79 | +Used when: |
106 | 80 |
|
107 |
| -| <<spark-internal-io-HadoopMapReduceWriteConfigUtil.md#, HadoopMapReduceWriteConfigUtil>> |
108 |
| -| [[HadoopMapReduceWriteConfigUtil]] |
| 81 | +* `SparkHadoopWriter` is requested to [executeTask](SparkHadoopWriter.md#executeTask) |
109 | 82 |
|
110 |
| -| <<spark-internal-io-HadoopMapRedWriteConfigUtil.md#, HadoopMapRedWriteConfigUtil>> |
111 |
| -| [[HadoopMapRedWriteConfigUtil]] |
| 83 | +## Implementations |
112 | 84 |
|
113 |
| -|=== |
| 85 | +* [HadoopMapReduceWriteConfigUtil](HadoopMapReduceWriteConfigUtil.md) |
| 86 | +* [HadoopMapRedWriteConfigUtil](HadoopMapRedWriteConfigUtil.md) |
0 commit comments