diff --git a/packages/google-cloud-dataproc/.gitattributes b/packages/google-cloud-dataproc/.gitattributes new file mode 100644 index 000000000000..33739cb74e44 --- /dev/null +++ b/packages/google-cloud-dataproc/.gitattributes @@ -0,0 +1,4 @@ +*.ts text eol=lf +*.js text eol=lf +protos/* linguist-generated +**/api-extractor.json linguist-language=JSON-with-Comments diff --git a/packages/google-cloud-dataproc/.prettierignore b/packages/google-cloud-dataproc/.prettierignore new file mode 100644 index 000000000000..9340ad9b86d3 --- /dev/null +++ b/packages/google-cloud-dataproc/.prettierignore @@ -0,0 +1,6 @@ +**/node_modules +**/coverage +test/fixtures +build/ +docs/ +protos/ diff --git a/packages/google-cloud-dataproc/.prettierrc.js b/packages/google-cloud-dataproc/.prettierrc.js new file mode 100644 index 000000000000..7649ee3c254e --- /dev/null +++ b/packages/google-cloud-dataproc/.prettierrc.js @@ -0,0 +1,22 @@ +// Copyright 2026 Google LLC +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// https://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. +// +// ** This file is automatically generated by gapic-generator-typescript. ** +// ** https://github.com/googleapis/gapic-generator-typescript ** +// ** All changes to this file may be overwritten. ** + + +module.exports = { + ...require('gts/.prettierrc.json') +} diff --git a/packages/google-cloud-dataproc/CODE_OF_CONDUCT.md b/packages/google-cloud-dataproc/CODE_OF_CONDUCT.md new file mode 100644 index 000000000000..2add2547a812 --- /dev/null +++ b/packages/google-cloud-dataproc/CODE_OF_CONDUCT.md @@ -0,0 +1,94 @@ + +# Code of Conduct + +## Our Pledge + +In the interest of fostering an open and welcoming environment, we as +contributors and maintainers pledge to making participation in our project and +our community a harassment-free experience for everyone, regardless of age, body +size, disability, ethnicity, gender identity and expression, level of +experience, education, socio-economic status, nationality, personal appearance, +race, religion, or sexual identity and orientation. + +## Our Standards + +Examples of behavior that contributes to creating a positive environment +include: + +* Using welcoming and inclusive language +* Being respectful of differing viewpoints and experiences +* Gracefully accepting constructive criticism +* Focusing on what is best for the community +* Showing empathy towards other community members + +Examples of unacceptable behavior by participants include: + +* The use of sexualized language or imagery and unwelcome sexual attention or + advances +* Trolling, insulting/derogatory comments, and personal or political attacks +* Public or private harassment +* Publishing others' private information, such as a physical or electronic + address, without explicit permission +* Other conduct which could reasonably be considered inappropriate in a + professional setting + +## Our Responsibilities + +Project maintainers are responsible for clarifying the standards of acceptable +behavior and are expected to take appropriate and fair corrective action in +response to any instances of unacceptable behavior. + +Project maintainers have the right and responsibility to remove, edit, or reject +comments, commits, code, wiki edits, issues, and other contributions that are +not aligned to this Code of Conduct, or to ban temporarily or permanently any +contributor for other behaviors that they deem inappropriate, threatening, +offensive, or harmful. + +## Scope + +This Code of Conduct applies both within project spaces and in public spaces +when an individual is representing the project or its community. Examples of +representing a project or community include using an official project e-mail +address, posting via an official social media account, or acting as an appointed +representative at an online or offline event. Representation of a project may be +further defined and clarified by project maintainers. + +This Code of Conduct also applies outside the project spaces when the Project +Steward has a reasonable belief that an individual's behavior may have a +negative impact on the project or its community. + +## Conflict Resolution + +We do not believe that all conflict is bad; healthy debate and disagreement +often yield positive results. However, it is never okay to be disrespectful or +to engage in behavior that violates the project’s code of conduct. + +If you see someone violating the code of conduct, you are encouraged to address +the behavior directly with those involved. Many issues can be resolved quickly +and easily, and this gives people more control over the outcome of their +dispute. If you are unable to resolve the matter for any reason, or if the +behavior is threatening or harassing, report it. We are dedicated to providing +an environment where participants feel welcome and safe. + +Reports should be directed to *googleapis-stewards@google.com*, the +Project Steward(s) for *Google Cloud Client Libraries*. It is the Project Steward’s duty to +receive and address reported violations of the code of conduct. They will then +work with a committee consisting of representatives from the Open Source +Programs Office and the Google Open Source Strategy team. If for any reason you +are uncomfortable reaching out to the Project Steward, please email +opensource@google.com. + +We will investigate every complaint, but you may not receive a direct response. +We will use our discretion in determining when and how to follow up on reported +incidents, which may range from not taking action to permanent expulsion from +the project and project-sponsored spaces. We will notify the accused of the +report and provide them an opportunity to discuss it before any action is taken. +The identity of the reporter will be omitted from the details of the report +supplied to the accused. In potentially harmful situations, such as ongoing +harassment or threats to anyone's safety, we may take action without notice. + +## Attribution + +This Code of Conduct is adapted from the Contributor Covenant, version 1.4, +available at +https://www.contributor-covenant.org/version/1/4/code-of-conduct.html \ No newline at end of file diff --git a/packages/google-cloud-dataproc/CONTRIBUTING.md b/packages/google-cloud-dataproc/CONTRIBUTING.md new file mode 100644 index 000000000000..a8b18d88f3a5 --- /dev/null +++ b/packages/google-cloud-dataproc/CONTRIBUTING.md @@ -0,0 +1,76 @@ +# How to become a contributor and submit your own code + +**Table of contents** + +* [Contributor License Agreements](#contributor-license-agreements) +* [Contributing a patch](#contributing-a-patch) +* [Running the tests](#running-the-tests) +* [Releasing the library](#releasing-the-library) + +## Contributor License Agreements + +We'd love to accept your sample apps and patches! Before we can take them, we +have to jump a couple of legal hurdles. + +Please fill out either the individual or corporate Contributor License Agreement +(CLA). + + * If you are an individual writing original source code and you're sure you + own the intellectual property, then you'll need to sign an [individual CLA](https://developers.google.com/open-source/cla/individual). + * If you work for a company that wants to allow you to contribute your work, + then you'll need to sign a [corporate CLA](https://developers.google.com/open-source/cla/corporate). + +Follow either of the two links above to access the appropriate CLA and +instructions for how to sign and return it. Once we receive it, we'll be able to +accept your pull requests. + +## Contributing A Patch + +1. Submit an issue describing your proposed change to the repo in question. +1. The repo owner will respond to your issue promptly. +1. If your proposed change is accepted, and you haven't already done so, sign a + Contributor License Agreement (see details above). +1. Fork the desired repo, develop and test your code changes. +1. Ensure that your code adheres to the existing style in the code to which + you are contributing. +1. Ensure that your code has an appropriate set of tests which all pass. +1. Title your pull request following [Conventional Commits](https://www.conventionalcommits.org/) styling. +1. Submit a pull request. + +### Before you begin + +1. [Select or create a Cloud Platform project][projects]. +1. [Enable billing for your project][billing]. +1. [Enable the Dataproc API][enable_api]. +1. [Set up authentication with a service account][auth] so you can access the + API from your local workstation. + + +## Running the tests + +1. [Prepare your environment for Node.js setup][setup]. + +1. Install dependencies: + + npm install + +1. Run the tests: + + # Run unit tests. + npm test + + # Run sample integration tests. + npm run samples-test + + # Run all system tests. + npm run system-test + +1. Lint (and maybe fix) any changes: + + npm run fix + +[setup]: https://cloud.google.com/nodejs/docs/setup +[projects]: https://console.cloud.google.com/project +[billing]: https://support.google.com/cloud/answer/6293499#enable-billing +[enable_api]: https://console.cloud.google.com/flows/enableapi?apiid=dataproc.googleapis.com +[auth]: https://cloud.google.com/docs/authentication/getting-started \ No newline at end of file diff --git a/packages/google-cloud-dataproc/LICENSE b/packages/google-cloud-dataproc/LICENSE new file mode 100644 index 000000000000..d64569567334 --- /dev/null +++ b/packages/google-cloud-dataproc/LICENSE @@ -0,0 +1,202 @@ + + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright [yyyy] [name of copyright owner] + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/packages/google-cloud-dataproc/protos/google/cloud/dataproc/v1/batches.proto b/packages/google-cloud-dataproc/protos/google/cloud/dataproc/v1/batches.proto index bc57cb52344b..04bec6acd990 100644 --- a/packages/google-cloud-dataproc/protos/google/cloud/dataproc/v1/batches.proto +++ b/packages/google-cloud-dataproc/protos/google/cloud/dataproc/v1/batches.proto @@ -1,4 +1,4 @@ -// Copyright 2025 Google LLC +// Copyright 2026 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. @@ -256,6 +256,10 @@ message Batch { // Optional. SparkSql batch config. SparkSqlBatch spark_sql_batch = 7 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. PySpark notebook batch config. + PySparkNotebookBatch pyspark_notebook_batch = 19 + [(google.api.field_behavior) = OPTIONAL]; } // Output only. Runtime information about batch execution. @@ -406,3 +410,27 @@ message SparkSqlBatch { // Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. repeated string jar_file_uris = 3 [(google.api.field_behavior) = OPTIONAL]; } + +// A configuration for running a PySpark Notebook batch workload. +message PySparkNotebookBatch { + // Required. The HCFS URI of the notebook file to execute. + string notebook_file_uri = 1 [(google.api.field_behavior) = REQUIRED]; + + // Optional. The parameters to pass to the notebook. + map params = 2 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. HCFS URIs of Python files to pass to the PySpark framework. + repeated string python_file_uris = 3 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. HCFS URIs of jar files to be added to the Spark CLASSPATH. + repeated string jar_file_uris = 4 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. HCFS URIs of files to be placed in the working directory of + // each executor + repeated string file_uris = 5 [(google.api.field_behavior) = OPTIONAL]; + + // Optional. HCFS URIs of archives to be extracted into the working directory + // of each executor. Supported file types: + // `.jar`, `.tar`, `.tar.gz`, `.tgz`, and `.zip`. + repeated string archive_uris = 6 [(google.api.field_behavior) = OPTIONAL]; +} diff --git a/packages/google-cloud-dataproc/protos/protos.d.ts b/packages/google-cloud-dataproc/protos/protos.d.ts index d143de6926b0..0f88e851d408 100644 --- a/packages/google-cloud-dataproc/protos/protos.d.ts +++ b/packages/google-cloud-dataproc/protos/protos.d.ts @@ -1914,6 +1914,9 @@ export namespace google { /** Batch sparkSqlBatch */ sparkSqlBatch?: (google.cloud.dataproc.v1.ISparkSqlBatch|null); + /** Batch pysparkNotebookBatch */ + pysparkNotebookBatch?: (google.cloud.dataproc.v1.IPySparkNotebookBatch|null); + /** Batch runtimeInfo */ runtimeInfo?: (google.cloud.dataproc.v1.IRuntimeInfo|null); @@ -1975,6 +1978,9 @@ export namespace google { /** Batch sparkSqlBatch. */ public sparkSqlBatch?: (google.cloud.dataproc.v1.ISparkSqlBatch|null); + /** Batch pysparkNotebookBatch. */ + public pysparkNotebookBatch?: (google.cloud.dataproc.v1.IPySparkNotebookBatch|null); + /** Batch runtimeInfo. */ public runtimeInfo?: (google.cloud.dataproc.v1.IRuntimeInfo|null); @@ -2006,7 +2012,7 @@ export namespace google { public stateHistory: google.cloud.dataproc.v1.Batch.IStateHistory[]; /** Batch batchConfig. */ - public batchConfig?: ("pysparkBatch"|"sparkBatch"|"sparkRBatch"|"sparkSqlBatch"); + public batchConfig?: ("pysparkBatch"|"sparkBatch"|"sparkRBatch"|"sparkSqlBatch"|"pysparkNotebookBatch"); /** * Creates a new Batch instance using the specified properties. @@ -2690,6 +2696,133 @@ export namespace google { public static getTypeUrl(typeUrlPrefix?: string): string; } + /** Properties of a PySparkNotebookBatch. */ + interface IPySparkNotebookBatch { + + /** PySparkNotebookBatch notebookFileUri */ + notebookFileUri?: (string|null); + + /** PySparkNotebookBatch params */ + params?: ({ [k: string]: string }|null); + + /** PySparkNotebookBatch pythonFileUris */ + pythonFileUris?: (string[]|null); + + /** PySparkNotebookBatch jarFileUris */ + jarFileUris?: (string[]|null); + + /** PySparkNotebookBatch fileUris */ + fileUris?: (string[]|null); + + /** PySparkNotebookBatch archiveUris */ + archiveUris?: (string[]|null); + } + + /** Represents a PySparkNotebookBatch. */ + class PySparkNotebookBatch implements IPySparkNotebookBatch { + + /** + * Constructs a new PySparkNotebookBatch. + * @param [properties] Properties to set + */ + constructor(properties?: google.cloud.dataproc.v1.IPySparkNotebookBatch); + + /** PySparkNotebookBatch notebookFileUri. */ + public notebookFileUri: string; + + /** PySparkNotebookBatch params. */ + public params: { [k: string]: string }; + + /** PySparkNotebookBatch pythonFileUris. */ + public pythonFileUris: string[]; + + /** PySparkNotebookBatch jarFileUris. */ + public jarFileUris: string[]; + + /** PySparkNotebookBatch fileUris. */ + public fileUris: string[]; + + /** PySparkNotebookBatch archiveUris. */ + public archiveUris: string[]; + + /** + * Creates a new PySparkNotebookBatch instance using the specified properties. + * @param [properties] Properties to set + * @returns PySparkNotebookBatch instance + */ + public static create(properties?: google.cloud.dataproc.v1.IPySparkNotebookBatch): google.cloud.dataproc.v1.PySparkNotebookBatch; + + /** + * Encodes the specified PySparkNotebookBatch message. Does not implicitly {@link google.cloud.dataproc.v1.PySparkNotebookBatch.verify|verify} messages. + * @param message PySparkNotebookBatch message or plain object to encode + * @param [writer] Writer to encode to + * @returns Writer + */ + public static encode(message: google.cloud.dataproc.v1.IPySparkNotebookBatch, writer?: $protobuf.Writer): $protobuf.Writer; + + /** + * Encodes the specified PySparkNotebookBatch message, length delimited. Does not implicitly {@link google.cloud.dataproc.v1.PySparkNotebookBatch.verify|verify} messages. + * @param message PySparkNotebookBatch message or plain object to encode + * @param [writer] Writer to encode to + * @returns Writer + */ + public static encodeDelimited(message: google.cloud.dataproc.v1.IPySparkNotebookBatch, writer?: $protobuf.Writer): $protobuf.Writer; + + /** + * Decodes a PySparkNotebookBatch message from the specified reader or buffer. + * @param reader Reader or buffer to decode from + * @param [length] Message length if known beforehand + * @returns PySparkNotebookBatch + * @throws {Error} If the payload is not a reader or valid buffer + * @throws {$protobuf.util.ProtocolError} If required fields are missing + */ + public static decode(reader: ($protobuf.Reader|Uint8Array), length?: number): google.cloud.dataproc.v1.PySparkNotebookBatch; + + /** + * Decodes a PySparkNotebookBatch message from the specified reader or buffer, length delimited. + * @param reader Reader or buffer to decode from + * @returns PySparkNotebookBatch + * @throws {Error} If the payload is not a reader or valid buffer + * @throws {$protobuf.util.ProtocolError} If required fields are missing + */ + public static decodeDelimited(reader: ($protobuf.Reader|Uint8Array)): google.cloud.dataproc.v1.PySparkNotebookBatch; + + /** + * Verifies a PySparkNotebookBatch message. + * @param message Plain object to verify + * @returns `null` if valid, otherwise the reason why it is not + */ + public static verify(message: { [k: string]: any }): (string|null); + + /** + * Creates a PySparkNotebookBatch message from a plain object. Also converts values to their respective internal types. + * @param object Plain object + * @returns PySparkNotebookBatch + */ + public static fromObject(object: { [k: string]: any }): google.cloud.dataproc.v1.PySparkNotebookBatch; + + /** + * Creates a plain object from a PySparkNotebookBatch message. Also converts values to other types if specified. + * @param message PySparkNotebookBatch + * @param [options] Conversion options + * @returns Plain object + */ + public static toObject(message: google.cloud.dataproc.v1.PySparkNotebookBatch, options?: $protobuf.IConversionOptions): { [k: string]: any }; + + /** + * Converts this PySparkNotebookBatch to JSON. + * @returns JSON object + */ + public toJSON(): { [k: string]: any }; + + /** + * Gets the default type url for PySparkNotebookBatch + * @param [typeUrlPrefix] your custom typeUrlPrefix(default "type.googleapis.com") + * @returns The default type url + */ + public static getTypeUrl(typeUrlPrefix?: string): string; + } + /** Properties of a RuntimeConfig. */ interface IRuntimeConfig { diff --git a/packages/google-cloud-dataproc/protos/protos.js b/packages/google-cloud-dataproc/protos/protos.js index 573c43086bca..6dbcef9c0aca 100644 --- a/packages/google-cloud-dataproc/protos/protos.js +++ b/packages/google-cloud-dataproc/protos/protos.js @@ -4348,6 +4348,7 @@ * @property {google.cloud.dataproc.v1.ISparkBatch|null} [sparkBatch] Batch sparkBatch * @property {google.cloud.dataproc.v1.ISparkRBatch|null} [sparkRBatch] Batch sparkRBatch * @property {google.cloud.dataproc.v1.ISparkSqlBatch|null} [sparkSqlBatch] Batch sparkSqlBatch + * @property {google.cloud.dataproc.v1.IPySparkNotebookBatch|null} [pysparkNotebookBatch] Batch pysparkNotebookBatch * @property {google.cloud.dataproc.v1.IRuntimeInfo|null} [runtimeInfo] Batch runtimeInfo * @property {google.cloud.dataproc.v1.Batch.State|null} [state] Batch state * @property {string|null} [stateMessage] Batch stateMessage @@ -4433,6 +4434,14 @@ */ Batch.prototype.sparkSqlBatch = null; + /** + * Batch pysparkNotebookBatch. + * @member {google.cloud.dataproc.v1.IPySparkNotebookBatch|null|undefined} pysparkNotebookBatch + * @memberof google.cloud.dataproc.v1.Batch + * @instance + */ + Batch.prototype.pysparkNotebookBatch = null; + /** * Batch runtimeInfo. * @member {google.cloud.dataproc.v1.IRuntimeInfo|null|undefined} runtimeInfo @@ -4518,12 +4527,12 @@ /** * Batch batchConfig. - * @member {"pysparkBatch"|"sparkBatch"|"sparkRBatch"|"sparkSqlBatch"|undefined} batchConfig + * @member {"pysparkBatch"|"sparkBatch"|"sparkRBatch"|"sparkSqlBatch"|"pysparkNotebookBatch"|undefined} batchConfig * @memberof google.cloud.dataproc.v1.Batch * @instance */ Object.defineProperty(Batch.prototype, "batchConfig", { - get: $util.oneOfGetter($oneOfFields = ["pysparkBatch", "sparkBatch", "sparkRBatch", "sparkSqlBatch"]), + get: $util.oneOfGetter($oneOfFields = ["pysparkBatch", "sparkBatch", "sparkRBatch", "sparkSqlBatch", "pysparkNotebookBatch"]), set: $util.oneOfSetter($oneOfFields) }); @@ -4587,6 +4596,8 @@ if (message.stateHistory != null && message.stateHistory.length) for (var i = 0; i < message.stateHistory.length; ++i) $root.google.cloud.dataproc.v1.Batch.StateHistory.encode(message.stateHistory[i], writer.uint32(/* id 17, wireType 2 =*/138).fork()).ldelim(); + if (message.pysparkNotebookBatch != null && Object.hasOwnProperty.call(message, "pysparkNotebookBatch")) + $root.google.cloud.dataproc.v1.PySparkNotebookBatch.encode(message.pysparkNotebookBatch, writer.uint32(/* id 19, wireType 2 =*/154).fork()).ldelim(); return writer; }; @@ -4651,6 +4662,10 @@ message.sparkSqlBatch = $root.google.cloud.dataproc.v1.SparkSqlBatch.decode(reader, reader.uint32()); break; } + case 19: { + message.pysparkNotebookBatch = $root.google.cloud.dataproc.v1.PySparkNotebookBatch.decode(reader, reader.uint32()); + break; + } case 8: { message.runtimeInfo = $root.google.cloud.dataproc.v1.RuntimeInfo.decode(reader, reader.uint32()); break; @@ -4797,6 +4812,16 @@ return "sparkSqlBatch." + error; } } + if (message.pysparkNotebookBatch != null && message.hasOwnProperty("pysparkNotebookBatch")) { + if (properties.batchConfig === 1) + return "batchConfig: multiple values"; + properties.batchConfig = 1; + { + var error = $root.google.cloud.dataproc.v1.PySparkNotebookBatch.verify(message.pysparkNotebookBatch); + if (error) + return "pysparkNotebookBatch." + error; + } + } if (message.runtimeInfo != null && message.hasOwnProperty("runtimeInfo")) { var error = $root.google.cloud.dataproc.v1.RuntimeInfo.verify(message.runtimeInfo); if (error) @@ -4900,6 +4925,11 @@ throw TypeError(".google.cloud.dataproc.v1.Batch.sparkSqlBatch: object expected"); message.sparkSqlBatch = $root.google.cloud.dataproc.v1.SparkSqlBatch.fromObject(object.sparkSqlBatch); } + if (object.pysparkNotebookBatch != null) { + if (typeof object.pysparkNotebookBatch !== "object") + throw TypeError(".google.cloud.dataproc.v1.Batch.pysparkNotebookBatch: object expected"); + message.pysparkNotebookBatch = $root.google.cloud.dataproc.v1.PySparkNotebookBatch.fromObject(object.pysparkNotebookBatch); + } if (object.runtimeInfo != null) { if (typeof object.runtimeInfo !== "object") throw TypeError(".google.cloud.dataproc.v1.Batch.runtimeInfo: object expected"); @@ -5065,6 +5095,11 @@ for (var j = 0; j < message.stateHistory.length; ++j) object.stateHistory[j] = $root.google.cloud.dataproc.v1.Batch.StateHistory.toObject(message.stateHistory[j], options); } + if (message.pysparkNotebookBatch != null && message.hasOwnProperty("pysparkNotebookBatch")) { + object.pysparkNotebookBatch = $root.google.cloud.dataproc.v1.PySparkNotebookBatch.toObject(message.pysparkNotebookBatch, options); + if (options.oneofs) + object.batchConfig = "pysparkNotebookBatch"; + } return object; }; @@ -6862,6 +6897,428 @@ return SparkSqlBatch; })(); + v1.PySparkNotebookBatch = (function() { + + /** + * Properties of a PySparkNotebookBatch. + * @memberof google.cloud.dataproc.v1 + * @interface IPySparkNotebookBatch + * @property {string|null} [notebookFileUri] PySparkNotebookBatch notebookFileUri + * @property {Object.|null} [params] PySparkNotebookBatch params + * @property {Array.|null} [pythonFileUris] PySparkNotebookBatch pythonFileUris + * @property {Array.|null} [jarFileUris] PySparkNotebookBatch jarFileUris + * @property {Array.|null} [fileUris] PySparkNotebookBatch fileUris + * @property {Array.|null} [archiveUris] PySparkNotebookBatch archiveUris + */ + + /** + * Constructs a new PySparkNotebookBatch. + * @memberof google.cloud.dataproc.v1 + * @classdesc Represents a PySparkNotebookBatch. + * @implements IPySparkNotebookBatch + * @constructor + * @param {google.cloud.dataproc.v1.IPySparkNotebookBatch=} [properties] Properties to set + */ + function PySparkNotebookBatch(properties) { + this.params = {}; + this.pythonFileUris = []; + this.jarFileUris = []; + this.fileUris = []; + this.archiveUris = []; + if (properties) + for (var keys = Object.keys(properties), i = 0; i < keys.length; ++i) + if (properties[keys[i]] != null) + this[keys[i]] = properties[keys[i]]; + } + + /** + * PySparkNotebookBatch notebookFileUri. + * @member {string} notebookFileUri + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.notebookFileUri = ""; + + /** + * PySparkNotebookBatch params. + * @member {Object.} params + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.params = $util.emptyObject; + + /** + * PySparkNotebookBatch pythonFileUris. + * @member {Array.} pythonFileUris + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.pythonFileUris = $util.emptyArray; + + /** + * PySparkNotebookBatch jarFileUris. + * @member {Array.} jarFileUris + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.jarFileUris = $util.emptyArray; + + /** + * PySparkNotebookBatch fileUris. + * @member {Array.} fileUris + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.fileUris = $util.emptyArray; + + /** + * PySparkNotebookBatch archiveUris. + * @member {Array.} archiveUris + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + */ + PySparkNotebookBatch.prototype.archiveUris = $util.emptyArray; + + /** + * Creates a new PySparkNotebookBatch instance using the specified properties. + * @function create + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {google.cloud.dataproc.v1.IPySparkNotebookBatch=} [properties] Properties to set + * @returns {google.cloud.dataproc.v1.PySparkNotebookBatch} PySparkNotebookBatch instance + */ + PySparkNotebookBatch.create = function create(properties) { + return new PySparkNotebookBatch(properties); + }; + + /** + * Encodes the specified PySparkNotebookBatch message. Does not implicitly {@link google.cloud.dataproc.v1.PySparkNotebookBatch.verify|verify} messages. + * @function encode + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {google.cloud.dataproc.v1.IPySparkNotebookBatch} message PySparkNotebookBatch message or plain object to encode + * @param {$protobuf.Writer} [writer] Writer to encode to + * @returns {$protobuf.Writer} Writer + */ + PySparkNotebookBatch.encode = function encode(message, writer) { + if (!writer) + writer = $Writer.create(); + if (message.notebookFileUri != null && Object.hasOwnProperty.call(message, "notebookFileUri")) + writer.uint32(/* id 1, wireType 2 =*/10).string(message.notebookFileUri); + if (message.params != null && Object.hasOwnProperty.call(message, "params")) + for (var keys = Object.keys(message.params), i = 0; i < keys.length; ++i) + writer.uint32(/* id 2, wireType 2 =*/18).fork().uint32(/* id 1, wireType 2 =*/10).string(keys[i]).uint32(/* id 2, wireType 2 =*/18).string(message.params[keys[i]]).ldelim(); + if (message.pythonFileUris != null && message.pythonFileUris.length) + for (var i = 0; i < message.pythonFileUris.length; ++i) + writer.uint32(/* id 3, wireType 2 =*/26).string(message.pythonFileUris[i]); + if (message.jarFileUris != null && message.jarFileUris.length) + for (var i = 0; i < message.jarFileUris.length; ++i) + writer.uint32(/* id 4, wireType 2 =*/34).string(message.jarFileUris[i]); + if (message.fileUris != null && message.fileUris.length) + for (var i = 0; i < message.fileUris.length; ++i) + writer.uint32(/* id 5, wireType 2 =*/42).string(message.fileUris[i]); + if (message.archiveUris != null && message.archiveUris.length) + for (var i = 0; i < message.archiveUris.length; ++i) + writer.uint32(/* id 6, wireType 2 =*/50).string(message.archiveUris[i]); + return writer; + }; + + /** + * Encodes the specified PySparkNotebookBatch message, length delimited. Does not implicitly {@link google.cloud.dataproc.v1.PySparkNotebookBatch.verify|verify} messages. + * @function encodeDelimited + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {google.cloud.dataproc.v1.IPySparkNotebookBatch} message PySparkNotebookBatch message or plain object to encode + * @param {$protobuf.Writer} [writer] Writer to encode to + * @returns {$protobuf.Writer} Writer + */ + PySparkNotebookBatch.encodeDelimited = function encodeDelimited(message, writer) { + return this.encode(message, writer).ldelim(); + }; + + /** + * Decodes a PySparkNotebookBatch message from the specified reader or buffer. + * @function decode + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {$protobuf.Reader|Uint8Array} reader Reader or buffer to decode from + * @param {number} [length] Message length if known beforehand + * @returns {google.cloud.dataproc.v1.PySparkNotebookBatch} PySparkNotebookBatch + * @throws {Error} If the payload is not a reader or valid buffer + * @throws {$protobuf.util.ProtocolError} If required fields are missing + */ + PySparkNotebookBatch.decode = function decode(reader, length, error) { + if (!(reader instanceof $Reader)) + reader = $Reader.create(reader); + var end = length === undefined ? reader.len : reader.pos + length, message = new $root.google.cloud.dataproc.v1.PySparkNotebookBatch(), key, value; + while (reader.pos < end) { + var tag = reader.uint32(); + if (tag === error) + break; + switch (tag >>> 3) { + case 1: { + message.notebookFileUri = reader.string(); + break; + } + case 2: { + if (message.params === $util.emptyObject) + message.params = {}; + var end2 = reader.uint32() + reader.pos; + key = ""; + value = ""; + while (reader.pos < end2) { + var tag2 = reader.uint32(); + switch (tag2 >>> 3) { + case 1: + key = reader.string(); + break; + case 2: + value = reader.string(); + break; + default: + reader.skipType(tag2 & 7); + break; + } + } + message.params[key] = value; + break; + } + case 3: { + if (!(message.pythonFileUris && message.pythonFileUris.length)) + message.pythonFileUris = []; + message.pythonFileUris.push(reader.string()); + break; + } + case 4: { + if (!(message.jarFileUris && message.jarFileUris.length)) + message.jarFileUris = []; + message.jarFileUris.push(reader.string()); + break; + } + case 5: { + if (!(message.fileUris && message.fileUris.length)) + message.fileUris = []; + message.fileUris.push(reader.string()); + break; + } + case 6: { + if (!(message.archiveUris && message.archiveUris.length)) + message.archiveUris = []; + message.archiveUris.push(reader.string()); + break; + } + default: + reader.skipType(tag & 7); + break; + } + } + return message; + }; + + /** + * Decodes a PySparkNotebookBatch message from the specified reader or buffer, length delimited. + * @function decodeDelimited + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {$protobuf.Reader|Uint8Array} reader Reader or buffer to decode from + * @returns {google.cloud.dataproc.v1.PySparkNotebookBatch} PySparkNotebookBatch + * @throws {Error} If the payload is not a reader or valid buffer + * @throws {$protobuf.util.ProtocolError} If required fields are missing + */ + PySparkNotebookBatch.decodeDelimited = function decodeDelimited(reader) { + if (!(reader instanceof $Reader)) + reader = new $Reader(reader); + return this.decode(reader, reader.uint32()); + }; + + /** + * Verifies a PySparkNotebookBatch message. + * @function verify + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {Object.} message Plain object to verify + * @returns {string|null} `null` if valid, otherwise the reason why it is not + */ + PySparkNotebookBatch.verify = function verify(message) { + if (typeof message !== "object" || message === null) + return "object expected"; + if (message.notebookFileUri != null && message.hasOwnProperty("notebookFileUri")) + if (!$util.isString(message.notebookFileUri)) + return "notebookFileUri: string expected"; + if (message.params != null && message.hasOwnProperty("params")) { + if (!$util.isObject(message.params)) + return "params: object expected"; + var key = Object.keys(message.params); + for (var i = 0; i < key.length; ++i) + if (!$util.isString(message.params[key[i]])) + return "params: string{k:string} expected"; + } + if (message.pythonFileUris != null && message.hasOwnProperty("pythonFileUris")) { + if (!Array.isArray(message.pythonFileUris)) + return "pythonFileUris: array expected"; + for (var i = 0; i < message.pythonFileUris.length; ++i) + if (!$util.isString(message.pythonFileUris[i])) + return "pythonFileUris: string[] expected"; + } + if (message.jarFileUris != null && message.hasOwnProperty("jarFileUris")) { + if (!Array.isArray(message.jarFileUris)) + return "jarFileUris: array expected"; + for (var i = 0; i < message.jarFileUris.length; ++i) + if (!$util.isString(message.jarFileUris[i])) + return "jarFileUris: string[] expected"; + } + if (message.fileUris != null && message.hasOwnProperty("fileUris")) { + if (!Array.isArray(message.fileUris)) + return "fileUris: array expected"; + for (var i = 0; i < message.fileUris.length; ++i) + if (!$util.isString(message.fileUris[i])) + return "fileUris: string[] expected"; + } + if (message.archiveUris != null && message.hasOwnProperty("archiveUris")) { + if (!Array.isArray(message.archiveUris)) + return "archiveUris: array expected"; + for (var i = 0; i < message.archiveUris.length; ++i) + if (!$util.isString(message.archiveUris[i])) + return "archiveUris: string[] expected"; + } + return null; + }; + + /** + * Creates a PySparkNotebookBatch message from a plain object. Also converts values to their respective internal types. + * @function fromObject + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {Object.} object Plain object + * @returns {google.cloud.dataproc.v1.PySparkNotebookBatch} PySparkNotebookBatch + */ + PySparkNotebookBatch.fromObject = function fromObject(object) { + if (object instanceof $root.google.cloud.dataproc.v1.PySparkNotebookBatch) + return object; + var message = new $root.google.cloud.dataproc.v1.PySparkNotebookBatch(); + if (object.notebookFileUri != null) + message.notebookFileUri = String(object.notebookFileUri); + if (object.params) { + if (typeof object.params !== "object") + throw TypeError(".google.cloud.dataproc.v1.PySparkNotebookBatch.params: object expected"); + message.params = {}; + for (var keys = Object.keys(object.params), i = 0; i < keys.length; ++i) + message.params[keys[i]] = String(object.params[keys[i]]); + } + if (object.pythonFileUris) { + if (!Array.isArray(object.pythonFileUris)) + throw TypeError(".google.cloud.dataproc.v1.PySparkNotebookBatch.pythonFileUris: array expected"); + message.pythonFileUris = []; + for (var i = 0; i < object.pythonFileUris.length; ++i) + message.pythonFileUris[i] = String(object.pythonFileUris[i]); + } + if (object.jarFileUris) { + if (!Array.isArray(object.jarFileUris)) + throw TypeError(".google.cloud.dataproc.v1.PySparkNotebookBatch.jarFileUris: array expected"); + message.jarFileUris = []; + for (var i = 0; i < object.jarFileUris.length; ++i) + message.jarFileUris[i] = String(object.jarFileUris[i]); + } + if (object.fileUris) { + if (!Array.isArray(object.fileUris)) + throw TypeError(".google.cloud.dataproc.v1.PySparkNotebookBatch.fileUris: array expected"); + message.fileUris = []; + for (var i = 0; i < object.fileUris.length; ++i) + message.fileUris[i] = String(object.fileUris[i]); + } + if (object.archiveUris) { + if (!Array.isArray(object.archiveUris)) + throw TypeError(".google.cloud.dataproc.v1.PySparkNotebookBatch.archiveUris: array expected"); + message.archiveUris = []; + for (var i = 0; i < object.archiveUris.length; ++i) + message.archiveUris[i] = String(object.archiveUris[i]); + } + return message; + }; + + /** + * Creates a plain object from a PySparkNotebookBatch message. Also converts values to other types if specified. + * @function toObject + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {google.cloud.dataproc.v1.PySparkNotebookBatch} message PySparkNotebookBatch + * @param {$protobuf.IConversionOptions} [options] Conversion options + * @returns {Object.} Plain object + */ + PySparkNotebookBatch.toObject = function toObject(message, options) { + if (!options) + options = {}; + var object = {}; + if (options.arrays || options.defaults) { + object.pythonFileUris = []; + object.jarFileUris = []; + object.fileUris = []; + object.archiveUris = []; + } + if (options.objects || options.defaults) + object.params = {}; + if (options.defaults) + object.notebookFileUri = ""; + if (message.notebookFileUri != null && message.hasOwnProperty("notebookFileUri")) + object.notebookFileUri = message.notebookFileUri; + var keys2; + if (message.params && (keys2 = Object.keys(message.params)).length) { + object.params = {}; + for (var j = 0; j < keys2.length; ++j) + object.params[keys2[j]] = message.params[keys2[j]]; + } + if (message.pythonFileUris && message.pythonFileUris.length) { + object.pythonFileUris = []; + for (var j = 0; j < message.pythonFileUris.length; ++j) + object.pythonFileUris[j] = message.pythonFileUris[j]; + } + if (message.jarFileUris && message.jarFileUris.length) { + object.jarFileUris = []; + for (var j = 0; j < message.jarFileUris.length; ++j) + object.jarFileUris[j] = message.jarFileUris[j]; + } + if (message.fileUris && message.fileUris.length) { + object.fileUris = []; + for (var j = 0; j < message.fileUris.length; ++j) + object.fileUris[j] = message.fileUris[j]; + } + if (message.archiveUris && message.archiveUris.length) { + object.archiveUris = []; + for (var j = 0; j < message.archiveUris.length; ++j) + object.archiveUris[j] = message.archiveUris[j]; + } + return object; + }; + + /** + * Converts this PySparkNotebookBatch to JSON. + * @function toJSON + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @instance + * @returns {Object.} JSON object + */ + PySparkNotebookBatch.prototype.toJSON = function toJSON() { + return this.constructor.toObject(this, $protobuf.util.toJSONOptions); + }; + + /** + * Gets the default type url for PySparkNotebookBatch + * @function getTypeUrl + * @memberof google.cloud.dataproc.v1.PySparkNotebookBatch + * @static + * @param {string} [typeUrlPrefix] your custom typeUrlPrefix(default "type.googleapis.com") + * @returns {string} The default type url + */ + PySparkNotebookBatch.getTypeUrl = function getTypeUrl(typeUrlPrefix) { + if (typeUrlPrefix === undefined) { + typeUrlPrefix = "type.googleapis.com"; + } + return typeUrlPrefix + "/google.cloud.dataproc.v1.PySparkNotebookBatch"; + }; + + return PySparkNotebookBatch; + })(); + v1.RuntimeConfig = (function() { /** diff --git a/packages/google-cloud-dataproc/protos/protos.json b/packages/google-cloud-dataproc/protos/protos.json index a1e23f9dad7a..317724581919 100644 --- a/packages/google-cloud-dataproc/protos/protos.json +++ b/packages/google-cloud-dataproc/protos/protos.json @@ -620,7 +620,8 @@ "pysparkBatch", "sparkBatch", "sparkRBatch", - "sparkSqlBatch" + "sparkSqlBatch", + "pysparkNotebookBatch" ] } }, @@ -674,6 +675,13 @@ "(google.api.field_behavior)": "OPTIONAL" } }, + "pysparkNotebookBatch": { + "type": "PySparkNotebookBatch", + "id": 19, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + }, "runtimeInfo": { "type": "RuntimeInfo", "id": 8, @@ -957,6 +965,57 @@ } } }, + "PySparkNotebookBatch": { + "fields": { + "notebookFileUri": { + "type": "string", + "id": 1, + "options": { + "(google.api.field_behavior)": "REQUIRED" + } + }, + "params": { + "keyType": "string", + "type": "string", + "id": 2, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + }, + "pythonFileUris": { + "rule": "repeated", + "type": "string", + "id": 3, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + }, + "jarFileUris": { + "rule": "repeated", + "type": "string", + "id": 4, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + }, + "fileUris": { + "rule": "repeated", + "type": "string", + "id": 5, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + }, + "archiveUris": { + "rule": "repeated", + "type": "string", + "id": 6, + "options": { + "(google.api.field_behavior)": "OPTIONAL" + } + } + } + }, "RuntimeConfig": { "fields": { "version": { diff --git a/packages/google-cloud-dataproc/samples/generated/v1/snippet_metadata_google.cloud.dataproc.v1.json b/packages/google-cloud-dataproc/samples/generated/v1/snippet_metadata_google.cloud.dataproc.v1.json index 97c601873750..6e8593086d41 100644 --- a/packages/google-cloud-dataproc/samples/generated/v1/snippet_metadata_google.cloud.dataproc.v1.json +++ b/packages/google-cloud-dataproc/samples/generated/v1/snippet_metadata_google.cloud.dataproc.v1.json @@ -1,7 +1,7 @@ { "clientLibrary": { "name": "nodejs-dataproc", - "version": "6.3.0", + "version": "0.1.0", "language": "TYPESCRIPT", "apis": [ { diff --git a/packages/google-cloud-dataproc/src/index.ts b/packages/google-cloud-dataproc/src/index.ts index 5db1abc6576a..49417fe6517c 100644 --- a/packages/google-cloud-dataproc/src/index.ts +++ b/packages/google-cloud-dataproc/src/index.ts @@ -1,4 +1,4 @@ -// Copyright 2026 Google LLC +// Copyright 2025 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. diff --git a/packages/google-cloud-dataproc/webpack.config.js b/packages/google-cloud-dataproc/webpack.config.js index e6ae2897925d..050b272c0e03 100644 --- a/packages/google-cloud-dataproc/webpack.config.js +++ b/packages/google-cloud-dataproc/webpack.config.js @@ -1,4 +1,4 @@ -// Copyright 2026 Google LLC +// Copyright 2021 Google LLC // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License.