Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
Original file line number Diff line number Diff line change
@@ -1,6 +1,8 @@
# Release History

## 1.0.0-beta.31 (Unreleased)
## 1.0.0-beta.1 (2024-10-01)

- Azure Resource Manager DataFactory client library for Java. This package contains Microsoft Azure SDK for DataFactory Management SDK. The Azure Data Factory V2 management API provides a RESTful set of web services that interact with Azure Data Factory V2 services. Package tag package-2018-06. For documentation on how to use this package, please see [Azure Management Libraries for Java](https://aka.ms/azsdk/java/mgmt).

### Features Added

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ Various documentation is available to help you get started
<dependency>
<groupId>com.azure.resourcemanager</groupId>
<artifactId>azure-resourcemanager-datafactory</artifactId>
<version>1.0.0-beta.30</version>
<version>1.0.0-beta.31</version>
</dependency>
```
[//]: # ({x-version-update-end})
Expand Down
17 changes: 6 additions & 11 deletions sdk/datafactory/azure-resourcemanager-datafactory/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -46,13 +46,9 @@
<jacoco.min.linecoverage>0</jacoco.min.linecoverage>
<jacoco.min.branchcoverage>0</jacoco.min.branchcoverage>
<revapi.skip>true</revapi.skip>
<spotless.skip>false</spotless.skip>
</properties>
<dependencies>
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-json</artifactId>
<version>1.3.0</version> <!-- {x-version-update;com.azure:azure-json;dependency} -->
</dependency>
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-core</artifactId>
Expand All @@ -63,17 +59,16 @@
<artifactId>azure-core-management</artifactId>
<version>1.15.4</version> <!-- {x-version-update;com.azure:azure-core-management;dependency} -->
</dependency>
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-json</artifactId>
<version>1.3.0</version> <!-- {x-version-update;com.azure:azure-json;dependency} -->
</dependency>
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-core-test</artifactId>
<version>1.27.0-beta.2</version> <!-- {x-version-update;com.azure:azure-core-test;dependency} -->
<scope>test</scope>
<exclusions>
<exclusion>
<groupId>com.azure</groupId>
<artifactId>azure-core-http-jdk-httpclient</artifactId>
</exclusion>
</exclusions>
</dependency>
<dependency>
<groupId>com.azure</groupId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,8 +11,8 @@
import com.azure.core.http.HttpPipelinePosition;
import com.azure.core.http.policy.AddDatePolicy;
import com.azure.core.http.policy.AddHeadersFromContextPolicy;
import com.azure.core.http.policy.HttpLoggingPolicy;
import com.azure.core.http.policy.HttpLogOptions;
import com.azure.core.http.policy.HttpLoggingPolicy;
import com.azure.core.http.policy.HttpPipelinePolicy;
import com.azure.core.http.policy.HttpPolicyProviders;
import com.azure.core.http.policy.RequestIdPolicy;
Expand Down Expand Up @@ -43,8 +43,8 @@
import com.azure.resourcemanager.datafactory.implementation.OperationsImpl;
import com.azure.resourcemanager.datafactory.implementation.PipelineRunsImpl;
import com.azure.resourcemanager.datafactory.implementation.PipelinesImpl;
import com.azure.resourcemanager.datafactory.implementation.PrivateEndpointConnectionOperationsImpl;
import com.azure.resourcemanager.datafactory.implementation.PrivateEndPointConnectionsImpl;
import com.azure.resourcemanager.datafactory.implementation.PrivateEndpointConnectionOperationsImpl;
import com.azure.resourcemanager.datafactory.implementation.PrivateLinkResourcesImpl;
import com.azure.resourcemanager.datafactory.implementation.TriggerRunsImpl;
import com.azure.resourcemanager.datafactory.implementation.TriggersImpl;
Expand All @@ -66,8 +66,8 @@
import com.azure.resourcemanager.datafactory.models.Operations;
import com.azure.resourcemanager.datafactory.models.PipelineRuns;
import com.azure.resourcemanager.datafactory.models.Pipelines;
import com.azure.resourcemanager.datafactory.models.PrivateEndpointConnectionOperations;
import com.azure.resourcemanager.datafactory.models.PrivateEndPointConnections;
import com.azure.resourcemanager.datafactory.models.PrivateEndpointConnectionOperations;
import com.azure.resourcemanager.datafactory.models.PrivateLinkResources;
import com.azure.resourcemanager.datafactory.models.TriggerRuns;
import com.azure.resourcemanager.datafactory.models.Triggers;
Expand Down Expand Up @@ -294,7 +294,7 @@ public DataFactoryManager authenticate(TokenCredential credential, AzureProfile
.append("-")
.append("com.azure.resourcemanager.datafactory")
.append("/")
.append("1.0.0-beta.30");
.append("1.0.0-beta.1");
if (!Configuration.getGlobalConfiguration().get("AZURE_TELEMETRY_DISABLED", false)) {
userAgentBuilder.append(" (")
.append(Configuration.getGlobalConfiguration().get("java.version"))
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -363,7 +363,6 @@ public AmazonRdsForSqlServerLinkedServiceTypeProperties withPooling(Object pooli
*/
@Override
public void validate() {
super.validate();
if (password() != null) {
password().validate();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -542,7 +542,6 @@ public AzureSqlDWLinkedServiceTypeProperties withPooling(Object pooling) {
*/
@Override
public void validate() {
super.validate();
if (password() != null) {
password().validate();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -569,7 +569,6 @@ public AzureSqlDatabaseLinkedServiceTypeProperties withPooling(Object pooling) {
*/
@Override
public void validate() {
super.validate();
if (password() != null) {
password().validate();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -568,7 +568,6 @@ public AzureSqlMILinkedServiceTypeProperties withPooling(Object pooling) {
*/
@Override
public void validate() {
super.validate();
if (password() != null) {
password().validate();
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -128,10 +128,15 @@ public AzureTableStorageLinkedServiceTypeProperties withEncryptedCredential(Stri
*/
@Override
public void validate() {
super.validate();
if (credential() != null) {
credential().validate();
}
if (accountKey() != null) {
accountKey().validate();
}
if (sasToken() != null) {
sasToken().validate();
}
}

/**
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@
package com.azure.resourcemanager.datafactory.fluent.models;

import com.azure.core.annotation.Fluent;
import com.azure.core.util.logging.ClientLogger;
import com.azure.json.JsonReader;
import com.azure.json.JsonToken;
import com.azure.json.JsonWriter;
Expand Down Expand Up @@ -172,7 +173,6 @@ public ExecutePowerQueryActivityTypeProperties withSourceStagingConcurrency(Obje
*/
@Override
public void validate() {
super.validate();
if (sinks() != null) {
sinks().values().forEach(e -> {
if (e != null) {
Expand All @@ -183,8 +183,29 @@ public void validate() {
if (queries() != null) {
queries().forEach(e -> e.validate());
}
if (dataFlow() == null) {
throw LOGGER.atError()
.log(new IllegalArgumentException(
"Missing required property dataFlow in model ExecutePowerQueryActivityTypeProperties"));
} else {
dataFlow().validate();
}
if (staging() != null) {
staging().validate();
}
if (integrationRuntime() != null) {
integrationRuntime().validate();
}
if (continuationSettings() != null) {
continuationSettings().validate();
}
if (compute() != null) {
compute().validate();
}
}

private static final ClientLogger LOGGER = new ClientLogger(ExecutePowerQueryActivityTypeProperties.class);

/**
* {@inheritDoc}
*/
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -46,19 +46,19 @@ public final class FactoryInner extends Resource {
private Map<String, Object> additionalProperties;

/*
* Fully qualified resource Id for the resource.
* The type of the resource.
*/
private String id;
private String type;

/*
* The name of the resource.
*/
private String name;

/*
* The type of the resource.
* Fully qualified resource Id for the resource.
*/
private String type;
private String id;

/**
* Creates an instance of FactoryInner class.
Expand Down Expand Up @@ -125,13 +125,13 @@ public FactoryInner withAdditionalProperties(Map<String, Object> additionalPrope
}

/**
* Get the id property: Fully qualified resource Id for the resource.
* Get the type property: The type of the resource.
*
* @return the id value.
* @return the type value.
*/
@Override
public String id() {
return this.id;
public String type() {
return this.type;
}

/**
Expand All @@ -145,13 +145,13 @@ public String name() {
}

/**
* Get the type property: The type of the resource.
* Get the id property: Fully qualified resource Id for the resource.
*
* @return the type value.
* @return the id value.
*/
@Override
public String type() {
return this.type;
public String id() {
return this.id;
}

/**
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,107 @@
// Copyright (c) Microsoft Corporation. All rights reserved.
// Licensed under the MIT License.
// Code generated by Microsoft (R) AutoRest Code Generator.

package com.azure.resourcemanager.datafactory.fluent.models;

import com.azure.core.annotation.Fluent;
import com.azure.core.util.logging.ClientLogger;
import com.azure.json.JsonReader;
import com.azure.json.JsonSerializable;
import com.azure.json.JsonToken;
import com.azure.json.JsonWriter;
import com.azure.resourcemanager.datafactory.models.DatasetLocation;
import java.io.IOException;

/**
* Iceberg dataset properties.
*/
@Fluent
public final class IcebergDatasetTypeProperties implements JsonSerializable<IcebergDatasetTypeProperties> {
/*
* The location of the iceberg storage. Setting a file name is not allowed for iceberg format.
*/
private DatasetLocation location;

/**
* Creates an instance of IcebergDatasetTypeProperties class.
*/
public IcebergDatasetTypeProperties() {
}

/**
* Get the location property: The location of the iceberg storage. Setting a file name is not allowed for iceberg
* format.
*
* @return the location value.
*/
public DatasetLocation location() {
return this.location;
}

/**
* Set the location property: The location of the iceberg storage. Setting a file name is not allowed for iceberg
* format.
*
* @param location the location value to set.
* @return the IcebergDatasetTypeProperties object itself.
*/
public IcebergDatasetTypeProperties withLocation(DatasetLocation location) {
this.location = location;
return this;
}

/**
* Validates the instance.
*
* @throws IllegalArgumentException thrown if the instance is not valid.
*/
public void validate() {
if (location() == null) {
throw LOGGER.atError()
.log(new IllegalArgumentException(
"Missing required property location in model IcebergDatasetTypeProperties"));
} else {
location().validate();
}
}

private static final ClientLogger LOGGER = new ClientLogger(IcebergDatasetTypeProperties.class);

/**
* {@inheritDoc}
*/
@Override
public JsonWriter toJson(JsonWriter jsonWriter) throws IOException {
jsonWriter.writeStartObject();
jsonWriter.writeJsonField("location", this.location);
return jsonWriter.writeEndObject();
}

/**
* Reads an instance of IcebergDatasetTypeProperties from the JsonReader.
*
* @param jsonReader The JsonReader being read.
* @return An instance of IcebergDatasetTypeProperties if the JsonReader was pointing to an instance of it, or null
* if it was pointing to JSON null.
* @throws IllegalStateException If the deserialized JSON object was missing any required properties.
* @throws IOException If an error occurs while reading the IcebergDatasetTypeProperties.
*/
public static IcebergDatasetTypeProperties fromJson(JsonReader jsonReader) throws IOException {
return jsonReader.readObject(reader -> {
IcebergDatasetTypeProperties deserializedIcebergDatasetTypeProperties = new IcebergDatasetTypeProperties();
while (reader.nextToken() != JsonToken.END_OBJECT) {
String fieldName = reader.getFieldName();
reader.nextToken();

if ("location".equals(fieldName)) {
deserializedIcebergDatasetTypeProperties.location = DatasetLocation.fromJson(reader);
} else {
reader.skipChildren();
}
}

return deserializedIcebergDatasetTypeProperties;
});
}
}
Loading