From 6c6fe90e7c6b63ade2c8058aaed7a5b24050522d Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Sat, 24 Nov 2018 10:38:48 -0800 Subject: [PATCH] Generated from 21fe5c8871099a3b7b6493afe486f5ce5d3b64fc (#3377) typo: datalake-store/data-plane/Microsoft.DataLakeStore/filesystem - seperated -> separated - lexographical -> lexicographical - contet -> content --- .../store/2016-11-01/filesystem/filesystem.go | 9 ++++++--- .../2015-10-01-preview/filesystem/filesystem.go | 16 ++++++++++------ .../2015-10-01-preview/filesystem/models.go | 2 +- 3 files changed, 17 insertions(+), 10 deletions(-) diff --git a/services/datalake/store/2016-11-01/filesystem/filesystem.go b/services/datalake/store/2016-11-01/filesystem/filesystem.go index 12d6a5d091f4..a70017566bf5 100644 --- a/services/datalake/store/2016-11-01/filesystem/filesystem.go +++ b/services/datalake/store/2016-11-01/filesystem/filesystem.go @@ -44,7 +44,8 @@ func NewClient() Client { // Parameters: // accountName - the Azure Data Lake Store account to execute filesystem operations on. // pathParameter - the Data Lake Store path (starting with '/') of the file to which to append. -// streamContents - the file contents to include when appending to the file. +// streamContents - the file contents to include when appending to the file. The maximum content size is 4MB. +// For content larger than 4MB you must append the content in 4MB chunks. // offset - the optional offset in the stream to begin the append operation. Default is to append at the end of // the stream. // syncFlag - optionally indicates what to do after completion of the concurrent append. DATA indicates that @@ -333,7 +334,8 @@ func (client Client) ConcatResponder(resp *http.Response) (result autorest.Respo // accountName - the Azure Data Lake Store account to execute filesystem operations on. // pathParameter - the Data Lake Store path (starting with '/') of the file to which to append using concurrent // append. -// streamContents - the file contents to include when appending to the file. +// streamContents - the file contents to include when appending to the file. The maximum content size is 4MB. +// For content larger than 4MB you must append the content in 4MB chunks. // appendMode - indicates the concurrent append call should create the file if it doesn't exist or just open // the existing file for append // syncFlag - optionally indicates what to do after completion of the concurrent append. DATA indicates that @@ -435,7 +437,8 @@ func (client Client) ConcurrentAppendResponder(resp *http.Response) (result auto // accountName - the Azure Data Lake Store account to execute filesystem operations on. // pathParameter - the Data Lake Store path (starting with '/') of the file to create. // streamContents - the file contents to include when creating the file. This parameter is optional, resulting -// in an empty file if not specified. +// in an empty file if not specified. The maximum content size is 4MB. For content larger than 4MB you must +// append the content in 4MB chunks. // overwrite - the indication of if the file should be overwritten. // syncFlag - optionally indicates what to do after completion of the create. DATA indicates that more data // will be sent immediately by the client, the file handle should remain open/locked, and file metadata diff --git a/services/preview/datalake/store/2015-10-01-preview/filesystem/filesystem.go b/services/preview/datalake/store/2015-10-01-preview/filesystem/filesystem.go index 589d149e5345..4d76e000bc4c 100644 --- a/services/preview/datalake/store/2015-10-01-preview/filesystem/filesystem.go +++ b/services/preview/datalake/store/2015-10-01-preview/filesystem/filesystem.go @@ -44,7 +44,8 @@ func NewClient() Client { // Parameters: // accountName - the Azure Data Lake Store account to execute filesystem operations on. // directFilePath - the Data Lake Store path (starting with '/') of the file to which to append. -// streamContents - the file contents to include when appending to the file. +// streamContents - the file contents to include when appending to the file. The maximum content size is 4MB. +// For content larger than 4MB you must append the content in 4MB chunks. // offset - the optional offset in the stream to begin the append operation. Default is to append at the end of // the stream. func (client Client) Append(ctx context.Context, accountName string, directFilePath string, streamContents io.ReadCloser, offset *int64) (result autorest.Response, err error) { @@ -215,11 +216,12 @@ func (client Client) CheckAccessResponder(resp *http.Response) (result autorest. } // Concat concatenates the list of source files into the destination file, removing all source files upon success. +// This operation has been deprecated, please use operation FileSystem_MsConcat instead. // Parameters: // accountName - the Azure Data Lake Store account to execute filesystem operations on. // destinationPath - the Data Lake Store path (starting with '/') of the destination file resulting from the // concatenation. -// sources - a list of comma seperated Data Lake Store paths (starting with '/') of the files to concatenate, +// sources - a list of comma separated Data Lake Store paths (starting with '/') of the files to concatenate, // in the order in which they should be concatenated. func (client Client) Concat(ctx context.Context, accountName string, destinationPath string, sources []string) (result autorest.Response, err error) { if tracing.IsEnabled() { @@ -313,7 +315,8 @@ func (client Client) ConcatResponder(resp *http.Response) (result autorest.Respo // accountName - the Azure Data Lake Store account to execute filesystem operations on. // filePath - the Data Lake Store path (starting with '/') of the file to which to append using concurrent // append. -// streamContents - the file contents to include when appending to the file. +// streamContents - the file contents to include when appending to the file. The maximum content size is 4MB. +// For content larger than 4MB you must append the content in 4MB chunks. // appendMode - indicates the concurrent append call should create the file if it doesn't exist or just open // the existing file for append func (client Client) ConcurrentAppend(ctx context.Context, accountName string, filePath string, streamContents io.ReadCloser, appendMode AppendModeType) (result autorest.Response, err error) { @@ -403,7 +406,8 @@ func (client Client) ConcurrentAppendResponder(resp *http.Response) (result auto // accountName - the Azure Data Lake Store account to execute filesystem operations on. // directFilePath - the Data Lake Store path (starting with '/') of the file to create. // streamContents - the file contents to include when creating the file. This parameter is optional, resulting -// in an empty file if not specified. +// in an empty file if not specified. The maximum content size is 4MB. For content larger than 4MB you must +// append the content in 4MB chunks. // overwrite - the indication of if the file should be overwritten. func (client Client) Create(ctx context.Context, accountName string, directFilePath string, streamContents io.ReadCloser, overwrite *bool) (result autorest.Response, err error) { if tracing.IsEnabled() { @@ -909,10 +913,10 @@ func (client Client) GetFileStatusResponder(resp *http.Response) (result FileSta // accountName - the Azure Data Lake Store account to execute filesystem operations on. // listFilePath - the Data Lake Store path (starting with '/') of the directory to list. // listSize - gets or sets the number of items to return. Optional. -// listAfter - gets or sets the item or lexographical index after which to begin returning results. For +// listAfter - gets or sets the item or lexicographical index after which to begin returning results. For // example, a file list of 'a','b','d' and listAfter='b' will return 'd', and a listAfter='c' will also return // 'd'. Optional. -// listBefore - gets or sets the item or lexographical index before which to begin returning results. For +// listBefore - gets or sets the item or lexicographical index before which to begin returning results. For // example, a file list of 'a','b','d' and listBefore='d' will return 'a','b', and a listBefore='c' will also // return 'a','b'. Optional. func (client Client) ListFileStatus(ctx context.Context, accountName string, listFilePath string, listSize *int32, listAfter string, listBefore string) (result FileStatusesResult, err error) { diff --git a/services/preview/datalake/store/2015-10-01-preview/filesystem/models.go b/services/preview/datalake/store/2015-10-01-preview/filesystem/models.go index 57cfa6a40667..341be7f8097c 100644 --- a/services/preview/datalake/store/2015-10-01-preview/filesystem/models.go +++ b/services/preview/datalake/store/2015-10-01-preview/filesystem/models.go @@ -1063,7 +1063,7 @@ type ContentSummary struct { DirectoryCount *int64 `json:"directoryCount,omitempty"` // FileCount - the number of files. FileCount *int64 `json:"fileCount,omitempty"` - // Length - the number of bytes used by the contet. + // Length - the number of bytes used by the content. Length *int64 `json:"length,omitempty"` // SpaceConsumed - the disk space consumed by the content. SpaceConsumed *int64 `json:"spaceConsumed,omitempty"`