Skip to content
This repository was archived by the owner on Nov 16, 2023. It is now read-only.
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
158 commits
Select commit Hold shift + click to select a range
fcc1924
Update readme with latest feedback (#39)
GalOshri Oct 31, 2018
e100472
Add THIRD-PARTY-NOTICES.txt and move CONTRIBUTING.md to root. (#40)
montebhoover Oct 31, 2018
8239922
Create CODE_OF_CONDUCT.md
ganik Nov 1, 2018
ad0af7c
Update issue templates
ganik Nov 1, 2018
19f4721
Create PULL_REQUEST_TEMPLATE.md
ganik Nov 1, 2018
1e16e64
Update issue templates
ganik Nov 1, 2018
007e624
Update issue templates
ganik Nov 1, 2018
110b0f9
Update issue templates
ganik Nov 1, 2018
0b5adef
Fixing link in CONTRIBUTING.md (#44)
justinormont Nov 2, 2018
a2ba6f5
Update contributing.md link. (#43)
montehoover Nov 4, 2018
45be3d7
Merge pull request #1 from Microsoft/master
ganik Nov 9, 2018
243325d
Initial checkin for ML.NET 0.7 upgrade
ganik Nov 9, 2018
cbfb439
fix tests
ganik Nov 9, 2018
653d8c1
put back columndropper
ganik Nov 10, 2018
1ae3060
fix tests
ganik Nov 10, 2018
10bd895
Update scikit-learn links to use https instead of http
Nov 19, 2018
bec566c
Merge pull request #56 from GalOshri/update-scikit-learn-https
justinormont Nov 20, 2018
725be2e
Merge pull request #2 from Microsoft/master
ganik Nov 23, 2018
f74b3c8
restart dotnetcore2 package work
ganik Nov 23, 2018
a9684bc
fix build
ganik Nov 23, 2018
0976828
fix mac & linux
ganik Nov 24, 2018
620d13d
fix build
ganik Nov 24, 2018
3e10cec
fix build
ganik Nov 24, 2018
04e87b7
dbg build
ganik Nov 24, 2018
98c8987
fix build
ganik Nov 24, 2018
d2e815f
fix build
ganik Nov 24, 2018
34c5f29
handle py 2.7
ganik Nov 24, 2018
bbb4c63
handle py27
ganik Nov 24, 2018
64da211
fix py27
ganik Nov 24, 2018
7ea0a25
fix build
ganik Nov 25, 2018
55308ec
fix build
ganik Nov 25, 2018
577d84e
fix build
ganik Nov 25, 2018
b571d22
ensure dependencies
ganik Nov 25, 2018
062d55a
ignore exceptions from ensure dependencies
ganik Nov 25, 2018
ace3cc9
Merge pull request #60 from ganik/ganik/dotnet2
ganik Nov 26, 2018
ab3d80d
up version
ganik Nov 26, 2018
36b4f48
Merge pull request #3 from Microsoft/master
ganik Nov 26, 2018
c2ce774
Merge branch 'master' into ganik/dotnet2
ganik Nov 26, 2018
d912ca5
Merge pull request #61 from ganik/ganik/dotnet2
ganik Nov 26, 2018
9fd5c3c
Update cv.py
zyw400 Nov 27, 2018
7c58875
Update cv.py
zyw400 Nov 27, 2018
1d02fc3
add a test for cv with data frame
zyw400 Nov 27, 2018
422bd8d
set DOTNET_SYSTEM_GLOBALIZATION_INVARIANT to true to fix app domain e…
ganik Nov 28, 2018
c6704fd
Merge pull request #62 from zyw400/yiwzh/fix_cv
ganik Nov 28, 2018
9d3376c
fix build
ganik Nov 28, 2018
83db226
Merge pull request #4 from Microsoft/master
ganik Nov 28, 2018
e54535d
Merge branch 'master' into ganik/amldocker
ganik Nov 28, 2018
4c68428
up version
ganik Nov 28, 2018
52ad2d0
Merge pull request #63 from ganik/ganik/amldocker
ganik Nov 28, 2018
341e01a
Add instructions for editing docstrings. (#51)
montebhoover Dec 6, 2018
9a0b50e
Fix build failures caused by dotnetcore2 module. (#67)
montebhoover Dec 7, 2018
0d2e4e6
Reduce number of build legs for PR validations and add nightly build …
montebhoover Dec 7, 2018
f7b7ded
Merge pull request #5 from Microsoft/master
ganik Dec 12, 2018
19b240e
Merge branch 'master' into ganik/mlnet.7
ganik Dec 12, 2018
b45a953
Increase version to 0.6.5. (#71)
montebhoover Dec 12, 2018
f3eb0bb
Update clr helper function to search multiple folders for clr binarie…
montebhoover Dec 15, 2018
155696c
fix drop column param name
ganik Dec 16, 2018
c94568d
Merge pull request #6 from Microsoft/master
ganik Dec 16, 2018
453a940
Merge branch 'master' into ganik/mlnet.7
ganik Dec 16, 2018
f95b3ba
Remove restricted permissions on build.sh script.
Dec 18, 2018
7a5e6d9
Fix lightgbm test failures by updating runtime dependencies.
montebhoover Dec 18, 2018
7a46ce1
fix TensorFlowScorer model_location paramter name
ganik Dec 18, 2018
7b7692c
Fix build.sh defaults so that it detects when running on a mac.
Dec 18, 2018
492751f
Since OneHotHashVectorizer is broken for output kind Key in ML.NET 0.…
ganik Dec 20, 2018
66cb189
Merge pull request #75 from montebhoover/improve_mac_ux
justinormont Dec 20, 2018
eb2b39f
fix tests
ganik Dec 20, 2018
d71a31e
Merge pull request #7 from Microsoft/master
ganik Dec 20, 2018
27d4a6a
Merge branch 'master' into ganik/mlnet.7
ganik Dec 20, 2018
af76d08
Merge pull request #8 from ganik/master
ganik Dec 20, 2018
c779510
fix pyproj test
ganik Dec 20, 2018
2bdfa41
Merge branch 'ganik/mlnet.7' of https://github.com/ganik/NimbusML int…
ganik Dec 20, 2018
d23d696
fix win 3.6 build
ganik Dec 20, 2018
172c1e8
fix comments
ganik Dec 20, 2018
80ce48f
Merge pull request #55 from ganik/ganik/mlnet.7
ganik Dec 20, 2018
b5f1c2e
Merge pull request #1 from Microsoft/master
zyw400 Jan 4, 2019
bfaf819
expose "parallel" to the fit/fit_transform function by including **pa…
zyw400 Jan 5, 2019
eaeb24c
add a test for the parallel
zyw400 Jan 5, 2019
a5997db
update parallel thread
zyw400 Jan 7, 2019
67530ff
fix tests comparison
zyw400 Jan 7, 2019
066469f
Update thread, retry build
zyw400 Jan 7, 2019
a9596ca
modify tests
zyw400 Jan 7, 2019
13d7b35
specify pytest-cov version
zyw400 Jan 7, 2019
af577c4
update pytest-cov version in build command for linux
zyw400 Jan 7, 2019
4dc79e1
for windows use the latest pytest-cov
zyw400 Jan 7, 2019
d2535be
Merge pull request #86 from zyw400/yiwzh/add_nthreads_to_graph
zyw400 Jan 8, 2019
3079d56
Enabled strong naming for DoNetBridge.dll (to be used for InternalsVi…
Jan 8, 2019
a556f39
Changed the keys to be the same as other internal repos
Jan 8, 2019
0fd4f0e
Changed the key filename
Jan 8, 2019
4f7f22b
Merge branch 'master' into strongname
Jan 9, 2019
b0c1e3a
Merge pull request #87 from shmoradims/strongname
Jan 9, 2019
9e57f19
Update to ML.NET 0.10.preview (#77)
montebhoover Jan 16, 2019
7c9a1c6
Simplify by using six.string_types (#89)
cclauss Jan 18, 2019
e5f2b65
Removed ISchema from DotNetBridge (#90)
Jan 24, 2019
dca1157
add configuration for python 3.7 (#101)
xadupre Apr 11, 2019
3616e73
Removing 3.7 for now as its not in PyPI
ganik May 7, 2019
210b220
Upgrade to ML.NET version 1.0.0 (#100)
ganik May 27, 2019
b5eb937
Fix latest Windows build issues. (#105)
pieths May 27, 2019
c35536d
Fixes #50 - summary() fails if called a second time. (#107)
pieths May 30, 2019
8da35e1
Fixes #99. Do not use hardcoded file separator. (#108)
pieths May 30, 2019
b4ec723
Delete the cached summaries when refitting a pipeline or a predictor.…
pieths Jun 1, 2019
91478d1
Fix signature import error when using latest version of scikit-learn.…
pieths Jun 2, 2019
a580331
Package System.Drawing.Common.dll as its missing in dotnetcore2 (#120)
ganik Jun 4, 2019
7848487
Upgrade the pytest-remotedata package to fix missing attribute error.…
pieths Jun 4, 2019
32e2d67
Upgrade version (#122)
ganik Jun 4, 2019
d09a5c5
Support quoted strings by default (#124)
ganik Jun 4, 2019
b57cfcc
upgrade to ML.NET 1.1 (#126)
ganik Jun 5, 2019
b4931e4
Put long running tests in to their own folder to shorten build times.…
pieths Jun 13, 2019
7863ca0
Expose ML.NET SSA & IID spike & changepoint detectors. (#135)
pieths Jun 14, 2019
3c689c6
Fix a few minor issues with time series unit tests and examples. (#139)
pieths Jun 18, 2019
207a6b6
Skip Image.py and Image_df.py tests for Ubuntu 14 (#149)
Stephen0620 Jun 18, 2019
0ca2b29
* Fixed the script for generating the documentation (#144)
Stephen0620 Jun 18, 2019
3b46629
Rename time_series package to timeseries. (#150)
pieths Jun 18, 2019
19b27f0
Fixed the issue of Ubuntu14 not skipping Image.py and Image_df.py (#161)
Stephen0620 Jun 28, 2019
c5153c2
Updated CharTokenizer.py example (#153)
Stephen0620 Jun 28, 2019
c45edfe
Skip CharTokenizer.py for extended tests (#163)
Stephen0620 Jul 1, 2019
7893bfd
Add support for returning custom values when overriding Pipeline.pred…
pieths Jul 1, 2019
c4b26d9
Initial creation of the release-next.md file. (#165)
pieths Jul 1, 2019
3993365
Initial implementation of the SsaForecaster entry point. (#164)
pieths Jul 2, 2019
29af47a
Final updates for release 1.2.0 (#167)
pieths Jul 3, 2019
4822871
Revert change b5eb9376dd14da606e91f7f94f1bec7b7609a7a1 to see if it (…
pieths Jul 3, 2019
a2c3e1f
Bring back build.cmd commit. It did not fix the signed build issue. (…
pieths Jul 3, 2019
8bb0c0c
Bring back the build.cmd change from b5eb9376dd14da606e91f7f94f1bec7b…
pieths Jul 3, 2019
4dddfda
Use restored dotnet CLI for signing (#171)
safern Jul 3, 2019
8da13e7
Update README.md
ganik Jul 4, 2019
08d8abf
Enable LinearSvmBinaryClassifier (#180)
najeeb-kazmi Jul 11, 2019
ab27816
Setup destructors for data passed to python (#184)
ganik Jul 12, 2019
c2f2b6b
Add azureml-dataprep support for dataflow objects (#181)
ganik Jul 12, 2019
4395c12
up version (#188)
ganik Jul 13, 2019
c0500d1
Save the model file when pickling a NimbusML Pipeline. (#189)
pieths Jul 18, 2019
266d27d
Remove stored references to X and y in BasePredictor. (#195)
pieths Jul 18, 2019
426fffe
Add observation level feature contributions to Pipeline and BasePredi…
najeeb-kazmi Jul 19, 2019
417bb35
Update release-next.md
najeeb-kazmi Jul 19, 2019
a36a6c0
Add classes_ to Pipeline and/or predictor when calling predict_proba.…
pieths Jul 25, 2019
5306833
Update Handler, Filter, and Indicator to automatically convert the in…
pieths Jul 31, 2019
1f97c9e
Combine models from transforms, predictors and pipelines in to one mo…
pieths Aug 4, 2019
47f8984
Fix build (#209)
ganik Aug 4, 2019
bea821e
Update release-next.md. (#211)
pieths Aug 5, 2019
68f9be1
Update release-next.md
ganik Aug 5, 2019
c4ebe0f
Update release-next.md
najeeb-kazmi Aug 5, 2019
51bdff2
Update release-next.md
najeeb-kazmi Aug 5, 2019
c655aad
Add classifier and FileDataStream unit tests to test_pipeline_combini…
pieths Aug 5, 2019
9dd9c11
Update release-next.md
najeeb-kazmi Aug 6, 2019
0458160
up version (#210)
ganik Aug 6, 2019
e257cf3
Enable EnsembleClassifier and EnsembleRegressor (#207)
najeeb-kazmi Aug 6, 2019
ecf456b
Create release notes for version 1.3.0. (#214)
pieths Aug 6, 2019
a3051aa
Update release-1.3.0.md
najeeb-kazmi Aug 6, 2019
ee136ff
Add --installPythonPackages flag to build scripts (#215)
najeeb-kazmi Aug 8, 2019
13844cc
Fix a bug with the classes_ attribute when no y input is specified du…
pieths Aug 8, 2019
8fa5878
Add NumSharp.Core.dll (#220)
ganik Aug 8, 2019
6f7cb41
Add timeseries documentation to the master branch. (#221)
pieths Aug 8, 2019
e348250
Docs update (#224)
najeeb-kazmi Aug 13, 2019
9c7c096
More doc fixes (#228)
najeeb-kazmi Aug 15, 2019
2baa87e
Pass python path to Dprep (#232)
ganik Aug 19, 2019
b08ea7b
Merge branch 'master' into temp/docs
Aug 19, 2019
9e38c5f
Remove all underscore files which are not getting recognized as renamed.
Aug 19, 2019
7d0acbf
Rename the files to start with an underscore character.
Aug 19, 2019
02623ce
Update the previously renamed files with underscores where required.
Aug 19, 2019
aca2d7a
Finish merge of nimbusml.pyproj from master.
Aug 19, 2019
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion build.cmd
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if /i [%1] == [--azureBuild] (
) else goto :Usage

:Usage
echo "Usage: build.cmd [--configuration <Configuration>] [--runTests] [--includeExtendedTests] [--buildDotNetBridgeOnly] [--skipDotNetBridge] [--azureBuild]"
echo "Usage: build.cmd [--configuration <Configuration>] [--runTests] [--installPythonPackages] [--includeExtendedTests] [--buildDotNetBridgeOnly] [--skipDotNetBridge] [--azureBuild]"
echo ""
echo "Options:"
echo " --configuration <Configuration> Build Configuration (DbgWinPy3.7,DbgWinPy3.6,DbgWinPy3.5,DbgWinPy2.7,RlsWinPy3.7,RlsWinPy3.6,RlsWinPy3.5,RlsWinPy2.7)"
Expand Down
2 changes: 1 addition & 1 deletion build.sh
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ mkdir -p "${DependenciesDir}"

usage()
{
echo "Usage: $0 --configuration <Configuration> [--runTests] [--includeExtendedTests]"
echo "Usage: $0 --configuration <Configuration> [--runTests] [--includeExtendedTests] [--installPythonPackages]"
echo ""
echo "Options:"
echo " --configuration <Configuration> Build Configuration (DbgLinPy3.7,DbgLinPy3.6,DbgLinPy3.5,DbgLinPy2.7,RlsLinPy3.7,RlsLinPy3.6,RlsLinPy3.5,RlsLinPy2.7,DbgMacPy3.7,DbgMacPy3.6,DbgMacPy3.5,DbgMacPy2.7,RlsMacPy3.7,RlsMacPy3.6,RlsMacPy3.5,RlsMacPy2.7)"
Expand Down
4 changes: 4 additions & 0 deletions src/DotNetBridge/Bridge.cs
Original file line number Diff line number Diff line change
Expand Up @@ -242,6 +242,10 @@ private struct EnvironmentBlock
// Call back to provide cancel flag.
[FieldOffset(0x28)]
public readonly void* checkCancel;

// Path to python executable.
[FieldOffset(0x30)]
public readonly sbyte* pythonPath;
#pragma warning restore 649 // never assigned
}

Expand Down
2 changes: 1 addition & 1 deletion src/DotNetBridge/DotNetBridge.csproj
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@
<PackageReference Include="Microsoft.ML.Dnn" Version="0.15.1" />
<PackageReference Include="Microsoft.ML.Ensemble" Version="0.15.1" />
<PackageReference Include="Microsoft.ML.TimeSeries" Version="1.3.1" />
<PackageReference Include="Microsoft.DataPrep" Version="0.0.1.5-preview" />
<PackageReference Include="Microsoft.DataPrep" Version="0.0.1.12-preview" />
<PackageReference Include="TensorFlow.NET" Version="0.10.10" />
<PackageReference Include="SciSharp.TensorFlow.Redist" Version="1.14.0" />
</ItemGroup>
Expand Down
5 changes: 4 additions & 1 deletion src/DotNetBridge/RunGraph.cs
Original file line number Diff line number Diff line change
Expand Up @@ -147,8 +147,11 @@ private static void RunGraphCore(EnvironmentBlock* penv, IHostEnvironment env, s
var extension = Path.GetExtension(path);
if (extension == ".txt")
dv = TextLoader.LoadFile(host, new TextLoader.Options(), new MultiFileSource(path));
else if(extension == ".dprep")
else if (extension == ".dprep")
{
DPrepSettings.Instance.PythonPath = BytesToString(penv->pythonPath);
dv = DataFlow.FromDPrepFile(path).ToDataView();
}
else
dv = new BinaryLoader(host, new BinaryLoader.Arguments(), path);
}
Expand Down
3 changes: 2 additions & 1 deletion src/NativeBridge/ManagedInterop.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@ EnvironmentBlock::~EnvironmentBlock()
FillDead(_vset[i]);
}

EnvironmentBlock::EnvironmentBlock(int verbosity, int maxThreadsAllowed, int seed)
EnvironmentBlock::EnvironmentBlock(int verbosity, int maxThreadsAllowed, int seed, const char* pythonPath)
{
// Assert that this class doesn't have a vtable.
assert(offsetof(EnvironmentBlock, verbosity) == 0);
Expand All @@ -86,6 +86,7 @@ EnvironmentBlock::EnvironmentBlock(int verbosity, int maxThreadsAllowed, int see
this->verbosity = verbosity;
this->maxThreadsAllowed = maxThreadsAllowed;
this->seed = seed;
this->pythonPath = pythonPath;
this->_kindMask = (1 << Warning) | (1 << Error);
if (verbosity > 0)
this->_kindMask |= (1 << Info);
Expand Down
5 changes: 4 additions & 1 deletion src/NativeBridge/ManagedInterop.h
Original file line number Diff line number Diff line change
Expand Up @@ -81,8 +81,11 @@ class CLASS_ALIGN EnvironmentBlock
// Check cancellation flag.
CHECKCANCEL checkCancel;

// Path to python executable
const char* pythonPath;

public:
EnvironmentBlock(int verbosity = 0, int maxThreadsAllowed = 0, int seed = 42);
EnvironmentBlock(int verbosity = 0, int maxThreadsAllowed = 0, int seed = 42, const char* pythonPath = NULL);
~EnvironmentBlock();
PyErrorCode GetErrorCode() { return _errCode; }
std::string GetErrorMessage() { return _errMessage; }
Expand Down
11 changes: 7 additions & 4 deletions src/NativeBridge/dllmain.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@
#define PARAM_MLNET_PATH "mlnetPath"
#define PARAM_DOTNETCLR_PATH "dotnetClrPath"
#define PARAM_DPREP_PATH "dprepPath"
#define PARAM_PYTHON_PATH "pythonPath"
#define PARAM_DATA "data"


Expand Down Expand Up @@ -74,13 +75,15 @@ bp::dict pxCall(bp::dict& params)
bp::extract<std::string> mlnetPath(params[PARAM_MLNET_PATH]);
bp::extract<std::string> dotnetClrPath(params[PARAM_DOTNETCLR_PATH]);
bp::extract<std::string> dprepPath(params[PARAM_DPREP_PATH]);
bp::extract<std::int32_t> verbose(params[PARAM_VERBOSE]);
bp::extract<std::string> pythonPath(params[PARAM_PYTHON_PATH]);
bp::extract<std::int32_t> verbose(params[PARAM_VERBOSE]);
std::int32_t i_verbose = std::int32_t(verbose);
std::string s_mlnetPath = std::string(mlnetPath);
std::string s_dotnetClrPath = std::string(dotnetClrPath);
std::string s_dprepPath = std::string(dprepPath);
std::string s_graph = std::string(graph);
const char *mlnetpath = s_mlnetPath.c_str();
std::string s_pythonPath = std::string(pythonPath);
std::string s_graph = std::string(graph);
const char *mlnetpath = s_mlnetPath.c_str();
const char *coreclrpath = s_dotnetClrPath.c_str();
const char *dpreppath = s_dprepPath.c_str();

Expand All @@ -93,7 +96,7 @@ bp::dict pxCall(bp::dict& params)
if (params.has_key(PARAM_SEED))
seed = bp::extract<int>(params[PARAM_SEED]);

EnvironmentBlock env(i_verbose, 0, seed);
EnvironmentBlock env(i_verbose, 0, seed, s_pythonPath.c_str());
int retCode;
if (params.has_key(PARAM_DATA) && bp::extract<bp::dict>(params[PARAM_DATA]).check())
{
Expand Down
2 changes: 1 addition & 1 deletion src/Platforms/build.csproj
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
<PackageReference Include="Microsoft.ML.Dnn" Version="0.15.1" />
<PackageReference Include="Microsoft.ML.Ensemble" Version="0.15.1" />
<PackageReference Include="Microsoft.ML.TimeSeries" Version="1.3.1" />
<PackageReference Include="Microsoft.DataPrep" Version="0.0.1.5-preview" />
<PackageReference Include="Microsoft.DataPrep" Version="0.0.1.12-preview" />
<PackageReference Include="TensorFlow.NET" Version="0.10.10" />
<PackageReference Include="SciSharp.TensorFlow.Redist" Version="1.14.0" />
</ItemGroup>
Expand Down
16 changes: 8 additions & 8 deletions src/python/docs/docstrings/EnsembleClassifier.txt
Original file line number Diff line number Diff line change
Expand Up @@ -30,14 +30,14 @@
* ``RandomFeatureSelector``: selects a random subset of the features
for each model.

:param num_models: indicates the number models to train, i.e. the number of
:param num_models: Indicates the number models to train, i.e. the number of
subsets of the training set to sample. The default value is 50. If
batches are used then this indicates the number of models per batch.

:param sub_model_selector_type: Determines the efficient set of models the
``output_combiner`` uses, and removes the least significant models. This is
used to improve the accuracy and reduce the model size. This is also called
pruning.
``output_combiner`` uses, and removes the least significant models.
This is used to improve the accuracy and reduce the model size. This is
also called pruning.

* ``ClassifierAllSelector``: does not perform any pruning and selects
all models in the ensemble to combine to create the output. This is
Expand All @@ -51,9 +51,9 @@
or ``"LogLossReduction"``.


:param output_combiner: indicates how to combine the predictions of the different
models into a single prediction. There are five available output
combiners for clasification:
:param output_combiner: Indicates how to combine the predictions of the
different models into a single prediction. There are five available
outputcombiners for clasification:

* ``ClassifierAverage``: computes the average of the scores produced by
the trained models.
Expand Down Expand Up @@ -92,7 +92,7 @@
and ``0 <= b <= 1`` and ``b - a = 1``. This normalizer preserves
sparsity by mapping zero to zero.

:param batch_size: train the models iteratively on subsets of the training
:param batch_size: Train the models iteratively on subsets of the training
set of this size. When using this option, it is assumed that the
training set is randomized enough so that every batch is a random
sample of instances. The default value is -1, indicating using the
Expand Down
16 changes: 8 additions & 8 deletions src/python/docs/docstrings/EnsembleRegressor.txt
Original file line number Diff line number Diff line change
Expand Up @@ -30,14 +30,14 @@
* ``RandomFeatureSelector``: selects a random subset of the features
for each model.

:param num_models: indicates the number models to train, i.e. the number of
:param num_models: Indicates the number models to train, i.e. the number of
subsets of the training set to sample. The default value is 50. If
batches are used then this indicates the number of models per batch.

:param sub_model_selector_type: Determines the efficient set of models the
``output_combiner`` uses, and removes the least significant models. This is
used to improve the accuracy and reduce the model size. This is also called
pruning.
``output_combiner`` uses, and removes the least significant models.
This is used to improve the accuracy and reduce the model size. This is
also called pruning.

* ``RegressorAllSelector``: does not perform any pruning and selects
all models in the ensemble to combine to create the output. This is
Expand All @@ -51,9 +51,9 @@
``"RSquared"``.


:param output_combiner: indicates how to combine the predictions of the different
models into a single prediction. There are five available output
combiners for clasification:
:param output_combiner: Indicates how to combine the predictions of the
different models into a single prediction. There are five available
output combiners for clasification:

* ``RegressorAverage``: computes the average of the scores produced by
the trained models.
Expand Down Expand Up @@ -86,7 +86,7 @@
and ``0 <= b <= 1`` and ``b - a = 1``. This normalizer preserves
sparsity by mapping zero to zero.

:param batch_size: train the models iteratively on subsets of the training
:param batch_size: Train the models iteratively on subsets of the training
set of this size. When using this option, it is assumed that the
training set is randomized enough so that every batch is a random
sample of instances. The default value is -1, indicating using the
Expand Down
10 changes: 4 additions & 6 deletions src/python/docs/docstrings/LinearSvmBinaryClassifier.txt
Original file line number Diff line number Diff line change
Expand Up @@ -5,12 +5,10 @@
.. remarks::
Linear SVM implements an algorithm that finds a hyperplane in the
feature space for binary classification, by solving an SVM problem.
For instance, with feature values $f_0, f_1,..., f_{D-1}$, the
prediction is given by determining what side of the hyperplane the
point falls into. That is the same as the sign of the feautures'
weighted sum, i.e. $\sum_{i = 0}^{D-1} \left(w_i * f_i \right) + b$,
where $w_0, w_1,..., w_{D-1}$ are the weights computed by the
algorithm, and *b* is the bias computed by the algorithm.
For instance, for a given feature vector, the prediction is given by
determining what side of the hyperplane the point falls into. That is
the same as the sign of the feautures' weighted sum (the weights being
computed by the algorithm) plus the bias computed by the algorithm.

This algorithm implemented is the PEGASOS method, which alternates
between stochastic gradient descent steps and projection steps,
Expand Down
2 changes: 1 addition & 1 deletion src/python/nimbusml/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@
Microsoft Machine Learning for Python
"""

__version__ = '1.3.0'
__version__ = '1.3.1'

# CoreCLR version of MicrosoftML is built on Windows.
# But file permissions are not preserved when it's copied to Linux.
Expand Down
2 changes: 0 additions & 2 deletions src/python/nimbusml/ensemble/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,5 +27,3 @@
'LightGbmRanker',
'LightGbmRegressor'
]


16 changes: 8 additions & 8 deletions src/python/nimbusml/ensemble/_ensembleclassifier.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,14 +57,14 @@ class EnsembleClassifier(core, BasePredictor, ClassifierMixin):
* ``RandomFeatureSelector``: selects a random subset of the features
for each model.

:param num_models: indicates the number models to train, i.e. the number of
:param num_models: Indicates the number models to train, i.e. the number of
subsets of the training set to sample. The default value is 50. If
batches are used then this indicates the number of models per batch.

:param sub_model_selector_type: Determines the efficient set of models the
``output_combiner`` uses, and removes the least significant models. This is
used to improve the accuracy and reduce the model size. This is also called
pruning.
``output_combiner`` uses, and removes the least significant models.
This is used to improve the accuracy and reduce the model size. This is
also called pruning.

* ``ClassifierAllSelector``: does not perform any pruning and selects
all models in the ensemble to combine to create the output. This is
Expand All @@ -77,9 +77,9 @@ class EnsembleClassifier(core, BasePredictor, ClassifierMixin):
``"AccuracyMicro"``, ``"AccuracyMacro"``, ``"LogLoss"``,
or ``"LogLossReduction"``.

:param output_combiner: indicates how to combine the predictions of the different
models into a single prediction. There are five available output
combiners for clasification:
:param output_combiner: Indicates how to combine the predictions of the
different models into a single prediction. There are five available
outputcombiners for clasification:

* ``ClassifierAverage``: computes the average of the scores produced by
the trained models.
Expand Down Expand Up @@ -123,7 +123,7 @@ class EnsembleClassifier(core, BasePredictor, ClassifierMixin):
:param train_parallel: All the base learners will run asynchronously if the
value is true.

:param batch_size: train the models iteratively on subsets of the training
:param batch_size: Train the models iteratively on subsets of the training
set of this size. When using this option, it is assumed that the
training set is randomized enough so that every batch is a random
sample of instances. The default value is -1, indicating using the
Expand Down
16 changes: 8 additions & 8 deletions src/python/nimbusml/ensemble/_ensembleregressor.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,14 +57,14 @@ class EnsembleRegressor(core, BasePredictor, RegressorMixin):
* ``RandomFeatureSelector``: selects a random subset of the features
for each model.

:param num_models: indicates the number models to train, i.e. the number of
:param num_models: Indicates the number models to train, i.e. the number of
subsets of the training set to sample. The default value is 50. If
batches are used then this indicates the number of models per batch.

:param sub_model_selector_type: Determines the efficient set of models the
``output_combiner`` uses, and removes the least significant models. This is
used to improve the accuracy and reduce the model size. This is also called
pruning.
``output_combiner`` uses, and removes the least significant models.
This is used to improve the accuracy and reduce the model size. This is
also called pruning.

* ``RegressorAllSelector``: does not perform any pruning and selects
all models in the ensemble to combine to create the output. This is
Expand All @@ -77,9 +77,9 @@ class EnsembleRegressor(core, BasePredictor, RegressorMixin):
can be ``"L1"``, ``"L2"``, ``"Rms"``, or ``"Loss"``, or
``"RSquared"``.

:param output_combiner: indicates how to combine the predictions of the different
models into a single prediction. There are five available output
combiners for clasification:
:param output_combiner: Indicates how to combine the predictions of the
different models into a single prediction. There are five available
output combiners for clasification:

* ``RegressorAverage``: computes the average of the scores produced by
the trained models.
Expand Down Expand Up @@ -117,7 +117,7 @@ class EnsembleRegressor(core, BasePredictor, RegressorMixin):
:param train_parallel: All the base learners will run asynchronously if the
value is true.

:param batch_size: train the models iteratively on subsets of the training
:param batch_size: Train the models iteratively on subsets of the training
set of this size. When using this option, it is assumed that the
training set is randomized enough so that every batch is a random
sample of instances. The default value is -1, indicating using the
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -57,14 +57,14 @@ class EnsembleClassifier(
* ``RandomFeatureSelector``: selects a random subset of the features
for each model.

:param num_models: indicates the number models to train, i.e. the number of
:param num_models: Indicates the number models to train, i.e. the number of
subsets of the training set to sample. The default value is 50. If
batches are used then this indicates the number of models per batch.

:param sub_model_selector_type: Determines the efficient set of models the
``output_combiner`` uses, and removes the least significant models. This is
used to improve the accuracy and reduce the model size. This is also called
pruning.
``output_combiner`` uses, and removes the least significant models.
This is used to improve the accuracy and reduce the model size. This is
also called pruning.

* ``ClassifierAllSelector``: does not perform any pruning and selects
all models in the ensemble to combine to create the output. This is
Expand All @@ -77,9 +77,9 @@ class EnsembleClassifier(
``"AccuracyMicro"``, ``"AccuracyMacro"``, ``"LogLoss"``,
or ``"LogLossReduction"``.

:param output_combiner: indicates how to combine the predictions of the different
models into a single prediction. There are five available output
combiners for clasification:
:param output_combiner: Indicates how to combine the predictions of the
different models into a single prediction. There are five available
outputcombiners for clasification:

* ``ClassifierAverage``: computes the average of the scores produced by
the trained models.
Expand Down Expand Up @@ -123,7 +123,7 @@ class EnsembleClassifier(
:param train_parallel: All the base learners will run asynchronously if the
value is true.

:param batch_size: train the models iteratively on subsets of the training
:param batch_size: Train the models iteratively on subsets of the training
set of this size. When using this option, it is assumed that the
training set is randomized enough so that every batch is a random
sample of instances. The default value is -1, indicating using the
Expand Down
Loading