See full description below
- Maven, version 3+
- Java JDK 1.8
- This settings.xml in your /.m2 directory
This will build, unit test, and package the whole project (all of the sub-modules). The artifact will be generated in: target
$ mvn clean install
Unit tests
This will run all tests in the project (and sub-modules).
$ mvn test
If you want to remote debug a single java unit test (default port is 5005):
$ cd core
$ mvn test -Dtest=<<YourTest>> -Dmaven.surefire.debug
Integration tests In addition to the unit tests, there are integration tests in the core project.
$ mvn verify -DrunITs
To run a single integration test:
$ mvn verify -DrunITs -Dit.test=<<YourIT>>
To run a single integration test in debug mode (for remote debugging in an IDE) on the default port of 5005:
$ mvn verify -DrunITs -Dit.test=<<YourIT>> -Dmaven.failsafe.debug
IntelliJ
- Don't use IntelliJ's built-in maven. Make it use the same one you use from the commandline.
- Project Preferences -> Build, Execution, Deployment -> Build Tools -> Maven ==> Maven home directory
Community Plugin Kick-starter
Note that the below information may be outdated
Pentaho is very well known for being a very good Business Analytics software, but is in fact much more than that; Pentaho is a great platform to build on top of.
Using an easy analogy, I see Pentaho acting as an operating sytem where people can build Application on top of
The goal of CPK is to provide a simple and easy way to develop pentaho plugins that behave like packaged applications, simplifying it's structure.
The UI is built using CDE, with a simple methodology to create new dashboards / pages and a sitemap that provides simple navigation and a default template
There are 3 options for doing server-side code:
- Kettle transformations / jobs
- Javascript server side code execution
- Java classes
The first two are recommended, since it's easier to register new endpoints just by dropping code in a directory and no compilation is necessary.
With this approach not only we expect to make it easier and faster to develop plugins, we also hope to lower down the specific technical requirements to build them. The end goal is that business consultants are able to build new plugins, not requiring specific java knowledge.
Kettle transformations or jobs of a CPK plugin are automatically exposed as rest endpoints. To execute the corresponding transformation/job simply perform an HTTP GET request to the endpoint.
http://{host}/pentaho/plugin/{cpkPluginId}/api/{kettleFileName}
If you have need of a private transformation/job, prefix its file name with an underscore and it won't be made available in an endpoint.
It is possible to specify from which step/job entry we want to fetch row results. To do so, simply include the stepName parameter in the query string of the request. For example, to get the transformation results from the "OUTPUT" step execute the request
http://{host}/pentaho/plugin/{cpkPluginId}/api/{kettleFileName}?stepName=OUTPUT
A kettle endpoint may have multiple steps or job entries from where we can fetch results. Because allowing to get results from any given intermediate step/job entry would be a possible security breach, it is necessary to explicitly identify which can be used as output. This is done by setting the name of the step/job entry to start with "OUTPUT". For example, a kettle endpoint for a transformation that has five steps with names "Step1", "Step2", "OutputStep", "OUTPUTStep1" and "OUTPUTStep2" will only allow getting results from "OUTPUTStep1" and "OUTPUTStep2".
If no stepName is specified in the HTTP request, or if the stepName is an invalid output step/job entry, CPK will use the default value of the parameter cpk.result.stepName defined in the transformation/job. If no stepName is specified in the HTTP request and no default value is set for cpk.result.stepName in the transformation/job, CPK will choose one valid output step to get the results from. If there is no valid output step/job entry, the transformation/job will still be executed but no row result information will be returned.
It is also possible to pass parameter values to a transformation/job at runtime. This is achieved by encoding the parameter names in the query string with a param prefix. For example, if a transformation expects a parameter named foo and you want to set bar as its value, your request should be
http://{host}/pentaho/plugin/{cpkPluginId}/api/{kettleFileName}?paramfoo=bar
The nature of a transformation/job result will vary and as such it is desirable to be able to format it differently. For example, a transformation may produce a tabular result set which we want to feed into a chart or it may select and filter some files which we want to zip and download.
CPK allows formatting a kettle endpoint result in four different ways:
- Json: returns the result rows in a standard CDA like result set format (metadata/queryinfo/resultset)
- ResultFiles: gets the files that were set in the result
- SingleCell: returns the content of the first cell of the first row in the result
- ResultOnly: returns status information about the execution
To select the desired format set the kettleOutput query string parameter to the chosen value. For example
http://{host}/pentaho/plugin/{cpkPluginId}/api/{kettleFileName}?kettleOutput=Json
It is possible to specify the default value to be used when the kettleOutput parameter is not defined in the query string. This is achieved by setting the default value of the transformation/job parameter cpk.response.kettleOutput.
If no query string or transformation/job parameter is defined for result formatting, CPK uses the following decision logic to determine which one best fits the endpoint result.
- Result has file names?
- Yes => ResultFiles
- No => Result from a kettle Job?
- Yes => Job's result contains rows?
- Yes => Json
- No => ResultOnly
- No => Result has only one row with a single cell?
- Yes => SingleCell
- No => Json
- Yes => Job's result contains rows?
This logic may also be explicitly triggered by setting kettleOutput value to Infered.
Some output formats may trigger the download of a file.
When ResultFiles is used and the result contains multiple files, CPK compresses them into a single Zip file that is returned as an attachment of the response. However, results that have a single file do not trigger a download by default, i.e. the file is not marked as an attachment of the response.
It is possible to change this default behaviour by setting the transformation/job parameter cpk.response.download to true. Also, the default value is always overrided when the value of the query string parameter download is defined in the request. For example, consider a transformation which result has one file and which cpk.response.download is set to true. If the following request is executed the file will not be marked as an attachment.
http://{host}/pentaho/plugin/{cpkPluginId}/api/{kettleFileName}?kettleOutput=ResultFiles&download=false
When the result has single file, CPK will try to determine the mime type from the file extension. If the mime type is known a priori you can override the default behaviour by setting the transformation/job parameter cpk.response.mimeType to the desired value (e.g. application/xml).
It is also possible to define the name of the downloaded file by setting the tranformation/job parameter cpk.response.attachmentName to the intended value.
In some cases it is useful to interpret the contents of a cell as the contents of a file. In this way the SingleCell output format behaves similarly to ResultFiles and as such it is sensible to the parameters download, cpk.response.download, cpk.response.mimeType and cpk.response.attachmentName.
For example, consider a transformation that writes xml to the single cell of the single row of the result with parameters cpk.response.mimeType=application/xml, cpk.response.attachmentName=myXml.xml and cpk.response.download=true. The following request will download a file named myXml.xml which content is the content of the single cell result.
http://{host}/pentaho/plugin/{cpkPluginId}/api/{transformationKettleFileName}
Each CPK plugin has a cache that stores the results obtained from the execution of its endpoints. Although there is one cache per CPK plugin, this cache can be enabled/disabled per kettle endpoint. By default caching is disabled and to enable it you set the value of the transformation/job parameter cpk.cache.isEnabled to true. You can also specify how long a result should be cached by setting the value of the transformation/job parameter cpk.cache.timeToLiveSeconds.
If at runtime you wish to bypass an enabled cache use the query string parameter bypassCache set to true. This will force the transformation/job to execute and update the previous cached valued.
##Dashboards
A Dashboard is used like the endpoints are. The URL to call is basically the same, the only change is the name after the "example" field on the address bar!
http://{host}/pentaho/plugin/{cpkPluginId}/api/{dashboardName}
Here you have a very useful parameter that is: "mode". It allows to change between "edit" and "render" mode on the fly! All you need to do is append "?mode=edit" to the URL like this:
http://{host}/pentaho/plugin/{cpkPluginId}/api/{dashboardName}?mode=edit
After you're finished editing the dashboard just hit "back" on your browser and refresh the page.
##CPK command list
status : Displays the status of the plugin and all it's endpoints
refresh : Reloads all the configurations, endpoints and dashboards, also clears the endpoints cache!
version : Returns the plugin version (Defined on the plugins "version.xml" file or through the Control Panel)
getSitemapJson : Returns a JSON with the plugins sitemap (Dashboards only!)
elementsList : Returns a JSON with the whole list of elements present on the plugin (dashboards and kettle endpoints)
To perform a command:
http://{host}/pentaho/pentaho/plugin/{pluginId}/api/{command}
This is the resulting plugin structure. Ideally, no compilation is necessary, so everything except maybe the lib directory could be stored in a VCS system.
This is the proposed stub configuration
to be completed
CPK_Plugin
|-- conf
|-- dashboards
|-- endpoints
|-- lib
`-- plugin.xml
Besides providing the regular templating for creating new plugins, CPK can have an administrative UI with the following features:
- List existing plugins
- Detect if the plugins are up to date with the latest version of CPK
- Allow the creation of new plugins
- Allow to change plugin metadata
- List and register new endpoints (UI and code)
Here's a list of stretch goals / nice to have
- Import UI from existing dashboards in solution
- Allow editing dashboards from this UI
- Submit marketplace metadata to Pentaho
- Generate distribution zip package
As the CPK framework or any of its dependencies gets improved, the plugins themselves can't stay outdated. There will be a version information attached to the CPK plugin version so that it's possible to upgrade to the latest version.
CPK will have as little code as possible, making it as simple as possible to develop plugins. However, it will need a few dependencies:
- Hitachi Vantara
- CPF - Community Plugin Framework, with the common set of code for the plugins
- CDE - Community Dashboard Editor
- CDF - Community Dashboard Framework
- CDA - Community Data Access
Once a plugin is developed, and the authors think it's in a state that can be shared, CPK will be able to generate a packaged plugin and metadata information so it can be integrated into Pentaho's marketplace.
Pentaho will then be able to categorize / approve the plugin so that it becomes available to other users through the marketplace
This project uses MPLv2