Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support defining a region for read, show, save, snap for visual automation #461

Closed
kensoh opened this issue Jun 21, 2019 · 1 comment
Closed
Labels

Comments

@kensoh
Copy link
Member

kensoh commented Jun 21, 2019

Following up on #457, where transparency (0% opacity) in images is supported in visual automation for advance use cases, raising an issue to create a new PR that allows defining regions for steps read, show, save, snap. This allows users to define a square/rectangular region on the screen to perform OCR for text extraction, and saving screenshot of that region.

@kensoh kensoh added the feature label Jun 21, 2019
kensoh added a commit to tebelorg/TagUI that referenced this issue Jun 21, 2019
kensoh added a commit that referenced this issue Jun 21, 2019
Following up on #457, where transparency (0% opacity) in images is supported in visual automation for advance use cases, raising a PR that allows defining regions for steps read, show, save, snap. This allows users to define a square/rectangular region on the screen to perform OCR for text extraction, and saving screenshot of that region.
@kensoh
Copy link
Member Author

kensoh commented Jun 21, 2019

Above commit implements this feature. Prior to packaged release, feature available from cutting edge version here - https://github.com/kelaberetiv/TagUI#set-up

Below is the updated section in visual automation to be updated to readme and tutorial on next packaged release. Adding Also for these steps, (x1,y1)-(x2,y2) can be used as the identifier to define the region of interest on the screen to perform OCR or capture snapshot.

Transparency (0% opacity) is supported in .png images, for eg using an image of an UI element with transparent background to enable clicking on an UI element that appears on different backgrounds on different occasions.

Another example is an image of the window or frame (PDF viewer, MS Word, textbox etc) with the center content of the image set as transparent. This allows using read, show, save, snap steps to perform OCR and save snapshots for application windows, containers, frames, textboxes with varying content. Also for these steps, (x1,y1)-(x2,y2) can be used as the identifier to define the region of interest on the screen to perform OCR or capture snapshot.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant