absolute mouse buttons on dock - macos

I have created an absolute mouse piece of hardware.
This is very much like a wacom tablet, but it just uses a regular mouse going into a piece of hardware that then increments or decrements the X and Y coordinates in order to specify the exact location of the mouse (yes I need to do this for another reason).
While everything works fine on multiple operating systems, it works fine on Mac OSX Yosemite except ---->
When I move to the dock, if I click on an icon it will not start the application, I can move the icon and once open I can close the app or scroll, anything you would expect from a mouse, it will not start properly. Also the back arrow in the system settings/preferences will only operate if I am on the exact right location on the button, i.e. normally there is no back action, but if I mess with it and get it in just the right location it will finally work.
This sounds like an oversight in the operating system, does anyone have any insight into this?

Related

What determines which screen the account control popup will appear on?

I'm running a Windows 10 system with one touch screen, one normal screen and no mouse or keyboard. I need to run an application with administrator privileges.
I'm worried about which screen the Account Control popup will appear on - if it appears on the non-touch screen then users will have no way to allow the program to run.
I know that:
It runs on screen 2 on my development PC (with mouse and keyboard, no touch screen), regardless of whether screen 2 is on the left or right
It runs on the same screen regardless of whether the application was launched by a shortcut on the left or right screen.
I tried to google this, but the only things that came back were links on how to disable User Account Control, which we really don't want to do.
What is the logic that determines which screen the popup appears on?
[Edit for new things I have tried]
Unplugging screen 2, running the application, closing the application, plug screen 2 back in.
Dragging the UAC popup to screen 1, clicking details, changing UAC settings and returning them to the original state.
[SOLVED]
I figured it out (at least partially) but I can't mark it as answered for at least two days.
Basically, after unplugging and replugging screens the other way around, changing the primary display worked. This didn't work until after I did the plug/unplug, but there's no guarantee that was the thing that did it, since I'd tried a few other things.
So after physically unplugging and replugging monitors so that screen one became screen 2 and the reverse, switching which screen was the primary display works to control the popup now.
Changing the primary display didn't work before. But it does now. I have no idea why - there's no guarantee that it was to do with switching them around.
I would suspect it to pop up by default on the primary display, but remember its last position. This would appear to be confirmed by TechNet: UAC Prompt displays on 2nd monitor only RRS feed:
escknx:
Regardless of what I try, UAC prompts display only on 2nd monitor, while everything else works fine on primary monitor.
Stereoman:
drag the UAC Prompt to your Primary Monitor and click show more details on the prompt then click change when these notifications appear, another UAC window will open up with a vertical bar with 4 different settings, just drag the bar setting down and then back up to it's original setting position then click ok, this will re trigger UAC's screen position to your primary monitor.
So: upon deployment, make sure the touch screen is your primary display, and position the UAC popup there if it isn't.

App don't work correctly on another Computer

I've two PC with dual monitors, with the same configuration (USB to VGA converter, same USB port, same VGA port on the second monitor), same drivers, same version, same Windows (Seven Professional), sames monitors with sames resolution, same on everything I know how to access to.
I've an app on which when I click on a button --> display another maximized window.
On the first PC, the app work correctly, the other maximized window display in full screen on the second monitor.
On the second PC, the app doesn't work correctly ! ):
When I click on the button the other maximized window just open (maximized) on the first monitor, the principal where the app is displayed (as same as the first PC).
So I've tried to
change the screen order (put 1 on left of 2 or 2 on left of 1)
check if the drivers have a problems, but no (no warning or interrogation sign in the "Uninstall a program" program
change size resolution on both screens but nothing
un-maximize the second window's app and close it on the correct screen, nothing, the window just maximize and display on the bad screen like before
Sooooooo I've no idea of where the problem could came...
Sorry if my English is bad and thanks in advance for your answers you're my last resort.
Have a nice day

Secondary touch monitor not moving with the cursor but still pressing buttons (or emitting events for Qt app)

I intend to make an app for Windows in Qt with multi-monitor support, one (secondary) monitor being touch. I want one person to be able to work with the touch screen independently on a second person, which would be working on the main screen with mouse and keyboard.
I don't have the the touch monitor yet, so I don't know how it really works, but I am afraid that touching the monitor would move the cursor (mouse pointer), making the work with mouse very hard.
So my question:
Is it possible somehow to make the touch screen not to affect the cursor in any way (not interrupting drag&drop counts, too), but still to be able to push buttons n' stuff, and that on either Windows or Qt level?
No buttons pushing, but generating QTouchEvents (or similar) would be sufficent, too.
Thanks for your responses.

Win7 and multiple monitors: switch focus when moving mouse to different monitor?

I recently upgraded to a dual monitor setup at work, and while the extra real estate is very nice, there's one annoyance: my intuitive reaction is that there are two "active" windows now, namely the topmost window in each monitor -- and I frequently get surprised when keyboard events go to the actual active window, rather than the one I've moused over and am looking at.
There's a setting in the control panel that gives this effect (ease of access -> make the mouse easier to use -> activate a window by hovering over it with the mouse) but it also acts on windows within the same monitor, which I don't want.
I frequently use my ThinkPad's scrolling function on unfocused windows which I don't want to receive focus, which come to think of it probably adds to my confusion, since I can scroll my email in the other window but my keyboard shortcuts don't go there.
Is there any way to achieve this effect or am I just wishing?
Thanks,
Ryan
Yeah, get a Mac :-p
In all seriousness OS X does provide this functionality. It might be worth searching for an add on that does the same sort of thing. I know of Wizmouse -- http://antibody-software.com/web/software/software/wizmouse-makes-your-mouse-wheel-work-on-the-window-under-the-mouse/
There might be more though.
AT LAST!!! Windows 10 has this support :-)
SM
You can change the settings to use classic windows appearance etc. and try to focus on the border color of the window. The board changes on the active window.
I use two monitors and there really isn't much you can do besides change your behavior.
Select things from the taskbar, drag active windows to the same screen and always refer to inactive windows by moving them to the inactive windows monitor and remember to go back to the window you want to be active.

Xcode window positions when switching from dual to single monitor

When I'm working at home I plug my MacBook in to my 20" monitor as a second (right side) monitor. I do all my editing in Xcode on the larger monitor, and leave the menu bar, debugging and documentation on the laptop's monitor.
The problem is when I disconnect from the second monitor and want to work on code in "laptop" mode. Now, whenever I open a file for editing, it shows up almost entirely off the screen and I have to drag it over to edit it. I understand (sort of) why this is happening, since I last had it open in a different monitor. What I'd like to do is reset the window positions in Xcode so the edit windows show up completely on the laptop monitor. Any ideas on how to do this?
I'd even be willing to nuke entries out of the preferences (or set up an applescript to do it) but the xcode plist is inscrutable to me.
Right click the project in finder, select show package contents, and delete the two username.* files. Its not automated... but it works.
Additionally if your using some kind of version control system you probably want to add username.* to your ignore pattern.

Resources